SciPhi-self-rag works properly only when running inference using the sciphi framework. In that case it works wonderfully well
this post was submitted on 22 Nov 2023
1 points (100.0% liked)
LocalLLaMA
3 readers
1 users here now
Community to discuss about Llama, the family of large language models created by Meta AI.
founded 1 year ago
MODERATORS
Can you share the code?
Some updates:
- I changed to Jina-small-en-v2, base model crash due to lack of RAM under WSL2
- Make a parents retriever (chunk 2000) and input it child retriever (chunk 400), 0 overlap (will share the method)
- Still use sciphy model but this time using the right template (From the indication from The bloke) by adding a template prompt rather than Alpaca prompt and it resolves the problem of hallucination
- Put text oobabooga on instruct by default, loader exllamav2hf
I got a strong 90% of success with the PDF, will send the code when this will be cleaned and optimized, thank you all for the help 😊
Make a parents retriever (chunk 2000) and input it child retriever (chunk 400), 0 overlap
Can you please elaborate.
You can find the method here :
Have you tried to chunk it based on headings or did you just chunk the pdfs into chunks with fixed token sizes?