I'm working on a RAG application. I'm using a pdf file as a text data and I have clean the data as well using NLTK. I already have ollama install in my local system and also llama2 model. I'm trying to embed the data in to vector form using ollama embedded model mxbai-embed-large and store it on chromabd, but I can't embed the whole data because there are too many of tokens so I embed some data. but if I give the prompt and get response, it give me the response but not according to my data. and how do I finetune the model? I just need guidance and sequence so I work on the project. It's important to me. I need your guidance. please reply. I'm using windows10. Here's my code where I'm having issues: my code image
I'm trying to get all data embedded and store it in chromadb and I also want to fine-tune the llama2 mode. I also get a clue that if I use UNSLOTH then it can be done. but don't know how. I don't to use hugging face because if I want to do with that then it can be done quickly also with langchain. But I want to do this with Ollama along my local system.
Just so you don't have to wait for an answer from others. Here's a tutorial for RAG implementation using Langchain with Ollama (base model and embedding) also using Chroma by [pixegami][1]
ps, i just thought it matched your specifics [1]: https://youtu.be/tcqEUSNCn8I?si=pd4ij85CJqHC4XoR