Question about the RAG process #7649
Unanswered
hamzachebbigit
asked this question in
Questions
Replies: 2 comments 1 reply
-
Are you running into a problem with the max tokens allowed? Or like you want a general description of a typical RAG pipeline? |
Beta Was this translation helpful? Give feedback.
1 reply
-
It is mentioned in the tutorial notebook |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hello :) I'm new to Haystack and want to understand how RAG works.
So every step I completely understood in your tutorial except the part about the prompt: You see in your documentation you said that the prompt should have 2 inputs: the documents and the question.... but I don't understand how it works exactly because when I tested it with Pinecone, it fetches the content of the chunks inside the Pinecone vector dataset and puts them into the prompt, then passes them to chatgpt (the llm in this case); If my understanding is correct, and please explain further if I'm wrong hhh what's the point of all this hassle to embed the chunks and store them just to give the exact text in the prompt? I could do the same thing, just copy and paste the text and I have a problem with the maximum number of tokens in the prompt. Maybe Iam missing something. I thought the idea of RAG is to avoid all this and tell the llm to fetch the content for each time with ID maybe and take the content of it...would this be possible with defining new pipelines ? Thank you in advance :)
Beta Was this translation helpful? Give feedback.
All reactions