Implementing RAG, some questions on llama.cpp #12125
gnusupport
started this conversation in
General
Replies: 0 comments
# for free
to join this conversation on GitHub.
Already have an account?
# to comment
-
While working with Emacs Lisp, I have so far imlpemented text splitting into chunks and using the embeddings model
That works fine and well and embeddings are recorded in the PostgreSQL database with the pgvector extension, this works well. Searching by embeddings works well.
I can quickly implement listing of documents, or people, whatever I am searching. This is done by PostgreSQL database.
And then according to what I learned, I am supposed to insert that information into the context of the LLM prompt in order to get the RAG functionality.
Sure I have some clue how to do it by curl, or Emacs Lisp over the API endpoint.
Though I would like to know is there, or would be there any way of implementing it in background so that I can somehow inject that stuff and get the responses over the Llama.cpp web UI?
Beta Was this translation helpful? Give feedback.
All reactions