Description:
This issue proposes the addition of RAG (Retrieval-Augmented Generation) to Polyglut multimodel LLM chat to enhances the app's memory with responses from local knowledge base, aligning with Polyglut's main goals of being a model experimentation tool that develops an almost human-like memory.
Motivation
- Improve response quality by grounding answers in up-to-date or domain-specific knowledge.
- Enable users to upload or reference their own documents for context-aware chat.
- Reduce hallucinations and increase trust in the model’s outputs.
Proposed Solution
Indexing: Creating a vector-based knowledge base from the project's documentation, code, and other relevant files.
Retrieving: When a user asks a question, the system will search this knowledge base for the most relevant information.
Augmenting: The retrieved information will be injected into the LLM's prompt, providing the model with real-time, specific context before it generates a response.
This will enable the LLM to answer questions about the Polyglut codebase and architecture accurately and with references.
Tasks
Task 1: Select and Set Up a Vector Store.
Task 2: Data Ingestion Pipeline.
Task 3: Implement the Retrieval Logic.
Task 4: Integrate RAG into the Chat API.
Task 5: Display Source Information in the UI.
Acceptance Criteria
Labels
enhancement, AI/ML
Description:
This issue proposes the addition of RAG (Retrieval-Augmented Generation) to Polyglut multimodel LLM chat to enhances the app's memory with responses from local knowledge base, aligning with Polyglut's main goals of being a model experimentation tool that develops an almost human-like memory.
Motivation
Proposed Solution
Indexing: Creating a vector-based knowledge base from the project's documentation, code, and other relevant files.
Retrieving: When a user asks a question, the system will search this knowledge base for the most relevant information.
Augmenting: The retrieved information will be injected into the LLM's prompt, providing the model with real-time, specific context before it generates a response.
This will enable the LLM to answer questions about the Polyglut codebase and architecture accurately and with references.
Tasks
Task 1: Select and Set Up a Vector Store.
Task 2: Data Ingestion Pipeline.
Task 3: Implement the Retrieval Logic.
Task 4: Integrate RAG into the Chat API.
Task 5: Display Source Information in the UI.
Acceptance Criteria
Labels
enhancement, AI/ML