You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository was archived by the owner on Feb 15, 2025. It is now read-only.
One of the limitations of Large Language Models (LLMs) is that they are only able to respond to scenarios contained within training data - and training on new data is expensive given the size of the model. Retrieval Augmented Generation (RAG) is a technique to supplement the LLM with new data to enable it to provide more up-to-date responses.
Acceptance Criteria:
IDAM for managing access to RAG Data
API is compliant with OpenAI endpoints (Chat, Embeddings, Files, Assistants at a minimum).
Handle GPT2 (bottleneck) in RAG Backend
Better embeddings model (currently Instructor-XL)
Smarter / Better RAG
Definition of Done:
The content you are editing has changed. Please copy your edits and refresh the page.
Problem Statement:
One of the limitations of Large Language Models (LLMs) is that they are only able to respond to scenarios contained within training data - and training on new data is expensive given the size of the model. Retrieval Augmented Generation (RAG) is a technique to supplement the LLM with new data to enable it to provide more up-to-date responses.
Acceptance Criteria:
Definition of Done:
Tasks
The text was updated successfully, but these errors were encountered: