-
Notifications
You must be signed in to change notification settings - Fork 10
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Feat/add nvidia nim #100
Feat/add nvidia nim #100
Conversation
…gents4Pharma into feat/add-ollama
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
The integration for NVIDIA NIMs looks really cool 😎. The change in url is a neat trick to switch between API and local too.
I only had small comment on the logic implemented to switch between models on the front end that I think will also allow for better readability of the display names. Let me know your thoughts if you disagree.
@dmccloskey I have implemented your suggestion. While testing the application, I noticed that Streaming with ChatNVIDIA does not work in tool-calling. I see that others have also opened issues regarding this on GitHub. So for now, I have disabled streaming when using ChatNVIDIA (only invoke possible). I will get back to this when I have found a reliable solution. Until then streaming is enabled only for OpenAI models. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
That is unfortunate about NVIDIA streaming 🙁. Hopefully, they will fix the issue quickly.
🎉 This PR is included in version 1.18.0 🎉 The release is available on GitHub release Your semantic-release bot 📦🚀 |
For authors
Description
demo11.mp4
In this PR, I'd like to add open-source text-generation and text-to-embedding models by NVIDIA NIM to T2B.
Changes Introduced
Added a new key (
text_embedding_model
) in the state to store text-to-embedding models (typeEmbeddings
class in LangGraph).Updated the existing key (
llm_model
) in the state to store chat model objects (typeBaseChatModel
in LangGraph).base_url
parameter in theChatNVIDIA
object, users can switch from NVIDIA API to local deployment as described here.Updated the Streamlit app to provide users with LLM selection options:
gpt-40-mini
(OpenAI/Default)meta/llama-3.3-70b-instruct
(NVIDIA NIM)and Embedding model options:
nvidia/llama-3.2-nv-embedqa-1b-v2
(NVIDIA NIM/Default)text-embedding-ada-002
(OpenAI)A small majority of tests now use NVIDIA NIM models.
Future Enhancements:
Fixes # (issue) N/A
Type of change
Please delete options that are not relevant.
How Has This Been Tested?
Please describe the tests you conducted to verify your changes. These may involve creating new test scripts or updating existing ones.
tests
foldertests/testX.py
)Checklist
tests
folder) that prove my fix is effective or that my feature worksFor reviewers
Checklist pre-approval
Checklist post-approval
develop
intomain
? If so, please make sure to add a prefix (feat/fix/chore) and/or a suffix BREAKING CHANGE (if it's a major release) to your commit message.Checklist post-merge
develop
intomain
and is it suppose to run an automated release workflow (if applicable)? If so, please make sure to check under the "Actions" tab to see if the workflow has been initiated, and return later to verify that it has completed successfully.