-
Notifications
You must be signed in to change notification settings - Fork 2.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Mistral via Azure #1678
Comments
Non-OpenAI models from Azure AI Foundry use OpenAI chat completions template. Simply run it with the OpenAI Chat template from settings.yaml. I got DeepSeek-R1 from Azure AI Foundry to work that way. |
I still can't get it to work - I changed the endpoints to reflect the "target URIs" in my LLM and embeddings deployments, but now I get a |
@pholz I've been testing other models in Azure AI Foundry. I realised that while DeepSeek-R1 works fine, other models will not work due to the different API endpoint format. base_url in OpenAI Chat mode:
GraphRAG auto completes the base_url into the following:
Seeking help from the GraphRAG team to add support for these Azure AI Model Inference models. Appreciate any workaround! |
Managed to use LiteLLM to route access to Azure AI Services (Azure AI Studio / Azure AI Foundry). Can give it a try! |
Thank you, I managed using LiteLLM proxy and a serverless mistral-mini deployment. If anyone else tries this, note that you will probably need to set |
Should it be possible to use a non-openAI model which is however hosted on Azure? Specifically, I would like to use Mistral, and I have it deployed on Azure already. But when running indexing, I keep getting
Operation 'chat' failed
errors from thefnllm
package.If I look at the model URLs, OpenAI deployments take the shape
<baseURL>/openai/deployments/<deploymentname>/chat/completions?api-version=2024-08-01-preview
whereas the Mistrall deployment uses<baseURL>/models/chat/completions?api-version=2024-05-01-preview
. Since in the settings file I only set the baseURL, I assume that the rest of the API is somehow assumed? Or is there a way to change this too?The text was updated successfully, but these errors were encountered: