-
Notifications
You must be signed in to change notification settings - Fork 54
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Ollama with llama3.1 not working #13
Comments
hey @gavinblair - this primarily stems from the fact that the model was not able to generate a valid output. can you tell me which quantised version of llama 3.1 are you using? |
8B. I'm using Q4_0. I'll try with Q5_K_M once I figure out how to use a different base url. |
maybe try 8b-instruct-q4_0, folks in the community have been able to make it work with llama 3.1 8b models |
I had the Get RetryError[<Future at 0x182e2357a60 state=finished raised ValidationError>] with ollama issue filed. llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output. |
I tried llama3.1:8b-instruct-q4_0 and it gave me the same error ================================================== Current State: agentq_base ================================================== Error executing the command play shape of you on youtube: RetryError[<Future at 0x21faa7ade40 state=finished raised ValidationError>] |
I'm encountering the same issue on Windows 10 with ollama3.1:latest, and I’ve tried other models but faced the same problem. I believe the issue might be related to this code snippet: As far as I understand, Ollama doesn’t require an API key, and the base URL when installed locally should be http://localhost:11434. dditionally, I encountered an authentication error with the Groq API, which I resolved by modifying the provider.py file as follows:
I hope this gets resolved soon. If I find a solution, I’ll let you know. |
any update on this , the issue still not fixed |
Here is the output I get, running with Ollama locally (just the example from the README)
The text was updated successfully, but these errors were encountered: