-
Notifications
You must be signed in to change notification settings - Fork 1.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
ollama support? #1001
Comments
@txhno Sorry about that random weird comment...removed your reply too since it had a quote of the link in it, hope that's OK! On topic -- exploring Ollama support is a really good idea. My understanding is that they just use We can put this on our backlog to investigate, but if you (or anyone reading this!) have some knowledge about how Ollama works, I'd be happy to tag-team and support a PR here. @riedgar-ms @nking-1 for awareness |
Hi! I’ve implemented a thin wrapper for Ollama support in my fork. Can you give it a shot before I submit a PR? Thanks! |
When will Ollama support be available? |
This is likely because the model's chat template did not load. see the comment here |
From what I know, for a model to work with guidance, it needs to provide guidance role start and role end tags, e.g., Currently, guidance uses the templates of the models as keys to find the constructed chat template classes; otherwise, it uses the predefined chat template class, Ollama uses llamacpp as its backend, and the models that Ollama serves contain a template and modelfile, see the output of Ollama api
Supposedly, if the model Ollama serves contains a chat template and the corresponding chat template is implemented in guidance, guidance will work fine. But for all the models of Ollama to fully work, it needs a way for the forked Ollama models to locate their role tags. One approach is to implement chat templates for all Ollama models in I am not sure if there are any other ways to automatically retrieve the role tags based on the model information provided by Ollama. If I have misunderstood anything, please correct me. Fan |
You are spot on. There is #947 which attempts to extract a |
Is your feature request related to a problem? Please describe.
would want to reuse the models that I already have downloaded on ollama
Describe the solution you'd like
being able to use models.ollama(model_name_or_path)
Describe alternatives you've considered
llama cpp works as of now, but ollama would just make process of using this app a lot more user friendly having downloads automated and models stored centrally
Additional context
none
The text was updated successfully, but these errors were encountered: