Replies: 2 comments 3 replies
-
This is not an answer but more like a question if you had found a solution for this problem. |
Beta Was this translation helpful? Give feedback.
1 reply
-
I had the same issue initially but not anymore, I think the solution was to remove the max_tokens argument from parse_curl_args, but tbh I'm not sure, might just have been solved by switching model to qwen-7b |
Beta Was this translation helpful? Give feedback.
2 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I have ollama setup hosting the deepseek-v2 model. In my neovim config, I've kept the provider as openai but changed the endpoint to be my local ollama server.
What I'm finding is that while the context, instruction, and question are all being sent in the request payload, the responses I get back are just wildly wrong (e.g. Q: "What's this lua code do?" A: "Here's some python code for a web scraper"). I suspect it has to do with the message payload and/or prompts? I suspect that the prompt statements and maybe format need to be tweaked to get avante to work against other providers.
Curious what others are doing to combat this type of behavior. Do I need to configure the prompt via avanterules (for example modify the payload to match https://github.com/continuedev/continue/blob/3791774e45ae87821f6539f38da5c0752d63fe4f/core/llm/templates/edit.ts#L173)? Or do I need to create a custom vendor provider and override the payload construction functions.
Have others had success when dealing with non-standard or local models?
Beta Was this translation helpful? Give feedback.
All reactions