You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
如果我想临时部署v3,我根据issue方法注册后,如果加入--quantization moe_wna16会报错Model DeepSeek-V3-awq cannot be run on engine vLLM, with format awq, size 671 and quantization moe_wna16.
不加这个参数model可以found但是会OOM,
大家好,有两个问题请教下:
1,我用自定义注册Deepspeek-v3-awq后,无法选择vllm引擎,请问如何解决?
2,看到前面老师说v3和v2.5架构一致,但是用v2.5拉起时会报 Model not found, name: deepseek-v2.5, format: pytorch, size: 236, quantization: moe_wna16的错误,请问如果如何解决呢?
谢谢!
The text was updated successfully, but these errors were encountered: