diff --git a/config.ini b/config.ini index c7fa7f29..530d7f0c 100644 --- a/config.ini +++ b/config.ini @@ -36,11 +36,11 @@ client_url = "http://127.0.0.1:8888/inference" [llm.server] # local LLM configuration -# support "internlm/internlm2-chat-7b", "internlm2_5-7b-chat" and "qwen/qwen-7b-chat-int8" +# support "internlm/internlm2-chat-7b" and "qwen/qwen-7b-chat-int8" # support local path, for example -# local_llm_path = "/path/to/your/internlm2_5" +# local_llm_path = "/path/to/your/internlm2-chat-7b" -local_llm_path = "internlm/internlm2_5-7b-chat" +local_llm_path = "internlm/internlm2-chat-7b" local_llm_max_text_length = 3000 # llm server listen port local_llm_bind_port = 8888 @@ -209,4 +209,4 @@ introduction = "github https://github.com/InternLM/HuixiangDou 用户体验群" # github.com/tencent/ncnn contributors [frontend.wechat_wkteam.18356748488] name = "卷卷群" -introduction = "ncnn contributors group" \ No newline at end of file +introduction = "ncnn contributors group"