-
Notifications
You must be signed in to change notification settings - Fork 1
/
Copy pathhuggingface.preset
54 lines (51 loc) · 1.93 KB
/
huggingface.preset
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
--https://huggingface.co/playground
--https://huggingface.co/docs/api-inference/index
apibase="https://api-inference.huggingface.co/models/{model}/v1"
apikey=win.GetEnv"HF_API_TOKEN" or "set HF_API_TOKEN env variable!"
-- https://huggingface.co/models?other=text-generation-inference&inference=warm
-- https://huggingface.co/models?other=text-generation-inference&inference=cold
model={
"01-ai/Yi-1.5-34B-Chat",
"codellama/CodeLlama-34b-Instruct-hf",
"deepseek-ai/DeepSeek-R1-Distill-Qwen-32B",
"google/gemma-1.1-7b-it",
"google/gemma-2-27b-it",
"google/gemma-2-2b-it",
"google/gemma-2-9b-it",
"HuggingFaceH4/starchat2-15b-v0.1",
"HuggingFaceH4/zephyr-7b-alpha",
"HuggingFaceH4/zephyr-7b-beta",
--"meta-llama/Llama-2-7b-chat-hf",
--"meta-llama/Llama-3.1-70B-Instruct",
--"meta-llama/Llama-3.1-8B-Instruct",
"meta-llama/Llama-3.2-11B-Vision-Instruct",
"meta-llama/Llama-3.2-1B-Instruct",
"meta-llama/Llama-3.2-3B-Instruct",
--"meta-llama/Llama-3.3-70B-Instruct",
--"meta-llama/Meta-Llama-3-70B-Instruct",
"meta-llama/Meta-Llama-3-8B-Instruct",
"microsoft/DialoGPT-large",
"microsoft/Phi-3-mini-4k-instruct",
"microsoft/Phi-3.5-mini-instruct",
"mistralai/Mistral-7B-Instruct-v0.2",
"mistralai/Mistral-7B-Instruct-v0.3",
"mistralai/Mistral-Nemo-Instruct-2407",
"mistralai/Mixtral-8x7B-Instruct-v0.1",
"NousResearch/Hermes-3-Llama-3.1-8B",
"NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO",
"Qwen/Qwen2-VL-7B-Instruct",
"Qwen/Qwen2.5-0.5B-Instruct",
"Qwen/Qwen2.5-1.5B-Instruct",
"Qwen/Qwen2.5-72B-Instruct",
"Qwen/Qwen2.5-Coder-32B-Instruct",
"Qwen/Qwen2.5-Coder-3B-Instruct",
"Qwen/QwQ-32B-Preview",
"tiiuae/falcon-7b-instruct",
}
modelsMeta="none"
local useCache = '"x-use-cache":"false"'
headers = function()
local ans = far.Message("Wait for model?","HuggingFace preset","No;Yes")
local waitModel = ans==2 and '"x-wait-for-model":"true"' or nil
return table.concat({useCache,waitModel},",")
end