Skip to content

Commit

Permalink
minor adjustment to llama.cpp server call
Browse files Browse the repository at this point in the history
  • Loading branch information
Fabian Wolf committed Nov 11, 2024
1 parent 8fc7c88 commit 12ae811
Show file tree
Hide file tree
Showing 2 changed files with 4 additions and 1 deletion.
1 change: 1 addition & 0 deletions .gitignore
Original file line number Diff line number Diff line change
Expand Up @@ -5,4 +5,5 @@ logs*
__pycache__

test_*.py
test_*
output.pdf
4 changes: 3 additions & 1 deletion webapp/llm_processing/routes.py
Original file line number Diff line number Diff line change
Expand Up @@ -227,7 +227,9 @@ def extract_from_report(
"-b",
"2048",
"-ub",
"2048",
"512",
"-t",
"16",
] + (["--verbose"] if verbose_llama else []) + (["--mlock"] if mlock else []) +
(["-ctk", kv_cache_type, "-ctv", kv_cache_type] if kv_cache_type != "" else []) +
(["-sm", "none", "-mg", str(gpu)] if gpu not in ["all", "ALL", "mps", ""] else [])+
Expand Down

0 comments on commit 12ae811

Please sign in to comment.