From 384201d4c8e2ef9a3219a70d65b9161258490df5 Mon Sep 17 00:00:00 2001 From: davidmezzetti <561939+davidmezzetti@users.noreply.github.com> Date: Mon, 25 Nov 2024 16:18:33 -0500 Subject: [PATCH] Update documentation --- docs/pipeline/text/llm.md | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/docs/pipeline/text/llm.md b/docs/pipeline/text/llm.md index 2a5fbf17a..10990be56 100644 --- a/docs/pipeline/text/llm.md +++ b/docs/pipeline/text/llm.md @@ -41,7 +41,7 @@ llm( """ ) -# Chat messages are also supported +# Chat messages automatically handle templating llm([ {"role": "system", "content": "You are a friendly assistant."}, {"role": "user", "content": "Answer the following question..."} @@ -61,8 +61,8 @@ llm = LLM("meta-llama/Meta-Llama-3.1-8B-Instruct") llm = LLM("meta-llama/Meta-Llama-3.1-8B-Instruct", method="transformers") # llama.cpp -llm = LLM("microsoft/Phi-3-mini-4k-instruct-gguf/Phi-3-mini-4k-instruct-gguf") -llm = LLM("microsoft/Phi-3-mini-4k-instruct-gguf/Phi-3-mini-4k-instruct-gguf", +llm = LLM("microsoft/Phi-3-mini-4k-instruct-gguf/Phi-3-mini-4k-instruct-q4.gguf") +llm = LLM("microsoft/Phi-3-mini-4k-instruct-gguf/Phi-3-mini-4k-instruct-q4.gguf", method="llama.cpp") # LiteLLM