From f2604c9b12e0ac3264c4c6276ce9166ad5cf85ad Mon Sep 17 00:00:00 2001 From: leon Date: Tue, 3 Dec 2024 17:14:36 +0100 Subject: [PATCH] Increase default max_tokens to 4000 in OpenAI model configuration --- llm_core/llm_core/models/openai.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/llm_core/llm_core/models/openai.py b/llm_core/llm_core/models/openai.py index d12df22d..7bcc0f11 100644 --- a/llm_core/llm_core/models/openai.py +++ b/llm_core/llm_core/models/openai.py @@ -67,7 +67,7 @@ class OpenAIModelConfig(ModelConfig): model_name: OpenAIModel = Field(default=default_openai_model, # type: ignore description="The name of the model to use.") - max_tokens: PositiveInt = Field(1000, description="""\ + max_tokens: PositiveInt = Field(4000, description="""\ The maximum number of [tokens](https://platform.openai.com/tokenizer) to generate in the chat completion. The total length of input tokens and generated tokens is limited by the model's context length. \