From a65e20e085e787013da2ba2285ab352fcb065457 Mon Sep 17 00:00:00 2001 From: berkecanrizai <63911408+berkecanrizai@users.noreply.github.com> Date: Fri, 20 Oct 2023 11:19:02 +0300 Subject: [PATCH] fix: linter format (#4832) GitOrigin-RevId: c5195206f02b13286677421bbe8c27f8df6bebc9 --- llm_app/model_wrappers/huggingface_wrapper/pipelines.py | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/llm_app/model_wrappers/huggingface_wrapper/pipelines.py b/llm_app/model_wrappers/huggingface_wrapper/pipelines.py index 31faa18..4f482e0 100644 --- a/llm_app/model_wrappers/huggingface_wrapper/pipelines.py +++ b/llm_app/model_wrappers/huggingface_wrapper/pipelines.py @@ -118,7 +118,9 @@ def __call__(self, text, **kwargs): }, {"role": "user", "content": text}, ] - prompt = self.tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True) + prompt = self.tokenizer.apply_chat_template( + messages, tokenize=False, add_generation_prompt=True + ) output = self.pipeline(prompt, max_new_tokens=max_new_tokens, **kwargs) return output[0]["generated_text"]