Skip to content

Commit

Permalink
tests: update model
Browse files Browse the repository at this point in the history
Signed-off-by: Tomas Dvorak <[email protected]>
  • Loading branch information
Tomas2D committed Nov 30, 2024
1 parent 853905f commit 561301f
Show file tree
Hide file tree
Showing 18 changed files with 42 additions and 38 deletions.
2 changes: 1 addition & 1 deletion examples/extensions/huggingface/huggingface_agent.py
Original file line number Diff line number Diff line change
Expand Up @@ -49,7 +49,7 @@ def __call__(self):

agent = IBMGenAIAgent(
client=client,
model="meta-llama/llama-3-70b-instruct",
model="meta-llama/llama-3-1-70b-instruct",
parameters=TextGenerationParameters(min_new_tokens=10, max_new_tokens=200, random_seed=777, temperature=0),
additional_tools=[BitcoinPriceFetcher()],
)
Expand Down
2 changes: 1 addition & 1 deletion examples/extensions/langchain/langchain_agent.py
Original file line number Diff line number Diff line change
Expand Up @@ -78,7 +78,7 @@ def _run(self, word: str, run_manager: Optional[CallbackManagerForToolRun] = Non
client = Client(credentials=Credentials.from_env())
llm = LangChainChatInterface(
client=client,
model_id="meta-llama/llama-3-70b-instruct",
model_id="meta-llama/llama-3-1-70b-instruct",
parameters=TextGenerationParameters(
max_new_tokens=250, min_new_tokens=20, temperature=0, stop_sequences=["\nObservation"]
),
Expand Down
2 changes: 1 addition & 1 deletion examples/extensions/langchain/langchain_chat_generate.py
Original file line number Diff line number Diff line change
Expand Up @@ -27,7 +27,7 @@ def heading(text: str) -> str:

llm = LangChainChatInterface(
client=Client(credentials=Credentials.from_env()),
model_id="meta-llama/llama-3-70b-instruct",
model_id="meta-llama/llama-3-1-70b-instruct",
parameters=TextGenerationParameters(
decoding_method=DecodingMethod.SAMPLE,
max_new_tokens=100,
Expand Down
2 changes: 1 addition & 1 deletion examples/extensions/langchain/langchain_chat_stream.py
Original file line number Diff line number Diff line change
Expand Up @@ -21,7 +21,7 @@ def heading(text: str) -> str:
print(heading("Stream chat with langchain"))

llm = LangChainChatInterface(
model_id="meta-llama/llama-3-70b-instruct",
model_id="meta-llama/llama-3-1-70b-instruct",
client=Client(credentials=Credentials.from_env()),
parameters=TextGenerationParameters(
decoding_method=DecodingMethod.SAMPLE,
Expand Down
2 changes: 1 addition & 1 deletion examples/extensions/langchain/langchain_sql_agent.py
Original file line number Diff line number Diff line change
Expand Up @@ -75,7 +75,7 @@ def create_llm():
client = Client(credentials=Credentials.from_env())
return LangChainChatInterface(
client=client,
model_id="meta-llama/llama-3-70b-instruct",
model_id="meta-llama/llama-3-1-70b-instruct",
parameters=TextGenerationParameters(
max_new_tokens=250, min_new_tokens=20, temperature=0, stop_sequences=["\nObservation"]
),
Expand Down
2 changes: 1 addition & 1 deletion examples/extensions/llama_index/llama_index_llm.py
Original file line number Diff line number Diff line change
Expand Up @@ -22,7 +22,7 @@ def heading(text: str) -> str:

llm = IBMGenAILlamaIndex(
client=client,
model_id="meta-llama/llama-3-70b-instruct",
model_id="meta-llama/llama-3-1-70b-instruct",
parameters=TextGenerationParameters(
decoding_method=DecodingMethod.SAMPLE,
max_new_tokens=100,
Expand Down
2 changes: 1 addition & 1 deletion examples/text/chat.py
Original file line number Diff line number Diff line change
Expand Up @@ -31,7 +31,7 @@ def heading(text: str) -> str:
)

client = Client(credentials=Credentials.from_env())
model_id = "meta-llama/llama-3-70b-instruct"
model_id = "meta-llama/llama-3-1-70b-instruct"

prompt = "What is NLP and how it has evolved over the years?"
print(heading("Generating a chat response"))
Expand Down
2 changes: 1 addition & 1 deletion src/genai/extensions/langchain/chat_llm.py
Original file line number Diff line number Diff line change
Expand Up @@ -93,7 +93,7 @@ class LangChainChatInterface(BaseChatModel):
client = Client(credentials=Credentials.from_env())
llm = LangChainChatInterface(
client=client,
model_id="meta-llama/llama-3-70b-instruct",
model_id="meta-llama/llama-3-1-70b-instruct",
parameters=TextGenerationParameters(
max_new_tokens=250,
)
Expand Down
2 changes: 1 addition & 1 deletion src/genai/extensions/langchain/llm.py
Original file line number Diff line number Diff line change
Expand Up @@ -62,7 +62,7 @@ class LangChainInterface(LLM):
client = Client(credentials=Credentials.from_env())
llm = LangChainInterface(
client=client,
model_id="meta-llama/llama-3-70b-instruct",
model_id="meta-llama/llama-3-1-70b-instruct",
parameters=TextGenerationParameters(max_new_tokens=50)
)
Expand Down
4 changes: 2 additions & 2 deletions src/genai/text/chat/chat_generation_service.py
Original file line number Diff line number Diff line change
Expand Up @@ -79,7 +79,7 @@ def create(
# Create a new conversation
response = client.text.chat.create(
model_id="meta-llama/llama-3-70b-instruct",
model_id="meta-llama/llama-3-1-70b-instruct",
messages=[HumanMessage(content="Describe the game Chess?")],
parameters=TextGenerationParameters(max_token_limit=100)
)
Expand Down Expand Up @@ -150,7 +150,7 @@ def create_stream(
# Create a new conversation
for response in client.text.chat.create_stream(
model_id="meta-llama/llama-3-70b-instruct",
model_id="meta-llama/llama-3-1-70b-instruct",
messages=[HumanMessage(content="Describe the game Chess?")],
parameters=TextGenerationParameters(max_token_limit=100)
):
Expand Down
3 changes: 3 additions & 0 deletions tests/e2e/test_examples.py
Original file line number Diff line number Diff line change
Expand Up @@ -16,6 +16,9 @@
"parallel_processing.py",
"chroma_db_embedding.py",
"deployment.py",
# no available model
"lm_eval_cli.py",
"lm_eval_model.py",
}
skip_for_python_3_12 = {
# These files are skipped for python >= 3.12 because transformers library cannot be installed
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,7 @@ interactions:
explain why instead of answering something incorrectly.\n If you don''t know
the answer to a question, please don''t share false information.\n ", "role":
"system"}, {"content": "What is NLP and how it has evolved over the years?",
"role": "user"}], "model_id": "meta-llama/llama-3-70b-instruct", "parameters":
"role": "user"}], "model_id": "meta-llama/llama-3-1-70b-instruct", "parameters":
{}}'
headers:
accept:
Expand All @@ -25,7 +25,7 @@ interactions:
uri: https://api.com/v2/text/chat?version=2024-03-19
response:
body:
string: '{"id":"01f28b95-5ecd-4a01-9a2b-7803c6db824f","model_id":"meta-llama/llama-3-70b-instruct","created_at":"2024-05-20T14:12:28.750Z","results":[{"generated_text":"Natural
string: '{"id":"01f28b95-5ecd-4a01-9a2b-7803c6db824f","model_id":"meta-llama/llama-3-1-70b-instruct","created_at":"2024-05-20T14:12:28.750Z","results":[{"generated_text":"Natural
Language Processing (NLP) is a subfield of artificial intelligence (AI) that
deals with","generated_token_count":20,"input_token_count":134,"stop_reason":"max_tokens","seed":1025128500}],"conversation_id":"90edb70b-c4e7-45f5-81c3-fab231227b7a"}'
headers:
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,7 @@ interactions:
explain why instead of answering something incorrectly.\n If you don''t know
the answer to a question, please don''t share false information.\n ", "role":
"system"}, {"content": "What is NLP and how it has evolved over the years?",
"role": "user"}], "model_id": "meta-llama/llama-3-70b-instruct", "parameters":
"role": "user"}], "model_id": "meta-llama/llama-3-1-70b-instruct", "parameters":
{}}'
headers:
accept:
Expand All @@ -25,7 +25,7 @@ interactions:
uri: https://api.com/v2/text/chat?version=2024-03-19
response:
body:
string: '{"id":"d8160c0f-36a3-49a0-b001-a3f9cbdb9643","model_id":"meta-llama/llama-3-70b-instruct","created_at":"2024-05-20T14:12:27.354Z","results":[{"generated_text":"NLP,
string: '{"id":"d8160c0f-36a3-49a0-b001-a3f9cbdb9643","model_id":"meta-llama/llama-3-1-70b-instruct","created_at":"2024-05-20T14:12:27.354Z","results":[{"generated_text":"NLP,
or Natural Language Processing, is a subfield of artificial intelligence (AI)
that deals","generated_token_count":20,"input_token_count":134,"stop_reason":"max_tokens","seed":2032949123}],"conversation_id":"bcc5d917-5e64-48e0-9047-ac5e032e8f21"}'
headers:
Expand Down
2 changes: 1 addition & 1 deletion tests/integration/extensions/test_langchain_chat.py
Original file line number Diff line number Diff line change
Expand Up @@ -18,7 +18,7 @@
@pytest.mark.integration
class TestLangChainChat:
def setup_method(self):
self.model_id = "meta-llama/llama-3-70b-instruct"
self.model_id = "meta-llama/llama-3-1-70b-instruct"

@pytest.fixture
def parameters(self):
Expand Down
1 change: 1 addition & 0 deletions tests/integration/extensions/test_lm_eval.py
Original file line number Diff line number Diff line change
Expand Up @@ -7,6 +7,7 @@


@pytest.mark.integration
@pytest.mark.skip
class TestLMEval:
@pytest.fixture(autouse=True)
def load_credentials(self):
Expand Down
Original file line number Diff line number Diff line change
@@ -1,7 +1,7 @@
interactions:
- request:
body: '{"messages": [{"content": "Do you want to destroy the world?", "role":
"user"}], "model_id": "meta-llama/llama-3-70b-instruct"}'
"user"}], "model_id": "meta-llama/llama-3-1-70b-instruct"}'
headers:
accept:
- '*/*'
Expand All @@ -17,7 +17,7 @@ interactions:
uri: https://api.com/v2/text/chat?version=2024-03-19
response:
body:
string: '{"id":"191d90ab-e7e0-4821-a66a-64dfdc0415b0","model_id":"meta-llama/llama-3-70b-instruct","created_at":"2024-05-20T14:12:31.312Z","results":[{"generated_text":"No,
string: '{"id":"191d90ab-e7e0-4821-a66a-64dfdc0415b0","model_id":"meta-llama/llama-3-1-70b-instruct","created_at":"2024-05-20T14:12:31.312Z","results":[{"generated_text":"No,
I do not want to destroy the world. I am designed to assist and provide helpful
information","generated_token_count":20,"input_token_count":18,"stop_reason":"max_tokens","seed":967002347}],"conversation_id":"02593997-bb70-47dd-97d4-209aa820aa98"}'
headers:
Expand Down Expand Up @@ -53,9 +53,9 @@ interactions:
response:
body:
string: '{"results":[{"id":"191d90ab-e7e0-4821-a66a-64dfdc0415b0","duration":759,"request":{"messages":[{"role":"user","content":"Do
you want to destroy the world?"}],"model_id":"meta-llama/llama-3-70b-instruct","conversation_id":"02593997-bb70-47dd-97d4-209aa820aa98"},"status":"success","created_at":"2024-05-20T14:12:31.000Z","response":{"id":"191d90ab-e7e0-4821-a66a-64dfdc0415b0","results":[{"seed":967002347,"stop_reason":"max_tokens","generated_text":"No,
you want to destroy the world?"}],"model_id":"meta-llama/llama-3-1-70b-instruct","conversation_id":"02593997-bb70-47dd-97d4-209aa820aa98"},"status":"success","created_at":"2024-05-20T14:12:31.000Z","response":{"id":"191d90ab-e7e0-4821-a66a-64dfdc0415b0","results":[{"seed":967002347,"stop_reason":"max_tokens","generated_text":"No,
I do not want to destroy the world. I am designed to assist and provide helpful
information","input_token_count":18,"generated_token_count":20}],"model_id":"meta-llama/llama-3-70b-instruct","created_at":"2024-05-20T14:12:31.312Z","conversation_id":"02593997-bb70-47dd-97d4-209aa820aa98"},"version":{"api":"v2","date":"2024-03-19"}}]}'
information","input_token_count":18,"generated_token_count":20}],"model_id":"meta-llama/llama-3-1-70b-instruct","created_at":"2024-05-20T14:12:31.312Z","conversation_id":"02593997-bb70-47dd-97d4-209aa820aa98"},"version":{"api":"v2","date":"2024-03-19"}}]}'
headers:
cache-control:
- private
Expand Down Expand Up @@ -86,7 +86,7 @@ interactions:
- request:
body: '{"conversation_id": "02593997-bb70-47dd-97d4-209aa820aa98", "messages":
[{"content": "What was my previous question?", "role": "user"}], "model_id":
"meta-llama/llama-3-70b-instruct"}'
"meta-llama/llama-3-1-70b-instruct"}'
headers:
accept:
- '*/*'
Expand All @@ -102,7 +102,7 @@ interactions:
uri: https://api.com/v2/text/chat?version=2024-03-19
response:
body:
string: '{"id":"098d2940-3e40-4b67-8239-b63e16ae53f2","model_id":"meta-llama/llama-3-70b-instruct","created_at":"2024-05-20T14:12:32.733Z","results":[{"generated_text":"Your
string: '{"id":"098d2940-3e40-4b67-8239-b63e16ae53f2","model_id":"meta-llama/llama-3-1-70b-instruct","created_at":"2024-05-20T14:12:32.733Z","results":[{"generated_text":"Your
previous question was \"Do you want to destroy the world?\"","generated_token_count":14,"input_token_count":54,"stop_reason":"eos_token","seed":934152098}],"conversation_id":"02593997-bb70-47dd-97d4-209aa820aa98"}'
headers:
content-length:
Expand Down Expand Up @@ -137,11 +137,11 @@ interactions:
response:
body:
string: '{"results":[{"id":"191d90ab-e7e0-4821-a66a-64dfdc0415b0","duration":759,"request":{"messages":[{"role":"user","content":"Do
you want to destroy the world?"}],"model_id":"meta-llama/llama-3-70b-instruct","conversation_id":"02593997-bb70-47dd-97d4-209aa820aa98"},"status":"success","created_at":"2024-05-20T14:12:31.000Z","response":{"id":"191d90ab-e7e0-4821-a66a-64dfdc0415b0","results":[{"seed":967002347,"stop_reason":"max_tokens","generated_text":"No,
you want to destroy the world?"}],"model_id":"meta-llama/llama-3-1-70b-instruct","conversation_id":"02593997-bb70-47dd-97d4-209aa820aa98"},"status":"success","created_at":"2024-05-20T14:12:31.000Z","response":{"id":"191d90ab-e7e0-4821-a66a-64dfdc0415b0","results":[{"seed":967002347,"stop_reason":"max_tokens","generated_text":"No,
I do not want to destroy the world. I am designed to assist and provide helpful
information","input_token_count":18,"generated_token_count":20}],"model_id":"meta-llama/llama-3-70b-instruct","created_at":"2024-05-20T14:12:31.312Z","conversation_id":"02593997-bb70-47dd-97d4-209aa820aa98"},"version":{"api":"v2","date":"2024-03-19"}},{"id":"098d2940-3e40-4b67-8239-b63e16ae53f2","duration":582,"request":{"messages":[{"role":"user","content":"What
was my previous question?"}],"model_id":"meta-llama/llama-3-70b-instruct","parent_id":"191d90ab-e7e0-4821-a66a-64dfdc0415b0","conversation_id":"02593997-bb70-47dd-97d4-209aa820aa98"},"status":"success","created_at":"2024-05-20T14:12:33.000Z","response":{"id":"098d2940-3e40-4b67-8239-b63e16ae53f2","results":[{"seed":934152098,"stop_reason":"eos_token","generated_text":"Your
previous question was \"Do you want to destroy the world?\"","input_token_count":54,"generated_token_count":14}],"model_id":"meta-llama/llama-3-70b-instruct","created_at":"2024-05-20T14:12:32.733Z","conversation_id":"02593997-bb70-47dd-97d4-209aa820aa98"},"parent_id":"191d90ab-e7e0-4821-a66a-64dfdc0415b0","version":{"api":"v2","date":"2024-03-19"}}]}'
information","input_token_count":18,"generated_token_count":20}],"model_id":"meta-llama/llama-3-1-70b-instruct","created_at":"2024-05-20T14:12:31.312Z","conversation_id":"02593997-bb70-47dd-97d4-209aa820aa98"},"version":{"api":"v2","date":"2024-03-19"}},{"id":"098d2940-3e40-4b67-8239-b63e16ae53f2","duration":582,"request":{"messages":[{"role":"user","content":"What
was my previous question?"}],"model_id":"meta-llama/llama-3-1-70b-instruct","parent_id":"191d90ab-e7e0-4821-a66a-64dfdc0415b0","conversation_id":"02593997-bb70-47dd-97d4-209aa820aa98"},"status":"success","created_at":"2024-05-20T14:12:33.000Z","response":{"id":"098d2940-3e40-4b67-8239-b63e16ae53f2","results":[{"seed":934152098,"stop_reason":"eos_token","generated_text":"Your
previous question was \"Do you want to destroy the world?\"","input_token_count":54,"generated_token_count":14}],"model_id":"meta-llama/llama-3-1-70b-instruct","created_at":"2024-05-20T14:12:32.733Z","conversation_id":"02593997-bb70-47dd-97d4-209aa820aa98"},"parent_id":"191d90ab-e7e0-4821-a66a-64dfdc0415b0","version":{"api":"v2","date":"2024-03-19"}}]}'
headers:
cache-control:
- private
Expand Down
Original file line number Diff line number Diff line change
@@ -1,7 +1,7 @@
interactions:
- request:
body: '{"messages": [{"content": "I want to kill them! There are my enemies.",
"role": "user"}], "model_id": "meta-llama/llama-3-70b-instruct", "moderations":
"role": "user"}], "model_id": "meta-llama/llama-3-1-70b-instruct", "moderations":
{"hap": {"input": {"enabled": true, "send_tokens": false, "threshold": 0.7},
"output": {"enabled": true, "send_tokens": true, "threshold": 0.7}}}, "parameters":
{"max_new_tokens": 10, "min_new_tokens": 3}}'
Expand All @@ -25,41 +25,41 @@ interactions:
string: 'retry: 3000
data: {"id":"3adaa5b8-38d7-41c8-bd26-2285a5ae8c9f","model_id":"meta-llama/llama-3-70b-instruct","created_at":"2024-05-20T14:12:33.754Z","conversation_id":"6a892fa0-6a3a-46e2-9a8d-c487fa172c52","moderations":{"hap":[{"score":0.8397554755210876,"flagged":true,"success":true,"position":{"start":0,"end":20}}]}}
data: {"id":"3adaa5b8-38d7-41c8-bd26-2285a5ae8c9f","model_id":"meta-llama/llama-3-1-70b-instruct","created_at":"2024-05-20T14:12:33.754Z","conversation_id":"6a892fa0-6a3a-46e2-9a8d-c487fa172c52","moderations":{"hap":[{"score":0.8397554755210876,"flagged":true,"success":true,"position":{"start":0,"end":20}}]}}
data: {"id":"3adaa5b8-38d7-41c8-bd26-2285a5ae8c9f","model_id":"meta-llama/llama-3-70b-instruct","created_at":"2024-05-20T14:12:33.826Z","results":[{"generated_text":"I
data: {"id":"3adaa5b8-38d7-41c8-bd26-2285a5ae8c9f","model_id":"meta-llama/llama-3-1-70b-instruct","created_at":"2024-05-20T14:12:33.826Z","results":[{"generated_text":"I
understan","generated_token_count":2,"input_token_count":0,"stop_reason":"not_finished"}],"conversation_id":"6a892fa0-6a3a-46e2-9a8d-c487fa172c52"}
data: {"id":"3adaa5b8-38d7-41c8-bd26-2285a5ae8c9f","model_id":"meta-llama/llama-3-70b-instruct","created_at":"2024-05-20T14:12:33.879Z","results":[{"generated_text":"d
data: {"id":"3adaa5b8-38d7-41c8-bd26-2285a5ae8c9f","model_id":"meta-llama/llama-3-1-70b-instruct","created_at":"2024-05-20T14:12:33.879Z","results":[{"generated_text":"d
tha","generated_token_count":3,"input_token_count":0,"stop_reason":"not_finished"}],"conversation_id":"6a892fa0-6a3a-46e2-9a8d-c487fa172c52"}
data: {"id":"3adaa5b8-38d7-41c8-bd26-2285a5ae8c9f","model_id":"meta-llama/llama-3-70b-instruct","created_at":"2024-05-20T14:12:33.917Z","results":[{"generated_text":"t
data: {"id":"3adaa5b8-38d7-41c8-bd26-2285a5ae8c9f","model_id":"meta-llama/llama-3-1-70b-instruct","created_at":"2024-05-20T14:12:33.917Z","results":[{"generated_text":"t
yo","generated_token_count":4,"input_token_count":0,"stop_reason":"not_finished"}],"conversation_id":"6a892fa0-6a3a-46e2-9a8d-c487fa172c52"}
data: {"id":"3adaa5b8-38d7-41c8-bd26-2285a5ae8c9f","model_id":"meta-llama/llama-3-70b-instruct","created_at":"2024-05-20T14:12:33.956Z","results":[{"generated_text":"u''r","generated_token_count":5,"input_token_count":0,"stop_reason":"not_finished"}],"conversation_id":"6a892fa0-6a3a-46e2-9a8d-c487fa172c52"}
data: {"id":"3adaa5b8-38d7-41c8-bd26-2285a5ae8c9f","model_id":"meta-llama/llama-3-1-70b-instruct","created_at":"2024-05-20T14:12:33.956Z","results":[{"generated_text":"u''r","generated_token_count":5,"input_token_count":0,"stop_reason":"not_finished"}],"conversation_id":"6a892fa0-6a3a-46e2-9a8d-c487fa172c52"}
data: {"id":"3adaa5b8-38d7-41c8-bd26-2285a5ae8c9f","model_id":"meta-llama/llama-3-70b-instruct","created_at":"2024-05-20T14:12:33.995Z","results":[{"generated_text":"e
data: {"id":"3adaa5b8-38d7-41c8-bd26-2285a5ae8c9f","model_id":"meta-llama/llama-3-1-70b-instruct","created_at":"2024-05-20T14:12:33.995Z","results":[{"generated_text":"e
feelin","generated_token_count":6,"input_token_count":0,"stop_reason":"not_finished"}],"conversation_id":"6a892fa0-6a3a-46e2-9a8d-c487fa172c52"}
data: {"id":"3adaa5b8-38d7-41c8-bd26-2285a5ae8c9f","model_id":"meta-llama/llama-3-70b-instruct","created_at":"2024-05-20T14:12:34.035Z","results":[{"generated_text":"g
data: {"id":"3adaa5b8-38d7-41c8-bd26-2285a5ae8c9f","model_id":"meta-llama/llama-3-1-70b-instruct","created_at":"2024-05-20T14:12:34.035Z","results":[{"generated_text":"g
","generated_token_count":7,"input_token_count":0,"stop_reason":"not_finished"}],"conversation_id":"6a892fa0-6a3a-46e2-9a8d-c487fa172c52"}
data: {"id":"3adaa5b8-38d7-41c8-bd26-2285a5ae8c9f","model_id":"meta-llama/llama-3-70b-instruct","created_at":"2024-05-20T14:12:34.075Z","results":[{"generated_text":"a
data: {"id":"3adaa5b8-38d7-41c8-bd26-2285a5ae8c9f","model_id":"meta-llama/llama-3-1-70b-instruct","created_at":"2024-05-20T14:12:34.075Z","results":[{"generated_text":"a
stron","generated_token_count":8,"input_token_count":0,"stop_reason":"not_finished"}],"conversation_id":"6a892fa0-6a3a-46e2-9a8d-c487fa172c52"}
data: {"id":"3adaa5b8-38d7-41c8-bd26-2285a5ae8c9f","model_id":"meta-llama/llama-3-70b-instruct","created_at":"2024-05-20T14:12:34.129Z","results":[{"generated_text":"g
data: {"id":"3adaa5b8-38d7-41c8-bd26-2285a5ae8c9f","model_id":"meta-llama/llama-3-1-70b-instruct","created_at":"2024-05-20T14:12:34.129Z","results":[{"generated_text":"g
sens","generated_token_count":9,"input_token_count":0,"stop_reason":"not_finished"}],"conversation_id":"6a892fa0-6a3a-46e2-9a8d-c487fa172c52"}
data: {"id":"3adaa5b8-38d7-41c8-bd26-2285a5ae8c9f","model_id":"meta-llama/llama-3-70b-instruct","created_at":"2024-05-20T14:12:34.171Z","results":[{"generated_text":"e
data: {"id":"3adaa5b8-38d7-41c8-bd26-2285a5ae8c9f","model_id":"meta-llama/llama-3-1-70b-instruct","created_at":"2024-05-20T14:12:34.171Z","results":[{"generated_text":"e
of","generated_token_count":10,"input_token_count":0,"stop_reason":"max_tokens","seed":401923414}],"conversation_id":"6a892fa0-6a3a-46e2-9a8d-c487fa172c52"}
Expand Down
Loading

0 comments on commit 561301f

Please sign in to comment.