AgentFinish llm token streaming #17842
Replies: 2 comments
-
I agree. Documentation states AgentFinish is not available as part of the streaming method. If this is something you’d like to be added, please start a discussion on github and explain why its needed.. I would argue that The With the current inference speed by GPT-4 Turbo (which anecdotally has become much slower the last month), the lack of streaming of |
Beta Was this translation helpful? Give feedback.
-
After some serious digging I found that stream events for the newest agent essentially solves this. |
Beta Was this translation helpful? Give feedback.
-
Checked
Feature request
When i create a stream endpoint through function agent and langserve, only
Agent Action
,Observation
, andFinal result
are streamed as shown on https://python.langchain.com/docs/modules/agents/how_to/streaming#using-agentactionobservation . However, for Final result, I would like to stream in units of tokens that llm gives as a response.Motivation
I thought it would be more user-friendly to stream
Final result(AgentFinish)
in llm token units thanAgent Action
andObservation
for ordinary users who asked the question. Because ordinary users can recognize that receiving llm responses in token units is currently responding to my question and they do not need to know about the intermediate steps of the agent.Proposal (If applicable)
No response
Beta Was this translation helpful? Give feedback.
All reactions