Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

refine streaming experience #15

Closed
simonpcouch opened this issue Oct 7, 2024 · 2 comments
Closed

refine streaming experience #15

simonpcouch opened this issue Oct 7, 2024 · 2 comments

Comments

@simonpcouch
Copy link
Owner

In the web interfaces for ChatGPT and (especially) Claude chats, the streaming feels super smooth, i.e. responses are pieced together in components even finer than tokens. It feels more "chunky" in pal. This may just be a function of responses from pal tending to be shorter than the few paragraphs typical of an unprompted ChatGPT or Claude, and thus being comprised of fewer total tokens, but I do think a really satisfying interface would make it feel like you're actually watching the model "type."

@jcheng5
Copy link

jcheng5 commented Oct 21, 2024

This might be because right now, stream_async is implemented using a polling mechanism under the hood. We're actively working on changing this to a more efficient, less chunky mechanism based on select or similar.

@simonpcouch
Copy link
Owner Author

Thanks for the context! Will leave this to elmer, then. :)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants