Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Catch token count issue while streaming with customized models #3241

Merged
merged 6 commits into from
Sep 25, 2024

Commits on Jul 28, 2024

  1. Catch token count issue while streaming with customized models

    If llama, llava, phi, or some other models are used for streaming (with stream=True), the current design would crash after fetching the response.
    
    A warning is enough in this case, just like the non-streaming use cases.
    BeibinLi authored Jul 28, 2024
    Configuration menu
    Copy the full SHA
    a40aa56 View commit details
    Browse the repository at this point in the history

Commits on Jul 29, 2024

  1. Configuration menu
    Copy the full SHA
    c27f0a9 View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    beb105a View commit details
    Browse the repository at this point in the history

Commits on Aug 6, 2024

  1. Configuration menu
    Copy the full SHA
    286d647 View commit details
    Browse the repository at this point in the history

Commits on Sep 25, 2024

  1. Configuration menu
    Copy the full SHA
    7d1a110 View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    86b9089 View commit details
    Browse the repository at this point in the history