Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add Support for Embeddings to Reduce Token Usage And reduce the risk of abnormal stop. #246

Open
Rain-Of-Stars opened this issue Dec 30, 2024 · 1 comment
Labels
bug Something isn't working

Comments

@Rain-Of-Stars
Copy link

Which API Provider are you using?

OpenAI Compatible

Which Model are you using?

deepseek

What happened?

I propose adding a feature using an embedding system to reduce tokens in requests. This could help manage and optimize the daily token. Leveraging embeddings may maintain or enhance functionality while reducing token usage and costs.

Steps to reproduce

Now the context consumption is too large, and multiple rounds of dialogue will overflow tokens.

Relevant API REQUEST output

No response

Additional context

No response

@Rain-Of-Stars Rain-Of-Stars added the bug Something isn't working label Dec 30, 2024
@Rain-Of-Stars Rain-Of-Stars changed the title Add Support for Embeddings to Reduce Token Usage Add Support for Embeddings to Reduce Token Usage And reduce the risk of abnormal stop. Dec 30, 2024
@mrubens
Copy link
Collaborator

mrubens commented Jan 9, 2025

@Rain-Of-Stars sounds great - any thoughts on the best way to do it?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

2 participants