Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Can this framework used in a production scenario with multiple users? #295

Open
MohammedShokr opened this issue Nov 13, 2024 · 3 comments
Open

Comments

@MohammedShokr
Copy link

I am asking about the scalability of this framework can it be used in production with too many users sending streams and the same time?

@makaveli10
Copy link
Collaborator

yes but with batching support which is not upstream as of now.

@yatoooon
Copy link

yatoooon commented Dec 5, 2024

Are there any standards for servers such as what gpu, cpu, and memory can accommodate the number of clients?

@makaveli10
Copy link
Collaborator

makaveli10 commented Jan 9, 2025

Not yet but whisper at inference doesnt take up a lot of memory so, with batching supported a single 4090 for whisper-small should support around 25-30 clients.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Development

No branches or pull requests

3 participants