We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
提供一个显卡部署多个模型的功能或最佳实践文档
能够支持众多类型的模型轻松完成部署让xinference在一众类似项目中显得非常突出,极大的方便了有一定算力基础的中小公司部署自主可控的推理服务。但像TTS或ASR类的服务显存需求往往不大,却只能单独占用一个显卡,不能充分利用显卡性能。希望官方能提供一个显卡部署多个模型的功能或最佳实践的范例文档。
目前了解到的情况:
The text was updated successfully, but these errors were encountered:
This issue is stale because it has been open for 7 days with no activity.
Sorry, something went wrong.
多运行几个 xinference 实例
No branches or pull requests
Feature request / 功能建议
提供一个显卡部署多个模型的功能或最佳实践文档
Motivation / 动机
能够支持众多类型的模型轻松完成部署让xinference在一众类似项目中显得非常突出,极大的方便了有一定算力基础的中小公司部署自主可控的推理服务。但像TTS或ASR类的服务显存需求往往不大,却只能单独占用一个显卡,不能充分利用显卡性能。希望官方能提供一个显卡部署多个模型的功能或最佳实践的范例文档。
Your contribution / 您的贡献
目前了解到的情况:
The text was updated successfully, but these errors were encountered: