Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

如何在实际应用中提升模型效率? #2252

Open
mzgcz opened this issue Nov 29, 2024 · 0 comments
Open

如何在实际应用中提升模型效率? #2252

mzgcz opened this issue Nov 29, 2024 · 0 comments
Labels
question Further information is requested

Comments

@mzgcz
Copy link

mzgcz commented Nov 29, 2024

Notice: In order to resolve issues more efficiently, please raise issue following the template.
(注意:为了更加高效率解决您遇到的问题,请按照模板提问,补充细节)

❓ Questions and Help

在实际应用中要怎样提升在线模型(Streaming)的效率呢?
语言模型可以通过batch size进行批量推理,来提升推理效率;可以使用多实例来应对推理请求并发的情况;可以使用TensorRT来优化推理速度。
请问对于FunASR在线模型,上面哪些措施是可行的,有没有更好的推荐?

Before asking:

  1. search the issues.
  2. search the docs.

What is your question?

Code

What have you tried?

What's your environment?

  • OS (e.g., Linux):
  • FunASR Version (e.g., 1.0.0):
  • ModelScope Version (e.g., 1.11.0):
  • PyTorch Version (e.g., 2.0.0):
  • How you installed funasr (pip, source):
  • Python version:
  • GPU (e.g., V100M32)
  • CUDA/cuDNN version (e.g., cuda11.7):
  • Docker version (e.g., funasr-runtime-sdk-cpu-0.4.1)
  • Any other relevant information:
@mzgcz mzgcz added the question Further information is requested label Nov 29, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
question Further information is requested
Projects
None yet
Development

No branches or pull requests

1 participant