Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Currently we use a single GPU for decoding. We have the plan to support Multi-GPU decoding and the script is on the way. #119

Open
Learneducn opened this issue Jul 20, 2024 · 5 comments

Comments

@Learneducn
Copy link

          Currently we use a single GPU for decoding. We have the plan to support Multi-GPU decoding and the script is on the way.

Originally posted by @ddlBoJack in #100 (comment)

@Learneducn
Copy link
Author

Hello, excuse me. When I run the inference and training scripts, I specified the CUDA ID, but it always defaulted to on CUDA=0. How to solve this?

@Learneducn
Copy link
Author

Hello, excuse me. When I ran the inference and training scripts, I specified the CUDA ID, but it always defaulted to cuda=0. How should I solve this? In short, an error was reported: torch.distributed.elastic .mutiprocessing.errors.ChildFailedError:torch.distributed.elastic.mutiprocessing.errors.ChildFailedError

@ddlBoJack
Copy link
Collaborator

Have you solved it? I think this may relate to a local problem with your GPU config.

@fclearner
Copy link

I believe there is a straightforward implementation for multi-GPU support. You can wrap the existing script with an outer script that handles the splitting of the test set and passes the GPU IDs accordingly. This approach is similar to what FunASR did previously.

@Learneducn
Copy link
Author

我相信多 GPU 支持有一个简单的实现。您可以使用外部脚本包装现有脚本,该脚本处理测试集的拆分并相应地传递 GPU ID。这种方法类似于 FunASR 之前所做的。
Thank you very much. The problem about specifying a certain card for testing has been solved. I have encountered another problem now. I directly used the SLAM framework to fine-tune the inference results. Why haven’t I directly used the whisper open source model to test the results?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants