-
Notifications
You must be signed in to change notification settings - Fork 15
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
The training of the model is very slow and the GPU power is very low. #9
Comments
Can you run your code through the profiler to see where the bottleneck lies? |
@nils-werner hi, i found that maybe the align_level() and preemphasize() work inefficiently. Among them, lfilter seems to be the culprit.It may be because lfilter requires a significant amount of GPU startup time. |
Now I am sure the problem lies with lfilter. For smaller batch sizes (like [2, 480000]), the parallelism is poor, and GPU computation efficiency is low. I might consider increasing the batch size and truncating the audio (e.g., [8, 160000]) to improve GPU efficiency. Additionally, I have tried making simple modifications to the functions align_level() and preemphasize() that involve calling lfilter:
Although the initial state information is lost due to the IIR filtering of the crudely folded time series, the error is relatively small. Do you have any good ways to improve lfilter parallelism through this external means? |
When i use this function as loss-function, my training is so slow and the power of gpu is only 70W/220W.
How to improve the GPU computing efficiency and accelerate training?
The text was updated successfully, but these errors were encountered: