-
Notifications
You must be signed in to change notification settings - Fork 2
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Training results are very poor. #4
Comments
Hi @xungeer29. Could you reproduce the metrics mentioned in the readme using the released checkpoint? My suggestion is to first ensure that the results of the official model inference can be reproduced before making any modifications. Additionally, according to the linear scaling rule, when the batch size changes, the learning rate also needs to be adjusted accordingly to achieve similar performance. |
I can reproduce the metrics mentioned in the readme using the released checkpoint, as shown below
LR only have a small impact on the results and cannot make the network completely non convergent. |
Have you solved this problem? |
No. |
Is there something wrong with the test code? |
The testing results using released checkpoints are correct. Have you also encountered the same problem? |
Have you trained the original model? What's the result? |
Thank you very much for making your training code public.
I used your default config file to train the model, just modify the batch_size to 256. The final results achieved 81+ for MPJPE and 88+ for MPVPE.
I don't know why the results are so bad.
The text was updated successfully, but these errors were encountered: