Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Some curiosity about the baseline training configuration. #18

Open
icecream-Tnak opened this issue Sep 26, 2022 · 0 comments
Open

Some curiosity about the baseline training configuration. #18

icecream-Tnak opened this issue Sep 26, 2022 · 0 comments

Comments

@icecream-Tnak
Copy link

icecream-Tnak commented Sep 26, 2022

Thank you for your contribution to the community, the creation of the Chinese recognition benchmark is critical to the advancement of the field.

I would like to know some more details about the training configuration of some baseline models for the sub-dataset "scene". Such as specific epoch, batchsize, lr, weight_decay, max_length, grad_clip, etc.

In particular, I noticed that the results of TransOCR reported in the paper (arXiv:2112.15093) are different from the results on GitHub. After the paper was submitted to arxiv, better experimental results were obtained based on different experimental hyperparameters?

Thank you again for your outstanding work and hope you can get back to me. Because I didn't find the specific configuration file, and the default parameters in TransOCR/main.py, such as epoch = 1000. Unfortunately, my current hardware cannot support such a long training time.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant