-
Notifications
You must be signed in to change notification settings - Fork 12
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
The inconsistent inference results #34
Comments
Thanks for your attention to our project! And sorry for the delay. Sorry for the misleading results. Could you please attach the corresponding log for each experiment for analysis and locate the issue for us? |
I located the issue, the data augmentations are inconsistent. For trained checkpoints, you have used resize with padding, for testing evaluation code in IMDLBenCo, you have used resize without padding, thus the data distribution is not aligned. On my custom dataset, I found that resize without padding works better than resize with padding. Did you observe similar results? |
Hi, IMDLBenCo/IMDLBenCo/statics/model_zoo/runs/demo_test_iml_vit.sh Lines 1 to 20 in e4f59d0
I am not sure how you train IML-ViT with resize without padding. Since the original code design only supports 1024x1024 input. Mostly we don't apply a traditional |
Oh, I was not using demo_test_iml_vit.sh for testing, but just use demo_train_iml_vit.sh for evaluation, where I put the evaluation code before the training code in the train.py. In the generated demo_train_iml_vit.sh, I believe the configs are like:
where if_resizing is set to True, and the data_transform would be resize without padding, like in the code below:
After I manually set the ablu_transform type to "pad", the results are consistent. I made the conclusion that it was not because On my custom dataset, however, I found "resize" mode yields better results by 1 or 2 percents. |
Thank you for your feedback. I see your points. Generally, the deep neural network fits a distribution as a function. Thus, keeping the training distribution similar to the testing distribution is essential. Just like the issue mentioned by you "I made the conclusion that it was not because Further, there are many possible explanations for the performance on your custom dataset. Such as:
Thanks again for your attention to our project. If you find the issue solved, please close the issue. You are also welcome to discuss further concerns and problems you met. |
Hi, I found that for the same checkpoints of IML-ViT, the inference results on CASIA v1 inferenced through this IMDB-IML-ViT framework is much lower (~12%) that computed within of the original code base framework IML-ViT (~70%).
The text was updated successfully, but these errors were encountered: