-
Notifications
You must be signed in to change notification settings - Fork 52
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Overfitting problem #5
Comments
hello,I get the same problem,I haven't saw the "Test Epoch"appeared worse ,it still training...now. |
@fxle Test log is saved in DFL_CNN/log/log_text.txt . |
@pengshiqi Oh,thank you very much! I find it .It seems to be improving.But the Loss2 value looks a little strange.In addition,Do you think the idea of 'filter bank' in this paper can improve rotation invariance at the same time? |
Sure, the I think a potential cause for this problem is that the 1x1 convolutional layer is not initialized randomly, as described in Section 3.3, which has not been implemented in this code. |
@pengshiqi Yes, you are right.It's not initialized .Do you know how to make it? I have some ideas to communicate with you.I suggest we add a qq.my qq numbers are :260730636 |
@pengshiqi @fxle I changed the model using a dropout layer. During training the model, I got loss2 decrease, loss 1 & 3 is basically not reduced. do you got same situation? |
@techzhou No,I didn't use a dropout layer.Maybe you can try to use regularization or Section 3.3 Layer Initialization to make it perform better, |
@techzhou |
@pengshiqi hi, how do you solve the overfitting problem? |
@pengshiqi hi, how do you solve the overfitting problem? @XieLeo @fxle @techzhou @Ien001 @ |
@pengshiqi |
Hello,
I cloned your repo and downloaded your dataset, but I could not get the same result as yours.
I train the model on 4 GeForce GTX 1080 Ti GPUs, and keep other arguments as the same. But it turns out to be overfitting.
After 100 epochs, I got prec@1 at 99% and prec@5 at 100% on the train set, but only prec@1 at 48% and prec@5 at 73% on the test set.
Here is part of the log:
During training the model, overfitting problem is inevitable. Because there are only about 6000 images for training, but there are too many parameters in VGG16.
Have you ever met the overfitting problem? And how did you get rid of it?
Looking forward to your reply!
Thank you very much!
The text was updated successfully, but these errors were encountered: