-
Notifications
You must be signed in to change notification settings - Fork 54
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Training on labeled only can achieve much better results than what is reported #21
Comments
Dear IssamLaradji first, thank you for considering GANBERT In our experiments, we just used the Tensorflow implementation of BERT that was available on the official repository (https://github.com/google-research/bert) during the early months of 2020. If you are interested in advanced parameters that you can use in more recent implementations (e.g., Huggingface, where, for example, you can use schedulers) you can take a look at https://github.com/crux82/ganbert-pytorch Hope it helps Bests Danilo |
Thanks for your reply! The implementation is the same as yours except that I used the
|
Hi! I've obtained similar results to @IssamLaradji! @crux82 did you consider that increasing training epochs number for BERT model can be better in terms of quality metric than applying whole GAN-BERT thing? Thanks in advance! |
Interesting idea of using an adversarial method for leveraging unlabeled data. I am trying to see how much unlabeled data can actually help.
In the plot below, I am comparing GanBert (Orange) that trains on both labeled and unlabeled data, and a basic model that uses Bert+Classifier (blue) that trains on the 109 labeled data only of Trec Data.
The paper reports that the basic model should achieve around 40%, but I am getting 60% which is very close to GanBert's. Are you sure that the baseline discussed in the paper is a reasonable one?
The text was updated successfully, but these errors were encountered: