-
Notifications
You must be signed in to change notification settings - Fork 191
Support for Distilbert #289
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
Codecov Report
@@ Coverage Diff @@
## master #289 +/- ##
==========================================
+ Coverage 45.23% 68.44% +23.2%
==========================================
Files 13 12 -1
Lines 1859 1255 -604
==========================================
+ Hits 841 859 +18
+ Misses 1018 396 -622
Continue to review full report at Codecov.
|
…ytorch_pretrained_bert from requirements
How did you train and dump "distilbert_qa.joblib"? it seems that many files for distilbert did not updated. |
Which files are you talking about? |
Thanks for the reply, Got it.
Oh, I checked the bertqa_sklearn.py, DistilBert Model and corresponding classes did not use. |
This PR adds support for Distilbert as mentioned in #197.
I will be also releasing a Distilbert trained on SQuAD 1.1 using Knowledge Distillation and
bert-large-uncased-whole-word-masking-finetuned-squad
as a teacher.From my experiments, this version of Distilbert achieves 80.1% EM and 87.5% F1-score (vs. 81.2% EM and 88.6% F1-score for our version of BERT), while being much faster and lighter.