You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
First of all, thank the author for open source. DKD is a very good job!
I trained a binary classification model and wanted to use DKD for distillation, but after one iteration of training, the output of the student model became very strange, causing the loss to become Inf.
Some details:
iteration 1:
iteration 2:
The text was updated successfully, but these errors were encountered:
NCKD loss can not be applied on binary classification tasks, since there is no "non-target class". The non-target logits will be always 1.0, may introduce some problems when calculating KLD.
First of all, thank the author for open source. DKD is a very good job!
I trained a binary classification model and wanted to use DKD for distillation, but after one iteration of training, the output of the student model became very strange, causing the loss to become Inf.
Some details:
iteration 1:
iteration 2:
The text was updated successfully, but these errors were encountered: