-
Notifications
You must be signed in to change notification settings - Fork 17
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
KL loss in validation #23
Comments
Hi, I think it's basically the same, you can think this KL-divergence as a cross-entropy function to multiple soft labels. |
@Jiacheng8 |
The two are equivalent except for a constant term, but of course the optimisation will be a little different for the lowest value, I will try to do an ablation result on this. |
@Jiacheng8 Thanks. My major concern is that using kl-loss results in stronger knowledge distillation enhancement, especially when temperature is also used in your case. I'm wondering if adopting this evaluation strategy is the main reason for performance improvement instead of the dataset itself having a higher quality. Yet, the paper didn't present ablation on this. |
Hi, may I ask why the KL loss is used during validation? This doesn't match equation 9 in the paper which is a cross-entropy loss.
The text was updated successfully, but these errors were encountered: