Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Distillation is broken #272

Closed
Tracked by #369 ...
eu9ene opened this issue Nov 20, 2023 · 4 comments · Fixed by #308
Closed
Tracked by #369 ...

Distillation is broken #272

eu9ene opened this issue Nov 20, 2023 · 4 comments · Fixed by #308
Assignees
Labels
bug Something is broken or not correct quality Improving robustness and translation quality

Comments

@eu9ene
Copy link
Collaborator

eu9ene commented Nov 20, 2023

Screenshot 2023-11-20 at 3 43 15 PM

The hypothesis that we also discussed in #161 is that distillation might not work as expected with the on-the-fly augmentation since the student is supposed to be train on the exact outputs of the teacher. The gap between the teacher and student models is way too large. See also #231. I'll try to disable augmentation on the fly for the student model first to see how much it helps.

If it's the case, the proper fix would be to do augmentation of the corpus before decoding by the teachers.

@eu9ene eu9ene added bug Something is broken or not correct quality Improving robustness and translation quality labels Nov 20, 2023
@eu9ene eu9ene self-assigned this Nov 20, 2023
@eu9ene
Copy link
Collaborator Author

eu9ene commented Nov 22, 2023

I removed the augmentation but the new student model is still training https://firefox-ci-tc.services.mozilla.com/tasks/eyHs_KRMRbONpKLXIgj2fA/runs/1

@eu9ene
Copy link
Collaborator Author

eu9ene commented Dec 13, 2023

Retraining with disabled augmentation didn't help, the results are almost the same. There's likely a bug somewhere else.

Screenshot 2023-12-12 at 4 04 59 PM

@eu9ene
Copy link
Collaborator Author

eu9ene commented Dec 14, 2023

It seems the issue is in the merged training corpus. Source and target sentences don't match in the output of merge-translated step: https://firefox-ci-tc.services.mozilla.com/tasks/QlmjiyNmTJ-fzGGM520NMQ/runs/0

@eu9ene eu9ene changed the title Distillation is broken after integrating OpusTrainer Distillation is broken Dec 18, 2023
@eu9ene
Copy link
Collaborator Author

eu9ene commented Dec 19, 2023

It's supposed to be fixed now. Retesting here https://firefox-ci-tc.services.mozilla.com/tasks/groups/V-OmRM1yS_GwESTVDPl0VQ

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something is broken or not correct quality Improving robustness and translation quality
Projects
None yet
Development

Successfully merging a pull request may close this issue.

1 participant