Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

optuna hyperparameter optimization for NER task on knowledge distillation #115

Open
1 task
Venkatesh3132003 opened this issue Aug 22, 2023 · 1 comment
Open
1 task

Comments

@Venkatesh3132003
Copy link

Venkatesh3132003 commented Aug 22, 2023

Information

The problem arises in chapter:

  • Making Transformers Efficient in Production

Describe the bug

while training i am getting proper F1 score of 0.755940
image

while finding best fit value of alpha and temperature value for NER task f1 score is 0.096029 which is less than 0.1
image

To Reproduce

Steps to reproduce the behavior:

1.compute metric is same as chapter 4 of NER
2.Hyperparameter are for alpha and temperature

def hp_space(trial):
return {"alpha": trial.suggest_float("alpha", 0, 1),
"temperature": trial.suggest_int("temperature", 2, 20)}

best_run = distil_roberta_trainer.hyperparameter_search(
n_trials=12, direction="maximize",backend="optuna", hp_space=hp_space)

Expected behavior

After the hyperparameter search the F1 score should be higher than baseline.

@Venkatesh3132003
Copy link
Author

When alpha is 1 F1 score is good and for any value of alpha between 0 and 1 F1 score is less than 0.1

@Venkatesh3132003 Venkatesh3132003 changed the title optuna hyperparameter optimnation for NER task on knowledge distillation optuna hyperparameter optimization for NER task on knowledge distillation Sep 30, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant