Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

About the experiment of MWPBert on math23k #21

Open
LzhinFdu opened this issue May 7, 2022 · 5 comments
Open

About the experiment of MWPBert on math23k #21

LzhinFdu opened this issue May 7, 2022 · 5 comments

Comments

@LzhinFdu
Copy link

LzhinFdu commented May 7, 2022

I got 'value accu=40.0' and found that the model uses 'bert-base-uncased' as the encoder by default. Could the reason be that I were not using a Chinese bert for math23k?

here is my instruction:
python run_mwptoolkit.py --model=MWPBert --dataset=math23k --task_type=single_equation --equation_fix=prefix --test_step=5 --gpu_id=0


I tried to change 'config["pretrained_model"]' to 'bert-base-chinese',but got some bugs which showed it doesn;t match the model......Is there any built-in method to change it?

@LYH-YF
Copy link
Owner

LYH-YF commented May 8, 2022

@LYH-YF LYH-YF closed this as completed May 8, 2022
@LYH-YF
Copy link
Owner

LYH-YF commented May 8, 2022

you can change it with command line like
--model=MWPBert --dataset=math23k --task_type=single_equation --gpu_id=0 --equation_fix=prefix --pretrained_model=bert-base-chinese

i hope this will help you

@LYH-YF LYH-YF reopened this May 8, 2022
@LzhinFdu
Copy link
Author

Very grateful for your help. The model works fine then.
However, I ended up with a score of 66.3, which is still lower than the result given in the paper.

@LYH-YF
Copy link
Owner

LYH-YF commented May 10, 2022

there may be something wrong with my code when i update v0.0.6, i will check it. and i'm so sorry for that.

@LYH-YF
Copy link
Owner

LYH-YF commented Aug 14, 2022

I got value acc 82.5, the latest result of MWPBert on math23k.

here is my here is my instruction:

python run_mwptoolkit.py --model=MWPBert --dataset=math23k --equation_fix=prefix --task_type=single_equation --pretrained_model=hfl/chinese-bert-wwm-ext --test_step=5 --gpu_id=0 --train_batch_size=32 --epoch_nums=85 --learning_rate=3e-4 --encoding_learning_rate=3e-5 --vocab_level=char

and I publish the result at the result table

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants