Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[BUG] Accelerator.__init__() got an unexpected keyword argument 'logging_dir' #3257

Open
2 of 4 tasks
as12138 opened this issue Nov 25, 2024 · 3 comments · May be fixed by #3261
Open
2 of 4 tasks

[BUG] Accelerator.__init__() got an unexpected keyword argument 'logging_dir' #3257

as12138 opened this issue Nov 25, 2024 · 3 comments · May be fixed by #3261

Comments

@as12138
Copy link

as12138 commented Nov 25, 2024

System Info

accelerate version: main
python version: 3.11
torch version: 2.4
numpy version: 1.26.4

Information

  • The official example scripts
  • My own modified scripts

Tasks

  • One of the scripts in the examples/ folder of Accelerate or an officially supported no_trainer script in the examples folder of the transformers repo (such as run_no_trainer_glue.py)
  • My own task or dataset (give details below)

Reproduction

When I run the accelerate/examples/megatron_1m_gpt_pretraining. py file
accelerate launch --config_file megatron_gpt_pretraining.py
--config_name "gpt2-large"
--tokenizer_name "gpt2-large"
--dataset_name wikitext
--dataset_config_name wikitext-2-raw-v1
--block_size 1024
--learning_rate 5e-5
--per_device_train_batch_size 24
--per_device_eval_batch_size 24
--num_train_epochs 5
--with_tracking
--report_to "wandb"
--output_dir "awesome_model"

Expected behavior

Normal training, but I found that in megatron_1m_gpt_pretraining on line 255 of the py file, there is an undefined parameter 'logging.dir' in the init method of the Accelerator function

@BenjaminBossan
Copy link
Member

Thanks for pointing this out. I think it should be project_dir instead. Are you interested in submitting a PR to fix this?

@muellerzr
Copy link
Collaborator

For clarity, the file is at https://github.com/huggingface/accelerate/blob/main/examples/by_feature/megatron_lm_gpt_pretraining.py :)

@as12138
Copy link
Author

as12138 commented Nov 25, 2024

of course

as12138 added a commit to as12138/accelerate that referenced this issue Nov 26, 2024
@as12138 as12138 linked a pull request Nov 26, 2024 that will close this issue
5 tasks
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging a pull request may close this issue.

3 participants