You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
One of the scripts in the examples/ folder of Accelerate or an officially supported no_trainer script in the examples folder of the transformers repo (such as run_no_trainer_glue.py)
My own task or dataset (give details below)
Reproduction
When I run the accelerate/examples/megatron_1m_gpt_pretraining. py file
accelerate launch --config_file megatron_gpt_pretraining.py
--config_name "gpt2-large"
--tokenizer_name "gpt2-large"
--dataset_name wikitext
--dataset_config_name wikitext-2-raw-v1
--block_size 1024
--learning_rate 5e-5
--per_device_train_batch_size 24
--per_device_eval_batch_size 24
--num_train_epochs 5
--with_tracking
--report_to "wandb"
--output_dir "awesome_model"
Expected behavior
Normal training, but I found that in megatron_1m_gpt_pretraining on line 255 of the py file, there is an undefined parameter 'logging.dir' in the init method of the Accelerator function
The text was updated successfully, but these errors were encountered:
System Info
Information
Tasks
no_trainer
script in theexamples
folder of thetransformers
repo (such asrun_no_trainer_glue.py
)Reproduction
When I run the accelerate/examples/megatron_1m_gpt_pretraining. py file
accelerate launch --config_file megatron_gpt_pretraining.py
--config_name "gpt2-large"
--tokenizer_name "gpt2-large"
--dataset_name wikitext
--dataset_config_name wikitext-2-raw-v1
--block_size 1024
--learning_rate 5e-5
--per_device_train_batch_size 24
--per_device_eval_batch_size 24
--num_train_epochs 5
--with_tracking
--report_to "wandb"
--output_dir "awesome_model"
Expected behavior
Normal training, but I found that in megatron_1m_gpt_pretraining on line 255 of the py file, there is an undefined parameter 'logging.dir' in the init method of the Accelerator function
The text was updated successfully, but these errors were encountered: