You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Config
I've done a few changes to the configuration file so that it fits my environment. I'm on a Mac computer, so I've tried running with CPU and MPS, but get the error with both ways.
This is the corresponding config YAML file, resembling the example config file.
name: "ASR_Custom_Finetuning"# use `init_from_nemo_model` or `init_from_pretrained_model` to initialize the model# We do not currently support `init_from_ptl_ckpt` to create a single script for all types of models.init_from_nemo_model: models/parakeet-ctc-1.1b.nemo # path to nemo model# init_from_pretrained_model: nvidia/parakeet-ctc-0.6b # path to nemo modelmodel:
sample_rate: 16000train_ds:
manifest_filepath: data/TRAIN_CLEAN_100.jsonsample_rate: ${model.sample_rate}batch_size: 16# you may increase batch_size if your memory allowsshuffle: truenum_workers: 8pin_memory: truemax_duration: 20min_duration: 0.1# tarred datasetsis_tarred: falsetarred_audio_filepaths: nullshuffle_n: 2048# bucketing paramsbucketing_strategy: "fully_randomized"bucketing_batch_size: nullvalidation_ds:
manifest_filepath: data/dev_clean.jsonsample_rate: ${model.sample_rate}batch_size: 16shuffle: falseuse_start_end_token: falsenum_workers: 8pin_memory: truetest_ds:
manifest_filepath: nullsample_rate: ${model.sample_rate}batch_size: 16shuffle: falseuse_start_end_token: falsenum_workers: 8pin_memory: truechar_labels: # use for char based modelsupdate_labels: falselabels: null # example list config: \[' ', 'a', 'b', 'c'\]tokenizer: # use for spe/bpe based tokenizer modelsupdate_tokenizer: falsedir: null # path to directory which contains either tokenizer.model (bpe) or vocab.txt (for wpe)type: bpe # Can be either bpe (SentencePiece tokenizer) or wpe (WordPiece tokenizer)spec_augment:
_target_: nemo.collections.asr.modules.SpectrogramAugmentationfreq_masks: 2# set to zero to disable ittime_masks: 10# set to zero to disable itfreq_width: 27time_width: 0.05optim:
name: adamwlr: 1e-4# optimizer argumentsbetas: [0.9, 0.98]weight_decay: 1e-3# scheduler setupsched:
name: CosineAnnealing# scheduler config overridewarmup_steps: 5000warmup_ratio: nullmin_lr: 5e-6trainer:
devices: 4# number of GPUs, -1 would use all available GPUsnum_nodes: 1max_epochs: 50max_steps: -1# computed at runtime if not setval_check_interval: 1.0# Set to 0.25 to check 4 times per epoch, or an int for number of iterationsaccelerator: cpustrategy:
_target_: lightning.pytorch.strategies.SingleDeviceStrategy# _target_: lightning.pytorch.strategies.DDPStrategygradient_as_bucket_view: trueaccumulate_grad_batches: 1gradient_clip_val: 0.0precision: 32# 16, 32, or bf16log_every_n_steps: 10# Interval of logging.enable_progress_bar: Truenum_sanity_val_steps: 0# number of steps to perform validation steps for sanity check the validation process before starting the training, setting to 0 disables itcheck_val_every_n_epoch: 1# number of evaluations on validation every n epochssync_batchnorm: trueenable_checkpointing: False # Provided by exp_managerlogger: false # Provided by exp_managerbenchmark: false # needs to be false for models with variable-length speech input as it slows down trainingexp_manager:
exp_dir: nullname: ${name}create_tensorboard_logger: truecreate_checkpoint_callback: truecheckpoint_callback_params:
# in case of multiple validation sets, first one is usedmonitor: "val_wer"mode: "min"save_top_k: 5always_save_nemo: True # saves the checkpoints as nemo files along with PTL checkpointsresume_if_exists: falseresume_ignore_no_checkpoint: falsecreate_wandb_logger: falsewandb_logger_kwargs:
name: nullproject: null
Environment overview (please complete the following information)
Environment location: Bare-metal
Method of NeMo install: From source (git commit 34f7408)
I was able to solve this by pulling the Nemo code from the 2.0.0 tag, and using that. I believe a mismatch in source code and the nemo library caused this. Probably because of the recent commit that changed from 'pytorch_lightning' -> 'lightning.pytorch'.
Describe the bug
I'm getting this error when trying to fine-tune the Parakeet model using the
/examples/asr/speech_to_text_finetune.py
file.Steps/Code to reproduce bug
Command
This is the command I use for fine-tuning
Data
So I've downloaded the Librispeech Datasets and processed it using
/scripts/dataset_processing/get_librispeech_data.py
Model
I've also downloaded the models from HuggingFace. https://huggingface.co/nvidia/parakeet-ctc-0.6b
Config
I've done a few changes to the configuration file so that it fits my environment. I'm on a Mac computer, so I've tried running with CPU and MPS, but get the error with both ways.
This is the corresponding config YAML file, resembling the example config file.
Environment overview (please complete the following information)
Environment details
The text was updated successfully, but these errors were encountered: