-
Notifications
You must be signed in to change notification settings - Fork 2.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Cannot run Mixer TTS colab with Mixer-TTS-X model #4803
Comments
Also I've added 'lm_tokens' in sup_data_types. It is shown in the command |
Also changed lm_tokens for tokens, since is the name of the parameter. Also changed the index for a number. Same results. |
Instead of passing in You can find the list of default arguments that the Note that the
to be this:
to make that part of the notebook run correctly. |
@redoctopus great help. Now that error dont come anymore. I have another error now:
Command:
I only added 1 parameter and also added in that after line, so don't know where is the mistake. Thanks! |
Now I've added 'lm_model' here, but it seems to not recognize it:
Error:
Thanks. |
Ah, yep. I ran into that too when reproducing--older versions of the dataset should be fine, but a fix is in the works in #4811 that updates the count to include |
Re |
Thanks. The unpack values error remains the same but thanks for the clarification on the pending fix. Just a question: Can I finetune this model with my own dataset right as in FastPitch right? This is, starting from the LJSpeech or HiFiTTS checkpoint and then adding my own audio, and select the speaker when inferencing. Thanks! |
No problem! I don't think anyone on our team has tried it yet, but yes, you should be able to fine-tune as usual. |
Hey, I saw the changes were merged into main, but I still getting the error:
The error:
Thanks! |
The fix was merged into the |
@redoctopus I think there is a problem with that. If I run
Error:
|
Have you tried reinstalling after switching branches? As of #4690 some preprocessing modules have been moved from the TTS collection to more general text processing, and if you haven't reinstalled the program can no longer find those classes. |
I'm using Google Colab PRO +...erasing and restarting each time the kernel, but the error is still there. Each time I reconnect to it I'm installing nemo |
Ahh, I see the problem now. #4690 was merged to Alternatively you could try cherrypicking the fix in another branch, if it is urgent. |
Oh, you should also be able to load the |
No problem! Thank you very much. I was able to run it loading directly the I have a question. After running:
The folders created under Why does this happen?It is correct? |
Also I get this in the following command (I know i've to change sup_data_types, but given the previous post...):
The error:
Thanks Jocelyn |
Yes, this should be fine--it will compute those on the fly during training. Regarding the error, it looks like it can't find |
Hi @redoctopus . Same error. Previous command:
Thank you |
I am not able to reproduce the error, this is the run command I have:
It looks like some parsing error, might be due to the whitespace (see facebookresearch/hydra#836). Can you use the same Once it does start running, I also found that you may run into a dim mismatch error--this is due to the fact that This can be resolved by replacing the cell that creates the normalizer/tokenizer with this:
|
Well I've to change
BTW, does the char tokenizer makes a difference in terms of final audio? Thanks Jocelyn! |
No problem! In this case I think the commented out lines are throwing it off, it seems to think that the command ends at that point. If you remove those lines it should see them again.
I believe Mixer-TTS-X uses the char tokenizer since it uses an external LM to get token embeddings. If you switch to a phoneme tokenizer, the LM model probably won't handle it well. |
Well It is progressing :) It seems like it is about to start training. I get the following now (changed tokenizer for char-based and the error is the same):
|
Ah yep, this is the error that occurs when it tries to load the old supplementary values. Can you remove that folder and re-run the supplementary data calculation with the new tokenizer? I think it's probably seeing the old data and still trying to load those values. I'd like to note that the The |
Well now it run without errors. I had not to run the I suppose I have to pass now all the parameters directly in the training command. One thing I'm seeing is that there is still no Will test the checkpoints and get back. Thank you very much!! |
Yes, that is expected--the Dataset does not save You're very welcome, good luck with training! |
Hi all. I'm struggling to run a training using Mixer-TTS-X model. I'm using the tutorial for training both Fastpitch and MixerTTS.
Modifications I've done:
pretrained_model = "tts_en_lj_mixerttsx"
Adding 'raw_texts' argument when generating a spectrogram:
spectrogram = spec_gen.generate_spectrogram(tokens=tokens, raw_texts=["Hey, this produces speech!"])
Correcting this:
from nemo.collections.tts.torch.data import MixerTTSXDataset
Just in case:
add lm_tokenizer parameter here:
Getting the right config file:
&& wget https://raw.githubusercontent.com/NVIDIA/NeMo/$BRANCH/examples/tts/conf/mixer-tts-x.yaml
Creating the lm_tokenizer object:
lm_tokenizer = LMTokens()
And after running the command:
I get the following error:
Any ideas are welcome.
Thanks,
The text was updated successfully, but these errors were encountered: