-
Notifications
You must be signed in to change notification settings - Fork 658
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
split and process valid set #25
Conversation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
What will happen when processing other datasets except for libritts and ljspeech? I suspect that line39 will cause a bug, since there is none valid.json
for others.
Besides, is there any corresponding design for the valid and test dataset in trainer?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Rewrite metadata.py to automatically adapt different json files.
The valid set is split to distinguish it from the test set and also used to compute validation loss in the trainer.
preprocessors/metadata.py
Outdated
@@ -8,7 +8,7 @@ | |||
from tqdm import tqdm | |||
|
|||
|
|||
def cal_metadata(cfg): | |||
def cal_metadata(cfg, dataset_types): |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
For SVC, TTA, and Vocoder, the existing calls to this funcion is cal_metadata(cfg)
. So the implementation here will cause a bug. Maybe you can set default values to dataset_types
.
The PR is to support the validation set split and process to standardize the dataset.