Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Question: Loss not decreasing #6

Open
jonashaag opened this issue May 7, 2020 · 3 comments
Open

Question: Loss not decreasing #6

jonashaag opened this issue May 7, 2020 · 3 comments

Comments

@jonashaag
Copy link
Contributor

Hi! I'd like to reproduce the very good results of your paper. When I run it on a 15,000 samples * 3s dataset, loss only decreases considerably at the very beginning (first few batches of epoch 1), and then stays the same. After 20 epochs of training, the results are not good. Any ideas what's wrong here? I have made no changes to the code except for what I submitted as PR, so using the default learning rate settings etc.

I'll upload a reconstruction sample after 20 epochs.

@jonashaag
Copy link
Contributor Author

out.wav.zip

@Andong-Li-speech
Copy link
Owner

Andong-Li-speech commented May 7, 2020

Hi! I'd like to reproduce the very good results of your paper. When I run it on a 15,000 samples * 3s dataset, loss only decreases considerably at the very beginning (first few batches of epoch 1), and then stays the same. After 20 epochs of training, the results are not good. Any ideas what's wrong here? I have made no changes to the code except for what I submitted as PR, so using the default learning rate settings etc.

I'll upload a reconstruction sample after 20 epochs.

Hi! First I want to appreciate your contributions as there is much room to improve the code quality. I think the problem is mainly due to self.Iter = 1. As we have to reuse the network multiple times. self.Iter=1 means only one stage is used and no correlation can be established between adjacent stages. In addition, as we only give the supervision to the final stage, we should output the estimation result of the last stage during the decoding process.

@Andong-Li-speech
Copy link
Owner

Hi! I'd like to reproduce the very good results of your paper. When I run it on a 15,000 samples * 3s dataset, loss only decreases considerably at the very beginning (first few batches of epoch 1), and then stays the same. After 20 epochs of training, the results are not good. Any ideas what's wrong here? I have made no changes to the code except for what I submitted as PR, so using the default learning rate settings etc.

I'll upload a reconstruction sample after 20 epochs.

Hi! I have uploaded the infos.zip, which contains the loss curves of different models and enhanced clips under -5dB SNR. I hope it is helpful for you!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants