-
Notifications
You must be signed in to change notification settings - Fork 40
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
when I replace activation function sigmoid with relu, but the project could not generate result. #5
Comments
I think its because the paper itself mentions that it will work with sigmoid activations. I am also looking for solutions. If you find any please keep me updated too ? |
I also tried this. But, in my opinion, this is not the matter of differentiability. I also tried SiLU = (x*sigmoid) activation function which is smooth, but I failed to generate. |
Thanks for your interests in our work! Our experiments show the feasibility to extract the ground-truth label and image in a simple setting. I think the problem with relu is caused by the difficult optimization. Better matching loss and optimizer (including learning rate...) are needed to realize the deep leakage with various activation functions. |
After a few experiments yesterday, I found that initial values of the model parameters are important. If the activation function changes, stds of initializer should be adjusted. In my setting, std=0.1 works for SiLU activation with other settings unchanged. ( lr of the lfbgs optimizer could be changed.) |
I am running the experiments to get the results on the mobilenet with the relu activation and sigmoid activation functions. |
Have you succeeded in deeper networks? In my experiments, loss is decreasing, but it stops at some points and does not converge. |
Hi, It didn't work with deep CNNs like VGG16. Also, when I changed the activation function to a different one from Sigmoid, it didn't converge. Has someone found a solution for that? Br, |
It works when I changed the activation functions. It is hard for ReLU, but it works for other invertible activations |
No description provided.
The text was updated successfully, but these errors were encountered: