Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

R1 loss - called with all params #3

Open
moshebeutel opened this issue Sep 7, 2024 · 2 comments
Open

R1 loss - called with all params #3

moshebeutel opened this issue Sep 7, 2024 · 2 comments

Comments

@moshebeutel
Copy link

Hi , reading your paper I understood that the first custom loss is regularized to minimize personalized parameters change. It seems that in line 112 in ours.py the diffs are calculated using all model.parameters() and w_glob and not just the personalized. at the end of iteration only personalized parameters are updated but I do not think this is equivalent to computing the diff on the personalized parameters as described in the paper. Am I missing something?
Thanks 🙏 Moshe

@moshebeutel
Copy link
Author

Same holds for the second loss / shared parameters

@xiyuanyang45
Copy link
Owner

xiyuanyang45 commented Sep 10, 2024

Then if there's any alternatives for only calculating updates for partial selected params?
I think that code is an available approach for implementation, welcome to discuss any other possible implementations of this algorithm

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants