Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add WeightNorm layer #2053

Closed
wants to merge 1 commit into from
Closed

Add WeightNorm layer #2053

wants to merge 1 commit into from

Conversation

mcabbott
Copy link
Member

@mcabbott mcabbott commented Aug 29, 2022

RFC as to whether this looks right, and whether the approach is a good one.

Needs tests.

Edit: earlier attempts are #993 (Dense only) and #1005. I had completely forgotten this, but the approach here is similar to @CarloLucibello's suggestion in #1005 (comment). (But I'm not sure why the forward pass is unchanged in the example there.)

Closes #942

Checklist

  • Tests are added
  • Entry in NEWS.md
  • Documentation, if applicable

@mcabbott mcabbott mentioned this pull request Sep 26, 2022
92 tasks
@bhvieira
Copy link
Contributor

I like the approach, would just advise to do performance tests as well. Due to compiler optimizations, I guess/hope this doesn't incur too much penalization. Also, would this work with recurrent layers? Perhaps define a new dispatch when it finds a Recur. That was also a big blocker in #1005, but I don't believe it should block merging this

@mcabbott mcabbott mentioned this pull request Dec 12, 2024
3 tasks
@CarloLucibello
Copy link
Member

Done in #2550

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

Successfully merging this pull request may close these issues.

Feature request: Weight normalization
3 participants