-
Notifications
You must be signed in to change notification settings - Fork 6.8k
[2.0] Adopt interleaved_matmul_* into npx namespace #20375
Conversation
Hey @barry-jin , Thanks for submitting the PR
CI supported jobs: [website, sanity, edge, clang, windows-gpu, miscellaneous, centos-cpu, unix-cpu, windows-cpu, unix-gpu, centos-gpu] Note: |
Would it be more beneficial to have the full multihead attention primitive exposed in the API instead (it could still be implemented by the interleaved matmuls)? Both Keras and pyTorch have it as a layer, and this could make the usage easier. What do you think @barry-jin? |
Can we have both? We'd like to upgrade Sockeye to MX2 without changing too many things at the same time. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
thanks for bring the interleaved matmul ops into npx! We have them in mx2onnx
Can this be merged? |
Description
Part1 of feature request #20373
Checklist
Essentials
Changes
Comments