-
Notifications
You must be signed in to change notification settings - Fork 97
Introduce layer-norm fusion #2492
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
Signed-off-by: Ganesan Ramalingam <grama@microsoft.com>
Signed-off-by: Ganesan Ramalingam <grama@microsoft.com>
Signed-off-by: Ganesan Ramalingam <grama@microsoft.com>
❌ 3 Tests Failed:
View the top 3 failed test(s) by shortest run time
To view more test analytics, go to the Test Analytics Dashboard |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Pull Request Overview
This PR introduces layer normalization fusion functionality to optimize ONNX models by fusing layer normalization patterns into the ONNX LayerNormalization operator. This is designed for ONNX opset 17+ where the LayerNormalization operator is available.
Key changes:
- Adds fusion rules to detect and fuse layer normalization patterns with and without bias
- Implements pattern matching for the mathematical sequence of operations that constitute layer normalization
- Includes comprehensive test cases to validate the fusion functionality
Reviewed Changes
Copilot reviewed 3 out of 3 changed files in this pull request and generated 2 comments.
| File | Description |
|---|---|
| onnxscript/rewriter/testing.py | Adds utility functions for generating random inputs and enhances numerical equality testing to support dictionary-based inputs |
| onnxscript/rewriter/onnx_fusions/_layer_norm.py | Implements the core layer normalization fusion logic with pattern matching and rewrite rules |
| onnxscript/rewriter/onnx_fusions/_layer_norm_test.py | Provides test cases for layer normalization fusion with and without bias scenarios |
Signed-off-by: Ganesan Ramalingam <grama@microsoft.com>
Signed-off-by: Ganesan Ramalingam <grama@microsoft.com>
Introduce layer-norm fusion rules, along with a couple of test cases.
This is just the first version. TO DO: