-
Notifications
You must be signed in to change notification settings - Fork 3.8k
[Relay][Topi] Hook up LayerNorm to new Topi #12935
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Relay][Topi] Hook up LayerNorm to new Topi #12935
Conversation
vinx13
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I'm wondering removing simplification for later norm affects non-metaschedule workflow. We didn't implement topi schedule for layer norm, I'm not sure if the default schedule_injective works well.
Btw there are also some unrelated (.csv, .md) files in this PR added by mistake.
|
I have removed the unneeded files. Good point about non-metaschedule workflows. I will see what I can do here. |
|
Any update? Also interested in using fused layer norm. |
|
Ah yes, I will try to get to this during the next week where I will have time.
|
332a388 to
21b8e2a
Compare
This makes layer_norm relay op dispatch to new topi committed in PR #12864.
It removes LayerNorm from simplify inference pass now that we actually have compute.
Using TIR CSE elimination pass with FP16 layernorm also necessitates the handling of FP16 type when packing args for CUDA.