-
Notifications
You must be signed in to change notification settings - Fork 23
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
tuned moe configs v2 #33
Conversation
Update the fused moe config.json files. These config files utilizes all the available Triton kernel parameters for tuning.
As per our discussion today, update this PR with the following:
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I will leave the final confirmation for Hari - but we are good to go as far as the state of our master branch is concerned.
d3eaa94
to
af93dba
Compare
af93dba
to
eb80843
Compare
Couple items we discussed to complete this PR:
|
463f981
to
fd04661
Compare
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks for the measurements!
Fixed. Reverted the revert. Added separate commit for init files fix. |
Update the fused moe config.json files. These config files utilizes all the available Triton kernel parameters for tuning. Used by both Prefill and Decode fused_moe kernels.