-
Notifications
You must be signed in to change notification settings - Fork 316
Add support for fbgemm int4 mm kernel #2255
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/ao/2255
Note: Links to docs will display an error until the docs builds have been completed. ✅ No FailuresAs of commit d2066dc with merge base b0cfeec ( This comment was automatically generated by Dr. CI and updates every 15 minutes. |
Thank you! community really needs this. |
9df9b49
to
3253e6a
Compare
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Okay everything looks pretty good but the API for the FBGEMM config feels gross imo I know its a thin wrapper around their op but I think we can do better than io string
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Looks good, can you also add a serialization test entry, want to ensure we can seralize str eums
Summary: we also plan to expose some other kernels like fp8xint4 and bf16xfp8, fp8xfp8 to compare with existing torchao kernels Test Plan: test/dtypes/test_fbgemm_int4_tensor.py Reviewers: Subscribers: Tasks: Tags:
1d8b558
to
d9fdf72
Compare
Summary:
we also plan to expose some other kernels like fp8xint4 and bf16xfp8, fp8xfp8 to compare with existing torchao kernels
Test Plan:
test/dtypes/test_fbgemm_int4_tensor.py
H100, with compile:
Note: fbgemm-int4-128 does not work with compile yet since the fbgemm op does not have meta device implementation.
Reviewers:
Subscribers:
Tasks:
Tags: