-
Notifications
You must be signed in to change notification settings - Fork 1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[GPU] Fp8 compute backports #2266
base: main
Are you sure you want to change the base?
Conversation
@@ -195,7 +207,13 @@ struct matmul_pd_t : public primitive_desc_t { | |||
sc.group_dims_[0] == 1 | |||
&& K() % sc.group_dims_[1] == 0); | |||
} else { | |||
ok = ok && (mask == 0); | |||
ok = ok |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Here must be a check for fp8 versus classic quantization.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I dont think it is required. With these changes gemm supports dst scales whenever dst is an integer or fp8 so there should be no difference in treatment. Attempts to use dst scales with non-fp8 float dst will be filtered in attr checks.
8aa5c64
to
accacad
Compare
accacad
to
606af26
Compare
make test |
Description
Backport of mixed fp8 support, additional scale support for compute primitivies.
Checklist
General
make test
andmake test_benchdnn_*
) pass locally for each commit?