You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
drisspg
changed the title
[MX] Create MX linear using new scaled_dot op in Triton
[MX | Triton] Create MX matmul op using new scaled_dot op in Triton
Oct 15, 2024
For others interested in the issue, while this feature has been merged to triton/main this does not yet seem to be released in triton version 2.1.0 (release 14th October, https://pypi.org/project/triton/3.1.0/#history).
Summary
Recently triton added the
scaled_dot
which consumes A, B in f8, f6, f4 packed in int32 format and u8m0 scales via int8 datatype. https://github.com/triton-lang/triton/pull/4795/files#diff-1d96a0ed473569188c00d6e16c54dd7050e0a66040438ac630c889aef7cbbbe8R1544Steps
upcast and matmul
approachao/torchao/prototype/mx_formats/mx_ops.py
Lines 64 to 68 in 48bc81c
The text was updated successfully, but these errors were encountered: