Skip to content

Conversation

@qjia7
Copy link
Contributor

@qjia7 qjia7 commented Mar 28, 2025

This PR uses 1d disptach group size and uses workgroup_idx instead of workgroup.x|workgroup.y in case they are normalized.

@guschmue guschmue added the ep:WebGPU ort-web webgpu provider label Mar 28, 2025
@snnn snnn closed this Apr 3, 2025
@snnn snnn reopened this Apr 3, 2025
fs-eire
fs-eire previously approved these changes Apr 4, 2025
@fs-eire fs-eire merged commit a1186f6 into main Apr 7, 2025
87 of 89 checks passed
@fs-eire fs-eire deleted the attention_1d_dispatch_groups branch April 7, 2025 13:36
zhaoxul-qti pushed a commit to CodeLinaro/onnxruntime that referenced this pull request Apr 17, 2025
This PR uses 1d disptach group size and uses workgroup_idx instead of
workgroup.x|workgroup.y in case they are normalized.
qjia7 added a commit that referenced this pull request Apr 25, 2025
Fixed the bug in #24228 which causes the incorrect result for phi models
when flash attention is disabled.
vraspar pushed a commit that referenced this pull request Apr 28, 2025
Fixed the bug in #24228 which causes the incorrect result for phi models
when flash attention is disabled.
ankitm3k pushed a commit to intel/onnxruntime that referenced this pull request May 12, 2025
Fixed the bug in microsoft#24228 which causes the incorrect result for phi models
when flash attention is disabled.
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

ep:WebGPU ort-web webgpu provider

Projects

None yet

Development

Successfully merging this pull request may close these issues.

5 participants