Skip to content

[CUDA] Run FlashAttention regression test only when FlashAttention is available#27206

Merged
hariharans29 merged 3 commits intomainfrom
hari/fix_GQA_build_errors
Feb 4, 2026
Merged

[CUDA] Run FlashAttention regression test only when FlashAttention is available#27206
hariharans29 merged 3 commits intomainfrom
hari/fix_GQA_build_errors

Conversation

@hariharans29
Copy link
Copy Markdown
Member

@hariharans29 hariharans29 commented Jan 29, 2026

Description

As title.

Checking if FlashAttention exists check includes if torch has CUDA support, the system has the right device to run FlashAttention, etc.

Motivation and Context

Fix Windows CUDA CI failures

Copy link
Copy Markdown
Contributor

@github-actions github-actions bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

You can commit the suggested changes from lintrunner.

Co-authored-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Copy link
Copy Markdown
Contributor

Copilot AI left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Pull request overview

This PR adds a conditional skip decorator to the TestGQARegressions test class to prevent FlashAttention regression tests from running when FlashAttention is not available on the system. This helps unblock CI build pipelines that may not have the necessary CUDA support or hardware requirements.

Changes:

  • Added @unittest.skipIf decorator to TestGQARegressions class to skip tests when FlashAttention is unavailable

💡 Add Copilot custom instructions for smarter, more guided reviews. Learn how to get started.

@hariharans29 hariharans29 enabled auto-merge (squash) January 29, 2026 19:15
@hariharans29 hariharans29 disabled auto-merge January 29, 2026 19:22
@hariharans29 hariharans29 enabled auto-merge (squash) January 30, 2026 23:41
@hariharans29 hariharans29 merged commit 260a48c into main Feb 4, 2026
92 of 101 checks passed
@hariharans29 hariharans29 deleted the hari/fix_GQA_build_errors branch February 4, 2026 01:10
tianleiwu pushed a commit that referenced this pull request Feb 4, 2026
… available (#27206)

### Description
As title.

Checking if FlashAttention exists check includes if torch has CUDA
support, the system has the right device to run FlashAttention, etc.

### Motivation and Context
Fix Windows CUDA CI failures

---------

Co-authored-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
tianleiwu added a commit that referenced this pull request Feb 4, 2026
#27021: [Disable matmul 1d
tests on
DML](1afc8bc)


#27206: [[CUDA] Run
FlashAttention regression test only when FlashAttention
is…](4d95d97)

#27120: [POWER : Fix build
failure due to unsupported cpuinfo on
ppc64le](2843ec0)

---------

Co-authored-by: Ti-Tai Wang <titaiwang@microsoft.com>
Co-authored-by: Hariharan Seshadri <shariharan91@gmail.com>
Co-authored-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: BODAPATIMAHESH <148746454+BODAPATIMAHESH@users.noreply.github.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

4 participants