Skip to content

Conversation

@anmyachev
Copy link
Contributor

@anmyachev anmyachev commented Jun 11, 2025

I noticed because (although I'm not sure if that's root cause):

# works
python benchmarks/dynamo/torchbench.py --accuracy --float32 -d xpu -n10 --inference --only detectron2_maskrcnn_r_50_c4  --backend=inductor --cold-start-latency

# doesn't work
# run it from .scripts_cache/pytorch
../../scripts/inductor_xpu_test.sh torchbench float32 inference accuracy xpu 0 static 1 0 detectron2_maskrcnn_r_50_c4

torch-xpu-ops sources in main: https://github.com/intel/torch-xpu-ops/blob/5b4d7444484576f721d2295761cf8fafa924ef36/.github/scripts/inductor_xpu_test.sh#L30

in custom branch (on which they made a reproducer for us): https://github.com/intel/torch-xpu-ops/blob/6e05eb3f3cb84bceb818a9f0461da773c56ac3ef/.github/scripts/inductor_xpu_test.sh#L31

Copy link
Contributor

@pbchekin pbchekin left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

There is mode "inference-no-freezing" in the script. You can set this mode via workflow input

@anmyachev
Copy link
Contributor Author

There is mode "inference-no-freezing" in the script. You can set this mode via workflow input

yes, it could, but it's not so obvious. For those who don't know (like me at the beginning :) ), it will seem like a separate special mod. And since the regression issue indicates a different mode, it's quite hard to think of using this one. Can we remove it as well?

@anmyachev
Copy link
Contributor Author

anmyachev commented Jun 11, 2025

@pbchekin
Copy link
Contributor

yes, it could, but it's not so obvious. For those who don't know (like me at the beginning :) ), it will seem like a separate special mod. And since the regression issue indicates a different mode, it's quite hard to think of using this one. Can we remove it as well?

I am not against it) @vlad-penkin suggested some time ago to keep both modes (inference with freezing, inference without freezing). Potentially we can make the latter default?

@anmyachev anmyachev force-pushed the anmyachev-patch-3 branch from 43fb9ae to 184a0a4 Compare June 11, 2025 15:12
@anmyachev
Copy link
Contributor Author

yes, it could, but it's not so obvious. For those who don't know (like me at the beginning :) ), it will seem like a separate special mod. And since the regression issue indicates a different mode, it's quite hard to think of using this one. Can we remove it as well?

I am not against it) @vlad-penkin suggested some time ago to keep both modes (inference with freezing, inference without freezing). Potentially we can make the latter default?

Added inference-with-freezing. Is it better now?

Copy link
Contributor

@pbchekin pbchekin left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Yep, works for me. Thanks!

@anmyachev anmyachev changed the title Update inductor_xpu_test.sh: don't add --freezing in reference mode Update inductor_xpu_test.sh: don't add --freezing in reference mode; add new mode: inference-with-freezing Jun 11, 2025
@anmyachev anmyachev merged commit 696dadf into main Jun 11, 2025
15 checks passed
@anmyachev anmyachev deleted the anmyachev-patch-3 branch June 11, 2025 17:50
david-hls pushed a commit to david-hls/intel-xpu-backend-for-triton that referenced this pull request Jun 18, 2025
There doesn't seem to be any test coverage for this.

Co-authored-by: Thomas Raoux <[email protected]>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

4 participants