-
Notifications
You must be signed in to change notification settings - Fork 185
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Benchmarking updates for semi-structured sparse training #398
Conversation
Summary: This PR does the following: - adds e2e ViT benchmarks for semi-structured sparse training - adds nn.Linear microbenchmarks - removes extra xformers benchmarking utils I copied over - removes MLP block benchmarks - updated README.md with new benchmarks + accuracy benchmarks Given we have nn.Linear microbenchmarks and e2e benchmarks, I felt that the MLP block benchmarks were unnecessary As a sanity check, I ran the MLP benchmarks with the new benchmarking suite and the old one, and got the same results: Test Plan: Reviewers: Subscribers: Tasks: Tags:
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/ao/398
Note: Links to docs will display an error until the docs builds have been completed. ✅ You can merge normally! (1 Unrelated Failure)As of commit 881ae2c with merge base 6b0ca2d (): BROKEN TRUNK - The following job failed but were present on the merge base:👉 Rebase onto the `viable/strict` branch to avoid these failures
This comment was automatically generated by Dr. CI and updates every 15 minutes. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Cool thank you! This is significantly clearer. I do want us to think a bit harder about the top line metric since 6% might not be super compelling to people not familiar with limitations of sparsity
@msaroufim We could compare to masking based approaches (which are slower than dense training) for a larger number, but I think it'd be a bit confusing since I'm assuming most users are coming with a dense model and not an existing sparse training script they want to accelerate. |
* Benchmarking updates for semi-structured sparse training Summary: This PR does the following: - adds e2e ViT benchmarks for semi-structured sparse training - adds nn.Linear microbenchmarks - removes extra xformers benchmarking utils I copied over - removes MLP block benchmarks - updated README.md with new benchmarks + accuracy benchmarks Given we have nn.Linear microbenchmarks and e2e benchmarks, I felt that the MLP block benchmarks were unnecessary As a sanity check, I ran the MLP benchmarks with the new benchmarking suite and the old one, and got the same results: Test Plan: Reviewers: Subscribers: Tasks: Tags: * update * add units
* cli * typos
Summary:
This PR does the following:
Given we have nn.Linear microbenchmarks and e2e benchmarks, I felt that
the MLP block benchmarks were unnecessary
As a sanity check, I ran the MLP benchmarks with the new benchmarking
suite and the old one, and got the same results:
NEW:
OLD:
Test Plan:
Reviewers:
Subscribers:
Tasks:
Tags: