python3Packages.vllm: 0.3.3->0.5.2#328258
Conversation
It was failing at runtime without new triton for me, I still maintain fork of nixpkgs with updated triton because of that. |
pkgs/development/python-modules/prometheus-fastapi-instrumentator/default.nix
Outdated
Show resolved
Hide resolved
b69e62d to
e30f59e
Compare
|
I think I've addressed everything, it should be good to go @samuela @SomeoneSerge @CertainLach. Let me know if you want me to squash the commits. |
|
Result of 4 packages marked as broken and skipped:
8 packages built:
|
|
thanks @SomeoneSerge, any idea why the vllm build is marked broken? Your config should have had cuda enabled |
b6bcace to
51c2fe6
Compare
broken on master (Ofborg rebases PRs on master before running any checks) |
|
outlines was broken in ff7d46d by an untested version bump, and marked as broken a few days ago when someone noticed. I fixed and tacked it on to this PR. |
pkgs/development/python-modules/prometheus-fastapi-instrumentator/default.nix
Outdated
Show resolved
Hide resolved
|
A new version of vllm has been released, seems like a small one so hopefully no need to modify the build other than bump version and hash. Trying now. |
|
Result of 8 packages marked as broken and skipped:
12 packages built:
|
|
sorry, removing the broken annotation on outlines (I'm working across two branches [a semi-stable one that builds and works for me and includes a vllm patch for MIG support, and this one] and I'm evidently failing to merge/cherrypick properly, apologies!) |
No worries. It's just Also you might want to look into lazygit |
|
Thanks @SomeoneSerge. Turns out this was not the issue I thought, this PR does have the broken annotation removed, not sure how it got reintroduced, checking now. |
0823906 to
949004f
Compare
|
will squash shortly, thanks @bcdarwin |
8bccdc3 to
a694582
Compare
There was a problem hiding this comment.
Side note: torch in nativeBuildInputs looks wrong
There was a problem hiding this comment.
yeah, not sure why that was put there.
There was a problem hiding this comment.
Maybe one of those 'let the user bring their own torch' ideas a la the current torchmetrics?
There was a problem hiding this comment.
Maybe one of those 'let the user bring their own torch' ideas a la the current torchmetrics?
I rather meant that vllm likely needs the host platform's torch, not the build platform's
There was a problem hiding this comment.
ok - I think this torch can be removed, torch is already in propagatedBuildInputs, shouldn't be needed twice.
a2e5cc8 to
067bbf0
Compare
|
Result of 2 packages failed to build:
14 packages built:
|
|
@SomeoneSerge do you have the build logs for python312Packages.vllm? |
|
I had to re-run nixpkgs-review because logs got GC-ed, but here: https://gist.github.com/SomeoneSerge/6a3344368c146171160a92d6865ffd57 TLDR: upstream marked vllm as only supporting up to 3.11; I highly doubt that was intentional though |
|
I'm tempted to patch this out and either replace the list with the current interpreter's version or replace the whole thing with an unwrapped |
|
So uh I'm not a vllm user, you probably know better: is there value in providing |
|
I'm using python311 so not having 312 doesn't bother me personally, but it looks like there isn't a good reason it's disabled. |
1d87675 to
1748d1e
Compare
vllm remove comment, minor version bump use refs/tags and format=pyproject use @Args and realphabetize inputs move stdenv, mv cuda_home, add/clean cuda deps, restore comment nixfmt rfc style remove stale comment
pep-517 attr names, no native torch, phase order Co-authored-by: cfhammill <christopher.hammill@unityhealth.to>
Co-authored-by: cfhammill <christopher.hammill@unityhealth.to>
Co-authored-by: cfhammill <christopher.hammill@unityhealth.to>
Co-authored-by: cfhammill <christopher.hammill@unityhealth.to>
Co-authored-by: cfhammill <christopher.hammill@unityhealth.to>
b8084fc to
17a50f1
Compare
|
@cfhammill I just added the stuff I said was optional, idk why; the python 3.12 build... doesn't immediately fail (it's been running for 40' now) so I intend to leave it at that; I'll check on Ofborg tomorrow and probably just merge, but please do run some checks and skim over the extra stuff I pushed that it makes sense. I don't use vllm so I can't reason much about this |
pkgs/development/python-modules/vllm/0002-setup.py-nix-support-respect-cmakeFlags.patch
Show resolved
Hide resolved
|
I'm happy with this! Thanks @SomeoneSerge! Merge when ready. |
|
Thanks @cfhammill |
Description of changes
Updates vllm. Requires two new python packages, lm-format-enforcer and prometheus-fastapi-instrumentator, this PR adds both.
I was able to get the cuda version to build on a slightly older revision of nixpkgs. Would have needed to rebuild the entire python/pytorch/cuda stack to try against master. Compiling vllm alone takes > 1h so this was slow going. Also have not built against rocm.
@CertainLach you mentioned that this need a special triton? It compiles ok for me with the version I fixed in #325843.
It compiles and can be imported, but since tests have not been enabled (and likely wouldn't work anyway) unsure if this works as expected.
CC @SomeoneSerge @ConnorBaker @samuela
Things done
nix.conf? (See Nix manual)sandbox = relaxedsandbox = truenix-shell -p nixpkgs-review --run "nixpkgs-review rev HEAD". Note: all changes have to be committed, also see nixpkgs-review usage./result/bin/)Add a 👍 reaction to pull requests you find important.