-
-
Notifications
You must be signed in to change notification settings - Fork 6k
recipe for vllm #24710
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Closed
Closed
recipe for vllm #24710
Changes from all commits
Commits
Show all changes
70 commits
Select commit
Hold shift + click to select a range
67a8dc7
recipe for vllm
mediocretech a3e6578
fixed ray conda package name ref
mediocretech de9e058
moved to ray-core
mediocretech ba1d0dd
Update build_steps.sh
mediocretech f62f10c
GPU requirement
mediocretech 68f4f5f
Build compiler deps
mediocretech d27b5eb
setting ninja max jobs in attempt to stop build cycling
mediocretech 4f19559
added overlink requirements
mediocretech 911519a
libtorch_python overlink
mediocretech b932232
latest source version
mediocretech e68a28c
sha mismatch
mediocretech 8c0a076
build test
mediocretech b5ae846
lint
mediocretech 2aeb452
debugging; trying remove build for yaml formating issue
mediocretech 942f98f
simplified build
mediocretech acaf85c
Update build_steps.sh
mediocretech a2d33d7
lowered build version due to pytorch version requirement not in conda…
mediocretech f9d9e2d
removed pip check due to conda-forge version differences
mediocretech 000f766
returned pip check, runaway build
mediocretech c2d4774
Merge branch 'main' into vllm
mediocretech bd0f153
No test for pip check
mediocretech 4307b81
Merge branch 'main' into vllm
mediocretech 2b0bce6
updated dependancy list
mediocretech 6e10d68
the more, the merrier
mediocretech bfba786
Merge branch 'main' into vllm
mediocretech a5ca474
no noarch
mediocretech 04064e1
Merge branch 'vllm' of github.com:mediocretech/staged-recipes into vllm
mediocretech dc76ebb
torch is needed for build
mediocretech 33b4b2f
add vllm
iamthebot d5ec5f2
Merge branch 'vllm' into al--vllm
iamthebot 35b4ff9
Merge branch 'main' into vllm
mediocretech 388a4fa
Merge branch 'vllm' into al--vllm
mediocretech ef0472c
Merge pull request #1 from iamthebot/al--vllm
mediocretech 43e2dbc
whitelist libcuda
mediocretech 4f5d04b
Merge branch 'main' into vllm
mediocretech a6b1e5f
force build
mediocretech 593f989
Merge branch 'main' into vllm
mediocretech 42b1ba3
Merge branch 'main' into vllm
mediocretech 671910b
Merge branch 'main' into vllm
mediocretech ae258f4
worth a shot
mediocretech 7a6e5e5
Merge branch 'main' into vllm
mediocretech 6e50aa8
Merge branch 'main' into vllm
mediocretech 6435f40
Merge branch 'main' into vllm
mediocretech 0a9a043
simplified build deps
mediocretech c41c3bf
libcusparse
mediocretech 2742768
removed libcuseparse from build
mediocretech 42350b4
libcusparse-dev
mediocretech 27ea238
versioning for cusparse
mediocretech 49a3b5e
touchvision dep
mediocretech b122ae9
version bump
mediocretech 2101b09
updated sha for newer version
mediocretech eae0ea1
triton runtime req
mediocretech 6b9c7fc
Merge branch 'main' into vllm
mediocretech 8738492
latest conda-build is incompatible
mediocretech 3a0dcdc
update to latest release
mediocretech 2a25f9b
unpinned dependancies for latest vllm release
mediocretech a5a87d9
corrected sha for 0.5.3
mediocretech bf73d7e
Update requirements.txt
mediocretech 7281d5f
Merge branch 'main' into vllm
mediocretech 89e18d0
hope for the future...0.6.2
mediocretech 2a8254d
hash for 0.6.2
mediocretech c5b616c
Merge branch 'main' into vllm
mediocretech ab8ff82
setuptools_scm
mediocretech c843952
changes to build
mediocretech a4f65c0
0.6.4 bump
mediocretech 72aa925
hash update
mediocretech d8ab547
build is now 3.12
mediocretech 328c5bd
Merge branch 'main' into vllm
mediocretech aabea76
latest version
mediocretech 36650d7
hash update
mediocretech File filter
Filter by extension
Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
There are no files selected for viewing
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
| Original file line number | Diff line number | Diff line change |
|---|---|---|
| @@ -0,0 +1,8 @@ | ||
| #!/bin/bash -e | ||
| export MAX_JOBS=1 | ||
| # needed to find libcrypt headers | ||
| export CPATH=${BUILD_PREFIX}/include | ||
| export VLLM_TARGET_DEVICE="cuda" | ||
|
|
||
| ${PYTHON} use_existing_torch.py | ||
| ${PYTHON} -m pip install . --no-deps -vv --no-deps --no-build-isolation |
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
| Original file line number | Diff line number | Diff line change |
|---|---|---|
| @@ -0,0 +1,99 @@ | ||
| {% set name = "vllm" %} | ||
| {% set version = "0.8.1" %} | ||
|
|
||
| package: | ||
| name: {{ name|lower }} | ||
| version: {{ version }} | ||
|
|
||
| source: | ||
| url: https://pypi.io/packages/source/{{ name[0] }}/{{ name }}/vllm-{{ version }}.tar.gz | ||
| sha256: 03d641f10d0181b630243c096484da0b5fdf608befe026af831ab87a51c09ab1 | ||
|
|
||
| build: | ||
| # noarch: python | ||
|
|
||
| rpaths: | ||
| - lib/ | ||
|
|
||
| script_env: | ||
| - TORCH_CUDA_ARCH_LIST=3.5;5.0;6.0;6.1;7.0;7.5;8.0;8.6+PTX # [cuda_compiler_version == "11.2"] | ||
| - TORCH_CUDA_ARCH_LIST=3.5;5.0;6.0;6.1;7.0;7.5;8.0;8.6;8.9+PTX # [cuda_compiler_version == "11.8"] | ||
| - TORCH_CUDA_ARCH_LIST=5.0;6.0;6.1;7.0;7.5;8.0;8.6;8.9;9.0+PTX # [(cuda_compiler_version or "").startswith("12")] | ||
|
|
||
| number: 0 | ||
| skip: True # [cuda_compiler_version == "None"] | ||
| skip: True # [osx or win] | ||
| skip: True # [py < 312] | ||
| missing_dso_whitelist: | ||
| - '*/libtorch_python.so' | ||
| - '*/libcuda.so.1' | ||
|
|
||
| requirements: | ||
| build: | ||
| - cmake | ||
| - make | ||
| - {{ stdlib("c") }} | ||
| - {{ compiler('c') }} | ||
| - {{ compiler('cxx') }} | ||
| - {{ compiler('cuda') }} | ||
|
|
||
| host: | ||
| - cuda-version {{ cuda_compiler_version }} | ||
| - python | ||
|
mediocretech marked this conversation as resolved.
|
||
| - ninja | ||
| - packaging | ||
| - setuptools | ||
| - setuptools-scm | ||
| - pytorch | ||
| - torchvision | ||
| - wheel | ||
| - pip | ||
| - libxcrypt | ||
|
|
||
| run: | ||
| - python | ||
| - ninja | ||
| - psutil | ||
| # - ray-core >=2.9.0 | ||
| - pandas | ||
| - pyarrow | ||
| - sentencepiece | ||
| - numpy | ||
| - einops | ||
| - pytorch | ||
| - transformers >=4.34.0 | ||
| - xformers | ||
| - fastapi | ||
| - uvicorn | ||
| - pydantic | ||
| - aioprometheus | ||
| - prometheus_client | ||
| - triton | ||
| # - libtorch | ||
| # - cuda-version {{ cuda_compiler_version }} # [cuda_compiler_version != "None"] | ||
| # - cuda-cupti-dev # [(cuda_compiler_version or "").startswith("12")] | ||
| # - cuda-cudart-dev # [(cuda_compiler_version or "").startswith("12")] | ||
| # - cuda-nvml-dev # [(cuda_compiler_version or "").startswith("12")] | ||
| # - cuda-nvtx-dev # [(cuda_compiler_version or "").startswith("12")] | ||
| # - libstdcxx-ng | ||
| # - libgcc-ng | ||
| # - libxcrypt | ||
|
|
||
| test: | ||
| imports: | ||
| - vllm | ||
| # commands: | ||
| # - pip check | ||
| requires: | ||
| - pip | ||
|
|
||
| about: | ||
| home: https://github.com/vllm-project/vllm | ||
| summary: A high-throughput and memory-efficient inference and serving engine for LLMs | ||
| license: Apache-2.0 | ||
| license_file: LICENSE | ||
|
|
||
| extra: | ||
| recipe-maintainers: | ||
| - mediocretech | ||
| - iamthebot | ||
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Uh oh!
There was an error while loading. Please reload this page.