Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
201 commits
Select commit Hold shift + click to select a range
990f522
update to transformers v5
hmellor Dec 12, 2025
42bc6a1
Merge branch 'main' into transformers-v5
hmellor Dec 17, 2025
dd261ff
Merge branch 'main' into transformers-v5
hmellor Dec 30, 2025
048a32c
Merge branch 'main' into transformers-v5
hmellor Jan 27, 2026
933bef9
Allow Transformer v5 in `common.txt`
hmellor Jan 27, 2026
12f6195
Merge branch 'main' into transformers-v5
hmellor Jan 27, 2026
769d436
Update PEFT pin to avoid bad import
hmellor Jan 27, 2026
214c373
Update lm-eval
hmellor Jan 27, 2026
ec4ffa9
`HF_HUB_ENABLE_HF_TRANSFER` -> `HF_XET_HIGH_PERFORMANCE`
hmellor Jan 28, 2026
94e1429
Skip custom model which uses old imports
hmellor Jan 28, 2026
fbb843a
Update some more lm-eval pins
hmellor Jan 29, 2026
954ea46
Merge branch 'main' into transformers-v5
hmellor Jan 29, 2026
352a274
Fix timtout issues from `huggingface-hub` v1
hmellor Jan 29, 2026
7c81a9c
Add `HF_HUB_DOWNLOAD_TIMEOUT` to other test images
hmellor Jan 29, 2026
1aa6c01
Merge branch 'main' into transformers-v5
hmellor Jan 30, 2026
eea0d7c
Update missed ROCM pin
hmellor Jan 30, 2026
30d8b3d
Install transformers from main temporarily
hmellor Jan 30, 2026
1767434
Merge branch 'main' into transformers-v5
hmellor Jan 30, 2026
17ad8ca
new main pin
hmellor Jan 30, 2026
3052b14
Merge branch 'main' into transformers-v5
hmellor Jan 30, 2026
6a8954e
Merge branch 'main' into transformers-v5
hmellor Feb 2, 2026
489d5d9
Add backward compatibility test as copy of nightly test
hmellor Feb 2, 2026
c3abbd7
Skip `MiniCPMV`
hmellor Feb 2, 2026
7b9016f
Merge branch 'main' into transformers-v5
hmellor Feb 3, 2026
97bdae0
bump huggingface-hub
hmellor Feb 3, 2026
ede39e6
Bump accelerate version
hmellor Feb 3, 2026
21917f4
Merge branch 'main' into transformers-v5
hmellor Feb 3, 2026
113b5ee
bump transformers main pin
hmellor Feb 3, 2026
8563dcc
Merge branch 'main' into transformers-v5
hmellor Feb 4, 2026
9ee40ac
bump main
hmellor Feb 4, 2026
57d0aad
Merge branch 'main' into transformers-v5
hmellor Feb 6, 2026
84447bd
bump main
hmellor Feb 6, 2026
ccc8b3e
Skip experimental Transformers backend features, fix later
hmellor Feb 6, 2026
ee4c25c
bump hf hub
hmellor Feb 6, 2026
d7dd270
bumpm hf experimental version
hmellor Feb 6, 2026
4da0a83
OpenCUA should be fixed now
hmellor Feb 6, 2026
873adc7
Merge branch 'main' into transformers-v5
hmellor Feb 6, 2026
6074afe
Merge branch 'main' into transformers-v5
hmellor Feb 9, 2026
fdacb70
Merge branch 'main' into transformers-v5
hmellor Feb 10, 2026
f7ac9c2
bump treansformers main
hmellor Feb 10, 2026
a6544bf
Merge branch 'main' into transformers-v5
hmellor Feb 10, 2026
093999b
bump transformers main
hmellor Feb 10, 2026
06a569f
Skip Molmo2
hmellor Feb 10, 2026
af97153
Skip openpangu
hmellor Feb 11, 2026
7682651
Merge branch 'main' into transformers-v5
hmellor Feb 11, 2026
c0ac4cd
bump transformers main
hmellor Feb 11, 2026
6e6fa6f
glmasr is no longer remote code in v5
hmellor Feb 11, 2026
ced047c
skip OpenCUA
hmellor Feb 11, 2026
148c40e
Skip HCXVisionForCausalLM
hmellor Feb 11, 2026
3daa6d8
Merge branch 'main' into transformers-v5
hmellor Feb 11, 2026
c46b56d
bump transformers main
hmellor Feb 11, 2026
f0f00aa
bump transformers main
hmellor Feb 13, 2026
12810a0
Merge branch 'main' into transformers-v5
hmellor Feb 13, 2026
afdf43f
Merge branch 'main' into transformers-v5
hmellor Feb 16, 2026
2e9fb90
Merge branch 'main' into transformers-v5
hmellor Feb 16, 2026
37c707d
Skip broken custom models for processor tests
hmellor Feb 16, 2026
567e00f
bump transformers main
hmellor Feb 16, 2026
c0f2e1b
Leave these version limits alone
hmellor Feb 16, 2026
a7d580a
Merge branch 'main' into transformers-v5
hmellor Feb 17, 2026
4159b7f
bump main
hmellor Feb 17, 2026
311ea04
Merge branch 'main' into transformers-v5
hmellor Feb 19, 2026
a1fb41b
bump transformers main
hmellor Feb 19, 2026
1f7f5fa
Merge branch 'main' into transformers-v5
hmellor Feb 20, 2026
b0d99c9
Fix Flamingo min versions
hmellor Feb 20, 2026
5f1d9f9
Fix Qwen3.5 min version and availability of checkpoints
hmellor Feb 20, 2026
a2fc272
Skip Plamo2 for HF (vLLM should still run ok)
hmellor Feb 20, 2026
6b563d4
Leave tensor schema skip alone and add another for hf reasons
hmellor Feb 20, 2026
fe32e65
Merge branch 'main' into transformers-v5
hmellor Feb 20, 2026
64fa2e2
Remove hf skip for tensor schema test
hmellor Feb 20, 2026
773ad0e
`MiniCPMV` version reason should stop it working in vLLM, not just HF
hmellor Feb 20, 2026
11672ab
Merge branch 'main' into transformers-v5
hmellor Feb 21, 2026
445c7fe
bump main
hmellor Feb 21, 2026
bf217bc
Merge branch 'main' into transformers-v5
hmellor Feb 23, 2026
cfaa2ed
Unskip models which should now work
hmellor Feb 23, 2026
04692c2
Ovis doesn't work in vLLM actually
hmellor Feb 23, 2026
743ff55
Merge branch 'main' into transformers-v5
hmellor Feb 23, 2026
3045933
Merge branch 'main' into transformers-v5
hmellor Feb 24, 2026
f7c7f5e
bump main
hmellor Feb 24, 2026
d99f3b5
Skip InternS1 properly
hmellor Feb 24, 2026
a7f676c
InternS1Pro can work
hmellor Feb 24, 2026
44b7504
Update OpenCUA skip
hmellor Feb 24, 2026
a6d4100
Update OpenPanguVL skip
hmellor Feb 24, 2026
6f6ee9e
Skip `ExaoneMoeMTP` because it's not compatible with the test harness...
hmellor Feb 24, 2026
27c7f10
Merge branch 'main' into transformers-v5
hmellor Feb 26, 2026
05cb1bc
Merge branch 'main' into transformers-v5
hmellor Feb 27, 2026
d35c05d
bump main
hmellor Feb 27, 2026
b3f8f01
Merge branch 'main' into transformers-v5
hmellor Mar 3, 2026
b0d6bb3
bump main
hmellor Mar 3, 2026
3176096
Merge branch 'main' into transformers-v5
hmellor Mar 4, 2026
65021f7
Merge branch 'main' of https://github.com/vllm-project/vllm into tran…
hmellor Mar 6, 2026
bd8cc8b
bump transformers
hmellor Mar 6, 2026
d19703c
Merge branch 'main' into transformers-v5
hmellor Mar 9, 2026
db2c800
bump transformers
hmellor Mar 9, 2026
2a36e1d
Merge branch 'main' into transformers-v5
hmellor Mar 10, 2026
91f54ac
bump transformers
hmellor Mar 10, 2026
121b681
Put ExaoneMoe back, we'll fix it another way
hmellor Mar 10, 2026
407fde0
Merge branch 'main' into transformers-v5
hmellor Mar 11, 2026
489aeda
bump transformers
hmellor Mar 11, 2026
aad386e
Merge branch 'main' into transformers-v5
hmellor Mar 13, 2026
fdbf94d
Merge branch 'main' into transformers-v5
hmellor Mar 13, 2026
4c138ee
bump transformers
hmellor Mar 13, 2026
1216a62
Merge branch 'main' into transformers-v5
hmellor Mar 14, 2026
b99bedc
bump transformers
hmellor Mar 14, 2026
07e1b40
Merge branch 'main' into transformers-v5
hmellor Mar 18, 2026
0c515b0
Bump main
hmellor Mar 19, 2026
b03a868
Merge branch 'main' into transformers-v5
hmellor Mar 19, 2026
1786f7f
bump main
hmellor Mar 19, 2026
698f4b1
Merge branch 'main' into transformers-v5
hmellor Mar 20, 2026
4da6603
bump main
hmellor Mar 20, 2026
f4b5c7d
Merge branch 'main' into transformers-v5
hmellor Mar 23, 2026
5b28c7d
Merge branch 'main' into transformers-v5
hmellor Mar 24, 2026
36460f8
bump main
hmellor Mar 24, 2026
3a2b517
bump main
hmellor Mar 25, 2026
1b81af3
Merge branch 'main' into transformers-v5
hmellor Mar 25, 2026
d978d72
Merge branch 'main' of https://github.com/vllm-project/vllm into tran…
hmellor Mar 26, 2026
b0fb9ec
bump main
hmellor Mar 26, 2026
43bbda5
bump main
hmellor Mar 27, 2026
e9e48f0
Merge branch 'main' of https://github.com/vllm-project/vllm into tran…
hmellor Mar 27, 2026
7b05415
bump main
hmellor Mar 27, 2026
740533f
skip broken models in VLM tests
hmellor Mar 27, 2026
5894f1d
More models not compatible with v5
hmellor Mar 27, 2026
3c393aa
Merge branch 'main' of https://github.com/vllm-project/vllm into tran…
hmellor Mar 30, 2026
3d46d90
bump main
hmellor Mar 30, 2026
49b5b2f
Merge branch 'main' into transformers-v5
hmellor Apr 1, 2026
ead016b
Merge branch 'main' of https://github.com/vllm-project/vllm into tran…
hmellor Apr 2, 2026
18dd0bd
Try try timeout fix
hmellor Apr 2, 2026
5c6f97a
bump main
hmellor Apr 2, 2026
b99d67d
Revert "Try try timeout fix"
hmellor Apr 3, 2026
19dd32d
Explicitly call `huggingface_hub.close_session` on shutdown
hmellor Apr 3, 2026
4dc0c85
Move close_session earlier
hmellor Apr 3, 2026
9887e8e
Merge branch 'main' of https://github.com/vllm-project/vllm into tran…
hmellor Apr 5, 2026
552e9e2
bump main
hmellor Apr 5, 2026
c48b942
Merge branch 'main' of https://github.com/vllm-project/vllm into tran…
hmellor Apr 7, 2026
77ca5a9
bump main
hmellor Apr 7, 2026
202134f
Merge branch 'main' of https://github.com/vllm-project/vllm into tran…
hmellor Apr 9, 2026
f9d42e1
pin to 5.5.1 and 0.15.0
hmellor Apr 9, 2026
59cfed0
Merge branch 'main' into transformers-v5
hmellor Apr 10, 2026
8877940
bump compressed tensors
hmellor Apr 10, 2026
cd78122
remove `--pre` from dockerfile installs
hmellor Apr 10, 2026
e9b8698
Revert change to rocm-test-in
hmellor Apr 10, 2026
7482524
pip-compile
hmellor Apr 10, 2026
139a83f
update all hf libs for best hub support
hmellor Apr 10, 2026
2da5970
Revert timeout change that didn't work
hmellor Apr 10, 2026
093aca6
test push
khluu Apr 10, 2026
a39173e
Merge branch 'main' into transformers-v5
hmellor Apr 10, 2026
6c8d30e
upgrade to transformers 5.5.3
khluu Apr 10, 2026
a6f6084
skip phi4 test
khluu Apr 10, 2026
b8463a2
skip sarvam
khluu Apr 10, 2026
eaa1e54
gemma4 fix
khluu Apr 10, 2026
1545c11
skip tarsier2
khluu Apr 10, 2026
5c3f5a5
skip minicpmv
khluu Apr 10, 2026
c6a4292
fix step3p5
khluu Apr 11, 2026
eb0479b
fix gguf loader
khluu Apr 11, 2026
24f77bf
fix music flamingo
khluu Apr 11, 2026
6ff178d
set shutdown timeout to 150s
khluu Apr 11, 2026
b5c68a3
lint
khluu Apr 11, 2026
5e8afc0
Merge branch 'main' into transformers-v5
khluu Apr 11, 2026
d9e6625
step3p5 fix
khluu Apr 11, 2026
86bc3f8
skip mteb tests
khluu Apr 11, 2026
1b51036
fix TransformersMultiModalMoEForCausalLM
khluu Apr 11, 2026
dae2db3
skip paddleocr, nemotron, voxtral
khluu Apr 11, 2026
4ce8ba8
fix gemma4 duplicate arg limit_mm_per_prompt
khluu Apr 11, 2026
407fc73
gemma4 video placement fix
khluu Apr 11, 2026
e648579
fix gemma4
khluu Apr 11, 2026
09f7c26
Update vllm/model_executor/models/gemma4_mm.py
khluu Apr 11, 2026
ffe85f5
fix ext pooling mm test
khluu Apr 11, 2026
814e130
Update vllm/model_executor/models/gemma4_mm.py
DarkLight1337 Apr 11, 2026
2393c1e
revert timeout change as it didn't fix the issue
hmellor Apr 13, 2026
16eb5f1
simpler test skip
hmellor Apr 13, 2026
bce473a
fix pre-commit
hmellor Apr 13, 2026
1b0635d
fix hf runner using vllm configs
hmellor Apr 13, 2026
57e7949
skip other phi4 tests
hmellor Apr 13, 2026
cfe4e32
skip failing ultravox test
hmellor Apr 13, 2026
9eb2d21
skip transformers backend eagle3 test because it's not urgent
hmellor Apr 13, 2026
b79aec0
Merge branch 'main' into transformers-v5
khluu Apr 13, 2026
3bedcc2
fix gemma4 image placeholder
khluu Apr 13, 2026
d42fd89
gemma4 tensor shape fix
khluu Apr 13, 2026
48a3199
skip fireredasr2 asr-nano-2512 fireredlid
khluu Apr 13, 2026
4c6cac1
use full gpu for basic models init test
khluu Apr 13, 2026
2d7903e
register custom config
khluu Apr 13, 2026
875c012
gc collect llm delete test
khluu Apr 14, 2026
6776f31
add todo comment
hmellor Apr 14, 2026
8f551d0
alternative fix for vllm config in get_tokenizer
hmellor Apr 14, 2026
e67530c
revert step3p5 test changes now that get_tokenizer is fixed
hmellor Apr 14, 2026
87f3a14
Bump `huggingface-hub` and remove delete workaround
hmellor Apr 14, 2026
40742ca
temp fix for tinymixtral test
khluu Apr 15, 2026
ea58ae3
Revert "temp fix for tinymixtral test"
khluu Apr 15, 2026
3693a95
fix tiny-mixtral CPU test: reduce bfloat16 rounding error
khluu Apr 15, 2026
f50bb9d
add back firered and funasr model back to registry
khluu Apr 15, 2026
6d40ca7
skip XverseForCausalLM tests on transformers v5
khluu Apr 15, 2026
e187e72
claude fix pretokenizer for step3p5 and tool parser
khluu Apr 15, 2026
cb03f5d
fix LoRA dual-stream defs guarded by import-time env check
khluu Apr 15, 2026
cc19a1b
fix get_tokenizer crash when tokenizer path has no model config
khluu Apr 15, 2026
d894c4b
Revert "claude fix pretokenizer for step3p5 and tool parser"
hmellor Apr 15, 2026
816db8b
better fix for bad tokenizer_class config
hmellor Apr 15, 2026
410ae69
Revert "fix LoRA dual-stream defs guarded by import-time env check"
hmellor Apr 15, 2026
962976d
test side fix for lora dual stream
hmellor Apr 15, 2026
2cba808
Merge branch 'main' into transformers-v5
hmellor Apr 15, 2026
f48f8ce
skip FireRedASR2, FireRedLID, FunASR tests on transformers >= 5.2
khluu Apr 15, 2026
79e9772
Merge branch 'main' into transformers-v5
khluu Apr 15, 2026
75efe07
fix FireRedASR2/FireRedLID/FunASR skip reason: hf -> vllm
khluu Apr 15, 2026
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion .buildkite/scripts/hardware_ci/run-cpu-test.sh
Original file line number Diff line number Diff line change
Expand Up @@ -16,5 +16,5 @@ echo "--- :docker: Building Docker image"
docker build --progress plain --tag "$IMAGE_NAME" --target vllm-test -f docker/Dockerfile.cpu .

# Run the image, setting --shm-size=4g for tensor parallel.
docker run --rm --cpuset-cpus="$CORE_RANGE" --cpuset-mems="$NUMA_NODE" -v ~/.cache/huggingface:/root/.cache/huggingface --privileged=true -e HF_TOKEN -e VLLM_CPU_KVCACHE_SPACE=16 -e VLLM_CPU_CI_ENV=1 -e VLLM_CPU_SIM_MULTI_NUMA=1 --shm-size=4g "$IMAGE_NAME" \
docker run --rm --cpuset-cpus="$CORE_RANGE" --cpuset-mems="$NUMA_NODE" -v ~/.cache/huggingface:/root/.cache/huggingface --privileged=true -e HF_TOKEN -e VLLM_CPU_KVCACHE_SPACE=16 -e VLLM_CPU_CI_ENV=1 -e VLLM_CPU_SIM_MULTI_NUMA=1 -e VLLM_CPU_ATTN_SPLIT_KV=0 --shm-size=4g "$IMAGE_NAME" \
timeout "$TIMEOUT_VAL" bash -c "set -euox pipefail; echo \"--- Print packages\"; pip list; echo \"--- Running tests\"; ${TEST_COMMAND}"
16 changes: 15 additions & 1 deletion .buildkite/test_areas/models_basic.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,6 @@ depends_on:
steps:
- label: Basic Models Tests (Initialization)
timeout_in_minutes: 45
device: h200_18gb
torch_nightly: true
source_file_dependencies:
- vllm/
Expand Down Expand Up @@ -73,3 +72,18 @@ steps:
- python3 examples/offline_inference/vision_language.py --model-type qwen2_5_vl
# Whisper needs spawn method to avoid deadlock
- VLLM_WORKER_MULTIPROC_METHOD=spawn python3 examples/offline_inference/audio_language.py --model-type whisper

- label: Transformers Backward Compatibility Models Test
working_dir: "/vllm-workspace/"
optional: true
soft_fail: true
commands:
- pip install transformers==4.57.5
- pytest -v -s tests/models/test_initialization.py
- pytest -v -s tests/models/test_transformers.py
- pytest -v -s tests/models/multimodal/processing/
- pytest -v -s tests/models/multimodal/test_mapping.py
- python3 examples/offline_inference/basic/chat.py
- python3 examples/offline_inference/vision_language.py --model-type qwen2_5_vl
# Whisper needs spawn method to avoid deadlock
- VLLM_WORKER_MULTIPROC_METHOD=spawn python3 examples/offline_inference/audio_language.py --model-type whisper
9 changes: 5 additions & 4 deletions docker/Dockerfile
Original file line number Diff line number Diff line change
Expand Up @@ -642,7 +642,7 @@ RUN --mount=type=cache,target=/root/.cache/uv \
else \
BITSANDBYTES_VERSION="${BITSANDBYTES_VERSION_X86}"; \
fi; \
uv pip install --system accelerate hf_transfer modelscope \
uv pip install --system accelerate modelscope \
"bitsandbytes>=${BITSANDBYTES_VERSION}" "timm${TIMM_VERSION}" "runai-model-streamer[s3,gcs,azure]${RUNAI_MODEL_STREAMER_VERSION}"

# ============================================================
Expand Down Expand Up @@ -756,9 +756,10 @@ RUN --mount=type=cache,target=/root/.cache/uv \
uv pip install --system -e tests/vllm_test_utils

# enable fast downloads from hf (for testing)
RUN --mount=type=cache,target=/root/.cache/uv \
uv pip install --system hf_transfer
ENV HF_HUB_ENABLE_HF_TRANSFER 1
ENV HF_XET_HIGH_PERFORMANCE 1

# increase timeout for hf downloads (for testing)
ENV HF_HUB_DOWNLOAD_TIMEOUT 60

# Copy in the v1 package for testing (it isn't distributed yet)
COPY vllm/v1 /usr/local/lib/python${PYTHON_VERSION}/dist-packages/vllm/v1
Expand Down
6 changes: 6 additions & 0 deletions docker/Dockerfile.cpu
Original file line number Diff line number Diff line change
Expand Up @@ -197,6 +197,12 @@ ADD ./.buildkite/ ./.buildkite/
RUN --mount=type=cache,target=/root/.cache/uv \
uv pip install -e tests/vllm_test_utils

# enable fast downloads from hf (for testing)
ENV HF_XET_HIGH_PERFORMANCE 1

# increase timeout for hf downloads (for testing)
ENV HF_HUB_DOWNLOAD_TIMEOUT 60

######################### RELEASE IMAGE #########################
FROM base AS vllm-openai

Expand Down
7 changes: 4 additions & 3 deletions docker/Dockerfile.nightly_torch
Original file line number Diff line number Diff line change
Expand Up @@ -272,9 +272,10 @@ RUN --mount=type=cache,target=/root/.cache/uv \
uv pip install --system -e tests/vllm_test_utils

# enable fast downloads from hf (for testing)
RUN --mount=type=cache,target=/root/.cache/uv \
uv pip install --system hf_transfer
ENV HF_HUB_ENABLE_HF_TRANSFER 1
ENV HF_XET_HIGH_PERFORMANCE 1

# increase timeout for hf downloads (for testing)
ENV HF_HUB_DOWNLOAD_TIMEOUT 60

RUN --mount=type=cache,target=/root/.cache/uv \
uv pip install --system -r requirements/test/nightly-torch.txt
Expand Down
7 changes: 4 additions & 3 deletions docker/Dockerfile.rocm
Original file line number Diff line number Diff line change
Expand Up @@ -365,9 +365,10 @@ RUN cd /vllm-workspace \
&& python3 -m pip install pytest-shard

# enable fast downloads from hf (for testing)
RUN --mount=type=cache,target=/root/.cache/uv \
uv pip install --system hf_transfer
ENV HF_HUB_ENABLE_HF_TRANSFER=1
ENV HF_XET_HIGH_PERFORMANCE=1

# increase timeout for hf downloads (for testing)
ENV HF_HUB_DOWNLOAD_TIMEOUT 60

# install audio decode package `torchcodec` from source (required due to
# ROCm and torch version mismatch) for tests with datasets package
Expand Down
2 changes: 1 addition & 1 deletion docs/getting_started/installation/gpu.rocm.inc.md
Original file line number Diff line number Diff line change
Expand Up @@ -240,7 +240,7 @@ uv pip install vllm==${VLLM_VERSION} \
# Install dependencies
pip install --upgrade numba \
scipy \
huggingface-hub[cli,hf_transfer] \
huggingface-hub[cli] \
setuptools_scm
pip install -r requirements/rocm.txt
Expand Down
4 changes: 2 additions & 2 deletions requirements/common.txt
Original file line number Diff line number Diff line change
Expand Up @@ -7,7 +7,7 @@ requests >= 2.26.0
tqdm
blake3
py-cpuinfo
transformers >= 4.56.0, < 5
transformers >= 4.56.0, != 5.0.*, != 5.1.*, != 5.2.*, != 5.3.*, != 5.4.*, != 5.5.0
tokenizers >= 0.21.1 # Required for fast incremental detokenization.
protobuf >= 5.29.6, !=6.30.*, !=6.31.*, !=6.32.*, !=6.33.0.*, !=6.33.1.*, !=6.33.2.*, !=6.33.3.*, !=6.33.4.* # Required by LlamaTokenizer, gRPC. CVE-2026-0994
fastapi[standard] >= 0.115.0 # Required by FastAPI's form models in the OpenAI API server's audio transcriptions endpoint.
Expand Down Expand Up @@ -37,7 +37,7 @@ pyyaml
six>=1.16.0; python_version > '3.11' # transitive dependency of pandas that needs to be the latest version for python 3.12
setuptools>=77.0.3,<81.0.0; python_version > '3.11' # Setuptools is used by triton, we need to ensure a modern version is installed for 3.12+ so that it does not try to import distutils, which was removed in 3.12
einops # Required for Qwen2-VL.
compressed-tensors == 0.14.0.1 # required for compressed-tensors
compressed-tensors == 0.15.0.1 # required for compressed-tensors
depyf==0.20.0 # required for profiling and debugging with compilation config
cloudpickle # allows pickling lambda functions in model_executor/models/registry.py
watchfiles # required for http server to monitor the updates of TLS files
Expand Down
6 changes: 3 additions & 3 deletions requirements/test/cuda.in
Original file line number Diff line number Diff line change
Expand Up @@ -18,7 +18,7 @@ httpx
librosa # required for audio tests
vector_quantize_pytorch # required for minicpmo_26 test
vocos # required for minicpmo_26 test
peft>=0.15.0 # required for phi-4-mm test
peft>=0.18.1 # required for phi-4-mm test
pqdm
ray[cgraph,default]>=2.48.0 # Ray Compiled Graph, required by pipeline parallelism tests
resampy # required for audio tests
Expand All @@ -39,8 +39,8 @@ opencv-python-headless >= 4.13.0 # required for video test
datamodel_code_generator # required for minicpm3 test
lm-eval[api]>=0.4.11 # required for model evaluation test
mteb[bm25s]>=2, <3 # required for mteb test
transformers==4.57.5
tokenizers==0.22.0
transformers==5.5.3
tokenizers==0.22.2
schemathesis>=3.39.15 # Required for openai schema test.
# quantization
bitsandbytes==0.49.2
Expand Down
20 changes: 10 additions & 10 deletions requirements/test/cuda.txt
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,7 @@ absl-py==2.1.0
# via
# rouge-score
# tensorboard
accelerate==1.0.1
accelerate==1.13.0
# via peft
aenum==3.1.16
# via lightly
Expand Down Expand Up @@ -248,7 +248,6 @@ filelock==3.16.1
# huggingface-hub
# ray
# torch
# transformers
# virtualenv
fiona==1.10.1
# via torchgeo
Expand Down Expand Up @@ -331,7 +330,7 @@ h5py==3.13.0
# via terratorch
harfile==0.3.0
# via schemathesis
hf-xet==1.1.7
hf-xet==1.4.3
# via huggingface-hub
hiredis==3.0.0
# via tensorizer
Expand All @@ -345,9 +344,10 @@ httpx==0.27.2
# via
# -r requirements/test/cuda.in
# diffusers
# huggingface-hub
# perceptron
# schemathesis
huggingface-hub==0.36.2
huggingface-hub==1.10.2
# via
# accelerate
# datasets
Expand Down Expand Up @@ -756,7 +756,7 @@ pathvalidate==3.2.1
# via pytablewriter
patsy==1.0.1
# via statsmodels
peft==0.16.0
peft==0.18.1
# via -r requirements/test/cuda.in
perceptron==0.1.4
# via -r requirements/test/cuda.in
Expand Down Expand Up @@ -982,7 +982,7 @@ referencing==0.35.1
# via
# jsonschema
# jsonschema-specifications
regex==2024.9.11
regex==2026.2.28
# via
# diffusers
# nltk
Expand All @@ -1002,7 +1002,6 @@ requests==2.32.3
# google-api-core
# google-cloud-storage
# gpt-oss
# huggingface-hub
# lightly
# lm-eval
# mistral-common
Expand All @@ -1015,7 +1014,6 @@ requests==2.32.3
# starlette-testclient
# tacoreader
# tiktoken
# transformers
# wandb
resampy==0.4.3
# via -r requirements/test/cuda.in
Expand Down Expand Up @@ -1216,7 +1214,7 @@ timm==1.0.17
# segmentation-models-pytorch
# terratorch
# torchgeo
tokenizers==0.22.0
tokenizers==0.22.2
# via
# -c requirements/common.txt
# -r requirements/test/cuda.in
Expand Down Expand Up @@ -1295,7 +1293,7 @@ tqdm==4.67.3
# tacoreader
# terratorch
# transformers
transformers==4.57.5
transformers==5.5.3
# via
# -c requirements/common.txt
# -r requirements/test/cuda.in
Expand All @@ -1317,7 +1315,9 @@ typepy==1.3.2
typer==0.15.2
# via
# fastsafetensors
# huggingface-hub
# perceptron
# transformers
types-python-dateutil==2.9.0.20241206
# via arrow
typeshed-client==2.8.2
Expand Down
4 changes: 2 additions & 2 deletions requirements/test/nightly-torch.txt
Original file line number Diff line number Diff line change
Expand Up @@ -29,8 +29,8 @@ opencv-python-headless >= 4.13.0 # required for video test
datamodel_code_generator # required for minicpm3 test
lm-eval[api]>=0.4.11 # required for model evaluation test
mteb[bm25s]>=2, <3 # required for mteb test
transformers==4.57.5
tokenizers==0.22.0
transformers==5.5.3
tokenizers==0.22.2
schemathesis>=3.39.15 # Required for openai schema test.
# quantization
bitsandbytes>=0.49.2
Expand Down
5 changes: 2 additions & 3 deletions requirements/test/rocm.in
Original file line number Diff line number Diff line change
Expand Up @@ -38,8 +38,8 @@ opencv-python-headless>=4.13.0 # required for video test
datamodel_code_generator # required for minicpm3 test
lm-eval[api]>=0.4.11 # required for model evaluation test
mteb[bm25s]>=2, <3 # required for mteb test
transformers==4.57.5
tokenizers==0.22.0
transformers==5.5.3
tokenizers==0.22.2
schemathesis>=3.39.15 # Required for openai schema test
# quantization
bitsandbytes==0.49.2
Expand Down Expand Up @@ -82,4 +82,3 @@ plotly # required for perf comparison html report
rapidfuzz
torchgeo==0.7.0
multiprocess==0.70.16
huggingface-hub==0.36.2
Loading
Loading