Skip to content

[Renderer] Move MM Hash parsing into Renderer#34711

Merged
vllm-bot merged 5 commits intovllm-project:mainfrom
DarkLight1337:mv-mm-hashes-parsing
Feb 18, 2026
Merged

[Renderer] Move MM Hash parsing into Renderer#34711
vllm-bot merged 5 commits intovllm-project:mainfrom
DarkLight1337:mv-mm-hashes-parsing

Conversation

@DarkLight1337
Copy link
Copy Markdown
Member

@DarkLight1337 DarkLight1337 commented Feb 17, 2026

Purpose

Follow-up to #34560

  • Normalize MultiModalUUIDDict -> MultiModalUUIDItems in Renderer instead of MM processor.
  • Remove tokenization_kwargs from MM hash calculation as the MM data isn't affected by that.
  • Update MultiModalProcessor.apply argument list to put UUIDs before miscellaneous kwargs.
  • Clean up MM data parsing. Replace _is_empty check with modality selection in get_dummy_mm_inputs.

Test Plan

Test Result


Essential Elements of an Effective PR Description Checklist
  • The purpose of the PR, such as "Fix some issue (link existing issues this PR will resolve)".
  • The test plan, such as providing test command.
  • The test results, such as pasting the results comparison before and after, or e2e results
  • (Optional) The necessary documentation update, such as updating supported_models.md and examples for a new model.
  • (Optional) Release notes update. If your change is user facing, please update the release notes draft in the Google Doc.

Signed-off-by: DarkLight1337 <tlleungac@connect.ust.hk>
@DarkLight1337 DarkLight1337 added the ready ONLY add when PR is ready to merge/full CI is needed label Feb 17, 2026
@mergify mergify Bot added deepseek Related to DeepSeek models multi-modality Related to multi-modality (#4194) labels Feb 17, 2026
Copy link
Copy Markdown
Contributor

@gemini-code-assist gemini-code-assist Bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request refactors the parsing of MultiModalUUIDDict to MultiModalUUIDItems from the multimodal processor to the renderer, which is a good architectural improvement. The changes are propagated through many files, and the refactoring itself seems solid. However, I've identified a critical issue where tokenization_kwargs are no longer included in the multimodal cache hash computation. This could lead to cache corruption and incorrect model outputs. I've provided a detailed comment and a suggested fix for this issue.

Comment thread vllm/multimodal/processing/processor.py
Signed-off-by: DarkLight1337 <tlleungac@connect.ust.hk>
@mergify
Copy link
Copy Markdown
Contributor

mergify Bot commented Feb 17, 2026

Hi @DarkLight1337, the pre-commit checks have failed. Please run:

uv pip install pre-commit
pre-commit install
pre-commit run --all-files

Then, commit the changes and push to your branch.

For future commits, pre-commit will run automatically on changed files before each commit.

Tip

Is mypy or markdownlint failing?
mypy and markdownlint are run differently in CI. If the failure is related to either of these checks, please use the following commands to run them locally:
# For mypy (substitute "3.10" with the failing version if needed)
pre-commit run --hook-stage manual mypy-3.10
# For markdownlint
pre-commit run --hook-stage manual markdownlint

Signed-off-by: DarkLight1337 <tlleungac@connect.ust.hk>
Signed-off-by: DarkLight1337 <tlleungac@connect.ust.hk>
Signed-off-by: DarkLight1337 <tlleungac@connect.ust.hk>
@DarkLight1337 DarkLight1337 enabled auto-merge (squash) February 17, 2026 17:14
@vllm-bot vllm-bot merged commit a0d8d94 into vllm-project:main Feb 18, 2026
57 of 60 checks passed
@DarkLight1337 DarkLight1337 deleted the mv-mm-hashes-parsing branch February 18, 2026 03:19
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

deepseek Related to DeepSeek models multi-modality Related to multi-modality (#4194) ready ONLY add when PR is ready to merge/full CI is needed

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants