[Core] Avoid using extra thread in UniProcExecutor#40891
Merged
zhuohan123 merged 1 commit intovllm-project:mainfrom May 7, 2026
Merged
[Core] Avoid using extra thread in UniProcExecutor#40891zhuohan123 merged 1 commit intovllm-project:mainfrom
UniProcExecutor#40891zhuohan123 merged 1 commit intovllm-project:mainfrom
Conversation
Contributor
There was a problem hiding this comment.
Code Review
This pull request replaces the ThreadPoolExecutor in UniProcExecutor with a custom AsyncOutputFuture class to manage asynchronous model outputs. Review feedback identifies a thread-safety issue in the AsyncOutputFuture.result method that could cause multiple calls to get_output(), violating its contract. Suggestions include implementing a double-checked locking pattern, using NotImplementedError for the timeout parameter, and adding the necessary threading import.
@zhuohan123's idea Signed-off-by: Nick Hill <nickhill123@gmail.com>
zhuohan123
approved these changes
May 7, 2026
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.
@zhuohan123's idea
Clear performance benefit, especially in low latency / high concurrency case.
Benchmark
--tensor-parallel-size 1 --distributed-executor-backend union a single NVIDIA GB200 GPU.Model:
Qwen/Qwen3-0.6B. Each side is mean ± population std across 3 timed runs sharing one server process; each run uses its own seed (1, 2, 3) and is preceded by a fresh warmup batch.Δ = relative change of with-mean vs. without-mean (✓ = improvement, ✗ = regression).
256 in / 2048 out; ignore-eos, no prefix cache.