Skip to content

Support configure skip_special_tokens in openai response api#32345

Merged
chaunceyjiang merged 1 commit intovllm-project:mainfrom
842974287:my_changes
Jan 15, 2026
Merged

Support configure skip_special_tokens in openai response api#32345
chaunceyjiang merged 1 commit intovllm-project:mainfrom
842974287:my_changes

Conversation

@842974287
Copy link
Contributor

@842974287 842974287 commented Jan 14, 2026

Purpose

Currently we always use default value which will skip decoding special tokens. Want to make it configurable.

Test Plan

Test Result


Essential Elements of an Effective PR Description Checklist
  • The purpose of the PR, such as "Fix some issue (link existing issues this PR will resolve)".
  • The test plan, such as providing test command.
  • The test results, such as pasting the results comparison before and after, or e2e results
  • (Optional) The necessary documentation update, such as updating supported_models.md and examples for a new model.
  • (Optional) Release notes update. If your change is user facing, please update the release notes draft in the Google Doc.

Signed-off-by: Shiyan Deng <dsy842974287@meta.com>
Copy link
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request adds support for configuring skip_special_tokens in the OpenAI-compatible response API. The change introduces the skip_special_tokens parameter to the ResponsesRequest model and correctly propagates it to the SamplingParams. The implementation is straightforward, follows existing patterns in the codebase, and the default value is consistent with other parts of the API. The change appears correct and I have no issues to report.

Copy link
Collaborator

@chaunceyjiang chaunceyjiang left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks~

@chaunceyjiang chaunceyjiang added the ready ONLY add when PR is ready to merge/full CI is needed label Jan 15, 2026
@chaunceyjiang chaunceyjiang enabled auto-merge (squash) January 15, 2026 02:10
@chaunceyjiang
Copy link
Collaborator

fix #20874 (comment)

@chaunceyjiang chaunceyjiang merged commit 375e598 into vllm-project:main Jan 15, 2026
54 checks passed
sammysun0711 pushed a commit to sammysun0711/vllm that referenced this pull request Jan 16, 2026
akh64bit pushed a commit to akh64bit/vllm that referenced this pull request Jan 16, 2026
dsuhinin pushed a commit to dsuhinin/vllm that referenced this pull request Jan 21, 2026
…oject#32345)

Signed-off-by: Shiyan Deng <dsy842974287@meta.com>
Signed-off-by: dsuhinin <suhinin.dmitriy@gmail.com>
ItzDEXX pushed a commit to ItzDEXX/vllm that referenced this pull request Feb 19, 2026
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

frontend ready ONLY add when PR is ready to merge/full CI is needed

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants