Skip to content

[Model] Support HyperCLOVAX-SEED-Omni-8B#585

Closed
effortprogrammer wants to merge 2 commits into
vllm-project:mainfrom
effortprogrammer:add-hyperclovax-omni
Closed

[Model] Support HyperCLOVAX-SEED-Omni-8B#585
effortprogrammer wants to merge 2 commits into
vllm-project:mainfrom
effortprogrammer:add-hyperclovax-omni

Conversation

@effortprogrammer
Copy link
Copy Markdown

PLEASE FILL IN THE PR DESCRIPTION HERE ENSURING ALL CHECKLIST ITEMS (AT THE BOTTOM) HAVE BEEN CONSIDERED.

Purpose

Add new model: https://huggingface.co/naver-hyperclovax/HyperCLOVAX-SEED-Omni-8B

Test Plan

Test Result


Essential Elements of an Effective PR Description Checklist
  • The purpose of the PR, such as "Fix some issue (link existing issues this PR will resolve)".
  • The test plan, such as providing test command.
  • The test results, such as pasting the results comparison before and after, or e2e results
  • (Optional) The necessary documentation update, such as updating supported_models.md and examples for a new model.
  • (Optional) Release notes update. If your change is user facing, please update the release notes draft.

BEFORE SUBMITTING, PLEASE READ https://github.com/vllm-project/vllm-omni/blob/main/CONTRIBUTING.md (anything written below this line will be removed by GitHub Actions)

@effortprogrammer effortprogrammer changed the title Add hyperclovax omni [Model] Support HyperCLOVAX-SEED-Omni-8B Jan 2, 2026
@effortprogrammer effortprogrammer marked this pull request as draft January 2, 2026 17:26
Copy link
Copy Markdown

@chatgpt-codex-connector chatgpt-codex-connector Bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

💡 Codex Review

Here are some automated review suggestions for this pull request.

ℹ️ About Codex in GitHub

Codex has been enabled to automatically review pull requests in this repo. Reviews are triggered when you

  • Open a pull request for review
  • Mark a draft as ready
  • Comment "@codex review".

If Codex has suggestions, it will comment; otherwise it will react with 👍.

When you sign up for Codex through ChatGPT, Codex can also answer questions or update the PR, like "@codex address that feedback".

Comment on lines +149 to +153
elif self.model_stage == "code2wav" and self.code2wav is not None:
# Input to Code2Wav stage are the Audio Tokens generated by Thinker
audio_tokens = input_ids
waveform = self.code2wav(audio_tokens)
return OmniOutput(
Copy link
Copy Markdown

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

P1 Badge OmniOutput returned without multimodal flag

The code2wav forward path returns an OmniOutput (lines 149‑153) but the class never sets have_multimodal_outputs. GPUGenerationModelRunner calls OmniGPUModelRunner.extract_multimodal_outputs (vllm_omni/worker/gpu_model_runner.py:314-331), which only accepts OmniOutput when the model advertises have_multimodal_outputs; otherwise it raises ValueError("Invalid hidden states type"). As soon as the code2wav stage runs via the generation worker, this missing flag causes a hard crash before any audio is produced.

Useful? React with 👍 / 👎.

@lishunyang12
Copy link
Copy Markdown
Collaborator

@effortprogrammer Hey, the HyperCLOVAX-SEED-Omni-8B draft doesn't have a test plan or results yet. Were you able to run inference with the Naver checkpoint? I see #869 and #613 are also working on the audio and vision decoder parts — are you coordinating with those PRs?

@effortprogrammer
Copy link
Copy Markdown
Author

@lishunyang12 I apologize for the delayed update. I heard from a Naver employee (the company behind this model) that there will be an update regarding the OMNI model. They mentioned that they plan to submit a new PR to support the OMNI model in the vllm-omni repo after the update.

I will close this PR now.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants