[Model]: add FLUX.2-dev model #1630
Conversation
Signed-off-by: wuzhongjian <wuzhongjian_yewu@cmss.chinamobile.com>
Signed-off-by: wuzhongjian <wuzhongjian_yewu@cmss.chinamobile.com>
Signed-off-by: wuzhongjian <wuzhongjian_yewu@cmss.chinamobile.com>
Signed-off-by: wuzhongjian <wuzhongjian_yewu@cmss.chinamobile.com>
# Conflicts: # docs/user_guide/diffusion_acceleration.md
Signed-off-by: wuzhongjian <wuzhongjian_yewu@cmss.chinamobile.com>
Signed-off-by: wuzhongjian <wuzhongjian_yewu@cmss.chinamobile.com>
There was a problem hiding this comment.
💡 Codex Review
Here are some automated review suggestions for this pull request.
Reviewed commit: a53145a246
ℹ️ About Codex in GitHub
Codex has been enabled to automatically review pull requests in this repo. Reviews are triggered when you
- Open a pull request for review
- Mark a draft as ready
- Comment "@codex review".
If Codex has suggestions, it will comment; otherwise it will react with 👍.
When you sign up for Codex through ChatGPT, Codex can also answer questions or update the PR, like "@codex address that feedback".
| """Taking only the first image for now.""", | ||
| ) | ||
| first_prompt = req.prompts[0] | ||
| prompt = first_prompt if isinstance(first_prompt, str) else (first_prompt.get("prompt") or "") |
There was a problem hiding this comment.
Allow embedding-only prompts by preserving None prompt
When a request supplies prompt_embeds without a text prompt, this line coerces missing text to "" instead of keeping it None. That makes check_inputs think both prompt and prompt_embeds were provided and raises a ValueError, so embedding-only inference for Flux2Pipeline is effectively broken.
Useful? React with 👍 / 👎.
| # compute the previous noisy sample x_t -> x_t-1 | ||
| latents = self.scheduler.step(noise_pred, t, latents, return_dict=False)[0] |
There was a problem hiding this comment.
Invoke step-end callback during denoising iterations
The pipeline accepts callback_on_step_end and validates callback_on_step_end_tensor_inputs, but the denoising loop never invokes the callback. Any caller relying on per-step hooks (progress reporting, custom stopping, or latent inspection) will silently lose that behavior in Flux2Pipeline compared with other diffusion pipelines in this repo.
Useful? React with 👍 / 👎.
PLEASE FILL IN THE PR DESCRIPTION HERE ENSURING ALL CHECKLIST ITEMS (AT THE BOTTOM) HAVE BEEN CONSIDERED.
Purpose
support https://huggingface.co/black-forest-labs/FLUX.2-dev
Test Plan
vLLM-Omni:
Test Result
vLLM-Omni:
Reproduced with 4xA800.
Essential Elements of an Effective PR Description Checklist
supported_models.mdandexamplesfor a new model. Please runmkdocs serveto sync the documentation editions to./docs.BEFORE SUBMITTING, PLEASE READ https://github.com/vllm-project/vllm-omni/blob/main/CONTRIBUTING.md (anything written below this line will be removed by GitHub Actions)