Conversation
Signed-off-by: vincentzed <207368749+vincentzed@users.noreply.github.com> Add debug instrumentation for LFM2-VL multimodal processing Adds comprehensive logging to diagnose why vision features may not be properly integrated in the LFM2-VL model. Signed-off-by: vincentzed <207368749+vincentzed@users.noreply.github.com>
Cherry-picked from JustinTong0323:update-transformers-v5 PR. Updates rope_theta and rope_scaling access to use config.rope_parameters dict instead of direct config attributes for transformers v5 compatibility. Signed-off-by: vincentzed <207368749+vincentzed@users.noreply.github.com>
Summary of ChangesHello @vincentzed, I'm Gemini Code Assist1! I'm currently reviewing this pull request and will post my feedback shortly. In the meantime, here's a summary to help you and other reviewers quickly get up to speed! This pull request introduces support for the LiquidAI LFM2.5-VL-1.6B multimodal model, enabling it to be run within the SGLang framework. It includes the necessary configuration, model architecture, and image processing logic. A significant refactoring effort was undertaken to standardize how Rotary Positional Embedding parameters are accessed across various language models, enhancing code maintainability and future compatibility. Additionally, the changes ensure alignment with newer versions of the HuggingFace Transformers library and incorporate detailed debugging capabilities for multimodal data handling. Highlights
🧠 New Feature in Public Preview: You can now enable Memory to help Gemini Code Assist learn from your team's feedback. This makes future code reviews more consistent and personalized to your project's style. Click here to enable Memory in your admin console. Using Gemini Code AssistThe full guide for Gemini Code Assist can be found on our documentation page, here are some quick tips. Invoking Gemini You can request assistance from Gemini at any point by creating a comment using either
Customization To customize Gemini Code Assist for GitHub experience, repository maintainers can create a configuration file and/or provide a custom code review style guide (such as PEP-8 for Python) by creating and adding files to a Limitations & Feedback Gemini Code Assist may make mistakes. Please leave feedback on any instances where its feedback is incorrect or counter productive. You can react with 👍 and 👎 on @gemini-code-assist comments. If you're interested in giving your feedback about your experience with Gemini Code Assist for Github and other Google products, sign up here. You can also get AI-powered code generation, chat, as well as code reviews directly in the IDE at no cost with the Gemini Code Assist IDE Extension. Footnotes
|
There was a problem hiding this comment.
Code Review
This pull request introduces support for the LFM-VL model and includes a significant refactoring across many model files to ensure compatibility with transformers>=5.0.0. The changes correctly adapt the way RoPE parameters are accessed from the model configuration.
However, there is a critical issue with leftover debugging code in several files. This code writes to hardcoded paths and includes session-specific information, which should not be part of the final codebase. I've added specific comments pointing out these instances. Please remove all debugging-related code before merging.
I've also noted a minor point about using a consistent logger in the new config file.
Signed-off-by: vincentzed <207368749+vincentzed@users.noreply.github.com>
0780f7d to
5855fe2
Compare
# Conflicts: # python/sglang/srt/configs/__init__.py # python/sglang/srt/model_executor/model_runner.py # python/sglang/srt/models/deepseek_v2.py # python/sglang/srt/models/gemma2.py # python/sglang/srt/models/gpt_oss.py # python/sglang/srt/models/qwen3_next.py # python/sglang/test/runners.py
Motivation
Tested on B300.
Modifications
Accuracy Tests
The branch already has patches to align with #17784 (necessary to make it work)
For the MMMU result, Please, see below.
Benchmarking and Profiling
Text model is fine.
Checklist
Review Process
/tag-run-ci-label,/rerun-failed-ci,/tag-and-rerun-ci