Override HF config.json via CLI#3722
Merged
Merged
Conversation
grimoire
reviewed
Jul 11, 2025
| tokenizer = Tokenizer(model_path).model.model | ||
| model_config = ModelConfig.from_pretrained(model_path, dtype=dtype, dist_config=dist_config) | ||
|
|
||
| if misc_config.hf_overrides is not None: |
Collaborator
There was a problem hiding this comment.
I think we can add hf_overrides as an argument of ModelConfig.from_pretrained and do the process in it.
lvhan028
reviewed
Jul 15, 2025
lvhan028
reviewed
Jul 15, 2025
lvhan028
approved these changes
Jul 15, 2025
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.
Objective
Allows users to change the HF config.json via CLI, which refers to
config.jsonvia CLI vllm-project/vllm#5836TODO
Usage
For example, dynamic rope scaling from CLI, either passes inline JSON
or passes arbitrary JSON keys individually
OC
{"rope_type": "yarn", "factor": 4.0, "original_max_position_embeddings": 32768}{"rope_type": "yarn", "factor": 4.0, "original_max_position_embeddings": 32768}default rope scalingBUG Fix
Modifications to
lmdeploy/turbomind/turbomind.py.When using OpenCompass to pass in nested dict parameters, they will be wrapped as mmegine
ConfigDicttype. If we use the lmdeploy TurboMind backend, the following errors will happen inyaml.safe_dump:Error Trace
Sample Code to Reproduce
Outputs: