-
Notifications
You must be signed in to change notification settings - Fork 487
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
bugfix: llava-hf/llava-interleave-qwen-7b-hf (#2497) #2657
Conversation
lmdeploy/serve/openai/api_server.py
Outdated
@@ -1054,13 +1054,15 @@ def serve(model_path: str, | |||
|
|||
_, pipeline_class = get_task(model_path) | |||
|
|||
vision_config = VisionConfig(kwargs.get("vision_max_batch_size", 1)) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
thanks, I will use the options to setting max_batch_size
I suggest just set |
- fix init raise exception because tie_word_embeddings config
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
@deepindeed2022 please resolve the linting error pip install pre-commit
cd lmdeploy # the root directory of the repo
pre-commit install
pre-commit run --all-files |
- fix init raise exception because tie_word_embeddings config
Motivation
issue #2497
python3 -m lmdeploy.serve.openai.api_server path/to/llava_hf/llava-interleave-qwen-7b-hf
Modification
vision-max-batch-size
for openai/api_server start config, reference fromlmdeploy/cli/serve.py
BC-breaking (Optional)
Does the modification introduce changes that break the backward-compatibility of the downstream repositories?
If so, please describe how it breaks the compatibility and how the downstream projects should modify their code to keep compatibility with this PR.
Use cases (Optional)
we have test
llava_hf/llava-interleave-qwen-7b-hf
, and support vision encoder with config in start command. such as:python3 -m lmdeploy.serve.openai.api_server path/to/llava_hf/llava-interleave-qwen-7b-hf --vision-max-batch-size 16