Skip to content

Add served model alias support#3

Closed
krystophny wants to merge 1 commit intomainfrom
feature/served-model-name-alias
Closed

Add served model alias support#3
krystophny wants to merge 1 commit intomainfrom
feature/served-model-name-alias

Conversation

@krystophny
Copy link
Copy Markdown
Collaborator

@krystophny krystophny commented Mar 24, 2026

Summary

  • add a dedicated --served-model-name server flag
  • expose a stable alias via the existing served_model_name load path
  • cover the new CLI option with parser tests

Why this is deployable on its own

  • it only changes how the server advertises and validates model names
  • it does not require the Responses branch to remain useful
  • it provides stable short aliases like qwen for local clients and benchmarks

Testing

  • PYTHONPATH=/Users/ert/code/vllm-mlx /Users/ert/code/.venv/bin/python -m pytest tests/test_server.py -q
  • python3 -m compileall vllm_mlx

@krystophny krystophny changed the title Add served model alias flag Add served model alias support Mar 24, 2026
@krystophny
Copy link
Copy Markdown
Collaborator Author

This PR is now superseded by upstream waybarrios/vllm-mlx#125, which merged --served-model-name support into main on March 21, 2026:
waybarrios#125

Our fork main is already synced to upstream main, so keeping this PR open no longer improves the stack. I am closing it to keep the fork PR set clean and non-duplicative.

@krystophny krystophny closed this Mar 24, 2026
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant