Skip to content

Conversation

@Fridge003
Copy link
Collaborator

@Fridge003 Fridge003 commented Apr 3, 2025

Motivation

Followup of #5005, this PR adds --mla-backend argument to better handle the difference between normal attention backend and mla backend. In default, --attention-backend is set to flashinfer, and --mla-backend is set to triton.
Arguments like --enable-flashinfer-mla and --enable-flashmla are still usable.

Modifications

Checklist

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant