Skip to content

Conversation

@nvchenghaoz
Copy link

@coderabbitai summary

@nvchenghaoz nvchenghaoz merged commit 4b50b3e into feat/ad_linear_attention Sep 26, 2025
2 of 3 checks passed
lucaslie pushed a commit that referenced this pull request Sep 29, 2025
* Fix the bamba unit test

Signed-off-by: Chenghao Zhang <[email protected]>

* none: Add triton backend for ssm_transform and cuda backend for conv

Signed-off-by: Chenghao Zhang <[email protected]>

* Fully Use the TRT LLM kernels

Signed-off-by: Chenghao Zhang <[email protected]>

* Add fake version for ssm transform op

Signed-off-by: Chenghao Zhang <[email protected]>

* Fix the datatype error in fake op

Signed-off-by: Chenghao Zhang <[email protected]>

* Fix the conv test error

Signed-off-by: Chenghao Zhang <[email protected]>

* Fix the triton ssm error

Signed-off-by: Chenghao Zhang <[email protected]>

* Fix the DemoLLM sampler mismatch

Signed-off-by: Chenghao Zhang <[email protected]>

* Update the implementation for triton/cuda kernels

Signed-off-by: Chenghao Zhang <[email protected]>

* Fix the d2d memcpy for decode

Signed-off-by: Chenghao Zhang <[email protected]>

* Revert the generator and remove the redundant code

Signed-off-by: Chenghao Zhang <[email protected]>

---------

Signed-off-by: Chenghao Zhang <[email protected]>
Signed-off-by: Suyog Gupta <[email protected]>
Co-authored-by: Suyog Gupta <[email protected]>
nvchenghaoz added a commit that referenced this pull request Oct 1, 2025
* Fix the bamba unit test

Signed-off-by: Chenghao Zhang <[email protected]>

* none: Add triton backend for ssm_transform and cuda backend for conv

Signed-off-by: Chenghao Zhang <[email protected]>

* Fully Use the TRT LLM kernels

Signed-off-by: Chenghao Zhang <[email protected]>

* Add fake version for ssm transform op

Signed-off-by: Chenghao Zhang <[email protected]>

* Fix the datatype error in fake op

Signed-off-by: Chenghao Zhang <[email protected]>

* Fix the conv test error

Signed-off-by: Chenghao Zhang <[email protected]>

* Fix the triton ssm error

Signed-off-by: Chenghao Zhang <[email protected]>

* Fix the DemoLLM sampler mismatch

Signed-off-by: Chenghao Zhang <[email protected]>

* Update the implementation for triton/cuda kernels

Signed-off-by: Chenghao Zhang <[email protected]>

* Fix the d2d memcpy for decode

Signed-off-by: Chenghao Zhang <[email protected]>

* Revert the generator and remove the redundant code

Signed-off-by: Chenghao Zhang <[email protected]>

---------

Signed-off-by: Chenghao Zhang <[email protected]>
Signed-off-by: Suyog Gupta <[email protected]>
Co-authored-by: Suyog Gupta <[email protected]>
nvchenghaoz added a commit that referenced this pull request Oct 3, 2025
* Fix the bamba unit test

Signed-off-by: Chenghao Zhang <[email protected]>

* none: Add triton backend for ssm_transform and cuda backend for conv

Signed-off-by: Chenghao Zhang <[email protected]>

* Fully Use the TRT LLM kernels

Signed-off-by: Chenghao Zhang <[email protected]>

* Add fake version for ssm transform op

Signed-off-by: Chenghao Zhang <[email protected]>

* Fix the datatype error in fake op

Signed-off-by: Chenghao Zhang <[email protected]>

* Fix the conv test error

Signed-off-by: Chenghao Zhang <[email protected]>

* Fix the triton ssm error

Signed-off-by: Chenghao Zhang <[email protected]>

* Fix the DemoLLM sampler mismatch

Signed-off-by: Chenghao Zhang <[email protected]>

* Update the implementation for triton/cuda kernels

Signed-off-by: Chenghao Zhang <[email protected]>

* Fix the d2d memcpy for decode

Signed-off-by: Chenghao Zhang <[email protected]>

* Revert the generator and remove the redundant code

Signed-off-by: Chenghao Zhang <[email protected]>

---------

Signed-off-by: Chenghao Zhang <[email protected]>
Signed-off-by: Suyog Gupta <[email protected]>
Co-authored-by: Suyog Gupta <[email protected]>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

4 participants