Skip to content

Update README.md#306

Merged
zhuohan123 merged 1 commit intovllm-project:mainfrom
bm777:patch-1
Jun 29, 2023
Merged

Update README.md#306
zhuohan123 merged 1 commit intovllm-project:mainfrom
bm777:patch-1

Conversation

@bm777
Copy link
Copy Markdown
Contributor

@bm777 bm777 commented Jun 29, 2023

Updated the README.md -> corrected the pip usage command

Updated the README.md -> corrected the pip usage command
Copy link
Copy Markdown
Member

@zhuohan123 zhuohan123 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM! Thanks for catching the errors.

@zhuohan123 zhuohan123 merged commit 9d27b09 into vllm-project:main Jun 29, 2023
michaelfeil pushed a commit to michaelfeil/vllm that referenced this pull request Jul 1, 2023
@bm777
Copy link
Copy Markdown
Contributor Author

bm777 commented Jul 9, 2023

It is a pleasure :)

hongxiayang pushed a commit to hongxiayang/vllm that referenced this pull request Feb 13, 2024
billishyahao pushed a commit to billishyahao/vllm that referenced this pull request Dec 31, 2024
Triton's autotuner had a bug (fixed in
triton-lang/triton@184fb53 )
that we happen to trigger - this PR is a temp. workaround before we
patch/update Triton.
wuhuikx pushed a commit to wuhuikx/vllm that referenced this pull request Mar 27, 2025
Remove useless code which wrongly introduced by multi step PR .

Signed-off-by: new-TonyWang <wangtonyyu222@gmail.com>
jikunshang added a commit to jikunshang/vllm that referenced this pull request Aug 15, 2025
* load gpt-oss int4

Signed-off-by: Yan Ma <yan.ma@intel.com>

* fix int4 path

* fix

* fix

* fix

* try padding at weight create

* fix

* test

* test

* use linear

* fix

* fix

* fix loading [1/N]

* fix loading

* fix o/qkv

* fix router

* remove log

* address comments

---------

Signed-off-by: Yan Ma <yan.ma@intel.com>
Co-authored-by: Yan Ma <yan.ma@intel.com>
yiliu30 pushed a commit to yiliu30/vllm-fork that referenced this pull request Aug 20, 2025
Update SHA for extension to: [Security] Fix: remove dead code (vllm-project#306)
(vllm-project#310)
yma11 pushed a commit to yma11/vllm that referenced this pull request Sep 2, 2025
* load gpt-oss int4

Signed-off-by: Yan Ma <yan.ma@intel.com>

* fix int4 path

* fix

* fix

* fix

* try padding at weight create

* fix

* test

* test

* use linear

* fix

* fix

* fix loading [1/N]

* fix loading

* fix o/qkv

* fix router

* remove log

* address comments

---------

Signed-off-by: Yan Ma <yan.ma@intel.com>
Co-authored-by: Yan Ma <yan.ma@intel.com>

add gpt oss tp support (vllm-project#307)

* add fused kernel

* always pack on k dim

* disbale padding

* fix interface

* fix shape

* qzero

* test padding

* hidden_size padding

* fix tp1

* ep

* layer index

* bias shape

* fix ep

* update tp

* update tp

* add log

* refine log

* log create weight

* fix ep rank/offset

* fix w2 bias, only add on rank 0

* address comments

fix group size 128 (vllm-project#318)

add env var for use_marlin (vllm-project#319)

* add env var for use_marlin

* use_marlin force true

* fix group size
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants