[Bugfix][Neuron] Fix soft prompt method error in NeuronExecutor#6313
[Bugfix][Neuron] Fix soft prompt method error in NeuronExecutor#6313WoosukKwon merged 1 commit intomainfrom
Conversation
|
@WoosukKwon Thanks for the prompt response to my issue #6269! When the PR is approved, can you please follow up with a procedure to update my install to run the patched vLLM on Neuron systems? I urgently need this for a project I am working on. |
|
Thanks for the fix. The current neuron CI only tests online inference, the offline inference capability is currently not tested for neuron backend. |
Well, the online inference also appears to be broken: python -m vllm.entrypoints.openai.api_server |
|
@liangfu As @areanddee pointed out, the error happens when |
…-project#6313) (cherry picked from commit 997df46)
…-project#6313) Signed-off-by: Alvant <alvasian@yandex.ru>
…-project#6313) Signed-off-by: LeiWang1999 <leiwang1999@outlook.com>
Fixes #6269
However, I'm still not sure how #4645 passed the Neuron CI test.