You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Install MindIE Turbo for Performance acceleration:
170
+
171
+
```{code-block} bash
172
+
:substitutions:
173
+
# Install MindIE Turbo
174
+
pip install mindie_turbo
175
+
```
157
176
:::
158
177
159
178
::::
@@ -194,6 +213,10 @@ docker run --rm \
194
213
-it $IMAGE bash
195
214
```
196
215
216
+
```{note}
217
+
vLLM and vLLM Ascend code are placed in `/workspace` in the docker image. And they are installed in develop mode so that the developer could easily modify the code.
218
+
```
219
+
197
220
::::
198
221
199
222
:::::
@@ -257,3 +280,10 @@ Prompt: 'The president of the United States is', Generated text: ' a very import
257
280
Prompt: 'The capital of France is', Generated text: ' Paris. The oldest part of the city is Saint-Germain-des-Pr'
258
281
Prompt: 'The future of AI is', Generated text: ' not bright\n\nThere is no doubt that the evolution of AI will have a huge'
259
282
```
283
+
284
+
### Perfermance Acceleration
285
+
286
+
Get more performance gains by optimizing Python and torch-npu with the Bisheng compiler, please follow these official turtorial:
287
+
288
+
[Optimizing Python with Bisheng](https://www.hiascend.com/document/detail/zh/Pytorch/600/ptmoddevg/trainingmigrguide/performance_tuning_0063.html)
289
+
[Optimizing torch-npu with Bisheng](https://www.hiascend.com/document/detail/zh/Pytorch/600/ptmoddevg/trainingmigrguide/performance_tuning_0058.html)
0 commit comments