Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion docs/getting_started/installation/cpu/arm.inc.md
Original file line number Diff line number Diff line change
Expand Up @@ -23,7 +23,7 @@ ARM CPU backend currently supports Float32, FP16 and BFloat16 datatypes.
# --8<-- [end:pre-built-wheels]
# --8<-- [start:build-wheel-from-source]

--8<-- "docs/getting_started/installation/cpu/cpu/build.inc.md"
--8<-- "docs/getting_started/installation/cpu/build.inc.md"

Testing has been conducted on AWS Graviton3 instances for compatibility.

Expand Down
2 changes: 1 addition & 1 deletion docs/getting_started/installation/cpu/x86.inc.md
Original file line number Diff line number Diff line change
Expand Up @@ -24,7 +24,7 @@ vLLM initially supports basic model inferencing and serving on x86 CPU platform,
# --8<-- [end:pre-built-wheels]
# --8<-- [start:build-wheel-from-source]

--8<-- "docs/getting_started/installation/cpu/cpu/build.inc.md"
--8<-- "docs/getting_started/installation/cpu/build.inc.md"

!!! note
- AVX512_BF16 is an extension ISA provides native BF16 data type conversion and vector product instructions, which brings some performance improvement compared with pure AVX512. The CPU backend build script will check the host CPU flags to determine whether to enable AVX512_BF16.
Expand Down