Skip to content

Commit dda2681

Browse files
committed
[None][doc] add blackwell supporting into doc.
Signed-off-by: nv-guomingz <[email protected]>
1 parent 991b83a commit dda2681

File tree

2 files changed

+4
-0
lines changed

2 files changed

+4
-0
lines changed

docs/source/overview.md

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -24,6 +24,9 @@ TensorRT-LLM supports the latest LLMs. Refer to the {ref}`support-matrix-softwar
2424

2525
TensorRT-LLM consists of pre– and post-processing steps and multi-GPU multi-node communication primitives in a simple, open-source Model Definition API for groundbreaking LLM inference performance on GPUs. Refer to the {ref}`multi-gpu-multi-node` section for more information.
2626

27+
### FP4 Support
28+
[NVIDIA B200 GPUs](https://www.nvidia.com/en-us/data-center/dgx-b200/) , when used with TensorRT-LLM, enable seamless loading of model weights in the new [FP4 format](https://developer.nvidia.com/blog/introducing-nvfp4-for-efficient-and-accurate-low-precision-inference/#what_is_nvfp4), allowing you to automatically leverage optimized FP4 kernels for efficient and accurate low-precision inference.
29+
2730
### FP8 Support
2831

2932
[NVIDIA H100 GPUs](https://www.nvidia.com/en-us/data-center/dgx-h100/) with TensorRT-LLM give you the ability to convert model weights into a new FP8 format easily and compile models to take advantage of optimized FP8 kernels automatically. This is made possible through [NVIDIA Hopper](https://blogs.nvidia.com/blog/h100-transformer-engine/) and done without having to change any model code.

docs/source/reference/support-matrix.md

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -158,6 +158,7 @@ The following table shows the supported software for TensorRT-LLM.
158158
- [10.11](https://docs.nvidia.com/deeplearning/tensorrt/release-notes/index.html)
159159
* - Precision
160160
-
161+
- Blackwell (SM100/SM120) - FP32, FP16, BF16, FP8, FP4, INT8, INT4
161162
- Hopper (SM90) - FP32, FP16, BF16, FP8, INT8, INT4
162163
- Ada Lovelace (SM89) - FP32, FP16, BF16, FP8, INT8, INT4
163164
- Ampere (SM80, SM86) - FP32, FP16, BF16, INT8, INT4[^smgte89]

0 commit comments

Comments
 (0)