diff --git a/docs/build/eps.md b/docs/build/eps.md index ab4d464db0eb6..1a78c9ec23cd4 100644 --- a/docs/build/eps.md +++ b/docs/build/eps.md @@ -454,7 +454,7 @@ The DirectML execution provider supports building for both x64 and x86 architect --- ## ARM Compute Library -See more information on the ACL Execution Provider [here](../execution-providers/ACL-ExecutionProvider.md). +See more information on the ACL Execution Provider [here](../execution-providers/community-maintained/ACL-ExecutionProvider.md). ### Prerequisites {: .no_toc } @@ -521,7 +521,7 @@ onnxruntime_test_all ## ArmNN -See more information on the ArmNN Execution Provider [here](../execution-providers/ArmNN-ExecutionProvider.md). +See more information on the ArmNN Execution Provider [here](../execution-providers/community-maintained/ArmNN-ExecutionProvider.md). ### Prerequisites {: .no_toc } @@ -569,7 +569,7 @@ The ARM Compute Library home directory and build directory must also be availabl --- ## RKNPU -See more information on the RKNPU Execution Provider [here](../execution-providers/RKNPU-ExecutionProvider.md). +See more information on the RKNPU Execution Provider [here](../execution-providers/community-maintained/RKNPU-ExecutionProvider.md). ### Prerequisites {: .no_toc } @@ -608,9 +608,9 @@ set(CMAKE_C_COMPILER aarch64-linux-gnu-gcc) --- ## Vitis-AI -See more information on the Xilinx Vitis-AI execution provider [here](../execution-providers/Vitis-AI-ExecutionProvider.md). +See more information on the Xilinx Vitis-AI execution provider [here](../execution-providers/community-maintained/Vitis-AI-ExecutionProvider.md). -For instructions to setup the hardware environment: [Hardware setup](../execution-providers/Vitis-AI-ExecutionProvider.md#hardware-setup) +For instructions to setup the hardware environment: [Hardware setup](../execution-providers/community-maintained/Vitis-AI-ExecutionProvider.md#hardware-setup) ### Linux {: .no_toc } @@ -629,7 +629,7 @@ The Vitis-AI execution provider is only supported on Linux. ## AMD MIGraphX -See more information on the MIGraphX Execution Provider [here](../execution-providers/MIGraphX-ExecutionProvider.md). +See more information on the MIGraphX Execution Provider [here](../execution-providers/community-maintained/MIGraphX-ExecutionProvider.md). ### Prerequisites {: .no_toc } @@ -774,7 +774,7 @@ Linux example: ## CANN -See more information on the CANN Execution Provider [here](../execution-providers/CANN-ExecutionProvider.md). +See more information on the CANN Execution Provider [here](../execution-providers/community-maintained/CANN-ExecutionProvider.md). ### Prerequisites {: .no_toc } diff --git a/docs/execution-providers/CUDA-ExecutionProvider.md b/docs/execution-providers/CUDA-ExecutionProvider.md index ed4896729765b..3bc802d5d7d1a 100644 --- a/docs/execution-providers/CUDA-ExecutionProvider.md +++ b/docs/execution-providers/CUDA-ExecutionProvider.md @@ -1,8 +1,8 @@ --- -title: CUDA (NVIDIA) +title: NVIDIA - CUDA description: Instructions to execute ONNX Runtime applications with CUDA parent: Execution Providers -nav_order: 5 +nav_order: 1 redirect_from: /docs/reference/execution-providers/CUDA-ExecutionProvider --- @@ -31,6 +31,7 @@ Please reference [Nvidia CUDA Minor Version Compatibility](https://docs.nvidia.c |ONNX Runtime|CUDA|cuDNN|Notes| |---|---|---|---| +|1.13|11.6|8.2.4 (Linux)
8.5.0.96 (Windows)|libcudart 11.4.43
libcufft 10.5.2.100
libcurand 10.2.5.120
libcublasLt 11.6.5.2
libcublas 11.6.5.2
libcudnn 8.2.4| |1.12
1.11|11.4|8.2.4 (Linux)
8.2.2.26 (Windows)|libcudart 11.4.43
libcufft 10.5.2.100
libcurand 10.2.5.120
libcublasLt 11.6.5.2
libcublas 11.6.5.2
libcudnn 8.2.4| |1.10|11.4|8.2.4 (Linux)
8.2.2.26 (Windows)|libcudart 11.4.43
libcufft 10.5.2.100
libcurand 10.2.5.120
libcublasLt 11.6.1.51
libcublas 11.6.1.51
libcudnn 8.2.4| |1.9|11.4|8.2.4 (Linux)
8.2.2.26 (Windows)|libcudart 11.4.43
libcufft 10.5.2.100
libcurand 10.2.5.120
libcublasLt 11.6.1.51
libcublas 11.6.1.51
libcudnn 8.2.4| diff --git a/docs/execution-providers/CoreML-ExecutionProvider.md b/docs/execution-providers/CoreML-ExecutionProvider.md index 918a286b95b1a..999c345bd0af5 100644 --- a/docs/execution-providers/CoreML-ExecutionProvider.md +++ b/docs/execution-providers/CoreML-ExecutionProvider.md @@ -1,8 +1,8 @@ --- -title: CoreML (Apple) +title: Apple - CoreML description: Instructions to execute ONNX Runtime with CoreML parent: Execution Providers -nav_order: 4 +nav_order: 8 redirect_from: /docs/reference/execution-providers/CoreML-ExecutionProvider --- {::options toc_levels="2" /} diff --git a/docs/execution-providers/DirectML-ExecutionProvider.md b/docs/execution-providers/DirectML-ExecutionProvider.md index 3f06e59f5aee4..cc1462a69bb2d 100644 --- a/docs/execution-providers/DirectML-ExecutionProvider.md +++ b/docs/execution-providers/DirectML-ExecutionProvider.md @@ -1,8 +1,8 @@ --- -title: DirectML (Windows) +title: Windows - DirectML description: Instructions to execute ONNX Runtime with the DirectML execution provider parent: Execution Providers -nav_order: 6 +nav_order: 5 redirect_from: /docs/reference/execution-providers/DirectML-ExecutionProvider --- diff --git a/docs/execution-providers/NNAPI-ExecutionProvider.md b/docs/execution-providers/NNAPI-ExecutionProvider.md index a846da19a1af9..458f847e06e23 100644 --- a/docs/execution-providers/NNAPI-ExecutionProvider.md +++ b/docs/execution-providers/NNAPI-ExecutionProvider.md @@ -1,8 +1,8 @@ --- -title: NNAPI (Android) +title: Android - NNAPI description: Instructions to execute ONNX Runtime with the NNAPI execution provider parent: Execution Providers -nav_order: 8 +nav_order: 7 redirect_from: /docs/reference/execution-providers/NNAPI-ExecutionProvider --- {::options toc_levels="2" /} diff --git a/docs/execution-providers/OpenVINO-ExecutionProvider.md b/docs/execution-providers/OpenVINO-ExecutionProvider.md index 4b4ccee938f99..7354240418a01 100644 --- a/docs/execution-providers/OpenVINO-ExecutionProvider.md +++ b/docs/execution-providers/OpenVINO-ExecutionProvider.md @@ -1,8 +1,8 @@ --- -title: OpenVINO™ (Intel) +title: Intel - OpenVINO™ description: Instructions to execute OpenVINO™ Execution Provider for ONNX Runtime. parent: Execution Providers -nav_order: 10 +nav_order: 3 redirect_from: /docs/reference/execution-providers/OpenVINO-ExecutionProvider --- diff --git a/docs/execution-providers/ROCm-ExecutionProvider.md b/docs/execution-providers/ROCm-ExecutionProvider.md index f1db133da44a1..aac4f7ba7aec4 100644 --- a/docs/execution-providers/ROCm-ExecutionProvider.md +++ b/docs/execution-providers/ROCm-ExecutionProvider.md @@ -1,8 +1,8 @@ --- -title: ROCm (AMD) +title: AMD - ROCm description: Instructions to execute ONNX Runtime with the AMD ROCm execution provider parent: Execution Providers -nav_order: 12 +nav_order: 10 redirect_from: /docs/reference/execution-providers/ROCm-ExecutionProvider --- diff --git a/docs/execution-providers/SNPE-ExecutionProvider.md b/docs/execution-providers/SNPE-ExecutionProvider.md index 1b0ff78093d1d..f40c951df3012 100644 --- a/docs/execution-providers/SNPE-ExecutionProvider.md +++ b/docs/execution-providers/SNPE-ExecutionProvider.md @@ -1,8 +1,8 @@ --- -title: SNPE (Qualcomm) +title: Qualcomm - SNPE description: Execute ONNX models with SNPE Execution Provider parent: Execution Providers -nav_order: 13 +nav_order: 6 redirect_from: /docs/reference/execution-providers/SNPE-ExecutionProvider --- diff --git a/docs/execution-providers/TensorRT-ExecutionProvider.md b/docs/execution-providers/TensorRT-ExecutionProvider.md index 62a3ffafefead..0d1a832815c5a 100644 --- a/docs/execution-providers/TensorRT-ExecutionProvider.md +++ b/docs/execution-providers/TensorRT-ExecutionProvider.md @@ -1,8 +1,8 @@ --- -title: TensorRT (NVIDIA) +title: NVIDIA - TensorRT description: Instructions to execute ONNX Runtime on NVIDIA GPUs with the TensorRT execution provider parent: Execution Providers -nav_order: 14 +nav_order: 2 redirect_from: /docs/reference/execution-providers/TensorRT-ExecutionProvider --- diff --git a/docs/execution-providers/Xnnpack-ExecutionProvider.md b/docs/execution-providers/Xnnpack-ExecutionProvider.md index 6beedd267beea..445e9a2d99f2d 100644 --- a/docs/execution-providers/Xnnpack-ExecutionProvider.md +++ b/docs/execution-providers/Xnnpack-ExecutionProvider.md @@ -2,7 +2,7 @@ title: XNNPACK description: Instructions to execute ONNX Runtime with the XNNPACK execution provider parent: Execution Providers -nav_order: 17 +nav_order: 9 --- {::options toc_levels="2" /} diff --git a/docs/execution-providers/add-execution-provider.md b/docs/execution-providers/add-execution-provider.md index 3844185773c3d..ad3d260f54064 100644 --- a/docs/execution-providers/add-execution-provider.md +++ b/docs/execution-providers/add-execution-provider.md @@ -1,8 +1,8 @@ --- -title: Add a new execution provider +title: Add a new provider description: Instructions to add a new execution provider to ONNX Runtime parent: Execution Providers -nav_order: 18 +nav_order: 12 redirect_from: /docs/how-to/add-execution-provider --- diff --git a/docs/execution-providers/ACL-ExecutionProvider.md b/docs/execution-providers/community-maintained/ACL-ExecutionProvider.md similarity index 77% rename from docs/execution-providers/ACL-ExecutionProvider.md rename to docs/execution-providers/community-maintained/ACL-ExecutionProvider.md index 33d11eb89d948..103dea94abebb 100644 --- a/docs/execution-providers/ACL-ExecutionProvider.md +++ b/docs/execution-providers/community-maintained/ACL-ExecutionProvider.md @@ -1,8 +1,9 @@ --- -title: ARM Compute Library (ACL) +title: Arm - ACL description: Instructions to execute ONNX Runtime with the ACL Execution Provider -parent: Execution Providers -nav_order: 1 +grand_parent: Execution Providers +parent: Community-maintained +nav_order: 2 redirect_from: /docs/reference/execution-providers/ACL-ExecutionProvider --- @@ -20,7 +21,7 @@ The integration of ACL as an execution provider (EP) into ONNX Runtime accelerat ## Build -For build instructions, please see the [build page](../build/eps.md#arm-compute-library). +For build instructions, please see the [build page](../../build/eps.md#arm-compute-library). ## Usage ### C/C++ @@ -32,9 +33,9 @@ Ort::SessionOptions sf; bool enable_cpu_mem_arena = true; Ort::ThrowOnError(OrtSessionOptionsAppendExecutionProvider_ACL(sf, enable_cpu_mem_arena)); ``` -The C API details are [here](../get-started/with-c.html). +The C API details are [here](../../get-started/with-c.html). ## Performance Tuning -For performance tuning, please see guidance on this page: [ONNX Runtime Perf Tuning](../performance/tune-performance.md) +For performance tuning, please see guidance on this page: [ONNX Runtime Perf Tuning](../../performance/tune-performance.md) When/if using [onnxruntime_perf_test](https://github.com/microsoft/onnxruntime/tree/master/onnxruntime/test/perftest){:target="_blank"}, use the flag -e acl diff --git a/docs/execution-providers/ArmNN-ExecutionProvider.md b/docs/execution-providers/community-maintained/ArmNN-ExecutionProvider.md similarity index 78% rename from docs/execution-providers/ArmNN-ExecutionProvider.md rename to docs/execution-providers/community-maintained/ArmNN-ExecutionProvider.md index 0030ffc0dbaa3..e148567f9e3da 100644 --- a/docs/execution-providers/ArmNN-ExecutionProvider.md +++ b/docs/execution-providers/community-maintained/ArmNN-ExecutionProvider.md @@ -1,7 +1,8 @@ --- -title: Arm NN +title: Arm - Arm NN description: Instructions to execute ONNX Runtime with Arm NN on Armv8 cores -parent: Execution Providers +grand_parent: Execution Providers +parent: Community-maintained nav_order: 2 redirect_from: /docs/reference/execution-providers/ArmNN-ExecutionProvider --- @@ -18,7 +19,7 @@ redirect_from: /docs/reference/execution-providers/ArmNN-ExecutionProvider Accelerate performance of ONNX model workloads across Armv8 cores with the ArmNN execution provider. [ArmNN](https://github.com/ARM-software/armnn) is an open source inference engine maintained by Arm and Linaro companies. ## Build -For build instructions, please see the [BUILD page](../build/eps.md#armnn). +For build instructions, please see the [BUILD page](../../build/eps.md#armnn). ## Usage ### C/C++ @@ -29,9 +30,9 @@ Ort::SessionOptions so; bool enable_cpu_mem_arena = true; Ort::ThrowOnError(OrtSessionOptionsAppendExecutionProvider_ArmNN(so, enable_cpu_mem_arena)); ``` -The C API details are [here](../get-started/with-c.md). +The C API details are [here](../../get-started/with-c.md). ## Performance Tuning -For performance tuning, please see guidance on this page: [ONNX Runtime Perf Tuning](../performance/tune-performance.md) +For performance tuning, please see guidance on this page: [ONNX Runtime Perf Tuning](../../performance/tune-performance.md) When/if using [onnxruntime_perf_test](https://github.com/microsoft/onnxruntime/tree/master/onnxruntime/test/perftest), use the flag -e armnn diff --git a/docs/execution-providers/CANN-ExecutionProvider.md b/docs/execution-providers/community-maintained/CANN-ExecutionProvider.md similarity index 93% rename from docs/execution-providers/CANN-ExecutionProvider.md rename to docs/execution-providers/community-maintained/CANN-ExecutionProvider.md index dc648cd5d1dbd..c1ca55fd9eae6 100644 --- a/docs/execution-providers/CANN-ExecutionProvider.md +++ b/docs/execution-providers/community-maintained/CANN-ExecutionProvider.md @@ -1,8 +1,9 @@ --- -title: CANN (Huawei) +title: Huawei - CANN description: Instructions to execute ONNX Runtime with the Huawei CANN execution provider -parent: Execution Providers -nav_order: 3 +grand_parent: Execution Providers +parent: Community-maintained +nav_order: 7 redirect_from: /docs/reference/execution-providers/CANN-ExecutionProvider --- @@ -32,11 +33,11 @@ Please reference table below for official CANN packages dependencies for the ONN ## Build -For build instructions, please see the [BUILD page](../build/eps.md#cann). +For build instructions, please see the [BUILD page](../../build/eps.md#cann). ## Install -Pre-built binaries of ONNX Runtime with CANN EP are published for most language bindings. Please reference [Install ORT](../install). +Pre-built binaries of ONNX Runtime with CANN EP are published for most language bindings. Please reference [Install ORT](../../install). ## Samples diff --git a/docs/execution-providers/MIGraphX-ExecutionProvider.md b/docs/execution-providers/community-maintained/MIGraphX-ExecutionProvider.md similarity index 82% rename from docs/execution-providers/MIGraphX-ExecutionProvider.md rename to docs/execution-providers/community-maintained/MIGraphX-ExecutionProvider.md index 758461ecb25cb..3914047f3c033 100644 --- a/docs/execution-providers/MIGraphX-ExecutionProvider.md +++ b/docs/execution-providers/community-maintained/MIGraphX-ExecutionProvider.md @@ -1,8 +1,9 @@ --- -title: MIGraphX (AMD) +title: AMD - MIGraphX description: Instructions to execute ONNX Runtime with the AMD MIGraphX execution provider -parent: Execution Providers -nav_order: 7 +grand_parent: Execution Providers +parent: Community-maintained +nav_order: 4 redirect_from: /docs/reference/execution-providers/MIGraphX-ExecutionProvider --- @@ -18,7 +19,7 @@ The [MIGraphX](https://github.com/ROCmSoftwarePlatform/AMDMIGraphX/) execution p {:toc} ## Build -For build instructions, please see the [BUILD page](../build/eps.md#amd-migraphx). +For build instructions, please see the [BUILD page](../../build/eps.md#amd-migraphx). ## Usage @@ -33,12 +34,12 @@ Ort::ThrowOnError(OrtSessionOptionsAppendExecutionProvider_MiGraphX(sf, device_i You can check [here](https://github.com/scxiao/ort_test/tree/master/char_rnn) for a specific c/c++ program. -The C API details are [here](../get-started/with-c.md). +The C API details are [here](../../get-started/with-c.md). ### Python When using the Python wheel from the ONNX Runtime build with MIGraphX execution provider, it will be automatically prioritized over the default GPU or CPU execution providers. There is no need to separately register the execution -provider. Python APIs details are [here](https://onnxruntime.ai/docs/api/python/api_summary.html). +provider. Python APIs details are [here](../../api/python/api_summary.html). *Note that the next release (ORT 1.10) will require explicitly setting the providers parameter if you want to use execution provider other than the default CPU provider when instantiating InferenceSession.* You can check [here](https://github.com/scxiao/ort_test/tree/master/python/run_onnx) for a python script to run an @@ -48,6 +49,6 @@ model on either the CPU or MIGraphX Execution Provider. MIGraphX providers an environment variable ORT_MIGRAPHX_FP16_ENABLE to enable the FP16 mode. ## Performance Tuning -For performance tuning, please see guidance on this page: [ONNX Runtime Perf Tuning](../performance/tune-performance.md) +For performance tuning, please see guidance on this page: [ONNX Runtime Perf Tuning](../../performance/tune-performance.md) When/if using [onnxruntime_perf_test](https://github.com/microsoft/onnxruntime/tree/master/onnxruntime/test/perftest#onnxruntime-performance-test), use the flag `-e migraphx` diff --git a/docs/execution-providers/RKNPU-ExecutionProvider.md b/docs/execution-providers/community-maintained/RKNPU-ExecutionProvider.md similarity index 88% rename from docs/execution-providers/RKNPU-ExecutionProvider.md rename to docs/execution-providers/community-maintained/RKNPU-ExecutionProvider.md index 049b60f22ca37..a4d0dd010ad5b 100644 --- a/docs/execution-providers/RKNPU-ExecutionProvider.md +++ b/docs/execution-providers/community-maintained/RKNPU-ExecutionProvider.md @@ -1,8 +1,9 @@ --- -title: RKNPU +title: Rockchip - RKNPU description: Instructions to execute ONNX Runtime on Rockchip NPUs with the RKNPU execution provider -parent: Execution Providers -nav_order: 11 +grand_parent: Execution Providers +parent: Community-maintained +nav_order: 5 redirect_from: /docs/reference/execution-providers/RKNPU-ExecutionProvider --- @@ -19,7 +20,7 @@ RKNPU DDK is an advanced interface to access Rockchip NPU. The RKNPU Execution P ## Build -For build instructions, please see the [BUILD page](../build/eps.md#rknpu). +For build instructions, please see the [BUILD page](../../build/eps.md#rknpu). ## Usage **C/C++** @@ -31,7 +32,7 @@ Ort::SessionOptions sf; Ort::ThrowOnError(OrtSessionOptionsAppendExecutionProvider_RKNPU(sf)); Ort::Session session(env, model_path, sf); ``` -The C API details are [here](../get-started/with-c.md). +The C API details are [here](../../get-started/with-c.md). ## Support Coverage diff --git a/docs/execution-providers/TVM-ExecutionProvider.md b/docs/execution-providers/community-maintained/TVM-ExecutionProvider.md similarity index 99% rename from docs/execution-providers/TVM-ExecutionProvider.md rename to docs/execution-providers/community-maintained/TVM-ExecutionProvider.md index d3b8b5e517a9c..04e58064f1267 100644 --- a/docs/execution-providers/TVM-ExecutionProvider.md +++ b/docs/execution-providers/community-maintained/TVM-ExecutionProvider.md @@ -1,8 +1,9 @@ --- -title: TVM (Apache) +title: Apache - TVM description: Instructions to execute ONNX Runtime with the Apache TVM execution provider -parent: Execution Providers -nav_order: 15 +grand_parent: Execution Providers +parent: Community-maintained +nav_order: 3 --- # TVM Execution Provider diff --git a/docs/execution-providers/Vitis-AI-ExecutionProvider.md b/docs/execution-providers/community-maintained/Vitis-AI-ExecutionProvider.md similarity index 97% rename from docs/execution-providers/Vitis-AI-ExecutionProvider.md rename to docs/execution-providers/community-maintained/Vitis-AI-ExecutionProvider.md index 3c9cb10822386..765d46b35a9ef 100644 --- a/docs/execution-providers/Vitis-AI-ExecutionProvider.md +++ b/docs/execution-providers/community-maintained/Vitis-AI-ExecutionProvider.md @@ -1,8 +1,9 @@ --- -title: Vitis AI +title: Xilinx - Vitis AI description: Instructions to execute ONNX Runtime on Xilinx devices with the Vitis AI execution provider -parent: Execution Providers -nav_order: 16 +grand_parent: Execution Providers +parent: Community-maintained +nav_order: 6 redirect_from: /docs/reference/execution-providers/Vitis-AI-ExecutionProvider --- @@ -39,7 +40,7 @@ The following table lists system requirements for running docker containers as w | Docker Version | 19\.03\.1 | ## Build -See [Build instructions](../build/eps.md#vitis-ai). +See [Build instructions](../../build/eps.md#vitis-ai). ### Hardware setup diff --git a/docs/execution-providers/community-maintained/index.md b/docs/execution-providers/community-maintained/index.md new file mode 100644 index 0000000000000..b62d67e3a0935 --- /dev/null +++ b/docs/execution-providers/community-maintained/index.md @@ -0,0 +1,10 @@ +--- +title: Community-maintained +parent: Execution Providers +has_children: true +nav_order: 11 +--- +# Community-maintained Providers +This list of providers for specialized hardware is contributed and maintained by ONNX Runtime community partners. + + diff --git a/docs/execution-providers/index.md b/docs/execution-providers/index.md index 4c2e4d102dc6b..34116ca5e61a8 100644 --- a/docs/execution-providers/index.md +++ b/docs/execution-providers/index.md @@ -28,12 +28,12 @@ ONNX Runtime supports many different execution providers today. Some of the EPs |CPU|GPU|IoT/Edge/Mobile|Other| ---|---|---|--- -|Default CPU|[NVIDIA CUDA](../execution-providers/CUDA-ExecutionProvider.md)|[Intel OpenVINO](../execution-providers/OpenVINO-ExecutionProvider.md)|[Rockchip NPU](../execution-providers/RKNPU-ExecutionProvider.md) (*preview*)| -|[Intel DNNL](../execution-providers/oneDNN-ExecutionProvider.md)|[NVIDIA TensorRT](../execution-providers/TensorRT-ExecutionProvider.md)|[ARM Compute Library](../execution-providers/ACL-ExecutionProvider.md) (*preview*)|[Xilinx Vitis-AI](../execution-providers/Vitis-AI-ExecutionProvider.md) (*preview*)| -|[TVM](../execution-providers/TVM-ExecutionProvider.md) (*preview*)|[DirectML](../execution-providers/DirectML-ExecutionProvider.md)|[Android Neural Networks API](../execution-providers/NNAPI-ExecutionProvider.md)|[Huawei CANN](../execution-providers/CANN-ExecutionProvider.md) (*preview*)| -|[Intel OpenVINO](../execution-providers/OpenVINO-ExecutionProvider.md)|[AMD MIGraphX](../execution-providers/MIGraphX-ExecutionProvider.md) (*preview*)|[ARM-NN](../execution-providers/ArmNN-ExecutionProvider.md) (*preview*)| +|Default CPU|[NVIDIA CUDA](../execution-providers/CUDA-ExecutionProvider.md)|[Intel OpenVINO](../execution-providers/OpenVINO-ExecutionProvider.md)|[Rockchip NPU](../execution-providers/community-maintained/RKNPU-ExecutionProvider.md) (*preview*)| +|[Intel DNNL](../execution-providers/oneDNN-ExecutionProvider.md)|[NVIDIA TensorRT](../execution-providers/TensorRT-ExecutionProvider.md)|[ARM Compute Library](../execution-providers/community-maintained/ACL-ExecutionProvider.md) (*preview*)|[Xilinx Vitis-AI](../execution-providers/community-maintained/Vitis-AI-ExecutionProvider.md) (*preview*)| +|[TVM](../execution-providers/community-maintained/TVM-ExecutionProvider.md) (*preview*)|[DirectML](../execution-providers/DirectML-ExecutionProvider.md)|[Android Neural Networks API](../execution-providers/NNAPI-ExecutionProvider.md)|[Huawei CANN](../execution-providers/community-maintained/CANN-ExecutionProvider.md) (*preview*)| +|[Intel OpenVINO](../execution-providers/OpenVINO-ExecutionProvider.md)|[AMD MIGraphX](../execution-providers/community-maintained/MIGraphX-ExecutionProvider.md) (*preview*)|[ARM-NN](../execution-providers/community-maintained/ArmNN-ExecutionProvider.md) (*preview*)| ||[AMD ROCm](../execution-providers/ROCm-ExecutionProvider.md) (*preview*)|[CoreML](../execution-providers/CoreML-ExecutionProvider.md) (*preview*)| -||[TVM](../execution-providers/TVM-ExecutionProvider.md) (*preview*)|[TVM](../execution-providers/TVM-ExecutionProvider.md) (*preview*)| +||[TVM](../execution-providers/community-maintained/TVM-ExecutionProvider.md) (*preview*)|[TVM](../execution-providers/community-maintained/TVM-ExecutionProvider.md) (*preview*)| ||[Intel OpenVINO](../execution-providers/OpenVINO-ExecutionProvider.md)|[Qualcomm SNPE](../execution-providers/SNPE-ExecutionProvider.md)| |[XNNPACK](../execution-providers/Xnnpack-ExecutionProvider.md)||[XNNPACK](../execution-providers/Xnnpack-ExecutionProvider.md)| diff --git a/docs/execution-providers/oneDNN-ExecutionProvider.md b/docs/execution-providers/oneDNN-ExecutionProvider.md index 353f8550c4e9d..baedeec2d3932 100644 --- a/docs/execution-providers/oneDNN-ExecutionProvider.md +++ b/docs/execution-providers/oneDNN-ExecutionProvider.md @@ -1,8 +1,8 @@ --- -title: oneDNN (Intel) +title: Intel - oneDNN description: Instructions to execute ONNX Runtime with the Intel oneDNN execution provider parent: Execution Providers -nav_order: 9 +nav_order: 4 redirect_from: /docs/reference/execution-providers/oneDNN-ExecutionProvider ---