Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
26 commits
Select commit Hold shift + click to select a range
d45304b
NXP backend: Update user guide and docs Readme (#14852)
roman-janik-nxp Oct 8, 2025
7505b16
Update top-level README.md file (#15049)
mergennachin Oct 13, 2025
15ccdf5
Fix documentation link for Core ATen operators (#15050)
mergennachin Oct 13, 2025
3c8f647
Fix various minor links in top-level README.md (#15052)
mergennachin Oct 13, 2025
68e587a
NXP Backend: Update Readme files (#14896)
robert-kalmar Oct 14, 2025
15be82f
[Samsung][docs] Update to the new template (#15087)
SS-JIA Oct 16, 2025
6108d6b
Add Pico2 Tutorials on Raspberry Pi (#15188)
psiddh Oct 17, 2025
c00cd39
Update mps docs and fix coreml/mps doc references (#15179)
metascroy Oct 17, 2025
a566c09
[ET-VK][docs] Update to the new template (#14996)
SS-JIA Oct 18, 2025
56ee96b
Success Stories page initial stage (#15236)
mergennachin Oct 18, 2025
923761c
Android Documentation Improvements and other fixes (#15260)
psiddh Oct 20, 2025
3d6b5d1
Updated Android doc with proper 1.0.0 backend links to executorch (#1…
psiddh Oct 20, 2025
4a73a87
Android Docs: Fix stale backend link (android-samsung-exynos) (#15287)
psiddh Oct 21, 2025
11ff6e9
Export LLMs with Optimum docs (#15062)
jackzhxng Oct 21, 2025
8d6d4d2
[ET-VK] Add redirect for backends-vulkan (#15305)
SS-JIA Oct 21, 2025
7aa15fe
Revise ExecuTorch documentation for Apple runtime (#15293)
shoumikhin Oct 21, 2025
146c8cb
Update docs on LMM runner Apple API (#15307)
shoumikhin Oct 21, 2025
8711ebd
Add Metal backend documentation to Voxtral README (#15273)
manuelcandales Oct 21, 2025
2e63fba
Update success-stories.md (#15309)
metascroy Oct 21, 2025
5f6167f
Add gemma to supported models (#15328)
lucylq Oct 21, 2025
cb63da3
Update build from source and getting started docs (#15311)
GregoryComer Oct 21, 2025
261eb2d
Fix more typos and broken links (#15331)
abhinaykukkadapu Oct 21, 2025
14e19e8
Update XNNPACK doc structure and add template (#14873)
GregoryComer Oct 20, 2025
571f925
Remove extra demo line from sucess-stories page (#15337)
psiddh Oct 22, 2025
5301a32
update backend cadence md for branch cut (#15277)
zonglinpeng Oct 21, 2025
89b5071
Minor doc fixes (#15336)
GregoryComer Oct 22, 2025
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
1 change: 0 additions & 1 deletion .gitignore
Original file line number Diff line number Diff line change
Expand Up @@ -62,7 +62,6 @@ xcuserdata/
/include/
/share/
/version.py
*.csv
*_etdump

# Android
Expand Down
8 changes: 4 additions & 4 deletions CONTRIBUTING.md
Original file line number Diff line number Diff line change
Expand Up @@ -24,17 +24,17 @@ For Apple, please refer to the [iOS documentation](docs/source/using-executorch-
executorch
├── <a href="backends">backends</a> - Backend delegate implementations for various hardware targets. Each backend uses partitioner to split the graph into subgraphs that can be executed on specific hardware, quantizer to optimize model precision, and runtime components to execute the graph on target hardware. For details refer to the <a href="docs/source/backend-delegates-integration.md">backend documentation</a> and the <a href="docs/source/using-executorch-export.md">Export and Lowering tutorial</a> for more information.
│ ├── <a href="backends/apple">apple</a> - Apple-specific backends.
│ │ ├── <a href="backends/apple/coreml">coreml</a> - CoreML backend for Apple devices. See <a href="docs/source/backends-coreml.md">doc</a>.
│ │ └── <a href="backends/apple/mps">mps</a> - Metal Performance Shaders backend for Apple devices. See <a href="docs/source/backends-mps.md">doc</a>.
│ │ ├── <a href="backends/apple/coreml">coreml</a> - CoreML backend for Apple devices. See <a href="docs/source/backends/coreml/coreml-overview.md">doc</a>.
│ │ └── <a href="backends/apple/mps">mps</a> - Metal Performance Shaders backend for Apple devices. See <a href="docs/source/backends/mps/mps-overview.md">doc</a>.
│ ├── <a href="backends/arm">arm</a> - ARM architecture backends. See <a href="docs/source/backends-arm-ethos-u.md">doc</a>.
│ ├── <a href="backends/cadence">cadence</a> - Cadence-specific backends. See <a href="docs/source/backends-cadence.md">doc</a>.
│ ├── <a href="backends/example">example</a> - Example backend implementations.
│ ├── <a href="backends/mediatek">mediatek</a> - MediaTek-specific backends. See <a href="docs/source/backends-mediatek.md">doc</a>.
│ ├── <a href="backends/openvino">openvino</a> - OpenVINO backend for Intel hardware.
│ ├── <a href="backends/qualcomm">qualcomm</a> - Qualcomm-specific backends. See <a href="docs/source/backends-qualcomm.md">doc</a>.
│ ├── <a href="backends/transforms">transforms</a> - Transformations for backend optimization.
│ ├── <a href="backends/vulkan">vulkan</a> - Vulkan backend for cross-platform GPU support. See <a href="docs/source/backends-vulkan.md">doc</a>.
│ └── <a href="backends/xnnpack">xnnpack</a> - XNNPACK backend for optimized neural network operations. See <a href="docs/source/backends-xnnpack.md">doc</a>.
│ ├── <a href="backends/vulkan">vulkan</a> - Vulkan backend for cross-platform GPU support. See <a href="docs/source/backends/vulkan/vulkan-overview.md">doc</a>.
│ └── <a href="backends/xnnpack">xnnpack</a> - XNNPACK backend for optimized neural network operations. See <a href="docs/source/backends/xnnpack/xnnpack-overview.md">doc</a>.
├── <a href="codegen">codegen</a> - Tooling to autogenerate bindings between kernels and the runtime.
├── <a href="configurations">configurations</a> - Configuration files.
├── <a href="devtools">devtools</a> - Model profiling, debugging, and inspection. Please refer to the <a href="docs/source/devtools-overview.md">tools documentation</a> for more information.
Expand Down
4 changes: 2 additions & 2 deletions README-wheel.md
Original file line number Diff line number Diff line change
Expand Up @@ -11,8 +11,8 @@ The `executorch` pip package is in beta.
The prebuilt `executorch.runtime` module included in this package provides a way
to run ExecuTorch `.pte` files, with some restrictions:
* Only [core ATen operators](docs/source/ir-ops-set-definition.md) are linked into the prebuilt module
* Only the [XNNPACK backend delegate](docs/source/backends-xnnpack.md) is linked into the prebuilt module.
* \[macOS only] [Core ML](docs/source/backends-coreml.md) and [MPS](docs/source/backends-mps.md) backend
* Only the [XNNPACK backend delegate](docs/source/backends/xnnpack/xnnpack-overview.md) is linked into the prebuilt module.
* \[macOS only] [Core ML](docs/source/backends/coreml/coreml-overview.md) and [MPS](docs/source/backends/mps/mps-overview.md) backend
are also linked into the prebuilt module.

Please visit the [ExecuTorch website](https://pytorch.org/executorch) for
Expand Down
12 changes: 4 additions & 8 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -104,16 +104,14 @@ outputs = method.execute([torch.randn(1, 3, 224, 224)])

Module module("model.pte");
auto tensor = make_tensor_ptr({2, 2}, {1.0f, 2.0f, 3.0f, 4.0f});
auto outputs = module.forward(tensor);
auto outputs = module.forward({tensor});
```

**[Swift (iOS)](https://docs.pytorch.org/executorch/main/ios-section.html)**
```swift
import ExecuTorch

let module = Module(filePath: "model.pte")
let input = Tensor<Float>([1.0, 2.0, 3.0, 4.0], shape: [2, 2])
let outputs = try module.forward(input)
let input = Tensor<Float>([1.0, 2.0, 3.0, 4.0])
let outputs: [Value] = try module.forward([input])
```

**[Kotlin (Android)](https://docs.pytorch.org/executorch/main/android-section.html)**
Expand Down Expand Up @@ -153,8 +151,6 @@ runner->generate("Hello, how are you?", config);

**[Swift (iOS)](https://docs.pytorch.org/executorch/main/llm/run-on-ios.html)**
```swift
import ExecuTorchLLM

let runner = TextRunner(modelPath: "llama.pte", tokenizerPath: "tiktoken.bin")
try runner.generate("Hello, how are you?", Config {
$0.sequenceLength = 128
Expand Down Expand Up @@ -202,7 +198,7 @@ ExecuTorch powers on-device AI at scale across Meta's family of apps, VR/AR devi

**LLMs:** [Llama 3.2/3.1/3](examples/models/llama/README.md), [Qwen 3](examples/models/qwen3/README.md), [Phi-4-mini](examples/models/phi_4_mini/README.md), [LiquidAI LFM2](examples/models/lfm2/README.md)

**Multimodal:** [Llava](examples/models/llava/README.md) (vision-language), [Voxtral](examples/models/voxtral/README.md) (audio-language)
**Multimodal:** [Llava](examples/models/llava/README.md) (vision-language), [Voxtral](examples/models/voxtral/README.md) (audio-language), [Gemma](examples/models/gemma3) (vision-language)

**Vision/Speech:** [MobileNetV2](https://github.com/meta-pytorch/executorch-examples/tree/main/mv2), [DeepLabV3](https://github.com/meta-pytorch/executorch-examples/tree/main/dl3), [Whisper](https://github.com/meta-pytorch/executorch-examples/tree/main/whisper/android/WhisperApp)

Expand Down
2 changes: 1 addition & 1 deletion backends/apple/coreml/README.md
Original file line number Diff line number Diff line change
@@ -1,7 +1,7 @@
# ExecuTorch Core ML Delegate

This subtree contains the Core ML Delegate implementation for ExecuTorch.
Core ML is an optimized framework for running machine learning models on Apple devices. The delegate is the mechanism for leveraging the Core ML framework to accelerate operators when running on Apple devices. To learn how to use the CoreML delegate, see the [documentation](https://github.com/pytorch/executorch/blob/main/docs/source/backends-coreml.md).
Core ML is an optimized framework for running machine learning models on Apple devices. The delegate is the mechanism for leveraging the Core ML framework to accelerate operators when running on Apple devices. To learn how to use the CoreML delegate, see the [documentation](https://github.com/pytorch/executorch/blob/main/docs/source/backends/coreml/coreml-overview.md).

## Layout
- `compiler/` : Lowers a module to Core ML backend.
Expand Down
2 changes: 1 addition & 1 deletion backends/cadence/build_cadence_fusionG3.sh
Original file line number Diff line number Diff line change
Expand Up @@ -9,7 +9,7 @@ set -euo pipefail

unset CMAKE_PREFIX_PATH
unset XTENSA_CORE
export XTENSA_CORE=FCV_FG3GP
export XTENSA_CORE=VANILLA_G3
git submodule sync
git submodule update --init
./backends/cadence/install_requirements.sh
Expand Down
2 changes: 1 addition & 1 deletion backends/cadence/build_cadence_hifi4.sh
Original file line number Diff line number Diff line change
Expand Up @@ -9,7 +9,7 @@ set -euo pipefail

unset CMAKE_PREFIX_PATH
unset XTENSA_CORE
export XTENSA_CORE=nxp_rt600_RI23_11_newlib
export XTENSA_CORE=VANILLA_HIFI
git submodule sync
git submodule update --init
./backends/cadence/install_requirements.sh
Expand Down
18 changes: 9 additions & 9 deletions backends/nxp/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -5,14 +5,14 @@ This subtree contains the ExecuTorch Backend implementation for the

The eIQ® Neutron NPU is a highly scalable accelerator core architecture providing machine learning (ML) acceleration,
able to support common and critical tasks for edge AI such as anomaly detection, speech recognition,
image classification, object detection, facial recognition, image segmentation, and generative AI use cases like
image classification, object detection, facial recognition, image segmentation, and generative AI use cases like
large and small language models (LLMs & SLMs) and text-to-speech (TTS).
The architecture provides power and performance optimized NPUs integrated with NXP's broad portfolio of
The architecture provides power and performance optimized NPUs integrated with NXP's broad portfolio of
microcontrollers and applications processors.

The eIQ Neutron NPUs offer support for a wide variety of neural network types such as CNN, RNN, TCN and Transformer
The eIQ Neutron NPUs offer support for a wide variety of neural network types such as CNN, RNN, TCN and Transformer
networks, as well as the ability to adapt and scale to new model architectures, topologies and layer types introduced
to AI workloads. ML application development with the eIQ Neutron NPU is fully supported by the
to AI workloads. ML application development with the eIQ Neutron NPU is fully supported by the
[eIQ machine learning software development environment](https://www.nxp.com/design/design-center/software/eiq-ml-development-environment/eiq-toolkit-for-end-to-end-model-development-and-deployment:EIQ-TOOLKIT).
The eIQ AI SW Stack provides a streamlined development experience for developers and end-users of NXP products.

Expand All @@ -22,7 +22,7 @@ At this moment following eIQ® Neutron NPU variants and NXP platforms are suppor

* **eIQ Neutron N3-64**, available on [i.MX RT700](https://www.nxp.com/products/i.MX-RT700)

In the future the NXP eIQ Neutron Backend will be extended to support [i.MX 9 Application Processors](https://www.nxp.com/products/processors-and-microcontrollers/arm-processors/i-mx-applications-processors/i-mx-9-processors:IMX9-PROCESSORS)
In the future the NXP eIQ Neutron Backend will be extended to support [i.MX 9 Application Processors](https://www.nxp.com/products/processors-and-microcontrollers/arm-processors/i-mx-applications-processors/i-mx-9-processors:IMX9-PROCESSORS)
with eIQ Neutron NPU, like the [i.MX 95](https://www.nxp.com/products/iMX95).


Expand All @@ -33,7 +33,7 @@ The eIQ Neutron NPU Backend should be considered as prototype quality at this mo
improvements. NXP and the ExecuTorch community is actively developing this codebase.

## Neutron Backend implementation and SW architecture
Neutron Backend uses the eIQ Neutron Converter as ML compiler to compile the delegated subgraph to Neutron microcode.
Neutron Backend uses the eIQ Neutron Converter as ML compiler to compile the delegated subgraph to Neutron microcode.
The Neutron Converter accepts the ML model in LiteRT format, for the **eIQ Neutron N3** class therefore the Neutron Backend
uses the LiteRT flatbuffers format as IR between the ExecuTorch and Neutron Converter ML compiler.

Expand All @@ -44,10 +44,10 @@ uses the LiteRT flatbuffers format as IR between the ExecuTorch and Neutron Conv
`node_conveters` is structured as single module for each Edge operator.
* `backend/ir/lib` - automatically generated handlers from LiteRT flatbuffers schema.
* `backend/ir/tflite_generator` and `backend/ir/tflite_optimizer` handle the serialization
of the in-memory built subgraph for delegation into LiteRT/TFLite flatbuffers
of the in-memory built subgraph for delegation into LiteRT/TFLite flatbuffers
representation. Code taken from the onnx2tflite tool.
* `edge_passes` - Various passes operating on Edge dialect level.
* `quantizer` - Neutron Backend quantizer implementation.
* `edge_passes` - Various passes operating on Edge dialect level.
* `quantizer` - Neutron Backend quantizer implementation.
* `runtime` - Neutron Backend runtime implementation. For running compiled on device.
* `tests/` - Unit tests for Neutron backend.
* `tests/converter/node_converter` - Operator level unit tests.
Expand Down
Loading
Loading