Rate this Page

NXP eIQ Neutron Backend#

This manual page is dedicated to introduction NXP eIQ Neutron backend. NXP offers accelerated machine learning models inference on edge devices. To learn more about NXP’s machine learning acceleration platform, please refer to the official NXP website.

For up-to-date status about running ExecuTorch on Neutron backend please visit the manual page.

Features#

ExecuTorch v1.0 supports running machine learning models on selected NXP chips (for now only i.MXRT700). Among currently supported machine learning models are:

  • Convolution-based neutral networks

  • Full support for MobileNetV2 and CifarNet

Target Requirements#

  • Hardware with NXP’s i.MXRT700 chip or a evaluation board like MIMXRT700-EVK.

Development Requirements#

$ pip install --index-url https://eiq.nxp.com/repository neutron_converter_SDK_25_12

Instead of manually installing requirements, except MCUXpresso IDE and SDK, you can use the setup script:

$ ./examples/nxp/setup.sh

Using NXP eIQ Backend#

To test the eIQ Neutron Backend, both AoT flow for model preparation and Runtime for execution, refer to the Getting started with eIQ Neutron NPU ExecuTorch backend

For a quick overview how to convert a custom PyTorch model, take a look at our example python script.

Runtime Integration#

An example runtime application using the eIQ NSYS (eIQ Neutron Simulator) is available examples/nxp/executor_runner, described in the tutorial Getting started with eIQ Neutron NPU ExecuTorch backend

To learn how to run the converted model on the NXP hardware, use one of our example projects on using ExecuTorch runtime from MCUXpresso IDE example projects list. For more finegrained tutorial, visit this manual page.

Reference#

Partitioner API — Partitioner options.

NXP eIQ Neutron Quantization — Supported quantization schemes.

NXP Tutorials — Tutorials.

NXP eIQ Dim Order Support — Dim order support (channels last inputs).