Bring your own model (BYOM)
Upload your own model directly into your Edge Impulse project (TensorFlow SavedModel, ONNX, or TensorFlow Lite)
Bring your own model or BYOM allows you to optimize and deploy your own pretrained model (TensorFlow SavedModel, ONNX, or TensorFlow Lite) to any edge device, directly from your Edge Impulse project.
Getting Started
First, create a new project in Edge Impulse.
Also make sure you have your own pretrained model available locally on your computer, in one of the following formats: TensorFlow SavedModel (saved_model.zip
), ONNX model (.onnx
) or TensorFlow Lite model (.tflite
)
For this guide, we will be uploading a pretrained image classification TFLite model for plant disease classification, downloaded from the TensorFlow Dev Hub.
Then, from the Dashboard of your Edge Impulse project under "Getting started", select Upload your model:
Step 1: Upload your model
Upload your trained model: Upload a TensorFlow SavedModel (
saved_model.zip
), ONNX model (.onnx
) or TensorFlow Lite model (.tflite
) to get started.Model performance: Do you want performance characteristics (latency, RAM and ROM) for a specific device? Select "No" to show the performance for a range of device types, or "Yes" to run performance profiling for any of our available officially supported Edge Impulse development platforms.
After configuring the settings for uploading your model, select Upload your model and wait for your model to upload, you can check the upload status via the "Upload progress" section.
ONNX models
When selecting an ONNX model, you can also upload a .npy
file to Upload representative features (Optional). If you upload a set of representative features - for example, your validation set - as an .npy
file we can automatically quantize this model for better on-device performance.
Step 2: Process your model
Depending on the model you have uploaded in Step 1, the configuration settings available for Step 2 will change.
For this guide, we have selected the following configuration model settings for optimal processing for an image classification model with input shape (300, 300, 3)
in RGB format, Classification model output and 16 output labels: Tomato Healthy, Tomato Septoria Leaf Spot, Tomato Bacterial Spot, Tomato Blight, Cabbage Healthy, Tomato Spider Mite, Tomato Leaf Mold, Tomato_Yellow Leaf Curl Virus, Soy_Frogeye_Leaf_Spot, Soy_Downy_Mildew, Maize_Ravi_Corn_Rust, Maize_Healthy, Maize_Grey_Leaf_Spot, Maize_Lethal_Necrosis, Soy_Healthy, Cabbage Black Rot
After configuring your model settings, select Save model to view your model's on-device performance information for both MCUs and microprocessors (if applicable, depending on your model's arena size).
Check model behavior
Optionally upload test data to ensure correct model settings and proper model processing:
Limitations
There are a couple of restrictions to converting models with our tooling:
The model must have 1 input tensor.
You need to have a batch dimension (e.g. if your input is an image of 320x320x3 then your input shape should be
(1,320,320,3)
), and the batch size must be equal to 1. For ONNX models you can use a variable batch size (we'll set it to 1).For classification and regression models: The model must have 1 output tensor, and the output tensor should have a batch dimension as well. E.g. a classification model with 3 classes should have output shape
(1,3)
.
Troubleshooting
--saved-model /tmp/saved_model does not exist:
If you encountered the following error:
Make sure to upload a .zip
archive containing at minimum a saved_model
directory that contains your saved_model.pb
.
Could not profile: No uploaded model yet
If you encounter the following error:
This often means that the model you are attempting to upload is unsupported. Only the following model formats are supported at this time:
TensorFlow SavedModel (in .zip archive)
ONNX (.onnx)
TensorFlow Lite (.tflite or .lite)
This model won’t run on MCUs.
If you see a message along these lines:
Then that means that part of your network is not supported by EON Compiler, and cannot run as-is on MCUs. A list of all supported ops can be found in our SDK; but there might be certain input/output types that are not supported even if the op is in this list (you'll get a proper error code if this is the case). If you're an enterprise customer please contact your solutions engineer, most of the times we can add the missing op within a few days.
Preprocessing your data using Edge Impulse DSP blocks
If you want to use our built-in processing blocks to preprocess your data, for example to turn your audio files into spectrograms before training your network, then you should:
Use the code in the edgeimpulse/processing-blocks repo to preprocess your data before training your model. This contains the Python version of the processing blocks.
Then, on the embedded side, call
extract_XXXX_features
to preprocess your sensor data, and pass the resulting features into theei_run_classifier
function.
Here's an end-to-end example of the embedded code to preprocess using an MFCC block:
Last updated