Add the library to your application via:
$ npm install edge-impulse-linux
Before you can classify data you'll first need to collect it. If you want to collect data from the camera or microphone on your system you can use the Edge Impulse CLI, and if you want to collect data from different sensors (like accelerometers or proprietary control systems) you can do so in a few lines of code.
To collect data from the camera or microphone, follow the getting started guide for your development board.
To collect data from other sensors you'll need to write some code where you instantiate a
DataForwarder object, write data samples, and finally call
finalize() which uploads the data to Edge Impulse. Here's an end-to-end example.
To classify data (whether this is from the camera, the microphone, or a custom sensor) you'll need a model file. This model file contains all signal processing code, classical ML algorithms and neural networks - and typically contains hardware optimizations to run as fast as possible. To grab a model file:
Train your model in Edge Impulse.
Download the model file via:
$ edge-impulse-linux-runner --download modelfile.eim
This downloads the file into
modelfile.eim. (Want to switch projects? Add
Then you can start classifying realtime sensor data. We have examples for:
- Audio - grabs data from the microphone and classifies it in realtime.
- Audio (moving average filter) - as above, but shows how to use the moving-average filter to smooth your data and reduce false positives.
- Camera - grabs data from a webcam and classifies it in realtime.
- Custom data - classifies custom sensor data.
Updated about a month ago