By default, the edge-impulse-linux-runnner uses the unoptimized version of your model.
If you want to enable the quantized version, you can use edge-impulse-linux-runnner --quantized.
Feel free to check the help:
$> edge-impulse-linux-runner --help
Usage: edge-impulse-linux-runner [options]
Edge Impulse Linux runner 1.2.5
-V, --version output the version number
--model-file <file> Specify model file, if not provided the model will be fetched from Edge Impulse
--api-key <key> API key to authenticate with Edge Impulse (overrides current credentials)
--download <file> Just download the model and store it on the file system
--clean Clear credentials
--silent Run in silent mode, don't prompt for credentials
--quantized Download int8 quantized neural networks, rather than the float32 neural networks. These might run
faster on some architectures, but have reduced accuracy.
--enable-camera Always enable the camera. This flag needs to be used to get data from the microphone on some USB
--dev List development servers, alternatively you can use the EI_HOST environmental variable to specify the
Edge Impulse instance.
--verbose Enable debug logs
-h, --help output usage information
I hope this will help you get better speed performances. (And I can see on your project that the quantized version has also a better accuracy).
Hi and thanks for these tips I think I managed to upload both models, but I did it by changing the options in Edge Impulse Studio – Deployment – Linux boards – Options. The default selection was ‘Quantized’. Is this not the correct way to select which model to upload to the board?