I am trying to get a model (SSD Mobilenet V2 fpnlite) to run exclusively on the NPU of the i.MX8M Plus. I am getting the following output:
INFO: Vx delegate: allowed_cache_mode set to 0.
INFO: Vx delegate: device num set to 0.
INFO: Vx delegate: allowed_builtin_code set to 0.
INFO: Vx delegate: error_during_init set to 0.
INFO: Vx delegate: error_during_prepare set to 0.
INFO: Vx delegate: error_during_invoke set to 0.
WARNING: Fallback unsupported op 32 to TfLite
INFO: Created TensorFlow Lite XNNPACK delegate for CPU.
I am not sure what part of my model isn't compatible as it is fully quantized. I tried the ssd_mobilenet_v2_coco_quant_postprocess.tflite model from the following page:
https://github.com/nxp-imx-support/imx-lane-detection?tab=readme-ov-file
According to that page, it runs on the NPU:
| i.MX 8M Plus | NPU | 11.03 ms | ./benchmark_model --graph=mobilenet_ssd_v2_coco_quant_postprocess.tflite --external_delegate_path=/usr/lib/libvx_delegate.so |
Unfortunately, I get the same output with this model as I do with the model I trained. Any suggestions on what might be causing the need to use the CPU?