Hi, NXP supporter,
I'm tried running model mobilenet-v1 with tensorflow 1.15 and quantized with eIQ-tool. However, when I run on NPU it gets fallback error to CPU but still recognizes the object and I want it to run on NPU. I tried with exporting tflite_graph.pb to saved_model and quantization with tensorflow 2.16.2 environment. I used the following method to quant but it still doesn't work, how can I run it completely on NPU without falling back to CPU for processing.
Hello @The_Nguyen
I hope you are doing very well.
Please refer to chapter 2.7.1 Running the example on the i.MX 8 platform hardware accelerator of i.MX Machine Learning User's Guide.
There is described how you can run a mobilnet with hardware acceleration GPU/NPU.
Best regards,
Salas.
Hello Salas,
I followed the instructions in the document in 2.7.1, I was able to run the example model, but for the mobilenet-v1 model, I quantized it with eIQ-tool and it could not run completely on the NPU. So the reason is that the model has not been fully quantized to int8 or is my configuration incorrect.
I read in the document that in iMX8MP, it is required to be fully quantized to int8 to be able to run completely on the NPU without falling back to the CPU, is that correct?
Thanks and Regards,
TheNguyen