Hi nxf60449,
Thanks for taking your time to run these tests.
Yes, the models I used are the same as you mentioned.
The results I mentioned are obtained by running them on dual A72 cores using "taskset -c 4-5". I am sorry that I missed out to mention the detail.
So I ran similar tests as you without using Taskset. The results I got are as follows:
*************************************************
Acceleration : CpuAcc
*************************************************
Model: mobilenet_v1_1.0_224_quant.tflite
ArmNN ---> 138.2 ms
TfLite ---> 116.6 ms
Model: mobilenet_v1_0.25_128_quant.tflite
ArmNN ---> 9.338 ms
TfLite ---> 6.44ms
*************************************************
Acceleration : VsiNpu
*************************************************
Model: mobilenet_v1_1.0_224_quant.tflite
ArmNN ---> 14.23 ms
TfLite ---> 12.28 ms
Model: mobilenet_v1_0.25_128_quant.tflite
ArmNN ---> 5.20 ms
TfLite ---> 2.25 ms
**************************************************
1. When you do not run them exclusievely on A72 cores, you can see improvement in performance with VsiNpu.
But on A72 with mobilenet_v1_0.25_128_quant.tflite, I do not see improvement with VsiNpu using ArmNN. May I know if you can also see this behavior and probably know the cause?
2. However, TfLite Interpreter always performs better than ArmNN. Should not ArmNN be more optimized compared to TfLite?
Maybe I am wrong, I believed ArmNN to perform better than TfLite.
Best Regards