To run the inference using the GPU/NPU hardware accelerator, add the --use_nnapi=true command line argument:
For NPU: adb shell /data/local/tmp/benchmark_model --graph=/data/local/tmp/mobilenet_v2_1.0_224_quant.tflite --use_nnapi=true --nnapi_accelerator_name=vsi-npu
For GPU: adb root adb shell setprop vendor.USE_GPU_INFERENCE 1 adb shell /data/local/tmp/benchmark_model --graph=/data/local/tmp/mobilenet_v2_1.0_224_quant.tflite --use_nnapi=true
more detailed information, you can refer to the link as below,
"https://www.nxp.com/docs/en/user-guide/IMX_ANDROID_TENSORFLOWLITE_USERS_GUIDE.pdf"