Hello everyone,
I’m looking for answers on deploying a model to the FRDM-MCXN947 board. I have a model already trained in TensorFlow Lite (float32). I used the eIQ environment to convert it to C source, and that worked well. However, I understand that the NPU only accepts int8 models—is that correct?
When I converted my model to TensorFlow Lite int8, I encountered an error during conversion, so it didn’t work. I tried manually adding quantization and dequantization nodes, but that also failed.
From the examples I’ve seen, when you use your dataset in eIQ to generate the model, it automatically adds quantization and dequantization nodes.
Finally, I used a Python script to convert my TensorFlow Lite model to int8 C source, and that worked, but the model’s output differs from my TensorFlow Lite int8 tests in Python.
I’d like to know: is there a way to convert a model to int8 using eIQ, or to add quantization/dequantization nodes?
Thank you very much.