Hello
While reading the GEN-AI-RAG-WHITEPAPER.pdf provided by NXP, I found that the TinyLlama-1B quantized model was ported to the i.MX95 device using the Neutron Execution Provider within ONNXRT runtime.
I have a few questions regarding this:
A detailed explanation would be greatly appreciated. Thank you in advance for your support!
Hello,
There is no public information available to explain this process, so please wait for the eIQ tool and BSP to be officially released.
Best Regards,
Zhiming