LLM on i.MX93

取消
显示结果 
显示  仅  | 搜索替代 
您的意思是: 

LLM on i.MX93

467 次查看
Behnam
Contributor I

I recently received my FRDM iMX93 board and am interested in running a minimal LLM or Transformer on it. I found a document by NXP (GEN-AI-RAG-WHITEPAPER) that states TinyLlama has successfully run on iMX93, but there is no more technical information or resources available.

Behnam_0-1753062161206.png

I know I need to ensure the model is compatible with TFLite, then quantize and compile it using eIQ and the Vela compiler, and finally run it on the iMX93 board. However, this seems like a long and complex process. I would prefer to have some kind of tutorial or at least a reliable reference to guide me through the steps.

 

0 项奖励
回复
4 回复数

60 次查看
Behnam
Contributor I

I was able to run the MobileBERT quantized model on i. MX93, but since the MatMul is not yet supported by the Vela Compiler (and u65), the performance does not seem very good. 

0 项奖励
回复

434 次查看
Zhiming_Liu
NXP TechSupport
NXP TechSupport

Hi,

The related document and toolkit are not ready to release.

Best Regards,
Zhiming

0 项奖励
回复

420 次查看
Behnam
Contributor I

Hi Zhimming

I see, so then can you share some high-level information that I can use to try to do it myself? For example, it would be helpful to know if I can use the eIQ for that or not, because I’ve tested a couple of models and it wasn’t successful for different reasons. I don’t need a full LLM right now, but a smaller language model like MobileNet would be helpful as well.

I’m looking forward to learning about the best practice for testing a small model and any details that could help me.

Best,
Behnam

0 项奖励
回复

365 次查看
Zhiming_Liu
NXP TechSupport
NXP TechSupport

Hi,

You can refer the eIQ user guide to export tflite model.

For model deploy, please refer this guide:

https://www.nxp.com/docs/en/user-guide/UG10166.pdf

Best Regards,
Zhiming

0 项奖励
回复