Text decoder on i.MX8 plus

取消
显示结果 
显示  仅  | 搜索替代 
您的意思是: 

Text decoder on i.MX8 plus

1,085 次查看
donnadamus
Contributor I

Hello everyone, 

I was wondering whether or not it is feasible to deploy text decoders based on attention mechanisms on the MX8 plus. 

Does the NPU support those layers and operations? 

Is it feasible, alternatively, to deploy it on the cpu? 

We are talking about a text decoder of 60M params

Thank you in advance

0 项奖励
回复
5 回复数

1,051 次查看
Zhiming_Liu
NXP TechSupport
NXP TechSupport

Hello,

We verified such model with 1.1 billion parameters on i.MX8MPlus.

For more detail, please refer this WHITEPAPER

https://www.nxp.com/webapp/Download?colCode=GEN-AI-RAG-WHITEPAPER


Best Regards,
Zhiming

0 项奖励
回复

1,045 次查看
donnadamus
Contributor I

Hi Zhiming,

thank you for you answer. I will read the paper.

One question: if you were able to deploy such models on the i.MX 8M Plus, then why here https://www.nxp.com/docs/en/user-guide/IMX-MACHINE-LEARNING-UG.pdf (Chapter 11) I can't seem to find support for MultiHeadAttention layer? 

Maybe I'm not looking in the right place?

Thank you in advance for your time.

Kind Regards,

Marco Donnarumma

 

 

 

0 项奖励
回复

1,008 次查看
Zhiming_Liu
NXP TechSupport
NXP TechSupport

Hello,

The LLM project and LMM finetune tool with eIQ is not released. NXP will release the eIQ that support deploying LLM model.

Best Regards,
Zhiming

0 项奖励
回复

991 次查看
donnadamus
Contributor I

Hello,

 

do we know if the release will happen in the near future?

Thank you in advance.

 

Marco

 

0 项奖励
回复

980 次查看
Zhiming_Liu
NXP TechSupport
NXP TechSupport

Hello,

The demo demo is expected to be released at the end of 2025Q1, and then later for eIQ, the actual release date depends on the project schedule.

Best Regards,
Zhiming

0 项奖励
回复