Text decoder on i.MX8 plus

キャンセル
次の結果を表示 
表示  限定  | 次の代わりに検索 
もしかして: 

Text decoder on i.MX8 plus

1,103件の閲覧回数
donnadamus
Contributor I

Hello everyone, 

I was wondering whether or not it is feasible to deploy text decoders based on attention mechanisms on the MX8 plus. 

Does the NPU support those layers and operations? 

Is it feasible, alternatively, to deploy it on the cpu? 

We are talking about a text decoder of 60M params

Thank you in advance

0 件の賞賛
返信
5 返答(返信)

1,069件の閲覧回数
Zhiming_Liu
NXP TechSupport
NXP TechSupport

Hello,

We verified such model with 1.1 billion parameters on i.MX8MPlus.

For more detail, please refer this WHITEPAPER

https://www.nxp.com/webapp/Download?colCode=GEN-AI-RAG-WHITEPAPER


Best Regards,
Zhiming

0 件の賞賛
返信

1,063件の閲覧回数
donnadamus
Contributor I

Hi Zhiming,

thank you for you answer. I will read the paper.

One question: if you were able to deploy such models on the i.MX 8M Plus, then why here https://www.nxp.com/docs/en/user-guide/IMX-MACHINE-LEARNING-UG.pdf (Chapter 11) I can't seem to find support for MultiHeadAttention layer? 

Maybe I'm not looking in the right place?

Thank you in advance for your time.

Kind Regards,

Marco Donnarumma

 

 

 

0 件の賞賛
返信

1,026件の閲覧回数
Zhiming_Liu
NXP TechSupport
NXP TechSupport

Hello,

The LLM project and LMM finetune tool with eIQ is not released. NXP will release the eIQ that support deploying LLM model.

Best Regards,
Zhiming

0 件の賞賛
返信

1,009件の閲覧回数
donnadamus
Contributor I

Hello,

 

do we know if the release will happen in the near future?

Thank you in advance.

 

Marco

 

0 件の賞賛
返信

998件の閲覧回数
Zhiming_Liu
NXP TechSupport
NXP TechSupport

Hello,

The demo demo is expected to be released at the end of 2025Q1, and then later for eIQ, the actual release date depends on the project schedule.

Best Regards,
Zhiming

0 件の賞賛
返信