Does eIQ NPU on MCXN947 support transformer based model?

キャンセル
次の結果を表示 
表示  限定  | 次の代わりに検索 
もしかして: 

Does eIQ NPU on MCXN947 support transformer based model?

290件の閲覧回数
TomC818
Contributor III

Hello,
As mentioned in the title, can the NPU in the mcu support transformer based model? From the official website, it was claimed that the NPU can support transformer models.

TomC818_0-1755162965843.png
Assuming with compression, quantization and external RAM through flexspi, a small transformer based model can be fitted on board. Can the NPU actually be used to accelerate the inference? 

An example would be GitHub - maxbbraun/llama4micro: A "large" language model running on a microcontroller
Instead of running on the CPU, can the model be ran on the NPU? 
How can such a use case be adapted to MCXN?

 

ラベル(1)
タグ(2)
0 件の賞賛
返信
1 返信

239件の閲覧回数
Harry_Zhang
NXP Employee
NXP Employee

Hi @TomC818 

Full “LLM on NPU” isn’t a supported path yet on MCX.

LLMs rely on dynamic sequences, KV caches, etc.—many are outside the current eIQ Neutron TFLite op set.

BR

Harry

0 件の賞賛
返信