Custom Convolution Quantization to run on NPU

キャンセル
次の結果を表示 
表示  限定  | 次の代わりに検索 
もしかして: 

Custom Convolution Quantization to run on NPU

558件の閲覧回数
taklause
Contributor II

Hello, I ll try to convert certain networks to the NPU with the best possible performance. 

Therefore I saw in the NXP IMX.8 ML Guide that the NPU provides faster inference for "per-tensor" quanized models. 

I saw in some example (ex. Posenet) that all the conv. layers have been quantized Layerwise ,not Channel-wise as a Tensorflow Default. 

I have yet not been able to quantize the convolution in a simple example in a Layer wise manner. Do you might have an example how to achieve this?

Thanks Daniel

 

 

 

0 件の賞賛
返信
2 返答(返信)

498件の閲覧回数
taklause
Contributor II

It can be done by using the eIQ converter tool. Thanks

0 件の賞賛
返信

537件の閲覧回数
Bio_TICFSL
NXP TechSupport
NXP TechSupport

Hello,

 

Do you see that examples with the NPU? It appears that do not work with ARM architectures.

 

Regards

0 件の賞賛
返信