ONNX quantised Model

キャンセル
次の結果を表示 
表示  限定  | 次の代わりに検索 
もしかして: 

ONNX quantised Model

785件の閲覧回数
mahanad
Contributor I

Hi,

How can I reduce the inference time for an onnx model..  It's taking roughly 6 seconds.. 

I tried to quantise the model using eiq toolkit but when I tried to load the model it's giving me the following error...

terminate called after throwing an instance of 'Ort::Exception'
  what(): Fatal error: QLinearAdd is not a registered function/op
Aborted

 

onnx.png

 

 

 

 

 

Thanks in advance...

0 件の賞賛
返信
2 返答(返信)

767件の閲覧回数
mahanad
Contributor I

Can it be used with onnx? According to the documentation, it's only for tflite (I might be wrong)

0 件の賞賛
返信

776件の閲覧回数
Zhiming_Liu
NXP TechSupport
NXP TechSupport

You need use vx delegrate in your inference code.

Please see this guide :

https://www.nxp.com.cn/docs/en/user-guide/IMX-MACHINE-LEARNING-UG.pdf

0 件の賞賛
返信