<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic very slow inference with pyEIQ with ONNX parser in eIQ Machine Learning Software</title>
    <link>https://community.nxp.com/t5/eIQ-Machine-Learning-Software/very-slow-inference-with-pyEIQ-with-ONNX-parser/m-p/1194927#M328</link>
    <description>&lt;P&gt;Hi,&lt;/P&gt;&lt;P&gt;Previously I used armnn 20.x built myself.&amp;nbsp; I converted ONNX to armmn first. Then I measured time and it was 80ms-300ms in depending on model and input size. Backend was 'CpuAcc'&lt;/P&gt;&lt;P&gt;Now I'm trying to use pyEIQ from BSP and get time ~3 seconds with 'VsiNpu' and ~1.5 second with 'CpuAcc' on middle model.&lt;/P&gt;&lt;P&gt;I found it strange. Any ideas why is it and how to solve it?&lt;/P&gt;&lt;P&gt;(I'm not sure but I remember that I run my models with 'VsiNpu' and there was a significant gain in inference speed. I wrote down time but forget way I got it)&lt;/P&gt;&lt;P&gt;I suppose issue may be with ONNXParser, but I can't check .armnn model directly cause libarmnnSerializer.so hasn't built.&lt;/P&gt;&lt;P&gt;so few more questions:&lt;/P&gt;&lt;P&gt;1) how load .armnn model?&lt;/P&gt;&lt;P&gt;2) what armnn version and patches you uses. I'm think it's not 19.08, cause pyarmmn started from 20.x version&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Thank in advance&lt;/P&gt;</description>
    <pubDate>Mon, 07 Dec 2020 18:54:13 GMT</pubDate>
    <dc:creator>korabelnikov</dc:creator>
    <dc:date>2020-12-07T18:54:13Z</dc:date>
    <item>
      <title>very slow inference with pyEIQ with ONNX parser</title>
      <link>https://community.nxp.com/t5/eIQ-Machine-Learning-Software/very-slow-inference-with-pyEIQ-with-ONNX-parser/m-p/1194927#M328</link>
      <description>&lt;P&gt;Hi,&lt;/P&gt;&lt;P&gt;Previously I used armnn 20.x built myself.&amp;nbsp; I converted ONNX to armmn first. Then I measured time and it was 80ms-300ms in depending on model and input size. Backend was 'CpuAcc'&lt;/P&gt;&lt;P&gt;Now I'm trying to use pyEIQ from BSP and get time ~3 seconds with 'VsiNpu' and ~1.5 second with 'CpuAcc' on middle model.&lt;/P&gt;&lt;P&gt;I found it strange. Any ideas why is it and how to solve it?&lt;/P&gt;&lt;P&gt;(I'm not sure but I remember that I run my models with 'VsiNpu' and there was a significant gain in inference speed. I wrote down time but forget way I got it)&lt;/P&gt;&lt;P&gt;I suppose issue may be with ONNXParser, but I can't check .armnn model directly cause libarmnnSerializer.so hasn't built.&lt;/P&gt;&lt;P&gt;so few more questions:&lt;/P&gt;&lt;P&gt;1) how load .armnn model?&lt;/P&gt;&lt;P&gt;2) what armnn version and patches you uses. I'm think it's not 19.08, cause pyarmmn started from 20.x version&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Thank in advance&lt;/P&gt;</description>
      <pubDate>Mon, 07 Dec 2020 18:54:13 GMT</pubDate>
      <guid>https://community.nxp.com/t5/eIQ-Machine-Learning-Software/very-slow-inference-with-pyEIQ-with-ONNX-parser/m-p/1194927#M328</guid>
      <dc:creator>korabelnikov</dc:creator>
      <dc:date>2020-12-07T18:54:13Z</dc:date>
    </item>
    <item>
      <title>Re: very slow inference with pyEIQ with ONNX parser</title>
      <link>https://community.nxp.com/t5/eIQ-Machine-Learning-Software/very-slow-inference-with-pyEIQ-with-ONNX-parser/m-p/1195467#M329</link>
      <description>&lt;P&gt;upd.&lt;/P&gt;&lt;P&gt;The maximum speed was when I used quantized tflite models.&lt;/P&gt;&lt;P&gt;As instance, resnet 18 takes ~0.5 seconds on tflite and VsiNpu but on armnn and VsiNpu take several seconds.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;So, can such huge performance difference be because of armnn engine?&lt;/P&gt;</description>
      <pubDate>Tue, 08 Dec 2020 11:45:43 GMT</pubDate>
      <guid>https://community.nxp.com/t5/eIQ-Machine-Learning-Software/very-slow-inference-with-pyEIQ-with-ONNX-parser/m-p/1195467#M329</guid>
      <dc:creator>korabelnikov</dc:creator>
      <dc:date>2020-12-08T11:45:43Z</dc:date>
    </item>
  </channel>
</rss>

