<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Which device is supported to run Quantized Gemma Model Inference in i.MX Solutions</title>
    <link>https://community.nxp.com/t5/i-MX-Solutions/Which-device-is-supported-to-run-Quantized-Gemma-Model-Inference/m-p/2087368#M609</link>
    <description>&lt;P&gt;especially has any of these devices&amp;nbsp;NPU ported by NXP for llama.cpp ?&lt;/P&gt;</description>
    <pubDate>Fri, 25 Apr 2025 14:41:26 GMT</pubDate>
    <dc:creator>ramkumarkoppu_p</dc:creator>
    <dc:date>2025-04-25T14:41:26Z</dc:date>
    <item>
      <title>Which device is supported to run Quantized Gemma Model Inference</title>
      <link>https://community.nxp.com/t5/i-MX-Solutions/Which-device-is-supported-to-run-Quantized-Gemma-Model-Inference/m-p/2087255#M608</link>
      <description>&lt;P&gt;Hi,&amp;nbsp;&lt;/P&gt;&lt;P&gt;Out of i.MX RT700 and i.MX 95 devices, which device has full software support to run inference of Gen AI models like&amp;nbsp;Quantized Google's Gemma Model — first in Python, then in C/C++ using the these devices NPU? Specifically:&lt;/P&gt;&lt;UL&gt;&lt;LI&gt;&lt;P class=""&gt;Which device NPU support transformer-based architectures, or is it limited to CNNs?&lt;/P&gt;&lt;/LI&gt;&lt;LI&gt;&lt;P class=""&gt;Which inference frameworks are supported for GenAI on this e.IQ platform?&lt;/P&gt;&lt;/LI&gt;&lt;/UL&gt;</description>
      <pubDate>Fri, 25 Apr 2025 10:41:20 GMT</pubDate>
      <guid>https://community.nxp.com/t5/i-MX-Solutions/Which-device-is-supported-to-run-Quantized-Gemma-Model-Inference/m-p/2087255#M608</guid>
      <dc:creator>ramkumarkoppu_p</dc:creator>
      <dc:date>2025-04-25T10:41:20Z</dc:date>
    </item>
    <item>
      <title>Re: Which device is supported to run Quantized Gemma Model Inference</title>
      <link>https://community.nxp.com/t5/i-MX-Solutions/Which-device-is-supported-to-run-Quantized-Gemma-Model-Inference/m-p/2087368#M609</link>
      <description>&lt;P&gt;especially has any of these devices&amp;nbsp;NPU ported by NXP for llama.cpp ?&lt;/P&gt;</description>
      <pubDate>Fri, 25 Apr 2025 14:41:26 GMT</pubDate>
      <guid>https://community.nxp.com/t5/i-MX-Solutions/Which-device-is-supported-to-run-Quantized-Gemma-Model-Inference/m-p/2087368#M609</guid>
      <dc:creator>ramkumarkoppu_p</dc:creator>
      <dc:date>2025-04-25T14:41:26Z</dc:date>
    </item>
  </channel>
</rss>

