eIQ Transfer Learning Lab with i.MX RT

Document created by Anthony Huereca Employee on Jun 24, 2019Last modified by Anthony Huereca Employee on Feb 26, 2020
Version 20Show Document
  • View in full screen mode

This lab will cover how to take an existing TensorFlow image classification model, and re-train it to categorize images of flowers. This is known as transfer learning. This updated model will then be converted into a TensorFlow Lite file. By using that file with the TensorFlow Lite inference engine that is part of NXPs eIQ package, the model can be ran on an i.MX RT embedded device. A camera attached to the board can then be used to look at photos of flowers and the model will determine what type of flowers the camera is looking at.


This lab can also be used without a camera+LCD, but the flowers images will need to be converted to a C array and loaded at compile time. 



Attached to this post you will find:

  • Photos to test out the new model
  • A lab document on how to do 'transfer learning' on a TensorFlow model and then run that model on the i.MX RT family.
  • The use of the camera+LCD is optional.
    • If have camera+LCD use: eIQ Transfer Learning Lab - With Camera.pdf
    • If do not have camera or LCD use: eIQ Transfer Learning Lab - Without Camera.pdf




This lab supports the following boards:




Updated January 2020 now that camera+LCD support are now in the SDK builder. 

4 people found this helpful