Solved! Go to Solution.
Oh, I have solved this Err by add necessary MODEL_RegisterOps
Hi,
Thanks for your reply and I'm glad to hear that your issue is solved.
TIC
-------------------------------------------------------------------------------
Note:
- If this post answers your question, please click the "Mark Correct" button. Thank you!
- We are following threads for 7 weeks after the last post, later replies are ignored
Please open a new thread and refer to the closed one, if you have a related question at a later point in time.
-------------------------------------------------------------------------------
Hi,
Thank you for your interest in NXP Semiconductor products and for the opportunity to serve you.
1) what is the maximum size of model that my kit can support? Thanks in advance.
-- In my opinion, the maximum size of the runnable model is determined by the memory source of the MCU or GPU, not the TensorFlow lite library.
Maybe you can use the model. summary() command to list the number of weights of the model to estimate the tensor arena, to be honest, different model architectures have different sizes and numbers of input, output, and intermediate tensors, so it’s difficult to know how much memory we’ll need. The number doesn’t need to be exact—we can reserve more memory than we need and we usually do this work through trial and error.
TIC
-------------------------------------------------------------------------------
Note:
- If this post answers your question, please click the "Mark Correct" button. Thank you!
- We are following threads for 7 weeks after the last post, later replies are ignored
Please open a new thread and refer to the closed one, if you have a related question at a later point in time.
-------------------------------------------------------------------------------
Hi, thanks for quick response,
I received this error when try to embedd my model into NXP kit. Is this cause by my config or my tensor model? Or I need to config maximum size of model or number of output?
Thanks
My model: