2025-02-28 9:39 PM - edited 2025-02-28 9:40 PM
I am a new user of the STM32MP257F-EV1 board and am not very familiar with the ST Edge AI tool. I am currently trying to extract the TensorFlow Whisper encoder, convert it to an INT8 .tflite model using post-training quantization (PTQ), and then use the ST Edge AI tool to convert it into the .nb format for NPU acceleration.
I followed this procedure to perform the quantization:
After generating the whisper_encoder_int8.tflite model, I used the following command to convert it to .nb format:
However, I encountered the following error:
I am unsure what is causing this error. Could it be an issue with how I generated the .tflite model, maybe because the encoder has 4 layers? Or is there a limitation with the ST Edge AI tool regarding tensor shapes or dimensions?
I would appreciate any insights or guidance on resolving this issue.
Thank you!
2025-03-04 6:46 AM
Hello @Justin_wu ,
At first glance, the issue could be related to shape mismatches or dynamic shapes that are not interpretable.
It could be due to post-processing layers being removed by ST Edge AI, leading to errors, or certain layers not being supported on this model.
Additionally, it seems that the model quantization is done per-channel rather than per-tensor, which can significantly impact performance by running the model on the GPU instead of the NPU.
I will try to replicate to find where it comes from and update you.
Have a good day,
Julian