INT16 Quantization for YOLOv8 models on STM32N6570-DK
Hi everyone,
I'm working on deploying an object detection model on the STM32N6570-DK board. While I have successfully deployed it, I noticed a significant drop in accuracy during live testing compared to the floating-point YOLOv8 model.
I'm now considering using INT16 quantization instead of INT8. My question is: How can I quantize the model to INT16? The tutorial I’m following states that quantization is only available for UINT8 and INT8.
Additionally, is it possible to deploy an object detection model using X-CUBE-AI in STM32CubeMX? I haven’t been able to find any tutorials on object detection with X-CUBE-AI.
Any guidance or resources would be greatly appreciated!
Thanks!
