I am trying to implement a YOLO model. I have customized the yolo26.yaml file to my own needs, where some of the operations are not supported by the TFLM library so I changed some of the conv blocks used. However, whenever I export it as tflite with int8 quantization using the export function, the output is always 0 for the detections. When I tested as .pt or .tflite_float32 the model performs as it should.
Could you please enlighten me on what could be the problem?
No, the output of int8 works normal. But I cannot use dynamic quantization because my main goal is to implement the model on an NPU equipped MCU. This dynamic quantization makes the input and output float32 and my converter does not accept it.
Is there any way to obtain static quantization with a working model? Like, should I use yolov8 or yolov5 to obtain this?