In Ultralytics HUB the COCO dataset includes segmentation polygons, not bounding boxes, but it can be used to train object detection dataset
are the polygon masks simply converted to bounding boxes during training
how are the Spatial Transforms Albumentations handled such as arbitrary rotations handled, is the bounding box used for training based on post transformer segmentation mask?
If Ultralytics’ affine transform is used on a dataset that contains segmentation masks, which is used to train an object detection model, is the conversion from segmentation masks to bounding boxes happening (1) before the affine transform, (2) after the affine transform
This is relevant because it is (2) then the bounding box could potentially be more accurate
note: btw Albumentations does support masks Targets by Transform, though I think it is on dense masks, not polygon masks
Just to help clarify (because I’m confused), are you asking about using segmentation annotations to train a standard “detect” (bounding boxes) model? I presume you’re asking about using segmentation annotations to train a segmentation model, but I want to be 100% certain in my understanding.