Hello everyone,
I’m currently working on a UAV-based object detection project and would really appreciate insights from the community.
I have several years of experience in AI, mainly focused on deep learning and computer vision. Over the past ~1 year, I’ve worked extensively with YOLO-based models across multiple object detection tasks.
Current Challenge
In this UAV scenario, I’m facing persistent issues with both false positives (FP) and false negatives (FN) under challenging conditions:
- Objects near image boundaries (edges/corners) are often missed or poorly localized
- Motion blur due to UAV movement significantly degrades detection performance
- Sensor noise / low-quality frames further reduce object visibility
Overall, the model struggles to maintain stable precision and recall in real flight conditions.
What I’ve Tried So Far
- Augmentations: mosaic, translation, scaling, HSV, rotation, etc.
- Adjusting confidence and IoU thresholds
- Negative sampling → significantly reduced false positives
- Experimented with multiple YOLO11-based model variants like MASF-Yolo
- Adjusted loss weights to emphasize primary objects
- Tried SAHI (slicing-based inference) for small objects → improved detection in some cases but introduced many additional false positives, making it less usable in my pipeline
Despite these efforts, performance remains inconsistent, especially in edge cases.
Now I 'd like to focus on improving two areas: Dataset, Loss function.
1. What are the most effective dataset strategies to improve robustness for edge cases like objects near image boundaries, motion blur, and noisy frames?
2. Which loss functions or training strategies are best for balancing false positives and false negatives in challenging UAV scenarios?
I know this is challenge but my goal is to develop a model that performs reliably in real UAV conditions, even with degraded image quality, while maintaining strong precision and recall.
Any suggestions, references, or shared experiences would be greatly appreciated.
Thanks in advance!