1. Gaussian YOLOv3: An Accurate and Fast Object Detector Using Localization Uncertainty for Autonomous Driving
- Summary: this paper designed a new loss function for YOLOv3 to predict localization uncertainty. Previous work built a uncertainty model for two-stage object detection models. But two-stage object detection approaches have large computational cost and brings too much latency for inference. Thus, the popular object detection models in autonomous driving or other camera-based environment are one-stage detection models. Unfortunately, one-stage detection models are less accurate than two-stage approaches. The accuracy in real world setting brings too many problems of security in driving. Thus, they reconstructed a new loss function for object detection models and improved its accuracy without afftecting speed. Inspired by prior work, they adopted gaussian models to predict the range of localization and utilized localization uncertainty to train a robust one-stage detector. In experiments, they reduced the FP by 41.40% and 40.62%, respectively, and increases the TP by 7.26% and 4.3%, respectively, on the KITTI and BDD datasets.
- Modifications on YOLOv3:
Summary: this paper is an application-based CV work. They analyzed the weakness of single-shot object detection on small objects and designed a new single-shot object detection through combining low-/mid-level semantic information. Previous SSD networks used a top-down pyramidal feature representation to combine high-level semantic features and performed bad on small objects. Thus, they targeted to improve SSD networks within low-/mid-level semantic features. In the design, they used a light-weight scratch network (LSN) to extract low-/mid-level features and a bi-directional network to combine both low-/mid-level and high-level features for SSD networks. In the experiment, they applied this new object detection approach on two public datasets (COCO and UAVDT) and got a better performance than other SSD networks.