Dense vehicle detection in rush hours is important for intelligent transportation systems. Most existing object detection methods can work well in off-peak vehicle detection for surveillance images. However, they may fail in dense vehicle detection in rush hours due to severe overlapping. To address this problem, a dense vehicle detection network is proposed by embedding the deformable channel-wise column transformer (DCCT) into the current you only look once (YOLO)-v5l network with a novel asymmetric focal loss (AF loss). The proposed DCCT fully extracts the column-wise occlusion information of vehicles in the images and guides the network to pay more attention to the visible area of partially occluded vehicles to improve the detection and positioning accuracy of weak feature targets. The proposed AF loss is used to balance the performance between easy and hard targets and address class imbalance. Extensive results demonstrate that the proposed network can accurately detect on-road densely located vehicles, even the minority classes in real time. Compared with the baseline YOLO-v5l, the mean average precision is improved by 3.93%, and it achieves comparable results with the existing state-of-the-art methods on the UA_Detrac dataset. |
ACCESS THE FULL ARTICLE
No SPIE Account? Create one
CITATIONS
Cited by 3 scholarly publications.
Transformers
Target detection
Surveillance
Detection and tracking algorithms
Visualization
Convolution
Intelligence systems