A pedestrian detection model based on YOLO for dense scenes
CSTR:
Author:
Affiliation:

1.College of Electronic Information and Automation, Civil Aviation University of China, Tianjin 300300, China;2. Party Committee Office (President’s Office), Tianjin University of Technology, Tianjin 300384, China

Clc Number:

Fund Project:

  • Article
  • |
  • Figures
  • |
  • Metrics
  • |
  • Reference
  • |
  • Related
  • |
  • Cited by
  • |
  • Materials
  • |
  • Comments
    Abstract:

    To address challenges in pedestrian detection within dense scenes, including high crowd density, severe occlusion, and overlapping individuals, an improved you only look once (YOLO)-based algorithm is proposed. First, deformable convolutions are employed to replace standard convolutions, enhancing the model’s adaptability to variations in shape and appearance under occlusions. Second, a multi-dimensional attention module is designed to emphasize critical local regions and extract more precise feature information. Lastly, a diagonal difference intersection-over-union (IoU) loss function is introduced, which incorporates a measure of the Euclidean distance difference between the main diagonal points of predicted and ground truth bounding boxes, thereby enhancing detection accuracy and regression performance. Experimental results demonstrate that the enhanced algorithm achieves a mean average precision at IoU=0.5 (mAP50) of 75.1% on the public dense pedestrian dataset WiderPerson, an improvement of 1.8% over the original YOLOv5 model, showcasing superior detection performance.

    Reference
    Related
    Cited by
Get Citation

Shujian XING, Furong WANG, Hong WANG. A pedestrian detection model based on YOLO for dense scenes[J]. Optoelectronics Letters,2026,(4):229-235

Copy
Related Videos

Share
Article Metrics
  • Abstract:
  • PDF:
  • HTML:
  • Cited by:
History
  • Received:November 14,2024
  • Revised:August 13,2025
  • Adopted:
  • Online: April 03,2026
  • Published:
Article QR Code