Abstract
The effective detection of high beam vehicles is useful to reduce the occurrence of traffic accidents and light pollution at night. It is time-consuming and laborious to detect the high beam vehicles by human. We propose a model Global Attention Content-Aware ReAssembly of FEatures Small Target Detection based on YOLOv8 (GCS-YOLOv8) to improve the efficiency of high beam vehicles detection. Firstly, a cross-stage feature fusion global attention module (CGM) is designed in the backbone to reduce the information loss during down-sampling and magnify global dimension-interactive features. Secondly, a Convolution-1*1 Content-Aware ReAssembly of FEatures Module (CCM) is designed in the neck to reconstruct subtle feature maps of high beam vehicles. Meanwhile, a CCM Small Target Detection Module (CSM) is designed in the neck to derive the feature maps with small receptive field. Finally, we set up a high beam vehicles dataset. The model GCS-YOLOv8 achieves the mean average precision (mAp) of 94% on the test set, and it is higher than other models. In addition, multiple ablation experiments are done to prove the effectiveness of our model GCS-YOLOv8.
Similar content being viewed by others
Data availability
No datasets were generated or analyzed during the current study.
References
Wu YQ, Hua JF, Chen Y et al. (2018) Difficulties and Counter-measures for illegal use of driving beam headlamp of motor vehicles. China Public Security (Academy Edition) (1):78–81
Su JY, Zhang WW, Wu XC et al (2018) The driving beam recognition of vehicle based on multi-structure feature extraction and trajectory tracking. J Electron Meas Instrum 32(10):103–110
Fernandez-Carrobles MM, Deniz O, Maroto F (2019) Gun and knife detection based on faster R-CNN for video surveillance. In: Recognition P, Analysis I (eds) Madrid. Springer, Spain, pp 441–452
Girshick R (2015) Fast r-cnn Proceedings of the IEEE International Conference on Computer Vision, pp 1440–1448
Girshick R, Donahue J, Darrell T, Malik J (2016) Region-based convolutional networks for accurate object detection and segmentation. IEEE Trans Pattern Anal Mach Intell 38(1):142–158. https://doi.org/10.1109/TPAMI.2015.2437384
Mittal U, Chawla P, Tiwari R (2023) EnsembleNet: A hybrid approach for vehicle detection and estimation of traffic density based on faster R-CNN and YOLO models. Neural Comput Appl 35(6):4755–4774
Jiang P, Ergu D, Liu F, Cai Y, Ma B (2022) A review of YOLO algorithm developments. Proc Comput Sci 199:1066–1073
Akshatha KR, Karunakar AK, Shenoy SB, Pai AK, Nagaraj NH, Rohatgi SS (2022) Human detection in aerial thermal images usingfaster R-CNN and SSD algorithms. Electronics 11(7):1151
Varghese R, Sambath M (2024) YOLOv8: A novel object detection algorithm with enhanced performance and robustness. In: 2024 International Conference on Advances in Data Engineering and Intelligent Computing Systems (ADICS). IEEE, pp 1–6
Mnih V, Heess N, Graves A, Kavukcuoglu K (2014) Recurrent models of visual attention. NeurIPS 2204–2212
Jaderberg M, Simonyan K, Zisserman A, Kavukcuoglu K (2015) Spatial transformer networks. NeurIPS 2017–2025
Hu J, Shen L, Albanie S, Sun G, Wu E (2020) Squeeze-and-excitation networks. IEEE TPAMI 42(8):2011–2023
Woo S, Park J, Lee J, Kweon IS (2018) CBAM: Convolutional block attention module. ECCV 3–19
Dosovitskiy A, Beyer L, Kolesnikov A, Weissenborn D, Zhai X, Unterthiner T, Dehghani M, Minderer M, Heigold G, Gelly S, Uszkoreit J, Houlsby N (2021) An image is worth 16 x 16 words: Transformers for image recognition at scale. ICLR 1–21
Liu Y, Shao Z, Hoffmann N (2021) Global attention mechanism: Retain information to enhance channel-spatial interactions. ar**v preprint ar**v:2112.05561
Dumoulin V, Visin F (2016) A guide to convolution arithmetic for deep learning. ar**v preprint ar**v:1603.07285
Shi W, Caballero J, Huszár F et al. (2016) Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. pp 1874–1883
Tian Z, He T, Shen C et al. (2019) Decoders matter for semantic segmentation: Data-dependent decoding enables flexible feature aggregation. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp 3126–3135
Hu X, Mu H, Zhang X et al. (2019) Meta-SR: A magnification-arbitrary network for super-resolution. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp 1575–1584
Wang J, Chen K, Xu R et al. (2019) Carafe: Content-aware reassembly of features. In: Proceedings of the IEEE/CVF international conference on computer vision. pp 3007–3016
Wang Q, Zhang H, Hong X, Zhou Q (2021) “Small object detection based on modified FSSD and model compression. In Proceedings IEEE 6th International Conference Signal Image Process. (ICSIP), pp 88–92
Lim J-S, Astrid M, Yoon H-J, Lee S-I (2021) Small object detection using context and attention. In Proceedings International Conference Artificial Intelligence Information Communication, (ICAIIC), pp 181–186
Li S et al. (2024) Improved YOLOv5s Algorithm for Small Target Detection in UAV Aerial Photography. In: IEEE Access, vol. 12, pp 9784–9791
Zhu X, Lyu S, Wang X, Zhao Q (2021) ‘TPH-YOLOv5: Improved YOLOv5 based on transformer prediction head for object detection on drone-captured scenarios. In: Proceedings IEEE/CVF International Conference Computing Vision. Workshops (ICCVW), pp 2778–2788
Fang H, Ding L, Wang L, Chang Y, Yan L, Han J (2022) Infrared small UAV target detection based on depthwise separable residual dense network and multiscale feature fusion. IEEE Trans Instrum Meas 71:1–20
Yang L, Song Q, Wang Z et al (2020) Hier R-CNN: Instance-level human parts detection and a new benchmark. IEEE Trans Image Process 30:39–54
Fei W, Xiaoping ZHU, Zhou Z et al (2024) Deep-reinforcement-learning-based UAV autonomous navigation and collision avoidance in unknown environments. Chin J Aeronaut 37(3):237–257
Zhou X, Xu X, Liang W et al (2021) Intelligent small object detection for digital twin in smart manufacturing with industrial cyber-physical systems. IEEE Trans Industr Inf 18(2):1377–1386
Wang F, Wang H, Qin Z et al. (2023) UAV target detection algorithm based on improved YOLOv8. IEEE Access
Li C, Li L, Jiang H et al. (2022) YOLOv6: A single-stage object detection framework for industrial applications. arXiv preprint arXiv:2209.02976
Wang CY, Bochkovskiy A, Liao HYM (2023) YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp 7464–7475
Wang CY, Yeh IH, Mark Liao HY (2025) Yolov9: Learning what you want to learn using programmable gradient information. In: European Conference on Computer Vision. Springer, Cham, pp 1–21
Wang A, Chen H, Liu L et al. (2024) Yolov10: Real-time end-to-end object detection. arXiv preprint arXiv:2405.14458
Khanam R, Hussain M (2024) YOLOv11: An overview of the key architectural enhancements. arXiv preprint arXiv:2410.17725
Zhao Y, Lv W, Xu S et al. (2024) Detrs beat yolos on real-time object detection. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp 16965–16974
Acknowledgements
This work is supported by R&D Program of Beijing Municipal Education Commission (KM202410017006, KM202210017006), Beijing Digital Education Research Project (BDEC2022619048), Ningxia Natural Science Foundation General Project (2022AAC03757, 2023AAC03889) and Ministry of Education Industry-School Cooperative Education Project (220607039172210, 22107153134955).
Funding
This study was funded by the General Project of Science and Technology Plan of Beijing Municipal Education Commission, KM202210017006.
Author information
Authors and Affiliations
Contributions
Conceptualization was performed by Lili Zhang; methodology was conducted by Ke Zhang; software and validation were developed by Li Jing; formal analysis and investigation were carried out by Wei Wei; resources were provided by Xudong Yang; data curation was contributed by Hongxin Tan; writing—original draft preparation was prepared by Yucheng Han; writing—review and editing were drafted by KangYang and Lili Zhang; visualization was analyzed by Kang Yang; project administration was approved by Pei Yu. All authors have read and agreed to the published version of the manuscript.
Corresponding author
Ethics declarations
Conflicts of interest
The authors declare no competing interests.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Zhang, L., Zhang, K., Yang, K. et al. Driving risks from light pollution: an improved YOLOv8 detection network for high beam vehicle image recognition. J Supercomput 81, 275 (2025). https://doi.org/10.1007/s11227-024-06809-z
Accepted:
Published:
Version of record:
DOI: https://doi.org/10.1007/s11227-024-06809-z