Fine Tuned YOLOv11-Based Road Sign Detection
Received: 16 April 2025 | Revised: 9 May 2025 and 28 May 2025 | Accepted: 31 May 2025 | Online: 2 August 2025
Corresponding author: Safa Teboulbi
Abstract
The rapid growth of autonomous driving and intelligent transportation systems has increased the need for accurate and efficient traffic sign detection. Recognizing traffic signs in real time plays a crucial role in enabling vehicles to understand and respond to dynamic road conditions, ensuring both safety and regulatory compliance. This study investigates the performance of three lightweight YOLOv11 variants, YOLOv11n, YOLOv11s, and YOLOv11m, for road sign detection, aiming to balance accuracy with computational efficiency for real-time deployment in resource-constrained environments. Each model was trained and evaluated using a consistent traffic sign dataset, with performance metrics including precision, recall, mean Average Precision (mAP), and F1 score. The YOLOv11n model demonstrated stable training behavior and achieved a peak mAP@0.5 of 0.52, with a mean F1 score of 0.47, indicating efficient detection of dominant classes but limited performance for underrepresented ones. The YOLOv11s showed improved generalization and localization abilities with a higher mAP@0.5 of 0.55 and a mean F1 score of 0.64, suggesting a balanced trade-off between speed and accuracy. The most advanced variant, YOLOv11m, achieved the highest mAP@0.5 of 0.70 and an F1 score of 0.63, demonstrating robust detection and convergence properties. However, all models exhibited difficulty in detecting rarely represented classes, such as "crosswalk," emphasizing the importance of dataset balancing. These findings confirm the suitability of these YOLOv11 variants for embedded traffic monitoring systems and highlight avenues for further improvement through data augmentation and fine-tuning.
Keywords:
road sign detection, fine-tuned YOLOv11, computer visionDownloads
References
T. Saidani, "Deep Learning Approach: YOLOv5-based Custom Object Detection," Engineering, Technology & Applied Science Research, vol. 13, no. 6, pp. 12158–12163, Dec. 2023. DOI: https://doi.org/10.48084/etasr.6397
A. Mogelmose, M. M. Trivedi, and T. B. Moeslund, "Vision-Based Traffic Sign Detection and Analysis for Intelligent Driver Assistance Systems: Perspectives and Survey," IEEE Transactions on Intelligent Transportation Systems, vol. 13, no. 4, pp. 1484–1497, Sep. 2012. DOI: https://doi.org/10.1109/TITS.2012.2209421
S. Saxena, S. Dey, M. Shah, and S. Gupta, "Traffic sign detection in unconstrained environment using improved YOLOv4," Expert Systems with Applications, vol. 238, Mar. 2024, Art. no. 121836. DOI: https://doi.org/10.1016/j.eswa.2023.121836
Y. Zhu, C. Zhang, D. Zhou, X. Wang, X. Bai, and W. Liu, "Traffic sign detection and recognition using fully convolutional network guided proposals," Neurocomputing, vol. 214, pp. 758–766, Nov. 2016. DOI: https://doi.org/10.1016/j.neucom.2016.07.009
R. Timofte, K. Zimmermann, and L. Van Gool, "Multi-view traffic sign detection, recognition, and 3D localisation," Machine Vision and Applications, vol. 25, no. 3, pp. 633–647, Apr. 2014. DOI: https://doi.org/10.1007/s00138-011-0391-3
D. Ciresan, U. Meier, and J. Schmidhuber, "Multi-column deep neural networks for image classification," in 2012 IEEE Conference on Computer Vision and Pattern Recognition, Jun. 2012, pp. 3642–3649. DOI: https://doi.org/10.1109/CVPR.2012.6248110
J. Stallkamp, M. Schlipsing, J. Salmen, and C. Igel, "Man vs. computer: Benchmarking machine learning algorithms for traffic sign recognition," Neural Networks, vol. 32, pp. 323–332, Aug. 2012. DOI: https://doi.org/10.1016/j.neunet.2012.02.016
P. Sanjeewani, B. Verma, and J. Affum, "Multi-stage Deep Learning Technique for Improving Traffic Sign Recognition," in 2021 36th International Conference on Image and Vision Computing New Zealand (IVCNZ), Tauranga, New Zealand, Dec. 2021, pp. 1–6. DOI: https://doi.org/10.1109/IVCNZ54163.2021.9653198
J. Redmon, S. Divvala, R. Girshick, and A. Farhadi, "You Only Look Once: Unified, Real-Time Object Detection," presented at the Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2016, pp. 779–788. DOI: https://doi.org/10.1109/CVPR.2016.91
Y. Peker, "Traffic Guider Object Detection Dataset by Yasin Peker." Roboflow Universe, 2023, [Online]. Available: https://universe.roboflow.com/yasin-peker-zeih0/traffic-guider.
M. Flores-Calero et al., "Traffic Sign Detection and Recognition Using YOLO Object Detection Algorithm: A Systematic Review," Mathematics, vol. 12, no. 2, Jan. 2024, Art. no. 297. DOI: https://doi.org/10.3390/math12020297
J. Redmon and A. Farhadi, "YOLOv3: An Incremental Improvement." arXiv, Apr. 08, 2018.
A. Bochkovskiy, C. Y. Wang, and H. Y. M. Liao, "YOLOv4: Optimal Speed and Accuracy of Object Detection." arXiv, Apr. 23, 2020.
M. Bakirci and I. Bayraktar, "The Cutting-Edge YOLO11 for Advanced Aircraft Detection in Synthetic Aperture Radar (SAR) Imagery," in 2024 8th International Symposium on Innovative Approaches in Smart Technologies (ISAS), İstanbul, Turkiye, Dec. 2024, pp. 1–6. DOI: https://doi.org/10.1109/ISAS64331.2024.10845222
Downloads
How to Cite
License
Copyright (c) 2025 Safa Teboulbi, Seifeddine Meesaoud, Mohamed Ali Hajjaji, Mohamed Atri, Abdellatif Mtibaa

This work is licensed under a Creative Commons Attribution 4.0 International License.
Authors who publish with this journal agree to the following terms:
- Authors retain the copyright and grant the journal the right of first publication with the work simultaneously licensed under a Creative Commons Attribution License that allows others to share the work with an acknowledgement of the work's authorship and initial publication in this journal.
- Authors are able to enter into separate, additional contractual arrangements for the non-exclusive distribution of the journal's published version of the work (e.g., post it to an institutional repository or publish it in a book), with an acknowledgement of its initial publication in this journal.
- Authors are permitted and encouraged to post their work online (e.g., in institutional repositories or on their website) after its publication in ETASR with an acknowledgement of its initial publication in this journal.
