Substation Danger Sign Detection and Recognition using Convolutional Neural Networks
Received: 14 November 2022 | Revised: 7 December 2022 | Accepted: 15 December 2022 | Online: 5 February 2023
Corresponding author: Guo Wang
Abstract
This paper focuses on the training of a deep neural network regarding danger sign detection and recognition in a substation. It involved applying the concepts of neural networks and computer vision to achieve results similar to traffic sign and number plate detection systems. The input data were captured in three distinct formats, i.e. grayscale, RGB, and YCbCr, which have been used as a base for comparison in this paper. The efficiency of the neural network was tested on a unique data set involving danger signs present in industrial and processing facilities. The data set was unique, consisting of four distinct symbols. The trained data were selected so that they would not facilitate overfitting and also would not be under fitted. The accuracy of the model varied with the input type and was tested with two distinct classifiers, CNN and SVM, and the results were compared. The model was designed to be fast and accurate, and it can be implemented on mobile devices.
Keywords:
CNN, Neural Network, comparative study, danger sign detectionDownloads
References
H. S. Lee and K. Kim, "Simultaneous Traffic Sign Detection and Boundary Estimation Using Convolutional Neural Network," IEEE Transactions on Intelligent Transportation Systems, vol. 19, no. 5, pp. 1652–1663, Feb. 2018. DOI: https://doi.org/10.1109/TITS.2018.2801560
"US smartphone ownership 2021," Statista. https://www.statista.com/statistics/219865/percentage-of-us-adults-who-own-a-smartphone/.
"Mobile Fact Sheet," Pew Research Center: Internet, Science & Tech. https://www.pewresearch.org/internet/fact-sheet/mobile/.
D. G. Gibson, A. Pereira, B. A. Farrenkopf, A. B. Labrique, G. W. Pariyo, and A. A. Hyder, "Mobile Phone Surveys for Collecting Population-Level Estimates in Low- and Middle-Income Countries: A Literature Review," Journal of Medical Internet Research, vol. 19, no. 5, May 2017, Art. no. e7428. DOI: https://doi.org/10.2196/jmir.7428
G. Montavon, W. Samek, and K.-R. Muller, "Methods for interpreting and understanding deep neural networks," Digital Signal Processing, vol. 73, pp. 1–15, Feb. 2018. DOI: https://doi.org/10.1016/j.dsp.2017.10.011
S. Song, Z. Que, J. Hou, S. Du, and Y. Song, "An efficient convolutional neural network for small traffic sign detection," Journal of Systems Architecture, vol. 97, pp. 269–277, Aug. 2019. DOI: https://doi.org/10.1016/j.sysarc.2019.01.012
A. Bouti, M. A. Mahraz, J. Riffi, and H. Tairi, "A robust system for road sign detection and classification using LeNet architecture based on convolutional neural network," Soft Computing, vol. 24, no. 9, pp. 6721–6733, May 2020. DOI: https://doi.org/10.1007/s00500-019-04307-6
R. Antar, S. Alghamdi, J. Alotaibi, and M. Alghamdi, "Automatic Number Plate Recognition of Saudi License Car Plates," Engineering, Technology & Applied Science Research, vol. 12, no. 2, pp. 8266–8272, Apr. 2022. DOI: https://doi.org/10.48084/etasr.4727
Y. Hatolkar, P. Agrawal, and S. Patil, "A Survey on Road Traffic Sign Recognition System using Convolution Neural Network," International Journal of Current Engineering and Technology, vol. 8, no. 1, pp. 104–108, Jan. 2018. DOI: https://doi.org/10.14741/ijcet/v.8.1.21
C. Qi et al., "An efficient pruning scheme of deep neural networks for Internet of Things applications," EURASIP Journal on Advances in Signal Processing, vol. 2021, no. 1, Jun. 2021, Art. no. 31. DOI: https://doi.org/10.1186/s13634-021-00744-4
M. Nagel, M. Fournarakis, R. A. Amjad, Y. Bondarenko, M. van Baalen, and T. Blankevoort, "A White Paper on Neural Network Quantization." arXiv, Jun. 15, 2021.
S. Lawrence, C. L. Giles, and A. C. Tsoi, "Lessons in Neural Network Training: Overfitting May be Harder than Expected.," in Fourteenth National Conference on Artificial Intelligence, Menlo Park, CA, USA, Jan. 1997, pp. 540–545.
S. Ren, K. He, R. Girshick, and J. Sun, "Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks," IEEE transactions on pattern analysis and machine intelligence, vol. 39, no. 6, pp. 1137–1149, Jun. 2017. DOI: https://doi.org/10.1109/TPAMI.2016.2577031
W. Liu et al., "SSD: Single Shot MultiBox Detector," in 14th European Conference on Computer Vision, Amsterdam, Netherlands, Oct. 2016, pp. 21–37. DOI: https://doi.org/10.1007/978-3-319-46448-0_2
J. Redmon, S. Divvala, R. Girshick, and A. Farhadi, "You Only Look Once: Unified, Real-Time Object Detection," in IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, Jun. 2016, pp. 779–788. DOI: https://doi.org/10.1109/CVPR.2016.91
M. Bhalekar and M. Bedekar, "D-CNN: A New model for Generating Image Captions with Text Extraction Using Deep Learning for Visually Challenged Individuals," Engineering, Technology & Applied Science Research, vol. 12, no. 2, pp. 8366–8373, Apr. 2022. DOI: https://doi.org/10.48084/etasr.4772
J. Dai, Y. Li, K. He, and J. Sun, "R-FCN: object detection via region-based fully convolutional networks," in 30th International Conference on Neural Information Processing Systems, Red Hook, NY, USA, Dec. 2016, pp. 379–387.
M. Njah and R. E. Hamdi, "A Constrained Multi-Objective Learning Algorithm for Feed-Forward Neural Network Classifiers," Engineering, Technology & Applied Science Research, vol. 7, no. 3, pp. 1685–1693, Jun. 2017. DOI: https://doi.org/10.48084/etasr.968
L. Alzubaidi et al., "Review of deep learning: concepts, CNN architectures, challenges, applications, future directions," Journal of Big Data, vol. 8, no. 1, Mar. 2021, Art. no. 53. DOI: https://doi.org/10.1186/s40537-021-00444-8
S. Sahel, M. Alsahafi, M. Alghamdi, and T. Alsubait, "Logo Detection Using Deep Learning with Pretrained CNN Models," Engineering, Technology & Applied Science Research, vol. 11, no. 1, pp. 6724–6729, Feb. 2021. DOI: https://doi.org/10.48084/etasr.3919
I. H. Sarker, "Deep Learning: A Comprehensive Overview on Techniques, Taxonomy, Applications and Research Directions," SN Computer Science, vol. 2, no. 6, Aug. 2021, Art. no. 420. DOI: https://doi.org/10.1007/s42979-021-00815-1
A. Shmilovici, "Support Vector Machines," in Data Mining and Knowledge Discovery Handbook, O. Maimon and L. Rokach, Eds. Boston, MA, USA: Springer US, 2005, pp. 257–276. DOI: https://doi.org/10.1007/0-387-25465-X_12
H. Basly, W. Ouarda, F. E. Sayadi, B. Ouni, and A. M. Alimi, "CNN-SVM Learning Approach Based Human Activity Recognition," in 9th International Conference on Image and Signal Processing, Marrakesh, Morocco, Jun. 2020, pp. 271–281. DOI: https://doi.org/10.1007/978-3-030-51935-3_29
S. Ahlawat and A. Choudhary, "Hybrid CNN-SVM Classifier for Handwritten Digit Recognition," Procedia Computer Science, vol. 167, pp. 2554–2560, Jan. 2020. DOI: https://doi.org/10.1016/j.procs.2020.03.309
M. Elleuch, R. Maalej, and M. Kherallah, "A New Design Based-SVM of the CNN Classifier Architecture with Dropout for Offline Arabic Handwritten Recognition," Procedia Computer Science, vol. 80, pp. 1712–1723, Jan. 2016. DOI: https://doi.org/10.1016/j.procs.2016.05.512
Y. Lai, N. Wang, Y. Yang, and L. Lin, "Traffic Signs Recognition and Classification based on Deep Feature Learning," in 7th International Conference on Pattern Recognition Applications and Methods, Funchal, Portugal, Jan. 2018, pp. 622–629. DOI: https://doi.org/10.5220/0006718806220629
C. Kanan and G. W. Cottrell, "Color-to-Grayscale: Does the Method Matter in Image Recognition?," PLOS ONE, vol. 7, no. 1, Jan. 2012, Art. no. e29740. DOI: https://doi.org/10.1371/journal.pone.0029740
"R. Zhu, D. Yu, S. Ji, and M. Lu, "Matching RGB and Infrared Remote Sensing Images with Densely-Connected Convolutional Neural Networks," Remote Sensing, vol. 11, no. 23, Jan. 2019, Art. no. 2836. DOI: https://doi.org/10.3390/rs11232836
K. B. Shaik, P. Ganesan, V. Kalist, B. S. Sathish, and J. M. M. Jenitha, "Comparative Study of Skin Color Detection and Segmentation in HSV and YCbCr Color Space," Procedia Computer Science, vol. 57, pp. 41–48, Jan. 2015. DOI: https://doi.org/10.1016/j.procs.2015.07.362
Z. S. G. Tari, J. Shah, and H. Pien, "Extraction of Shape Skeletons from Grayscale Images," Computer Vision and Image Understanding, vol. 66, no. 2, pp. 133–146, May 1997. DOI: https://doi.org/10.1006/cviu.1997.0612
J. Basilio, G. Torres, G. Sanchez-Perez, L. K. T. Medina, H. Meana, and G. S. Perez, "Explicit Image Detection using YCbCr Space Color Model as Skin Detection," in Applications of Mathematics and Computer Engineering, 2011, pp. 123–128.
S. Indolia, A. K. Goswami, S. P. Mishra, and P. Asopa, "Conceptual Understanding of Convolutional Neural Network- A Deep Learning Approach," Procedia Computer Science, vol. 132, pp. 679–688, Jan. 2018. DOI: https://doi.org/10.1016/j.procs.2018.05.069
K. Eckle and J. Schmidt-Hieber, "A comparison of deep networks with ReLU activation function and linear spline-type methods," Neural Networks, vol. 110, pp. 232–242, Feb. 2019. DOI: https://doi.org/10.1016/j.neunet.2018.11.005
S. Bruch, X. Wang, M. Bendersky, and M. Najork, "An Analysis of the Softmax Cross Entropy Loss for Learning-to-Rank with Binary Relevance," in ACM SIGIR International Conference on Theory of Information Retrieval, New York, NY, USA, Oct. 2015, pp. 75–78.
Z. Zhang and M. R. Sabuncu, "Generalized Cross Entropy Loss for Training Deep Neural Networks with Noisy Labels," May 2018. Available: https://ui.adsabs.harvard.edu/abs/2018arXiv180507836Z.
Downloads
How to Cite
License
Copyright (c) 2022 Wajid Ali, Guo Wang, Kalim Ullah, Muhammad Salman, Sajad Ali
This work is licensed under a Creative Commons Attribution 4.0 International License.
Authors who publish with this journal agree to the following terms:
- Authors retain the copyright and grant the journal the right of first publication with the work simultaneously licensed under a Creative Commons Attribution License that allows others to share the work with an acknowledgement of the work's authorship and initial publication in this journal.
- Authors are able to enter into separate, additional contractual arrangements for the non-exclusive distribution of the journal's published version of the work (e.g., post it to an institutional repository or publish it in a book), with an acknowledgement of its initial publication in this journal.
- Authors are permitted and encouraged to post their work online (e.g., in institutional repositories or on their website) after its publication in ETASR with an acknowledgement of its initial publication in this journal.