APPLICATION OF A CONVOLUTIONAL NEURAL NETWORK AND A KOHONEN NETWORK FOR ACCELERATED DETECTION AND RECOGNITION OF OBJECTS IN IMAGES
One of the most effective ways to improve the accuracy and speed of algorithms for searching and recognizing objects in images is to pre-select areas of interest in which it is likely to detect objects of interest. To determine areas of interest in a pre-processed radar or satellite image of the underlying surface, the Kohonen network was used. The found areas of interest are sent to the convolutional neural network, which provides the final detection and recognition of objects. The combination of the above methods allows to speed up the process of searching and recognizing objects in images, which is becoming more expedient due to the constantly growing volume of data for analysis. The process of preliminary processing of input data is described, the process of searching and recognizing patterns of aircraft against the underlying surface is presented, and the analysis of the results is carried out. The use of the Kohonen neural network makes it possible to reduce the amount of data analyzed by the convolutional network by 18–125 times, which accordingly accelerates the process of detection and recognition of the object of interest. The size of the parts of the input image fed into the convolutional network, into which the zones of interest are divided, is tied to the image scale and is equal to the size of the largest detectable object, which can significantly reduce the training sample. Application of the presented methods and centering of the object on training images allows accelerating the convolutional network training by more than 5 times and increasing the recognition accuracy by at least 10%, as well as minimizing the required minimum number of layers and neurons of the network by at least halving, respectively increasing its speed
Simard, P. Y., Steinkraus, D., Platt, J. C. (2003). Best practices for convolutional neural networks applied to visual document analysis. Seventh International Conference on Document Analysis and Recognition, 2003. Proceedings. doi: https://doi.org/10.1109/icdar.2003.1227801
Agayan, K. Yu., Hanzhin, V. G. (2018). Neyronnaya set' s arhitekturoy Kohonena dlya raspoznavaniya izobrazheniy. Sb. trudov IX-oy Evraziyskoy nauchno-prakticheskoy konferentsii «Prochnost' neodnorodnyh struktur» - PROST 2018. Moscow: OOO «Studio-Print», 153.
Gerasimova, N. I., Verkhoturova, A. E. (2014). Earch of the image fragment with application of Kohonen neural network. Informatsionnye tekhnologii v nauke, upravlenii, sotsial'noy sfere i meditsine: sbornik nauchnyh trudov Mezhdunarodnoy konferentsii s mezhdunarodnym uchastiem. Ch. 1. Tomsk: Izd-vo TPU, 68–70.
Soldatova, O. P., Chayka, P. D. (2015). Efficiency analysis of solution of classification using hybrid kohonen neural networks. Izvestiya Samarskogo nauchnogo tsentra Rossiyskoy akademii nauk, 17 (2 (5)), 1147–1152.
Kajan, S., Sekaj, I., Lajtman, M. (2011). Cluster Analysis Aplications in Matlab Using Kohonen. Network learning, 4 (6), 7.
Le Cun, Y., Bengio, Y. (1995). Convolutional networks for images, speech, and time series. The handbook of brain theory and neural networks, 3361 (10).
LeCun, Y., Kavukcuoglu, K., Farabet, C. (2010). Convolutional networks and applications in vision. Proceedings of 2010 IEEE International Symposium on Circuits and Systems. doi: https://doi.org/10.1109/iscas.2010.5537907
Girshick, R., Donahue, J., Darrell, T., Malik, J. (2014). Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation. 2014 IEEE Conference on Computer Vision and Pattern Recognition. doi: https://doi.org/10.1109/cvpr.2014.81
Girshick, R., Donahue, J., Darrell, T., Malik, J. (2016). Region-Based Convolutional Networks for Accurate Object Detection and Segmentation. IEEE Transactions on Pattern Analysis and Machine Intelligence, 38 (1), 142–158. doi: https://doi.org/10.1109/tpami.2015.2437384
Girshick, R. (2015). Fast R-CNN. 2015 IEEE International Conference on Computer Vision (ICCV). doi: https://doi.org/10.1109/iccv.2015.169
Ren, S., He, K., Girshick, R., Sun, J. (2017). Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Transactions on Pattern Analysis and Machine Intelligence, 39 (6), 1137–1149. doi: https://doi.org/10.1109/tpami.2016.2577031
He, K., Gkioxari, G., Dollar, P., Girshick, R. (2017). Mask R-CNN. 2017 IEEE International Conference on Computer Vision (ICCV). doi: https://doi.org/10.1109/iccv.2017.322
Redmon, J., Divvala, S., Girshick, R., Farhadi, A. (2016). You Only Look Once: Unified, Real-Time Object Detection. 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). doi: https://doi.org/10.1109/cvpr.2016.91
Liu, W., Anguelov, D., Erhan, D., Szegedy, C., Reed, S., Fu, C.-Y., Berg, A. C. (2016). SSD: Single Shot MultiBox Detector. Lecture Notes in Computer Science, 21–37. doi: https://doi.org/10.1007/978-3-319-46448-0_2
Haykin, S. (2008). Neural Networks: A Comprehensive Foundation. Izdatel'skiy dom Vil'yams, 1104.
Kohonen, T. (2001). Self-organizing maps. Vol. 30. Springer Science & Business Media, 501. doi: https://doi.org/10.1007/978-3-642-56927-2
Skuratov, V., Kuzmin, K., Nelin, I., Sedankin, M. (2019). Application of Kohonen neural networks to search for regions of interest in the detection and recognition of objects. Eastern-European Journal of Enterprise Technologies, 3 (9 (99)), 41–48. doi: https://doi.org/10.15587/1729-4061.2019.166887
Skuratov, V., Kuzmin, K., Nelin, I., Sedankin, M. (2020). Application of kohonen self-organizing map to search for region of interest in the detection of objects. EUREKA: Physics and Engineering, 1, 62–69. doi: https://doi.org/10.21303/2461-4262.2020.001133
Abstract views: 18 PDF Downloads: 16
Copyright (c) 2020 Victor Skuratov, Konstantin Kuzmin , Igor Nelin , Mikhail Sedankin
This work is licensed under a Creative Commons Attribution 4.0 International License.
Our journal abides by the CREATIVE COMMONS copyright rights and permissions for open access journals.
Authors, who are published in this journal, agree to the following conditions:
1. The authors reserve the right to authorship of the work and pass the first publication right of this work to the journal under the terms of a Creative Commons Attribution License, which allows others to freely distribute the published research with the obligatory reference to the authors of the original work and the first publication of the work in this journal.
2. The authors have the right to conclude separate supplement agreements that relate to non-exclusive work distribution in the form in which it has been published by the journal (for example, to upload the work to the online storage of the journal or publish it as part of a monograph), provided that the reference to the first publication of the work in this journal is included.