MODEL AND METHOD OF TRAINING THE CLASSIFIER OF OBSERVATION CONTEXT ON VIDEO INSPECTION IMAGES OF SEWER PIPES

В’ячеслав Васильович Москаленко, Микола Олександрович Зарецький, Ярослав Юрійович Ковальський, Сергій Сергійович Мартиненко

Abstract


Video inspection is often used to diagnose sewer pipe defects. To correctly encode founded defects according to existing standards, it is necessary to consider a lot of contextual information about the orientation and location of the camera from sewer pipe video inspection. A model for the classification of context on frames during observations in the video inspection of sewer pipes and a five-stage method of machine learning is proposed. The main idea of the proposed approach is to combine the methods of deep machine learning with the principles of information maximization and coding with self-correcting Hamming codes. The proposed model consists of a deep convolutional neural network with a sigmoid layer followed by the rounding output layer and information-extreme decision rules. The first stages of the method are data augmentation and training of the feature extractor in the Siamese model with softmax triplet loss function. The next steps involve calculating a binary code for each class of recognition that is used as a label in learning with a binary cross-entropy loss function to increase the compactness of the distribution of each class's observations in the Hamming binary space. At the last stage of the training method, it is supposed to optimize the parameters of radial-basis decision rules in the Hamming space for each class according to the existing information-extreme criterion. The information criterion, expressed as a logarithmic function of the accuracy characteristics of the decision rules, provides the maximum generalization and reliability of the model under the most difficult conditions in the statistical sense. The effectiveness of this approach was tested on data provided by Ace Pipe Cleaning (Kansas City, USA) and MPWiK (Wroclaw, Poland) by comparing learning results according to the proposed and traditional models and training schemes. The obtained model of the image frame classifier provides acceptable for practical use classification accuracy on the test sample, which is 96.8 % and exceeds the result of the traditional scheme of training with the softmax output layer by 6.8 %.

Keywords


sewer pipe inspection; convolutional neural networks; Siamese networks; information and extreme training; classification

References


Moradi, S., Zayed, T., Golkho, F. Review On Computer Aided Sewer Pipeline Defect Detection And Condition Assessment. Infrastructures, 2019, vol. 4, p. 10 p. DOI: 10.3390/infrastructures4010010.

Cadena, C., Carlone, L., Carrillo, H., Latif, Y., Scaramuzza, D., Neira, J., Reid, I., Leonard, J. J. Past, Present, and Future of Simultaneous Localization And Mapping: Towards the Robust-Perception Age. IEEE Transactions on Robotics, 2016, vol. 32, iss. 6, pp. 1309–1332. DOI: 10.1109/TRO.2016.2624754.

Syahrian, N. M., Risma, P., Dewi, T. Vision-Based Pipe Monitoring Robot For Crack Detection Using Canny Edge Detection Method As An Image Processing Technique. Kinetik, 2017, vol. 2, pp. 243–250. DOI: 10.22219/kinetik.v2i4.243.

Myrans, J., Everson, R. M., Kapelan, Z. Automated Detection Of Fault Types In CCTV Sewer Surveys. Journal Of Hydroinformatics,2018, vol. 21, pp. 153–163. DOI: 10.2166/hydro.2018.073.

Cheng, J. C. P., Wang, M. Automated Detection Of Sewer Pipe Defects In Closed-Circuit Television Images Using Deep Learning Techniques. Automation in Construction, 2018, vol. 95, pp. 155–171. DOI: 10.1016/j.autcon.2018.08.006.

Panella, F., Boehm, J., Loo, Y., Kaushik, A., Gonzalez, D. Deep learning and image processing for automated crack detection and defect measurement in underground structures. ISPRS - International Archives Of The Photogrammetry, Remote Sensing And Spatial Information Sciences, XLII-2. Riva del Garda, 4-7 June, 2018, pp. 829–835. DOI: 10.5194/isprs- archives-xlii-2-829-2018.

Zhan, H., Shi, B., Duan, L. Y., Kot, A. C., Deepshoe: An Improved Multi-Task View-Invariant CNN For Street-To-Shop Shoe Retrieval. Computer Vision And Image Understanding, 2019, vol. 180, pp. 23–33. DOI: 10.1016/j.cviu.2019.01.001.

Joutsijoki, H., Haponen, M., Rasku, J., Aalto- Setälä, K., Juhola, M. Error-Correcting Output Codes In Classification Of Human Induced Pluripotent Stem Cell Colony Images. Biomed Research International, 2016, vol. 2016, pp. 1–13. DOI: 10.1155/2016/3025057.

Moskalenko, V. Moskalenko, A., Korobov, A., Semashko, V. The Model And Training Algorithm Of Compact Drone Autonomous Visual Navigation System. Data, 2018, vol. 4, pp. 1–14. DOI: 10.3390/data4010004.

Michele, A., Colin, V., Santika, D. D. Mobilenet Convolutional Neural Networks And Support Vector Machines For Palmprint Recognition. Procedia Computer Science, 2019, vol. 157, pp. 110–117.DOI: 10.1016/j.procs.2019.08.147.

Dovbysh, A., Shelehov, I., Prylepa, D., Golub, I. Information Synthesis Of Adaptive System For Visual Diagnostics Of Emotional And Mental State Of A Person. Eastern-European Journal Of Enterprise Technologies, 2016, vol. 4, no. 9(82), pp. 11–17. DOI: 10.15587/1729-4061.2016.756.




DOI: https://doi.org/10.32620/reks.2020.3.06

Refbacks

  • There are currently no refbacks.