Next Article in Journal
SAR Image Classification Using Markov Random Fields with Deep Learning
Next Article in Special Issue
Isostatic Anomaly and Isostatic Additional Force Analysis by Multiple Geodetic Observations in Qinling Area
Previous Article in Journal
Subbasin Spatial Scale Effects on Hydrological Model Prediction Uncertainty of Extreme Stream Flows in the Omo Gibe River Basin, Ethiopia
Previous Article in Special Issue
Sea-Crossing Bridge Detection in Polarimetric SAR Images Based on Windowed Level Set Segmentation and Polarization Parameter Discrimination
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Technical Note

Automatic Damage Detection and Diagnosis for Hydraulic Structures Using Drones and Artificial Intelligence Techniques

1
State Key Laboratory of Hydrology-Water Resources and Hydraulic Engineering, Hohai University, Nanjing 210098, China
2
College of Water Conservancy and Hydropower Engineering, Hohai University, Nanjing 210098, China
3
National Dam Safety Research Center, Wuhan 430010, China
4
Yangtze Institute for Conservation and Development, Hohai University, Nanjing 210098, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(3), 615; https://doi.org/10.3390/rs15030615
Submission received: 16 December 2022 / Revised: 12 January 2023 / Accepted: 18 January 2023 / Published: 20 January 2023

Abstract

:
Large-volume hydraulic concrete structures, such as concrete dams, often suffer from damage due to the influence of alternating loads and material aging during the service process. The occurrence and further expansion of cracks will affect the integrity, impermeability, and durability of the dam concrete. Therefore, monitoring the changing status of cracks in hydraulic concrete structures is very important for the health service of hydraulic engineering. This study combines computer vision and artificial intelligence methods to propose an automatic damage detection and diagnosis method for hydraulic structures. Specifically, to improve the crack feature extraction effect, the Xception backbone network, which has fewer parameters than the ResNet backbone network, is adopted. With the aim of addressing the problem of premature loss of image detail information and small target information of tiny cracks in hydraulic concrete structures, an adaptive attention mechanism image semantic segmentation algorithm based on Deeplab V3+ network architecture is proposed. Crack images collected from concrete structures of different types of hydraulic structures were used to develop crack datasets. The experimental results show that the proposed method can realize high-precision crack identification, and the identification results have been obtained in the test set, achieving 90.537% Intersection over Union (IOU), 91.227% Precision, 91.301% Recall, and 91.264% F1_score. In addition, the proposed method has been verified on different types of cracks in actual hydraulic concrete structures, further illustrating the effectiveness of the method.

1. Introduction

Large-volume hydraulic concrete structures such as concrete dams must bear the loads caused by water pressure, temperature, and earthquakes. It may be affected by dry–wet cycles under the action of many unfavorable factors during the service process [1,2]. The tensile strength of concrete is low, and the complex working conditions and the material properties of concrete make hydraulic concrete prone to cracks [3]. The existence and development of cracks can lead to leakage and corrosion damage or cause dam failure, endangering the safety of people’s life and property [4].
At present, the dam-crack detection method at home and abroad adopts mainly the traditional manual inspection [5]. Not only is this method time-consuming and laborious, but also, it is difficult to reach many high structures. The inspection method using manual inspection can inspect only the cracks existing in the appearance of the structure, and it is difficult to find hidden cracks inside the dam body, and the judgment of the status of the cracks depends on subjective experience [6]. Moreover, the manual inspection of high dams still has certain safety risks. In recent years, dam safety monitoring systems have been introduced to achieve automatic sensing of structural changes [7,8]. At present, cracks in hydraulic structures such as concrete dams are generally monitored by in situ monitoring instruments such as joint gauges and strain gauges. Since the occurrence of cracks is spatially random and uncertain, the aging of internal observation instruments cannot be easy [9]. It is hard to monitor cracks with this conventional point monitoring method without a good solution.
In recent years, with the development of artificial intelligence and information technology, computer vision (CV)-based structural crack recognition and detection technology has been gradually applied to civil engineering operations and maintenance [10,11,12]. Inspection equipment represented by drones can be equipped with various customized cameras to carry out structural defect scanning inspection work [13,14]. CV-based technology has significant advantages such as a large image acquisition range, high spatial resolution, and short detection time. Due to the large size of the dam body and numerous ancillary buildings, the single-machine inspection task will generate a large amount of image or video data [15,16]. Because the occurrence of cracks is a probabilistic event, the proportion of information directly related to structural damage is extremely low. It is necessary to combine advanced digital image processing technology to extract information closely related to cracks from these data.
In recent years, artificial intelligence technology has been widely used in many practical engineering fields to replace labor and reduce costs [17,18]. Among them, the deep learning method represented by the convolutional neural network (CNN) is widely used in the structural damage identification of civil infrastructure. For example, Cha et al. [19] developed a vision-based method using a deep architecture of CNN for detecting concrete cracks without calculating the defect features. Khani et al. [20] proposed a novel crack detection framework that utilizes techniques from both classical image processing and deep learning methodologies. Kim et al. [21] developed shallow convolutional neural network (CNN)-based architecture for surface-concrete crack detection. However, most of the above research focuses on crack recognition and classification and targets detection, and few studies consider the feature extraction of hydraulic concrete structure cracks in drone-based application scenarios. Due to the interference of various factors in the process of drone aerial photography, there are many blurred and low-quality problems in the crack images collected, so it is necessary to combine deep learning and artificial intelligence algorithms.
In this paper, considering these limitations, this paper uses the Xception backbone network instead of the conventional kernels in DeepLab V3+, the internal convolution kernel of which uses depth-separable convolution, and the hole rate of the convolution kernel can be set by itself. The adaptive attention mechanism module is embedded into the backbone network by utilizing the residual network and the channel attention strategy. The purpose of this module is to supplement the detailed information of the unit input feature map to the output feature map, to delay the trend of loss of detailed information.
The main contributions of this work are the following three points.
(1)
The Xception backbone-based crack automatic segmentation network achieves faster detection efficiency and few parameters because it uses depthwise separable convolution for its internal convolution kernel, and the hole rate of the convolution kernel can be set by itself.
(2)
The combination of the attention mechanism module and the Deeplab V3+ backbone network can significantly improve the accuracy of the model for identifying small-scale concrete cracks.
(3)
The proposed method shows strong crack pixel-level detection performance on a variety of different types and background roughness crack images.
The rest of the paper is organized as follows: Section 2 presents an overview of the methodology developed in this paper, including the Deeplabv3+ neural network. Section 3 details the experimental setup, and the experimental results are presented and discussed in Section 4. Conclusions are provided in the final section of the paper.

2. Methodology and Materials

2.1. The Deeplabv3+ Semantic Segmentation Network

The structure diagram of the classic DeepLabv3+ network model is shown in Figure 1. The DeepLab series is developed by Google and is currently a widely used image segmentation model, using an end-to-end training method [22]. DeepLabv3+ consists of two parts, including the Encoder and Decoder at the encoding end. The backbone network at the encoding end selects the Xception model. The ASPP (Atrous Spatial Pyramid Pooling) is designed to combined with atrous convolution can expand the receptive field of the convolution kernel without losing resolution (no downsampling) [23]. Together with the ASPP module as a feature extraction module, the feature maps (Feature maps) from Xception in the backbone network are divided into two branches: one is the feature map output by the last layer of the backbone network, which is sent to the ASPP module; the other part is the backbone network output from the middle layer. The shallow Feature maps are sent to the Decoder module, and the Decoder at the decoding end introduces the shallow Feature maps and fuses the high-level semantic features output by the ASPP module. This makes the two input feature maps through convolution (Conv) or upsampling (Upsample) operations. If the size is the same, then use Concat to splice them together and send them to a set of 3×3 convolution blocks for processing. Finally, perform linear interpolation and upsampling again to obtain an output segmentation result image with the same resolution as the original image.

2.2. The Xception Backbone

Considering that devices such as drones are difficult to achieve high-performance computing, deep learning needs to be improved to improve its real-time detection capabilities [24]. Considering the performance of the device, this paper uses the Xception backbone network, the internal convolution kernel of which uses depth-separable convolution, and the hole rate of the convolution kernel can be set by itself. The improved Xception-65 network has fewer parameters than the ResNet backbone network [25], showing the improvement of the DeepLabv3+ neural network, as shown in Figure 2. Equation (2) is to illustrate the introduction of hole convolution, which significantly reduces the amount of model weight parameters. The specific calculation formula is as follows:
W 1 = K * K * C 1 * C 2
W 2 = K * K * C 1 + 1 * 1 * C 1 * C 2
where the size of the ordinary convolution kernel and the depth-separable convolution kernel with holes are both K * K , the number of channels of the input feature map is C 1 , the number of channels of the output feature map is C 2 , and the parameter quantity generated after the convolution of the ordinary convolution kernel and the feature map is W 1 .
Figure 2 shows the flowchart of the improved crack pixel-level segmentation and quantification framework. It can be seen that on the left is the calculation process of encoding and decoding construction, and on the right is the detailed composition of each component. Inspired by the above research methods, for the current Deeplab V3+ semantic segmentation framework, it is easy to cause the loss of detailed information or small targets. In this paper, the adaptive attention mechanism module is embedded into the backbone network by utilizing the residual network and the channel attention strategy. The purpose of this module is to supplement the detailed information of the unit input feature map to the output feature map in order to delay the trend of loss of detailed information.
The method proposed in this paper consists of the following four parts. The method of using channel compression is inspired by the above research methods and aims at the problem that the current Deeplab V3+ semantic segmentation framework is prone to loss of detailed information or small targets. In this paper, the adaptive attention mechanism module is embedded into the backbone network by utilizing the residual network and the channel attention strategy. The purpose of this module is to supplement the detailed information of the unit input feature map to the output feature map to delay the trend of loss of detailed information.
Figure 3 shows the comparison of the standard convolution process and the depthwise separable convolution process. It can be seen from Figure 3a that the process of convolving an ordinary convolution kernel with a feature map, and Figure 3b shows the process of convolving a depthwise separable convolution kernel with a hole and a feature map [26].

2.3. The Adaptive Attention Mechanism

Most of the cracks in hydraulic concrete structures are tiny and hidden in complex backgrounds, making them difficult to identify [27]. In image semantic segmentation, it is usually necessary to obtain high-level semantic information about the image to strengthen the understanding of the whole image. However, in the down-sampling process of obtaining the high-level semantic information of the image, the detailed information of the image will inevitably be lost, especially for some small objects. The number of pixels occupied in the image is small, and it is easy to lose in the process of multiple downsampling. However, this detailed information and these small objects will exist in the channels of low-level feature maps.
In this paper, the residual network is used in combination with the channel compression method and the attention mechanism to build an adaptive attention mechanism module, and it is embed into the Xception backbone network. Specifically, the attention mechanism module is embedded in the input layer, middle layer, and output layer of the Deeplab V3+ backbone network, and a weight value is introduced to multiply each attention mechanism module to achieve the purpose of constraining the attention mechanism module. First, a global average pooling process is performed on the input feature map of the attention mechanism module. Then, after that, a 1 × 1 × C feature map is output, where C is the number of channels, which is the same as the channel number of the output feature map of the residual block. Equation (3) is to illustrate the calculation process of the residual block output feature map to obtain richer high-level semantic information. The specific calculation formula is as follows:
g C = 1 H * W i = 1 H j = 1 W G C ( i , j )
where g c R c from a feature graph G of size H × W and number of channels C, C (i, j) generated; (i, j) represents the coordinates of the pixel on the feature map.
Based on the above-mentioned methods, the calculation process of the self-attention mechanism is described. Firstly, the attention mechanism module is embedded in the input layer, middle layer, and output layer of the Deeplab V3+ backbone network, and a weight value is introduced to multiply each attention mechanism module to achieve the purpose of constraining the attention mechanism module. Secondly, the Deeplab V3+ embedding the attention module is trained on the constructed concrete crack dataset, so as to manually obtain the weight value (empirical value) of the attention mechanism module; then, the input layer, the middle layer, and the output layer are explored. Finally, the weight value of the attention mechanism module is changed to automatically update by backpropagation, so as to obtain the optimal weight value of the attention mechanism module and the optimal segmentation model.

2.4. Evaluation Indicators

In this study, three evaluation metrics are used to calculate the model’s ability to crack segmentation, including recall, precision, and F1_score. The details about these evaluation indicators are shown as follows:
Recall   = T P T P + F N
Precision   = T P T P + F P
F 1 _ score   = 2   Precision     Recall     Precision   +   Recall  
where TP is the True Positives, FN is the False Negative, and FP is the False Positives.
To quantitatively evaluate the effect of crack identification, an evaluation index of intersection ratio, which is called the Intersection over Union (IOU), is introduced. Figure 4 demonstrates the visual display of IOU indicators.

3. Experimental Setup

3.1. Project Description

Figure 5 shows the overall layout of the water conservancy project. This project is composed of a dam, a flood discharge structure, a power station, a navigation structure, and an irrigation canal head. This project is located in the middle reaches of the Huangbai River East branch of the first tributary of the Yangtze River. The rainwater collection area above the dam site is 862 km2, the annual average rainfall is 1150 mm, and the annual average runoff is 388 million m3. The total storage capacity of the reservoir is 196 million m3. It is a large (2) type reservoir that focuses mainly on water storage and irrigation and also has comprehensive benefits such as flood control, power generation, and urban water supply. The reservoir is the main water source for farmland irrigation, and the average annual power generation of the power station is 55.1 million KW.h.

3.2. Drone Inspection

The project developed a set of high-precision UAV inspection systems for dam surface, which applies to dam surface from the aspects of dam surface inspection path planning and independent safety inspection control, dam surface defect identification, and health diagnosis system construction. Figure 6 shows the drone techniques used for dam inspection tasks. Table 1 demonstrates the related parameters of drones. Independent inspection of the development of the prototype. Before carrying out the inspection work, drone path planning is carried out according to the actual layout of the dam, so as to realize the shooting of the whole range of the dam. The relevant parameters of the drone equipped with the camera are shown in Table 1. In the data post-processing stage, real-time return of on-site data such as ultra-high-definition images, cloud computing, artificial intelligence in-depth data analysis, graphic image detection, and other technologies to implement intelligent risk diagnosis, realize independent inspection of the dam surface, and seamlessly acquire honey image information on the dam surface.

3.3. Dataset Label and Generation

In this study, images of surface cracks of relevant hydraulic concrete structures were extensively collected through on-site shooting, report review, and network retrieval. These hydraulic structures include dam surface, face rockfill dam face, sluice pier, channel lining, and other hydraulic concrete structures. The original resolution of the collected images is high, and it is difficult to directly input the network for training. In this study, the original acquired image is divided into 200×200 low-resolution small images for further model annotation and model training.
Figure 7 shows some images from the constructed hydraulic concrete structure crack dataset. Moreover, it can be seen from Figure 7 that crack images acquired from the hydraulic structures suffer from severe environmental interference noises, such as water stains, alkaline substances precipitation, etc. It is of great significance to carefully label the crack information and morphological features under the inference of these unfavorable factors.
Figure 8 shows the comparison of the crack image and refined annotation image. It can be seen from the figure that there are significant differences in the background of the cracks in the constructed hydraulic concrete structure crack dataset. There are both smooth concrete surfaces and rough concrete surfaces. As observed from Figure 8, the pixel-wise annotations of cracks include mainly two parts, including the original images and annotation images at the pixel level.

4. Result and Discussion

4.1. Model Training

All experiments were implemented based on Pytorch, and the operating system is Windows 10. The dataset is randomly divided into a certain proportion, with a total of 5000 images, which are divided into the training set, verification set, and test set according to the ratio of 6:2:2. Table 2 shows the hardware and software environment in which the comparative experiment was run. Specifically, the central processing unit (CPU) is Intel i7-12700KF, and the graphics card is NVIDIA GeForce RTX 3070. Note that the proposed and compared methods are trained on the same computing station. The number of model batches is uniformly set to 8, and the number of training rounds (Epoch) is set to 100. The learning rate is set to 0.0005.
Figure 9 shows the changes in the loss functions in both the training and validation sets during 100 iterations. As can be seen from the figure, with the increase in the number of iterations, the loss function of the model on the training set and verification set shows a gradual and smooth downward trend, indicating that the model effectively learns the rules from the dataset. After 100 iterations, the loss function of the model tends to converge, indicating that the model is fully trained.

4.2. Ablation Experiments

In this study, a series of ablation studies were implemented to analyze the importance of these proposed components. Table 3 shows the contribution comparison of different improvements of the proposed method. It can be seen from the table that the detection effect of the crack identification model applying the improved backbone network and the attention enhancement mechanism on the test set is significantly better than that only applying the improved backbone network and the attention enhancement mechanism.

4.3. Model Comparison with Other Algorithms

In this study, the proposed method is compared with several advanced semantic segmentation methods to further verify the model performance. On the proposed hydraulic structure crack dataset, the method is compared with other state-of-the-art (SOTA) semantic segmentation methods, including Unet [28], Deeplab V3+, fully convolutional network (FCN) [29], and Canny [30]. All implementations of the SOTA algorithms are based on the published paper and fine-tuned hyperparameters using only the developed crack dataset. Table 4 shows the comparative evaluation of crack detection results of different segmentation methods. Figure 10 demonstrates the crack detection effect of the proposed method and other SOTA algorithm. It can be seen from the figure that the crack identification effect of the proposed method is better than that of the digital image processing method. Specifically, the recognition result of the proposed crack detection method is smooth and continuous, whereas the traditional digital image processing method has fracture and discontinuity. Compared with other DL-based methods, the proposed method tends to have fewer fractures in the recognition results, which is more in line with the real crack distribution.
It can be seen from Table 4 and Figure 10 that the proposed method using the improved backbone network and attention mechanism has stronger feature extraction and crack identification capabilities and has achieved better performance on the test set. Specifically, the introduction of the Xception-based backbone network effectively reduces the number of model calculation parameters, which effectively improves the inference speed of the model on a single crack image with a resolution of 200 × 200. In addition, the deep learning method significantly outperforms traditional image processing algorithms in crack image inference efficiency.

4.4. Test Result Visualization

To evaluate the recognition effect of the proposed method in crack segmentation better and effectively, three kinds of crack images containing different backgrounds and crack shapes are selected to evaluate the identification performance effect. Figure 11 demonstrates the identification effect of the proposed method on different types of cracks for hydraulic concrete structures. It can be seen from the figure that the proposed method has achieved good performance on different types of cracks in the test set, and the results of neural network segmentation and crack identification are basically consistent.

5. Conclusions

Crack detection for hydraulic concrete structures is of great significance to ensure the safety of dam operation. At present, the manual inspection method mainly used in crack detection is not only inefficient but also high-risk. Moreover, the traditional image processing method has a slow detection speed, low degree of automation, and weak generalization ability. Considering these limitations, this paper develops an improved Deeplab V3+ network via the Xception backbone and the adaptive attention mechanism network for the crack identification and segmentation of hydraulic concrete structures. The proposed method can effectively identify cracks of different types and background complexities. The specific contributions of this study are as follows.
(a)
The experimental results show that the proposed method can realize high-precision crack identification, and the identification results have been obtained in the test set, achieving 90.537 IOU, 91.227 precision, 91.301 recall, and 91.264 F1_score.
(b)
The fusion of a lightweight backbone network and attention mechanism can improve the accuracy of model crack identification and improve the speed of model crack detection.
(c)
The proposed method can effectively identify different types of s types of cracks in hydraulic concretes. It can be seen from the experimental results that the proposed method has a good recognition effect on wide, narrow, transverse, and longitudinal cracks.
However, some issues require further elaboration and clarification. The types of defects in hydraulic concrete structures are highly complex and diverse. In addition to cracks, there are also many types of defects such as spalling, collapse, and pitting. Therefore, it is necessary to combine deep learning and semantic segmentation network models to further study the identification methods of different types of defects. In addition, it is necessary to further combine aerial photography technologies such as drones to study automatic identification methods for defects to guide drone navigation.

Author Contributions

Conceptualization, methodology, software, formal analysis, investigation, resources, data curation, writing, Y.Z.; original draft preparation, funding acquisition, H.T. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Key R&D Program of China (2022YFC3202602), National Natural Science Foundation of China (U2040205), China Postdoctoral Science Foundation (2022M720998), the Natural Science Foundation of Jiangsu Province (BK20220978), and Jiangsu Funding Program for Excellent Postdoctoral Talent (2022ZB176).

Data Availability Statement

The data used to support the findings of this study are included within the article. The open-access link about the crack dataset is https://share.weiyun.com/0C8xmt9r, and the password is “fdgg33”.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhu, Y.; Niu, X.; Gu, C.; Yang, D.; Sun, Q.; Rodriguez, E.F. Using the DEMATEL-VIKOR Method in Dam Failure Path Identification. Int. J. Environ. Res. Public Health 2020, 17, 1480. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Niu, X. The First Stage of the Middle-Line South-to-North Water-Transfer Project. Engineering 2022, 16, 21–28. [Google Scholar] [CrossRef]
  3. Su, H.; Li, J.; Wen, Z.; Guo, Z.; Zhou, R. Integrated Certainty and Uncertainty Evaluation Approach for Seepage Control Effectiveness of a Gravity Dam. Appl. Math. Model. 2019, 65, 1–22. [Google Scholar] [CrossRef]
  4. Zoubir, H.; Rguig, M.; El Aroussi, M.; Chehri, A.; Saadane, R.; Jeon, G. Concrete Bridge Defects Identification and Localization Based on Classification Deep Convolutional Neural Networks and Transfer Learning. Remote Sens. 2022, 14, 4882. [Google Scholar] [CrossRef]
  5. Ren, Q.; Li, M.; Kong, T.; Ma, J. Multi-Sensor Real-Time Monitoring of Dam Behavior Using Self-Adaptive Online Sequential Learning. Autom. Constr. 2022, 140, 104365. [Google Scholar] [CrossRef]
  6. Li, Y.; Bao, T.; Shu, X.; Gao, Z.; Gong, J.; Zhang, K. Data-Driven Crack Behavior Anomaly Identification Method for Concrete Dams in Long-Term Service Using Offline and Online Change Point Detection. J. Civ. Struct. Health Monit. 2021, 11, 1449–1460. [Google Scholar] [CrossRef]
  7. Dai, B.; Gu, C.; Zhao, E.; Qin, X. Statistical Model Optimized Random Forest Regression Model for Concrete Dam Deformation Monitoring. Struct. Control. Health Monit. 2018, 25, e2170. [Google Scholar] [CrossRef]
  8. Zhu, Y.; Niu, X.; Gu, C.; Dai, B.; Huang, L. A Fuzzy Clustering Logic Life Loss Risk Evaluation Model for Dam-Break Floods. Complexity 2021, 2021, 7093256. [Google Scholar] [CrossRef]
  9. Liang, J.; Chen, B.; Shao, C.; Li, J.; Wu, B. Time Reverse Modeling of Damage Detection in Underwater Concrete Beams Using Piezoelectric Intelligent Modules. Sensors 2020, 20, 7318. [Google Scholar] [CrossRef]
  10. Dong, C.Z.; Catbas, F.N. A Review of Computer Vision–Based Structural Health Monitoring at Local and Global Levels. Struct. Health Monit. 2020, 20, 692–743. [Google Scholar] [CrossRef]
  11. Spencer, B.F.; Hoskere, V.; Narazaki, Y. Advances in Computer Vision-Based Civil Infrastructure Inspection and Monitoring. Engineering 2019, 5, 199–222. [Google Scholar] [CrossRef]
  12. Shin, D.; Jin, J.; Kim, J. Enhancing Railway Maintenance Safety Using Open-Source Computer Vision. J. Adv. Transp. 2021, 2021, 5575557. [Google Scholar] [CrossRef]
  13. Feroz, S.; Dabous, S.A. Uav-Based Remote Sensing Applications for Bridge Condition Assessment. Remote Sens. 2021, 13, 1809. [Google Scholar] [CrossRef]
  14. Yao, H.; Qin, R.; Chen, X. Unmanned Aerial Vehicle for Remote Sensing Applications—A Review. Remote Sens. 2019, 11, 1443. [Google Scholar] [CrossRef] [Green Version]
  15. González-deSantos, L.M.; Martínez-Sánchez, J.; González-Jorge, H.; Navarro-Medina, F.; Arias, P. UAV Payload with Collision Mitigation for Contact Inspection. Autom. Constr. 2020, 115, 103200. [Google Scholar] [CrossRef]
  16. Zhao, J.; Zhang, X.; Yan, J.; Qiu, X.; Yao, X.; Tian, Y.; Zhu, Y.; Cao, W. A Wheat Spike Detection Method in Uav Images Based on Improved Yolov5. Remote Sens. 2021, 13, 3095. [Google Scholar] [CrossRef]
  17. Bui, K.T.T.; Tien Bui, D.; Zou, J.; Van Doan, C.; Revhaug, I. A Novel Hybrid Artificial Intelligent Approach Based on Neural Fuzzy Inference Model and Particle Swarm Optimization for Horizontal Displacement Modeling of Hydropower Dam. Neural Comput. Appl. 2018, 29, 1495–1506. [Google Scholar] [CrossRef]
  18. Li, Y.; Wang, H.; Dang, L.M.; Nguyen, T.N.; Han, D.; Lee, A.; Jang, I.; Moon, H. A Deep Learning-Based Hybrid Framework for Object Detection and Recognition in Autonomous Driving. IEEE Access 2020, 8, 194228–194239. [Google Scholar] [CrossRef]
  19. Cha, Y.J.; Choi, W.; Büyüköztürk, O. Deep Learning-Based Crack Damage Detection Using Convolutional Neural Networks. Comput. Civ. Infrastruct. Eng. 2017, 32, 361–378. [Google Scholar] [CrossRef]
  20. Mohtasham Khani, M.; Vahidnia, S.; Ghasemzadeh, L.; Ozturk, Y.E.; Yuvalaklioglu, M.; Akin, S.; Ure, N.K. Deep-Learning-Based Crack Detection with Applications for the Structural Health Monitoring of Gas Turbines. Struct. Health Monit. 2020, 19, 1440–1452. [Google Scholar] [CrossRef]
  21. Kim, B.; Yuvaraj, N.; Sri Preethaa, K.R.; Arun Pandian, R. Surface Crack Detection Using Deep Learning with Shallow CNN Architecture for Enhanced Computation. Neural Comput. Appl. 2021, 33, 9289–9305. [Google Scholar] [CrossRef]
  22. Tang, W.; Zou, D.; Yang, S.; Shi, J.; Dan, J.; Song, G. A Two-Stage Approach for Automatic Liver Segmentation with Faster R-CNN and DeepLab. Neural Comput. Appl. 2020, 32, 6769–6778. [Google Scholar] [CrossRef]
  23. Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 834–848. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Gopalakrishnan, K.; Gholami, H.; Vidyadharan, A.; Agrawal, A. Crack Damage Detection in Unmanned Aerial Vehicle Images of Civil Infrastructure Using Pre-Trained Deep Learning Model. Int. J. Traffic Transp. Eng. 2018, 8, 1–14. [Google Scholar] [CrossRef]
  25. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef] [Green Version]
  26. Lin, M.; Chen, Q.; Yan, S. Network In Network: Original Paper. In Proceedings of the 2nd International Conference on Learning Representations, ICLR 2014, Banff, AB, Canada, 14–16 April 2014; pp. 1–10. [Google Scholar]
  27. Zhong, H.; Li, H.; Ooi, E.T.; Song, C. Hydraulic Fracture at the Dam-Foundation Interface Using the Scaled Boundary Finite Element Method Coupled with the Cohesive Crack Model. Eng. Anal. Bound. Elem. 2018, 88, 41–53. [Google Scholar] [CrossRef]
  28. Zhang, L.; Shen, J.; Zhu, B. A Research on an Improved Unet-Based Concrete Crack Detection Algorithm. Struct. Health Monit. 2021, 20, 1864–1879. [Google Scholar] [CrossRef]
  29. Dung, C.V.; Anh, L.D.; Vu, C.; Duc, L.; Dung, C.V.; Anh, L.D. Autonomous Concrete Crack Detection Using Deep Fully Convolutional Neural Network. Autom. Constr. 2019, 99, 52–58. [Google Scholar] [CrossRef]
  30. Akagic, A.; Buza, E.; Omanovic, S.; Karabegovic, A. Pavement Crack Detection Using Otsu Thresholding for Image Segmentation. In Proceedings of the 41st International Convention on Information and Communication Technology, Electronics and Microelectronics, MIPRO 2018, Opatija, Croatia, 21–25 May 2018; pp. 1092–1097. [Google Scholar] [CrossRef]
Figure 1. The basic architecture of the DeepLabv3+ neural network.
Figure 1. The basic architecture of the DeepLabv3+ neural network.
Remotesensing 15 00615 g001
Figure 2. The flowchart of the improved framework.
Figure 2. The flowchart of the improved framework.
Remotesensing 15 00615 g002
Figure 3. The comparison of the standard convolution process and depthwise separable convolution process. (a) Standard convolution process; (b) depthwise separable convolution process with dilation.
Figure 3. The comparison of the standard convolution process and depthwise separable convolution process. (a) Standard convolution process; (b) depthwise separable convolution process with dilation.
Remotesensing 15 00615 g003
Figure 4. The visual display of IOU indicators.
Figure 4. The visual display of IOU indicators.
Remotesensing 15 00615 g004
Figure 5. Real photos from the project.
Figure 5. Real photos from the project.
Remotesensing 15 00615 g005
Figure 6. The drone used in this dam inspection task.
Figure 6. The drone used in this dam inspection task.
Remotesensing 15 00615 g006
Figure 7. Sample images of the constructed dataset.
Figure 7. Sample images of the constructed dataset.
Remotesensing 15 00615 g007
Figure 8. Comparison of crack image and refined annotation image.
Figure 8. Comparison of crack image and refined annotation image.
Remotesensing 15 00615 g008
Figure 9. Loss function changes during model training.
Figure 9. Loss function changes during model training.
Remotesensing 15 00615 g009
Figure 10. Comparison of calculation effects of different crack detection methods.
Figure 10. Comparison of calculation effects of different crack detection methods.
Remotesensing 15 00615 g010
Figure 11. The demonstration of the identification effect of the proposed method on different types of cracks.
Figure 11. The demonstration of the identification effect of the proposed method on different types of cracks.
Remotesensing 15 00615 g011
Table 1. The related parameters of drones and cameras.
Table 1. The related parameters of drones and cameras.
ParametersValues
weight570 g
size180 × 97 × 74 mm (fold); 183 × 253 × 77 mm (stretch)
flight time34 min
maximum ascent speed4 m/s
maximum descent speed3 m/s
maximum horizontal flight speed19 m/s
maximum flight altitude500 m
maximum wind resistance ratingLevel 5 wind
maximum transmission distance10,000 m
maximum supported storage8 G (airborne memory) + 256 GB
camera pixels3840 × 2160 pixels
equivalent focal length24 mm
aperturef/2.8
angle of view84°
Table 2. Experimental hardware and software environment configuration.
Table 2. Experimental hardware and software environment configuration.
Lab EnvironmentConfiguration Details
Hardware environmentCPU: Intel i7-12700KF
GPU: NVIDIA GeForce RTX 3070
Memory:32 GB
Software environmentWindow 10 system
Development environmentVS code
Model Computing EnvironmentPytorch
Table 3. Contribution comparison of different components.
Table 3. Contribution comparison of different components.
Xception BackboneThe Adaptive Attention Mechanism NetworkIOUPrecisionRecallF1
74.82084.35082.17083.246
86.09490.17881.07085.385
71.64082.00077.98879.948
90.53791.22791.30191.264
Table 4. Comparative evaluation of crack detection results of different segmentation methods.
Table 4. Comparative evaluation of crack detection results of different segmentation methods.
ModelsIOUInterference Speed/s
Proposed method90.5370.0192
Deeplabv3+74.8200.0224
FCN78.2040.0215
UNet82.0300.0201
Canny69.2420.5
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhu, Y.; Tang, H. Automatic Damage Detection and Diagnosis for Hydraulic Structures Using Drones and Artificial Intelligence Techniques. Remote Sens. 2023, 15, 615. https://doi.org/10.3390/rs15030615

AMA Style

Zhu Y, Tang H. Automatic Damage Detection and Diagnosis for Hydraulic Structures Using Drones and Artificial Intelligence Techniques. Remote Sensing. 2023; 15(3):615. https://doi.org/10.3390/rs15030615

Chicago/Turabian Style

Zhu, Yantao, and Hongwu Tang. 2023. "Automatic Damage Detection and Diagnosis for Hydraulic Structures Using Drones and Artificial Intelligence Techniques" Remote Sensing 15, no. 3: 615. https://doi.org/10.3390/rs15030615

APA Style

Zhu, Y., & Tang, H. (2023). Automatic Damage Detection and Diagnosis for Hydraulic Structures Using Drones and Artificial Intelligence Techniques. Remote Sensing, 15(3), 615. https://doi.org/10.3390/rs15030615

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop