Deep Convolutional Neural Networks for Image Reconstruction and Damage Recognition in UAV Bridge Inspection
Publicado en línea: 26 mar 2025
Recibido: 27 oct 2024
Aceptado: 23 feb 2025
DOI: https://doi.org/10.2478/amns-2025-0811
Palabras clave
© 2025 Shun Wang et al., published by Sciendo
This work is licensed under the Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.
Bridges are an important part of highways and also an important part of highway maintenance and management [1]. With the increase in traffic flow and automobile tonnage, highway bridges and municipal bridges with disease work are more common, which can easily cause traffic accidents [2-3]. Due to the lack of maintenance units for high-altitude operation, ships, and other related professional testing equipment, resulting in large, extra-large bridges, high-altitude tower columns, cables, steel pipe vaults, and other special important components of the detection work is difficult to carry out normally, resulting in special structural parts of the bridge disease can not be found in a timely manner and dealt with, in the long run for the management and maintenance of the important bridges is extremely detrimental to the maintenance of the bridge [4-8]. In addition to bridge damage also includes natural disasters and other factors [9-10]. Therefore, how to utilize scientific and technological means to monitor and inspect the safety of bridges during the service period to ensure that bridges are used for a long period has become a hot topic of research in the field of bridges at home and abroad [11-12].
At present, the mainstream bridge disease detection method is still using artificial bridge cars, telescopes, and other traditional means [13-14]. It has been proved by practice that with the traditional means of detection, there are many problems, and it is not easy to ensure an all-round detection of the bridge [15]. In recent years, due to the rapid development of drone camera technology, the technology has been utilized in the field of bridge appearance detection [16-17]. That is, the use of drones carrying image acquisition system to collect information, and then the use of image processing technology, noise removal, compression, and reduction of the image so that you can have a more clear intuitive understanding of the bridge’s appearance of the disease condition, so as to improve the detection efficiency and reduce the detection cost [18-22]. Therefore, the application of digital image processing technology in bridge detection, with the ground control equipment and all kinds of powerful information processing software, effectively solves the problem of difficulty in observing the special structural parts of the bridge so that the bridge management and maintenance work into a new height [23-27].
In this paper, the DJI Mavic Ar2i quadcopter UAV is selected as the bridge image acquisition device, and the ideal bridge image is successfully acquired by setting the shooting distance. For the problems of redundancy and useless information in the original bridge images, the wavelet transform is utilized to process the original bridge images to eliminate the repetition of the image information and, at the same time, remove the noise in the images. In order to improve the visual effect of the image, dynamic compression is performed on the bridge rail image after wavelet analysis, and nonlinear transformation is used to make the bridge image softer under different lighting environments to enhance the quality of the captured bridge image, and to improve the efficiency and accuracy of image reconstruction. A bridge image reconstruction model is constructed based on the local perception and other capabilities of the deep convolutional neural network, and the enhanced image is input into the deep convolutional neural network for training. A stochastic method is used to initialize the relevant parameters of the deep convolutional neural network, and the gradient descent method is used for updating. Bridge image reconstruction experiments are carried out using the established model to analyze the reconstruction effect of the model in this paper. The MobileNet-v2 lightweight network improves the traditional deep convolutional neural network, and the improved deep convolutional neural network is utilized to extract damaged features from the reconstructed bridge images. Based on the unique local inference ability of the spine neural network, the decision layer of the spine neural network is introduced, and the thought features after spreading are split into two equal parts and input into it, combined with the accelerating effect of activation function on the model training, the identification results of the model on the bridge damage are finally obtained. Different working condition positions and bridge damage indicators are selected to verify the performance of the bridge damage recognition model constructed in this paper in the training set and test set, respectively.
The bridge structure is complex and varied, so the UAV needs to be flexible enough when collecting bridge images. DJI Mavic Ar2i quadcopter is selected for image acquisition, which has a compact body, is equipped with a seven-eye vision system, can hover and fly stably outdoors, and is equipped with a three-axis gimbal camera that can stably take 48-megapixel photos. However, the angle of view of the camera is -90°~+24° (the angle of view is considered as 0°, downward is negative, and upward is positive). Therefore, a light conversion device and a corresponding stabilization device are designed in front of the camera in order to capture the image of the top of the UAV. The main component of the light conversion device is a trapezoidal reflector in front of the lens, which is at an angle of 45° to the vertical plane. When collecting images of the side of the bridge, the original UAV equipment can be used, but when collecting images of the bottom of the bridge, the light conversion device is installed under the fuselage, and the lens faces forward to capture images of the top of the UAV.
The shooting distance is the key factor that affects the pixel size and the actual size. When shooting, the camera is parallel to the shooting plane so that the distance from the objects in the acquisition range to the lens is as consistent as possible. In order to obtain higher quality images as much as possible under the premise of ensuring the safety of the UAV, the shooting distance should be determined according to the pixel accuracy and the requirements of bridge disease detection. According to the relevant technical specifications, the crack width of the concrete bridge and prestressed concrete bridge superstructure is limited to 0.3~0.8mm, so the shooting distance needs to be satisfied with the observation of cracks of 0.1mm wide. According to the pixel accuracy of 5000pixel×5000pixel, the shooting area of a single image is judged to be 50mm×100mm. The orthogonal straight lines and borders of 50mm×100mm are drawn on the white paper. The shooting lens is moved so that the straight lines are all over the lens. The shooting distance is recorded by repeating many times, and the average value of the shooting distance is 560mm, i.e., the shooting distance is required to be less than 560mm. The following formula can express the pixel size and shooting distance:
Where
The original bridge image has a three-dimensional structure, including a large amount of information. There is a certain degree of redundancy between the information, and it also contains a lot of useless information, which negatively affects the efficiency of the image reconstruction and, at the same time, makes the image transmission speed slower and the storage space of the bridge image large. Therefore, the wavelet transform is used for preprocessing of the original bridge image to eliminate the repetition of the image information and at the same time, to remove the image Noise. Let the original bridge image signal be
where
The estimation of
Where,
The bridge image signal is obtained after
After the above process, the bridge image information processing can be realized, and then wavelet analysis is used to denoise the bridge image operation.
In order to make the information in the bridge image better displayed and improve the visual effect of the image, the dynamic compression of the bridge image after wavelet analysis using logarithmic transformation is adopted so that the dark pixels in the image amplitude expanded, making the image closer to the human visual system. Let the bridge image light component be
After the bridge image is logarithmically transformed, it may produce some high-light areas, which makes the bridge image too bright. For this reason, the bridge image is then nonlinearly transformed to make the bridge image softer in different lighting environments:
Where,
After realizing dynamic compression for bridge images, the probability of causing bridge images to be more blurred is quite high, therefore, it is necessary to carry out inverse sharpening processing for bridge images to further enrich the detailed information of bridge images. Let the Gaussian blurring of bridge images [29] can be described as:
Where
Anti-sharpening can be described:
The inverse sharpening process can remove the low-frequency information in the bridge image and retain the high-frequency information of the bridge image so that the detailed information of the bridge image is significantly enhanced.
Deep Convolutional Neural Network (DCNN) belongs to the type of feedforward neural network [30], but deep convolutional neural networks have superior learning performance compared to traditional feedforward neural networks. Deep convolutional neural networks have the following advantages:
Local perception ability: each neuron can perceive the connection between the bridge image pixels, and these local connections can be composed together to get the global connection of the bridge image, laying the foundation for the bridge image reconstruction. Weight sharing ability, different bridge image regions, can be feature extracted by convolutional layer to extract bridge image reconstruction features from different perspectives, more comprehensively describing the connection between bridge image regions. Pooling layer: A deep convolutional neural network can increase the pooling layer in order to reduce the dimension of bridge image reconstruction features, and at the same time, it has strong robustness to the bridge image rotation and other transformations, which reduces the bridge image reconstruction workload and reconstruction time.
DCNN includes many convolutional layers and pooling layers, mapping the input and output layers through the intermediate layers, and its basic structure is shown in Figure 1.

Basic structure of deep convolutional neural network
Step 1: Acquire bridge images by UAV equipment, normalize the bridge images to be reconstructed so that they are of the same size and dimension, and also perform horizontal flipping and panning operations.
Step 2: A denoising method is designed using wavelet transform, which is adapted to perform the denoising operation of bridge images to reduce the negative impact of noise on bridge images.
Step 3: Enhance the denoised bridge image by using the nonlinear transform method to make the bridge image brightness and contrast more reasonable, improve the bridge image clarity, and remove ambiguity.
Step 4: Input the enhanced processed bridge image into the deep convolutional neural network for training, initialize the relevant parameters of the deep convolutional neural network by random method through the convolution, pooling, and other operations, and calculate the error of the output layer.
Step 5: When the output layer error of the deep convolutional neural network is larger than the practical application requirements, it enters the error feedback operation and adopts the gradient descent method to update the relevant parameters of the deep convolutional neural network so as to reduce the output error continuously.
Step 6: When the output layer error of the deep convolutional neural network is less than the actual application requirements, establish the optimal bridge image reconstruction model.
Step 7: Conduct bridge image reconstruction experiments using the established model and analyze its reconstruction effect.
In this section, a deep neural network is first utilized to achieve fast extraction of bridge damage features after image reconstruction. Then, the spine neural network with step-by-step input is constructed to accurately recognize the damage, and the overall framework of the model is shown in Fig. 2. Based on the model parameter migration and the characteristics of the spine neural network module, this paper proposes an improved lightweight deep convolutional neural network with step-by-step input capability. The network obtains 7*7*128 high-dimensional feature maps after deep separable convolution operation, and the redundant information in the shallow features is constructed through the maximum pooling layer to obtain 1*1*1280 highly robust feature maps as the input of the spine neural network. The input part is split into two parts and gradually fed into the spine neural network. Finally, in this paper, the outputs of multiple spine neural networks with local inference ability are fused and the features are input to the fully connected layer to get the final classification results.

Overall network framework
The disadvantage of deep convolutional neural network models is that they are data-dependent and require a large amount of data available for training. In the bridge damage recognition task under UAV inspection, there is a problem with small data samples. Migration learning reduces the sample dependence on data and improves the performance of the model by transferring the a priori knowledge learned by the model in a large number of datasets to the learning process of the target dataset. In this paper, we choose the migration learning strategy based on network parameters. Firstly, the ImageNet dataset is utilized to train the network in the source domain. Secondly, the feature extraction part of the pre-trained network is used for model design to train the target dataset. Finally, the parameters of the classifiers are fine-tuned by freezing the parameter updates of the feature extraction module. This results in a shared model structure and parameters, which improves accuracy while reducing configuration requirements and decreasing training time.
MobileNet-v2, a typical representative of lightweight networks [31], is widely used for image classification tasks. The model tends to have a smaller model space complexity compared to traditional deep convolutional neural networks, allowing it to work on mobile devices and computers with lower computing power. The depth-separable convolution in the model creatively separates the traditional convolution process into depth (DW) convolution and point-by-point (PW) convolution, effectively reducing the computational effort of the model.
One of the basic modules consists of two DW convolutions as well as one PW convolution. The computational procedure is shown below. Suppose the input sample is x. After the DW convolution, the feature mapping of the hidden layer is as in Equation (14):
In Equation (14),
And
In Eqs. (15), (16),
A short join section is added at the beginning and end of each base module to weaken the problem of gradient vanishing by superimposing the input x directly onto the final output when the sliding step of the convolution kernel is 1, and the input feature size is equal to the output feature size. The mapping of the output behaves as in Eq. (17):
Where
The overall feature extraction network input is shown in equation (18):
Where
Spinal neural networks are nonlinear decision-making layers capable of accepting partial information, with local inference capabilities not generally found in fully connected layers. It can reduce the model error rate and improve the model performance. Usually, the one-dimensional features after flattening are split into two equal parts, which are fed into the spine neural network decision layer, and the decision information of the bottom layer and the split inputs are spliced together as the higher spine neural network decision layer. Finally, all the decision information is spliced into the last fully connected layer to get the output.
After the feature extraction process, the feature
The spinal neural network used in this section consists of four ganglia, with the first layer of spinal ganglia having input Al and output
where
The second spinal ganglion layer accepts the outputs
The third spinal ganglion layer accepts the outputs
The last layer of the spinal ganglia accepts the outputs
The fully connected layer accepts the outputs from all the ganglia and self-learns to assign different weights and outputs the classification result
The essence of a deep convolutional neural network is a multilayer composite function. The role of the activation function in the neural network is to use the information in the neural network for nonlinear conversion and increase the expression ability of the neural network so as to better fit the nonlinear complex model. The activation module of the spinal neural network can directly affect the model training effect. The appropriate activation function can play a multiplier effect, while the inappropriate activation function will affect the model fitting effect. In the local decision-making model proposed in this section, the activation function, as an important part of the decision-making module, plays an important role in the response correction of the neural network, which is conducive to speeding up the training process and improving the classification performance. In general, Leaky-ReLU and its variants (PReLU, RReLU) outperform the ReLU activation function in general bridge datasets. In this paper, the UAV bridge inspection dataset is not suitable for the ReLU activation function used by conventional spine neural networks due to its characteristics, so the activation function with superior performance is selected by comparing Leaky-ReLU and its variants (PReLU, RReLU).
After determining the structure of the algorithm, it is necessary to construct a suitable dataset as the training dataset of the model. Concrete material has the characteristics of easy access, good fire resistance, not easy to weather, and is often used as a material for bridge construction. Therefore, this experiment utilizes UAV equipment to collect multiple concrete bridge images (including damaged and undamaged images) on a bridge, and by rotating panning cropping, and other operations on the acquired images, a dataset of more than 5000 concrete bridge images is obtained, which contains a variety of concrete bridge images. The bridge damage images in the dataset include different states of damage, such as transverse cracks, longitudinal cracks, bridge corrosion, and so on. After obtaining the dataset, the dataset is labeled using the Labeling image annotation tool with the type of Crack, and the training and test sets are divided according to the ratio of 7:3.
In order to verify the effectiveness of the wavelet analysis method for bridge image denoising, the peak signal-to-noise ratio (PSNR) and structural similarity (SSIM) are used to objectively evaluate the merits of the denoised image results, which are calculated as follows:
Where,
Four complex image denoising models, including wavelet analysis, full variational method (A), partial differential method (B), and median filtering method (C) in this paper, are respectively utilized to process the noisy bridge images collected by UAV in real-time. In order to reduce the error value of the bridge image denoising results, the signal-to-noise ratio PSNR of the processed bridge images is calculated after processing 20 different bridge images by the four denoising models, and the results are shown in Table 1. Analysis of the data in the table reveals that the wavelet analysis method used in this paper denoises the bridge images with the highest signal-to-noise ratio and the cleanest noise removal to obtain the ideal bridge image.
The bridge image peak signal-to-noise ratio of different methods
Image number | Ours | A | B | C |
---|---|---|---|---|
1 | 45.42 | 15.91 | 21.53 | 13.17 |
2 | 55.67 | 27.47 | 26.06 | 16.7 |
3 | 46.94 | 23.74 | 12.98 | 29.82 |
4 | 49.74 | 26.93 | 16.75 | 24.65 |
5 | 47.12 | 15.44 | 16.27 | 12.55 |
6 | 51.04 | 26.50 | 13.21 | 27.25 |
7 | 58.94 | 27.31 | 26.79 | 27.97 |
8 | 48.83 | 28.95 | 27.93 | 13.89 |
9 | 46.40 | 17.68 | 26.87 | 26.13 |
10 | 48.62 | 26.79 | 26.10 | 16.86 |
11 | 53.62 | 25.11 | 19.46 | 12.93 |
12 | 52.84 | 24.73 | 27.26 | 14.15 |
13 | 56.60 | 20.35 | 14.08 | 18.59 |
14 | 59.39 | 16.52 | 22.49 | 17.32 |
15 | 48.46 | 26.57 | 11.67 | 22.86 |
16 | 51.31 | 18.35 | 26.02 | 26.10 |
17 | 55.70 | 23.85 | 20.18 | 27.9 |
18 | 55.58 | 20.57 | 25.56 | 28.78 |
19 | 46.51 | 15.25 | 17.84 | 19.74 |
20 | 46.65 | 16.01 | 26.57 | 28.13 |
On the basis of the above experiments, further testing of different methods in the bridge image denoising after the mean square error, this indicator represents the size of the error between the processed image and the original image, the smaller the value, indicating that the processed image is closer to the original image, i.e., the better the denoising effect, the results are shown in Figure 3. Analysis can be obtained in this paper wavelet analysis method of bridge image denoising. The mean square error is always maintained within 0.0001, has superior performance, and can achieve a clear display of the bridge image.

Comparison of the mean square error results
Structural similarity (SSIM) is another evaluation index of bridge image denoising processing. This index can evaluate the degree of similarity between the denoised image grayscale and detailed texture features and the original image. The closer the value is to 1 indicates that the closer the image is to the original image, i.e., the better the denoising effect, the results are shown in Table 2. Analyzing the data in the table, the structural similarity of the bridge image after denoising by the wavelet analysis method in this paper is 0.974 on average, which is closer to 1 than other methods, indicating that the wavelet analysis method has a better denoising effect, and it has certain application value.
Comparison of SSIM results
Image number | Ours | A | B | C |
---|---|---|---|---|
1 | 0.987 | 0.777 | 0.683 | 0.750 |
2 | 0.984 | 0.770 | 0.678 | 0.754 |
3 | 0.958 | 0.788 | 0.682 | 0.745 |
4 | 0.975 | 0.753 | 0.656 | 0.747 |
5 | 0.987 | 0.764 | 0.669 | 0.751 |
6 | 0.987 | 0.766 | 0.661 | 0.755 |
7 | 0.984 | 0.790 | 0.664 | 0.737 |
8 | 0.988 | 0.782 | 0.662 | 0.760 |
9 | 0.969 | 0.780 | 0.667 | 0.732 |
10 | 0.987 | 0.769 | 0.666 | 0.760 |
11 | 0.963 | 0.773 | 0.679 | 0.743 |
12 | 0.964 | 0.774 | 0.645 | 0.730 |
13 | 0.966 | 0.764 | 0.683 | 0.723 |
14 | 0.975 | 0.786 | 0.657 | 0.730 |
15 | 0.967 | 0.763 | 0.641 | 0.728 |
16 | 0.983 | 0.775 | 0.660 | 0.749 |
17 | 0.952 | 0.777 | 0.644 | 0.742 |
18 | 0.973 | 0.769 | 0.681 | 0.746 |
19 | 0.969 | 0.755 | 0.645 | 0.741 |
20 | 0.956 | 0.759 | 0.686 | 0.764 |
In order to verify that the deep convolutional neural network model with completed training has good reconstruction effect and generalization ability, i.e., it can reconstruct untrained bridge images. From the test set, 20 bridge images that have completed the denoising process and contain both damaged and undamaged bridges are selected for image reconstruction using the DCNN algorithm. BP neural network (BP), convolutional neural network (CNN), and feedforward neural network (FNN) are selected as the comparison models of this paper’s algorithm to reconstruct the same images under the same experimental environment in order to show the superior performance of this paper’s algorithm.
Fig. 4 shows the loss rate and iteration process changes of each model trained on the dataset, from which it can be seen that when the number of iterations is more than 6000, the loss rate of the DCNN model in this paper tends to be stable and maintains at about 0.02. While BP, CNN and FNN have higher values of loss rate than this paper’s model when the loss rate tends to stabilize.

Model training loss rate
Figure 5 shows the reconstruction accuracy and iterative process variation of each model during training. The image reconstruction accuracy is calculated as:

Model training accuracy
Where
From the figure, it can be seen that when the number of model training times is more than 6000 times, the image reconstruction accuracy under the model of this paper tends to stabilize, which is about 0.995. At this time, the reconstruction accuracy of the BP neural network, convolutional neural network, and feed-forward neural network are 0.798, 0.682, and 0.603, respectively. Compared with the reconstruction accuracy of this paper’s model is improved by 24.69%, 45.89%, and 65.01%, respectively. It shows that the neural convolutional neural network in this paper performs better in the training set and has higher reconstruction accuracy for the bridge images in the training set.
The reconstruction accuracy and reconstruction time of each model on 20 bridge images selected from the test set are shown in Table 3. From the table, it can be clearly seen that the mean value of reconstruction accuracy of deep convolutional neural network in this paper on 20 images is 0.981, while the mean values of reconstruction accuracy of the other three algorithms are 0.823, 0.765, and 0.689, respectively. The reconstruction accuracy of this paper’s model for the bridge images is improved by 19.20%, 28.24%, and 42.38%, respectively. At the same time, the average time required for this model to complete the reconstruction task of 20 bridge images is 0.459s, while the average reconstruction time of the other three algorithms is 6.789s, 13.788s, and 20.570s, respectively, compared with the reduction of the reconstruction time of the bridge images in this paper’s model is 93.24%, 96.67%, and 97.77%, respectively. Combined with the experimental results of the training set, it can be concluded that the deep convolutional neural network used in this paper shows excellent performance of high reconstruction accuracy and short reconstruction time in the bridge image reconstruction task collected by UAVs and is able to efficiently complete the bridge image reconstruction task.
Image reconstruction comparison
Number | DCNN | BP | CNN | FNN | ||||
---|---|---|---|---|---|---|---|---|
Acc | Time/s | Acc | Time/s | Acc | Time/s | Acc | Time/s | |
1 | 0.976 | 0.288 | 0.837 | 7.772 | 0.778 | 13.771 | 0.754 | 19.265 |
2 | 0.965 | 0.066 | 0.747 | 6.232 | 0.819 | 14.257 | 0.632 | 24.307 |
3 | 0.970 | 0.071 | 0.909 | 7.430 | 0.779 | 13.28 | 0.726 | 23.820 |
4 | 0.985 | 0.670 | 0.832 | 7.237 | 0.792 | 15.718 | 0.639 | 19.348 |
5 | 0.998 | 0.793 | 0.789 | 6.208 | 0.823 | 13.643 | 0.641 | 18.086 |
6 | 0.993 | 0.318 | 0.877 | 7.765 | 0.795 | 11.642 | 0.642 | 15.660 |
7 | 0.998 | 0.309 | 0.822 | 5.572 | 0.795 | 12.924 | 0.703 | 20.258 |
8 | 0.979 | 0.059 | 0.768 | 5.949 | 0.769 | 14.637 | 0.696 | 24.699 |
9 | 0.99 | 0.590 | 0.760 | 7.991 | 0.790 | 15.454 | 0.753 | 19.772 |
10 | 0.985 | 0.404 | 0.890 | 5.503 | 0.738 | 14.562 | 0.625 | 16.051 |
11 | 0.977 | 0.876 | 0.745 | 7.476 | 0.771 | 11.004 | 0.665 | 16.611 |
12 | 0.980 | 0.161 | 0.838 | 6.495 | 0.742 | 12.908 | 0.680 | 20.618 |
13 | 0.999 | 0.702 | 0.762 | 5.825 | 0.762 | 15.696 | 0.735 | 24.489 |
14 | 0.985 | 0.779 | 0.812 | 6.284 | 0.780 | 15.526 | 0.706 | 17.989 |
15 | 0.981 | 0.456 | 0.844 | 5.138 | 0.702 | 14.072 | 0.686 | 23.111 |
16 | 0.963 | 0.314 | 0.842 | 6.920 | 0.729 | 14.944 | 0.678 | 20.063 |
17 | 0.986 | 0.779 | 0.906 | 7.680 | 0.756 | 13.074 | 0.756 | 20.746 |
18 | 0.964 | 0.203 | 0.784 | 7.806 | 0.766 | 11.540 | 0.712 | 18.323 |
19 | 0.971 | 0.733 | 0.816 | 6.618 | 0.707 | 12.782 | 0.714 | 23.987 |
20 | 0.982 | 0.599 | 0.873 | 7.869 | 0.710 | 14.328 | 0.644 | 24.188 |
Mean | 0.981 | 0.459 | 0.823 | 6.789 | 0.765 | 13.788 | 0.689 | 20.570 |
The selection of damage indicators is a difficult point in the work of bridge damage identification, the ideal selection of bridge damage indicators should have the following characteristics: sensitive to structural damage. Easy to obtain in the actual engineering. Strong robustness and low sensitivity to testing errors.
At present, the acceleration response signal is used as the damage indicator in damage identification research. In order to study the actual use of different characteristic parameters of bridges for damage identification work, a bridge monitoring parameter is selected as a damage indicator for comparative study.
Deflection is one of the most intuitive indicators to reflect the health condition of the bridge. When the bridge component damage will lead to a decrease in the main frequency of the vibration characteristics of the change, the deflection will become larger. Deflection is also easily accessible data in daily bridge monitoring work.
Therefore, this paper selects deflection and acceleration as damage indicators for damage identification.
Build MobileNet-v2 lightweight network and spinal neural network based on Matlab deep network designer. Bridge Damage Recognition Model Building Based on DCNN During the training process, the denoising processed sample set is re-divided into a training set and test set in the ratio of 7:3. Where the training set samples are used to train the model in order to make the model learn the feature parameters of the data. The test set sample data is used to test the generalization ability of the model.
Fig. 6 shows the model training plot, and (a) and (b) show the sample set model training plot based on deflection data and the sample set model training plot based on acceleration data, respectively. The network extracts the information features of the learning training dataset through three convolution operations and two pooling operations, minimizes the loss rate after several iterations, and then outputs the damage recognition results through the fully connected layer and the classification layer. The model is trained after 200 rounds of iterations to reach the optimal training results.

Model training results
At the end of training the model, Matlab was used to write the code to apply the model, and the test set data samples with the same dimension as the model training were input to verify the model’s damage recognition accuracy. Table 4 shows the results of model damage identification based on the deflection damage index, and Table 5 shows the results of model damage identification based on the acceleration damage index. 1 in the table indicates that there is no damage to the bridge, and 2-10 indicate the working condition positions of the north side beam end of one span, the north side span middle of one span, the south side span middle of one span, the south side beam end of one span, the north side beam end of the second span, the middle of two spans, the south side beam end of two spans, the middle of the main tower, and the bottom of the main tower, etc., respectively. From the table, it can be seen that although the bridge damage identification model constructed in this paper has the situation of misplaced damage identification, the overall model can reach more than 90% of the bridge damage identification rate, which fully demonstrates the good performance of this paper’s model in the bridge damage identification task.
Recognition results based on deflection damage index
Position | Damage degree/% | Recognition result | Recognition rate/% | Position | Damage degree/% | Recognition result | Recognition rate |
---|---|---|---|---|---|---|---|
1 | 0 | 1 | 100 | 6 | 20 | 6 | 100 |
0 | 1 | 100 | 40 | 6 | 100 | ||
0 | 1 | 100 | 60 | 6 | 100 | ||
0 | 1 | 100 | 80 | 6 | 100 | ||
2 | 20 | 2 | 100 | 7 | 20 | 1 | 90.82 |
40 | 2 | 100 | 40 | 7 | 100 | ||
60 | 2 | 100 | 60 | 5 | 91.48 | ||
80 | 2 | 100 | 80 | 8 | 93.56 | ||
3 | 20 | 3 | 100 | 8 | 20 | 1 | 92.38 |
40 | 3 | 100 | 40 | 7 | 95.66 | ||
60 | 3 | 100 | 60 | 8 | 100 | ||
80 | 3 | 100 | 80 | 8 | 100 | ||
4 | 20 | 1 | 92.17 | 9 | 20 | 9 | 100 |
40 | 4 | 100 | 40 | 9 | 100 | ||
60 | 4 | 100 | 60 | 9 | 100 | ||
80 | 4 | 100 | 80 | 9 | 100 | ||
5 | 20 | 1 | 92.36 | 10 | 20 | 10 | 100 |
40 | 5 | 100 | 40 | 10 | 100 | ||
60 | 5 | 100 | 60 | 10 | 100 | ||
80 | 5 | 100 | 80 | 10 | 100 |
Recognition results based on acceleration damage indicators
Position | Damage degree/% | Recognition result | Recognition rate/% | Position | Damage degree/% | Recognition result | Recognition rate |
---|---|---|---|---|---|---|---|
1 | 0 | 1 | 100 | 6 | 20 | 6 | 100 |
0 | 1 | 100 | 40 | 6 | 100 | ||
0 | 1 | 100 | 60 | 6 | 100 | ||
0 | 2 | 99.38 | 80 | 2 | 91.64 | ||
2 | 20 | 1 | 93.74 | 7 | 20 | 7 | 100 |
40 | 2 | 100 | 40 | 1 | 90.39 | ||
60 | 6 | 90.62 | 60 | 7 | 100 | ||
80 | 3 | 95.84 | 80 | 7 | 100 | ||
3 | 20 | 7 | 90.71 | 8 | 20 | 8 | 100 |
40 | 3 | 100 | 40 | 2 | 91.26 | ||
60 | 3 | 100 | 60 | 8 | 100 | ||
80 | 3 | 100 | 80 | 8 | 100 | ||
4 | 20 | 4 | 100 | 9 | 20 | 9 | 100 |
40 | 5 | 94.76 | 40 | 9 | 100 | ||
60 | 4 | 100 | 60 | 9 | 100 | ||
80 | 4 | 100 | 80 | 9 | 100 | ||
5 | 20 | 2 | 92.83 | 10 | 20 | 9 | 98.47 |
40 | 3 | 93.46 | 40 | 10 | 100 | ||
60 | 5 | 100 | 60 | 10 | 100 | ||
80 | 1 | 90.17 | 80 | 10 | 100 |
Fig. 7 shows the comparison between the real label and the predicted label of the model, and (a) and (b) show the predicted recognition results of the model with deflection data and acceleration as the dataset, respectively. From the figure, it can be seen that the accuracy of damage recognition of the model trained using the two damage metrics is high. Combined with the data in Table 4 and Table 5, the credibility of the model for damage identification generally improves with the aggravation of the damage degree, and the differences between the real label and the predicted label are concentrated in the middle of the second span, the middle of the south span of the first span and the south beam end of the first span, while the error identification of the model damage recognition with acceleration as the damage index is mainly concentrated in the north beam end of the first span and the south beam end of the first span.

Comparison of the real label and forecast label
In this paper, the reconstruction and damage recognition task of bridge images collected by UAV based on a deep convolutional neural network is performed, and wavelet analysis and nonlinear transformation are used to enhance the image quality and improve the accuracy of image reconstruction and damage recognition during the task.
The reconstruction accuracy of the bridge image tends to stabilize at about 0.995 after the model has been trained more than 6000 times, while the reconstruction accuracies of the BP neural network, convolutional neural network, and feedforward neural network at this time are 0.798, 0.682, and 0.603, respectively. Compared with the reconstruction accuracies of this paper, the reconstruction accuracies of this paper’s model are improved by 24.69%, 45.89%, and 65.01%, respectively. In the test set, the average reconstruction accuracy of this paper’s model is improved by 19.20%, 28.24%, and 42.38% compared with the other three models, and the reconstruction time is reduced by 93.24%, 96.67%, and 97.77% compared with the other three algorithms. It shows that the bridge image reconstruction model constructed based on deep neural networks in this paper can significantly reduce the image reconstruction time and efficiently complete the bridge image reconstruction task under the premise of ensuring high reconstruction accuracy. At the same time, the bridge damage recognition model maintains the accuracy of damage recognition in 9 working condition locations, such as the girder end on the north side of the first span of the bridge and the north side of the first span. Through the identification of bridge damage by the model, it can accurately and quickly locate the bridge damage position, significantly improve the efficiency of UAV bridge inspection, shorten the time of damage investigation in bridge maintenance, and extend the service life of the bridge.