Next Article in Journal
Degradation Tendency Prediction for Pumped Storage Unit Based on Integrated Degradation Index Construction and Hybrid CNN-LSTM Model
Previous Article in Journal
Color Sensor Accuracy Index Utilizing Metamer Mismatch Radii
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Improved YOLO-V3 with DenseNet for Multi-Scale Remote Sensing Target Detection

College of Electronic and Information Engineering, Nanjing University of Aeronautics and Astronautics, Nanjing 211106, China
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(15), 4276; https://doi.org/10.3390/s20154276
Submission received: 18 June 2020 / Revised: 25 July 2020 / Accepted: 28 July 2020 / Published: 31 July 2020
(This article belongs to the Section Remote Sensors)

Abstract

:
Remote sensing targets have different dimensions, and they have the characteristics of dense distribution and a complex background. This makes remote sensing target detection difficult. With the aim at detecting remote sensing targets at different scales, a new You Only Look Once (YOLO)-V3-based model was proposed. YOLO-V3 is a new version of YOLO. Aiming at the defect of poor performance of YOLO-V3 in detecting remote sensing targets, we adopted DenseNet (Densely Connected Network) to enhance feature extraction capability. Moreover, the detection scales were increased to four based on the original YOLO-V3. The experiment on RSOD (Remote Sensing Object Detection) dataset and UCS-AOD (Dataset of Object Detection in Aerial Images) dataset showed that our approach performed better than Faster-RCNN, SSD (Single Shot Multibox Detector), YOLO-V3, and YOLO-V3 tiny in terms of accuracy. Compared with original YOLO-V3, the mAP (mean Average Precision) of our approach increased from 77.10% to 88.73% in the RSOD dataset. In particular, the mAP of detecting targets like aircrafts, which are mainly made up of small targets increased by 12.12%. In addition, the detection speed was not significantly reduced. Generally speaking, our approach achieved higher accuracy and gave considerations to real-time performance simultaneously for remote sensing target detection.

1. Introduction

Recently, remote sensing images [1,2,3,4] have attracted more research in the field of computer version (CV) with the rapid development of satellite and imaging technology. There is a significant value on information extraction of remote sensing images. Remote sensing target detection [5,6,7] has important and extensive applications in military, navigation, salvage, and other aspects, which requires high speed and accuracy for target detection algorithms.
The rapid development of computer technology makes it possible for the applications of the convolutional neural network (CNN) [8,9,10,11], which requires high computing power. Compared with traditional target detection algorithms like HOG-SVM (Histogram of Oriented Gradients-Support Vector Machine) [12,13], DPM (Deformable Parts Model) [14,15], and HOG-Cascade [16,17], CNN-based target detection algorithms have great advantages in many aspects such as speed and accuracy. Convolutional neural network (CNN) is a kind of feed forward neural network with convolutional computing and it usually has a deep structure. It is one of the most important components of deep learning [18,19,20]. Recently, the research of deep learning in target detection has become a hot spot. The CNN-based target detection models can mainly be divided into two categories, which include the two-stage ones and the one-stage ones.
Currently, the two-stage ones are represented by R-CNN [21], and then Fast R-CNN [22,23], Faster R-CNN [24,25], and Mask R-CNN [26,27], which have been developed on the basis of it. As the name implies, the two-stage target detection algorithms divide the detection processes into two steps. First, the Region Proposed Network (RPN) [28,29,30] is used to extract the information of the targets and then the detection layers predict location and category information of the targets. The other ones are one-stage target detection algorithms including SSD (Single Shot Multibox Detector) [31,32,33], DSSD (Deconvolution Single Shot Multibox Detector) [34], FSSD (Feature Fusion Single Shot Multibox Detector) [35], YOLO [36], YOLO-V2 [37], and YOLO-V3 [38]. Instead of using the region proposed network (RPN), the one-stage algorithms obtain the predictive information of location and category directly. Therefore, they are also called the regression-based algorithms and they can usually achieve higher detection speed than the two-stage ones. At present, numerous state-of-the-art target detection models with higher speed are proposed based on YOLO such as YOLO-V3 tiny [39] and TF-YOLO [40]. Therefore, the accuracy of them is not satisfactory.
From the current research, remote sensing target detection usually faces the following challenges: one is that remote sensing targets are usually small and take up fewer pixels, which makes it difficult to extract features. The second challenge is that remote sensing images are usually disturbed by shadow, light, and other external factors. In addition, the scales of the remote sensing targets are usually different. To solve these problems, researchers have made unremitting efforts.
In order to realize remote sensing target detection, the inchoate research is mainly based on template matching, which is to match the target with a specific template for detection. For example, Weber et al. [41] proposed a method of making use of image analysis to extract coastline templates and adopted this method to detect oil tanks. It achieved good results. However, although the method of template matching is simple and effective, its overall robustness is poor and it is sensitive to the shapes of the targets and geometric deformation. The algorithms based on image analysis are to judge whether each region of the remote sensing image has a target by segmentation and classification. For example, Feng et al. [42] proposed the algorithm named multi-resolution segmentation, which segmented remote sensing images into multiple regions for detection by three parameters: shape, scale, and density. Compared with the method of template matching, this method is more flexible and can combine contextual semantic information, which has achieved good results in some tasks. However, this kind of algorithm still needs to be designed manually for segmentation, which is not universal.
Compared with the previous two algorithms, the remote sensing target detection algorithms based on deep learning have better accuracy and robustness because they no longer use the features of manual design. Sun et al. [43] extracted the region of interest with sliding windows, and then used the features of Bag-of-Words to detect the targets. Zhang et al. [44] and Yu et al. [45] combined the prior characteristics of the airports and coasts, respectively, with deep learning to conduct remote sensing target detection. More commonly, researchers use existing target detection algorithms such as Faster R-CNN in remote sensing target detection tasks. However, when these models are applied to remote sensing target detection tasks, their performance is poor due to the factors such as illumination, cloud cover, and complex background interference.
As an advanced target detection model, YOLO-V3 adopts a feature pyramid network (FPN) [46,47], ResNet (Residual Network) [48], and achieves good performance in speed and accuracy. YOLO-V3 predicts targets at three different scales. Compared with the previous two versions, YOLO-V3 enhances the capability of detecting multi-scale targets, especially small targets. Abundant improved algorithms have been proposed since YOLO-V3 came out. References [49,50] adopted four detection layers to enhance the performance of detecting small targets. Reference [51] adopted circular ground truth to realize tomato detection. Reference [52] increased another shortcut connection to concatenate 2 CBLs (Convolution-Batch Normalization-Leak ReLU) between two ‘residual units’ to enhance the performance for the feature extraction network of information transfer. Reference [40] simplified the feature extraction network to obtain faster detection speed and adopted multiple layers concatenation to enhance the performance of feature extraction. The above improved algorithms achieved a good detection effect. However, the resolution of remote sensing images is large. The scales of remote sensing targets are small and the backgrounds are complex. These algorithms, which have excellent performance on routine datasets, are not suitable for remote sensing target detection. Therefore, we need to design a feature extraction network and detection networks for our proposed algorithm elaborately.
According to the characteristics of remote sensing targets, the proposed method was improved based on the YOLO-V3 model. The main contributions in this paper include the following. (1) In order to reduce reliance on ResNet and enhance the ability of feature information extraction, which is inspired by DenseNet, improved densely connected units proposed to replace some of the residual units of Darknet53. (2) To further improve the ability of detecting multi-scale remote sensing targets, we extended the original three output layers of YOLO-V3 to 4. (3) In order to avoid gradient vanishing, instead of five convolutional layers in each detection layers, three residual units were adopted. The experimental results on remote sensing images show that the proposed method not only has good performance in accuracy, but also gives attention to real-time performance for remote sensing target detection.
The rest of this paper is as follows. In Section 2, we introduced the theory of YOLO and the framework of YOLO-V3. In Section 3, we described the improved method of our approach in details. Section 4 gives the experiments of the proposed algorithm on the RSOD dataset and compared the performance of our approach with other classical algorithms. Lastly, the conclusion is shown in Section 5.

2. The Theory of YOLO

YOLO (You Only Look Once) is a kind of one-stage algorithm, which transforms target detection as a regression problem. Compared with Faster R-CNN, YOLO obtain the predictive information of location and categories directly without a region proposed network (RPN). After continuous development, YOLO has been developed from YOLO-V1 to YOLO-V2 and the latest YOLO-V3.

2.1. The Principle of YOLO

At the beginning, the network divides each input image into S × S grid cells. The grid, which center on the ground truth (GT) of the target falls in, is responsible for detecting it. Each grid cell defines B bounding boxes as well as their corresponding confidence scores. Each bounding box contains C classes. We denote them as P ( C l a s s i | O b j e c t ) . If the center of the target falls in the grid cell, then P ( O b j e c t ) = 1 . Otherwise, P ( O b j e c t ) = 0 . The confidence score is defined as: P ( O b j e c t ) × IOU p r e d t r u t h . It reflects the probability that the grid cell contains targets and the accuracy that the bounding box predicts. IOU represents the overlap area between the bounding box and the ground truth (GT). The class-specific scores can be denoted in Equation (1).
P ( C l a s s i | O b j e c t ) × P ( O b j e c t ) × IOU p r e d t r u t h = P ( C l a s s ) × IOU p r e d t r u t h
YOLO has made greater achievements than Faster R-CNN in terms of speed, but it also brings the low accuracy of detection. On the basis of YOLO-V1, YOLO-V2 introduces the concept of the anchor box and runs k-means on the dataset to generate appropriate prediction boxes at the beginning. Instead of full connected layers (FC), YOLO-V2 introduces convolutional layers in the output end. In addition, YOLO-V2 also adopts Batch Normalized, New feature extraction network (Darknet19), which greatly improves the performance compared with YOLO-V1.
YOLO-V3 is a further improved version based on YOLO-V2 by upgrading the original Darknet19 to Darknet53 and adopts multi-scale detection layers (three scales) to detect the targets. This allows YOLO-V3 to detect small targets more effectively.

2.2. The Network of YOLO-V3

YOLO-V3 adopts Darknet53 as its feature extraction network. In order to prevent information loss caused by pooling layers, Darknet53 adopts a full convolutional network (FCN). The network is basically made up of convolutional kernels of 1 × 1 or 3 × 3. Since it contains 53 convolutional layers, it is called Darknet53. In order to extract deeper features and avoid gradient fading by drawing on the residual network, Darknet53 added five residual modules to the network in which each was composed of one or multiple residual units.
YOLO-V3 borrows the idea of the feature pyramid network (FPN). The network carries out five times of the down-sampling processing on each input image. The output feature map of the feature extraction is down-sampled by 32×, which means the output feature map is 1/32 of the size of the input image. Then YOLO-V3 transmits the last three down-sampled layers to the detection layers for target detection. The network of YOLO-V3 predicts at three scales. The sizes of the three scales are 13 × 13, 26 × 26, and 52 × 52, which are responsible to detection big targets, medium-sized targets, and small targets, respectively. The deep-level feature maps contain a mass of semantic information while the shallow-level feature maps contain a mass of fine-grained information. Therefore, to carry out feature fusion, the network uses up-sampling to keep the size of the feature map down-sampled by 32×, which is consistent with the feature map down-sampled by 16×, and then merges the feature maps by concatenation. Similarly, we do the same for the feature map down-sampled by 16× and the feature map down-sampled by 8×. The structure of YOLO-V3 and its feature extraction network are shown in Figure 1 and Table 1, respectively.

3. Related Work

3.1. Improved Densely Connected Network

The improvement of You Only Look Once (YOLO)-V3 is mainly based on the concept of a residual network. Darknet53 uses several residual units, and the ResNet made up of these residual units contains a large number of parameters and it is responsible for the main calculations for YOLO-V3 network. Unlike ResNet, which adds the values of the subsequent layers by constructing an identity map, DenseNet [53] connects all the layers for channel merging to achieve feature reuse. Compared with ResNet, the back propagation of the gradient is enhanced, which can make better use of feature information and improve the transmittance of the information between layers.The structure of DenseNet is shown in Figure 2.
In Figure 2, x 1 , x 2 , x 3 , and x 4 represent the feature maps of the output layers, while H 1 , H 2 , H 3 , and H 4 refers to the nonlinear transformations. The network contains l ( l + 1 ) / 2 connections with l layers. Each layer is connected to all the other layers. Thus, each layer can receive all the feature maps of the preceding ( l 1 ) layers. The feature map of each layer can be expressed in Equation (2).
x l = H l [ x 0 , x 1 , , x l 1 ]
The proposed densely connected network in this paper borrows from the idea of residual units in Figure 1. The convolution, Batch Normalization, and Leaky-ReLU make up the CBL module, while two CBL modules are cascaded into a Double-CBL (DCBL) module. We use the DCBL module as transport layer H i : Conv ( 1 × 1 × 32)-BN-ReLU-Conv (3 × 3 × 64)-BN-ReLU and Conv (1 × 1 × 64)-BN- ReLU- Conv (3 × 3 × 128)-BN-ReLU. Thus, too many layers of DenseNet will lead the feature maps getting redundant and decrease the speed of detection, we set four layers for each module. The increment of the feature maps for each layer in module ’DENSE 1st’ is 64 while the increment of the feature maps for each layer in module ’DENSE 2nd’ is 128.
With the aim of reducing the network’s dependence on residual units, a part of the lower resolution layers of the feature extraction network is replaced by the improved densely connected network. The structure diagram of the proposed feature extraction network is shown in Figure 3.
To show the structure of our approach in detail, Table 2 gives the feature extraction network of our approach.

3.2. The Proposed Algorithm with Multi-Scale Detection

For an input image of 416 × 416, the size of the feature maps of the three detection layers are 13 × 13, 26 × 26, and 52 × 52, respectively. The smaller the size of the feature map is, the larger the area in the input image is in which each grid cell will correspond. On the contrary, the larger the size of the feature map is, the smaller the area in the input image is in which each grid cell will correspond. It means the 13 × 13 detection layer is suitable for detecting large targets, while the 52 × 52 detection layer is suitable for detecting small targets. Generally speaking, remote sensing images contain a large amount of small targets. In order to further enhance the detection performance of remote sensing targets, we need a larger-sized detection scale. The size of the new scale is 104 × 104. Compared with the original three scales, the four detecting scales strategy is suitable for detecting smaller-sized targets.
Furthermore, in order to avoid gradient fading, we replace the five convolutional layers with three residual units, which is in front of each detection layer. The structure of residual units and the proposed network are shown in Table 3 and Figure 4, respectively.
Table 1 and Table 2 show the structure of the feature extraction network of YOLO-V3 and our approach, respectively. Table 3 shows the structure of the residual units, which is in the end of four detection layers of our proposed network. In the end, Figure 4 exhibits the massive structure of our proposed network.

3.3. K-Means for Anchor Boxes

Inspired by Faster-RCNN, YOLO-V2 and YOLO-V3 introduced the ideal of the anchor box to predict the bounding boxes more accurately. In our approach, we ran K-means to generate the anchor boxes. The function of the K-means algorithm is conducting latitude clustering to make anchor boxes and adjacent ground truth have larger IOU values, which is not directly related to the size of anchor boxes.
d ( b o x , c e n t r o i d ) = 1 IOU ( b o x , c e n t r o i d )
IOU refers to the intersection ratio and it is defined in Equation (4).
IOU = S o v e r l a p S u n i o n
S o v e r l a p refers to the overlap area between the predicted box and the ground truth and S u n i o n refers to the union area between them. The pseudocode of K-means in this paper is shown in Algorithm 1.
Algorithm 1: The pseudocode of K-means
1: Given K cluster center points: ( W i , H i ) , i { 1 , 2 , , k } , W i , H i refer to the width and height of each anchor box.
2: Calculate the distance between each ground truth and each cluster center:
d ( b o x , c e n t r o i d ) = 1 IOU ( b o x , c e n t r o i d ) . Since the position of the anchor box is not fixed, the center point of each ground truth is coincident with the clustering center.
3: Recalculate the cluster center for each cluster: W i = 1 N i w i , H i = 1 N i h i
4: Repeat step 2 and step 3 until the clusters converge.
We ran the K-means algorithm to get anchor boxes. In Figure 5, we can see the average IOU with a different number of clusters. The curve got more flat when the number increased. Since there are four detection layers in the network of our approach, we selected 12 clusters (anchor boxes). The sizes of the anchor boxes are as follows: (21, 24), (25, 31), (33, 41), (51, 54), (61, 88), (82, 91), (109, 114), (121, 153), (169, 173), (232, 214), (241, 203), (259, 271). Among them, (21, 24), (25, 31), (33, 41) are the anchor boxes for Scale 4. (51, 54), (61, 88), (82, 91) are the anchor boxes for Scale 3. (109, 114), (121, 153), (169, 173) are the anchor boxes for Scale 2 and (232, 214), (241, 203), (259, 271) are the anchor boxes for Scale 1.

3.4. Relative to the Grid Cell

When detecting the targets, we need to get the values of bounding boxes based on the predicted values. The process is shown in Figure 6. In Figure 6, t x , t y , t w , and t h represent the predicted values of the network. c x and c y represent the offset of the gird relative to the upper left. The values of bounding boxes can be represented as:
b x = σ ( t x ) + c x b y = σ ( t y ) + c y b w = p w e t w b h = p h e t h σ ( x ) = 1 / ( 1 + e x )

3.5. The NMS Algorithm for Merging Bounding Boxes

Since there may be several bounding boxes corresponding to one target, the last step of our approach is to conduct non-maximum suppression (NMS) of the bounding boxes, which is aimed at eliminating unnecessary boxes. The steps of NMS are below.
  • Step 1: Take the bounding box with the highest confidence as the target for comparison. Then we compare the IOU between the bounding box and remaining boxes.
  • Step 2: If the IOU is larger than the threshold we set, then remove the bounding box from the remaining bounding boxes.
  • Step 3: Take the bounding box with the second highest confidence as the target for comparison and repeat Step 1 and Step 2 until all the bounding boxes are left.
The pseudocode of the algorithm is summarized in Algorithm 2.
Algorithm 2: The pseudocode of non-maximum suppression (NMS) for our approach
Original Bounding Boxes:
   B = [ b 1 , , b M ] , C = [ c 1 , , c M ] , t h r e s h o l d = 0.6
   B refers to the set of original bounding boxes
   C refers to the set of confidences of B
Detection result:
   F refers to the set of the final bounding boxes
1:   F [ ]
2:   while B [ ] do:
3:     k arg max C
4:     F F . a p p e n d ( b k ) ; B d e l B [ b k ] ; C d e l C [ c k ]
5:     for b i B do:
6:       if IOU ( b i , b k ) t h r e s h o l d
7:        B d e l B [ b i ] ; C d e l C [ c i ]
8:       end
9:     end
10:  end

4. Experiment and Results

In order to verify the validity of our improved YOLO-V3 for remote sensing target detection, we compared our approach with original YOLO-V3, YOLO-V3 tiny, and other state-of-the-art algorithms on RSOD and the USC-AOD dataset. The conditions of our experiment are as follows: Framework: python 3.6.5 and tensorflow 1.13.1, Operating system: Windows 10, CPU: i7-7700k, and GPU: NVIDIA GeForce RTX 2070. We set 50,000 training steps in this experiment. The learning rate of the model decreased from 0.001 to 0.0001 after 30,000 steps and to 0.00001 after 40,000 steps. We set the same parameters for other comparison algorithms. The initialization parameters of training lies in Table 4.

4.1. Loss Function

When training the network, loss function is used to measure the error between the predicted and true value. The loss function of the network can be defined in Equation (6).
L o s s = E r r o r c o o r d + E r r o r i o u + E r r o r c l s
E r r o r c o o r d refers to a coordinate prediction error and it can be defined as:
E r r o r c o o r d = λ c o o r d i = 1 s 2 j = 1 B I i j o b j [ ( x i x ¯ i ) 2 + ( y i y ¯ i ) 2 ] + λ c o o r d i = 1 s 2 j = 1 B I i j o b j [ ( w i w ¯ i ) 2 + ( h i h ¯ i ) 2 ]
In Equation (7), λ c o o r d refers to the weight of the coordinate error and we selected λ c o o r d = 5 in our model. S 2 refers to the number of the grids ( S × S ). B refers to the number of bounding boxes per grid. Ι i j o b j refers to whether there is an object that falls in the jth bounding box of the ith grid cell. ( x ¯ i , y ¯ i , w ¯ i , h ¯ i ) and ( x i , y i , w i , h i ) refer to the center coordinate, height, and width of the predicted box and the ground truth, respectively.
E r r o r i o u refers to an IOU error and it is defined as:
E r r o r i o u = i = 1 s 2 j = 1 B I i j o b j ( C i C ¯ i ) 2 + λ n o o b j i = 1 s 2 j = 1 B I i j n o o b j ( C i C ¯ i ) 2
λ n o o b j refers to the confidence penalty when there is no object and we selected λ n o o b j = 0.5 in our model. C i And C ¯ i refer to the true and predicted confidence, respectively.
E r r o r c l s refers to the classification error and it is defined as:
E r r o r c l s = i = 1 s 2 j = 1 B I i j o b j c c l a s s e s ( p i ( c ) p ^ i ( c ) ) 2
where c refers to the number of classes of the targets.

4.2. The Evaluation Indicators

Based on the classification accuracy and prediction accuracy, the samples can be divided into four categories: TP (true positive), FP (fault positive), TN (true negative), and FN (fault negative). We define precision and recall in Equation (10) and Equation (11).
Precision = T P T P + F P
Recall = T P T P + F N
Mean average precision (mAP) is a performance metric for predicting target locations and categories. The accuracy and recall are mutually restricted in practice, and there will be ambiguity when compared separately. Therefore, in our experiment, we introduced mAP, which is one of the most important metrics to evaluate the performance of target detection algorithms.

4.3. Experiment on Remote Sensing Target Detection

The classifier trained based on a conventional dataset is not good at detecting remote sensing targets since remote sensing images have their particularities.
  • Scale diversity. Remote sensing images can be taken from hundreds of meters to nearly 10,000 meters in height, and ground targets may be of different sizes even if they are of the same kind. For example, ships in ports may be only tens of meters to more than 300 meters in size.
  • Perspective particularity. The perspective of remote sensing images is basically overhead, but most of the conventional datasets are still ground level, so the mode of the same target is usually different. The detector trained well on the conventional datasets, which may have a poor effect on the remote sensing images.
  • Problem of small targets. Most of the remote sensing targets are small in size. As a result, the target information is limited. The information of the targets has been lost due to the down sampling layers of the Convolutional Neural Network (CNN). After four times of down sampling, the feature map of the target with 24 × 24 pixels may take up only 1 pixel.
  • Problem of multi-directions. The viewing angle of remote sensing images are usually overhead, while the directions of the targets are uncertain while there is a degree of certainty in conventional datasets.
  • The high complexity of the background. The fields of remote sensing images are relatively large (usually covering several square kilometers). The fields of vision may contain various backgrounds, which will produce strong interference to the target detection.
Based on the above reasons, it is often difficult to train an ideal target detector from conventional datasets for target detection tasks of remote sensing images. A special remote sensing image database is needed.

4.3.1. Dataset Analysis

Taking everything into consideration, we selected the RSOD and UCS-AOD dataset in the experiment. RSOD is the dataset of aerial images. It contains the targets of four categories: aircraft, playground, overpass, and oil tank. UCS-AOD is the dataset of target detection in aerial images. We generally consider the target, which the ground truth takes up less than 0.12% of the whole image as a small target. The ground truth takes up 0.12–0.5%, which is a medium target, and the ground truth takes up more than 0.5%, which is a large target. Of the four categories, the aircraft targets are mostly small in size. The oil tank targets are major of a small or medium size. The playground and overpass targets are big in size. The dataset includes targets under different lighting conditions and at different heights, and the shooting angles of the targets are also different.
Table 5 and Table 6 show the statistics of our remote sensing datasets. The targets in the dataset are mainly small or medium in size, and the distribution of the targets is relatively dense, which increases the difficulty of target detection. Figure 7 contains eight samples of the datasets in this paper. The targets in these samples are under a complex background. After a series of convolutional layers and down sampling layers, the targets take up even fewer pixels, which makes it difficult to detect them.

4.3.2. Experimental Results and Analysis in RSOD and UCS-AOD Dataset

In order to compare the accuracy and real-time performance of the algorithms, the mAP and speed of our approach are evaluated. We compared our approach with the state-of-the-art target detection models in the RSOD dataset, and the comparison results are shown in Table 7. Furthermore, the comparison results of the targets with different sizes are shown in Table 8.
Table 7 shows that our approach is superior to other classical algorithms in the index of mAP. The detection speed is not significantly reduced relative to YOLO-V3. For aircrafts and oil tanks, which are mainly small and medium-sized targets, our approach has a clear improvement in detection accuracy compared to YOLO-V3. The experimental results show that our improved YOLO-V3 can effectively detect the remote sensing targets under the complex background in the condition of real-time detection. In Table 8, we divide target categories by size. We can see that our approach has more advantages than YOLO-V3 in detecting small-sized targets.
For the universality of our algorithm, we ran the experiment on the UCS-AOD dataset. The comparison results are shown in Table 9. In addition, from Table 8 and Table 9, we can see that the leak detection rate is significantly lower than YOLO-V3 and other state-of-the-art algorithms.
Under different backgrounds, partial detection results of our approach in RSOD and UCS-AOD dataset are shown in Figure 8. In the conditions of different illumination, different distributions, and different target sizes, our approach can detect the target accurately, which proves excellent detection performance for multi-scale remote sensing targets.

4.3.3. Ablation Experiments

In this section, we need to verify the effectiveness of each improved module we proposed. In order to analyze the influence of module ‘DENSE 1st’ and module ‘DENSE 2nd’ (Figure 3) on the detection accuracy, different combination modes were set up in the experiment under the condition of three detection scales. The experimental results of each combination in the RSOD dataset are shown in Table 10.
It can be seen from the first experiment and the fourth experiment that the feature extraction network of the fourth experiment introduced dense connection modules based on Darknet53. mAP of its model improved from 77.10% to 84.38%. On the other hand, the detection speed of the fourth experiment increased from 29.7 FPS to 32.3 FPS compared to the first experiment. The experimental results show that our proposed feature extraction network can improve the performance of remote sensing target detection and also has advantages in detection speed.
In addition, Table 11 compared the experimental results of each module at the detection end based on an improved feature extraction network. It can be found in the comparison of the first experiment and the second experiment, and the comparison between the third experiment and the fourth experiment, that the fourth detection scale increased and improved mAP up to 5.95% and 5.78%, respectively. Among them, for the small-sized targets like aircraft, the accuracy is improved by 8.72% and 7.04%, respectively. This shows that the increased detection scale can effectively improve the detection accuracy of small targets. Compared with six convolutional layers, the ‘Res 3’ module can avoid gradient fading and reduce the number of parameters. The comparison of experiment 1 and experiment 3, and the comparison between experiment 2 and experiment 4 show that the ‘Res 3’ module can slightly increase the detection speed.
The ablation experiments result in Table 9 and Table 10, which proved that the improved feature extraction network and detection end we proposed can improve the feature extraction ability of the network and enhanced the detection accuracy of multi-scale remote sensing targets, especially small-sized targets. In addition, the detection speed of our approach is not significantly reduced when compared to YOLO-V3 and meets the real-time requirements.

4.3.4. Expansion Experiment

In order to verify the effectiveness of our approach more intuitively, we selected several images and compared the detection results with YOLO-V3 and Faster RCNN. The comparison of the detection results are shown in Figure 9.
In Figure 9, a total of 24 detection results of eight groups were chosen in the RSOD dataset and UCS-AOD dataset to prove the superiority of the improved YOLO-V3. The pictures in the first list are the detection results of the YOLO-V3 network. The pictures in the second list are the detection results of Faster RCNN and the pictures in the third list are the detection results of our approach. It can be clearly seen that there are several small targets missed and detected by YOLO-V3. Although Faster RCNN performed better than YOLO-V3, leak detection still exists. On the other hand, all the targets were detected by our approach. The contrast experiments of eight groups and the ablation experiments showed that, by improving the feature extraction network and increasing the fourth detection scale, our approach enhanced the performance of detecting small targets with complex background conditions in remote sensing images.

5. Conclusions

In practical engineering applications, we need to consider both accuracy and speed of detection. The existing remote sensing target detection algorithms often fail to consider both of them. In this paper, we proposed an improved YOLO-V3-based model for multi-scale remote sensing target detection. On account of the complexity of the background of remote sensing targets, this puts forward a higher requirement for the ability of the network to extract features. In this paper, we focused on improving the original feature extraction network. Several improvements have been introduced to the original YOLO-V3 network. First, in order to extract feature information more effectively, a dense connection network (DenseNet) was introduced in the feature extraction network. Second, to enhance the performance of detecting small-sized targets, we extended the detection scales from 3 to 4. Third, we replaced three residual units with five convolutional layers, which is in each detection layer to avoid gradient fading. We can see from the ablation experiments that each improved module we proposed is effective in improving the detection accuracy. Experiments on RSOD and UCS-AOD datasets show that our approach achieves better performance on multi-scale remote sensing target detection. The improvement on the feature extraction network greatly improved the ability of extracting the features of the targets. The additional fourth detection scale strengthens the performance of detecting small targets. In the case of losing a portion of detection speed, the accuracy is greatly improved, especially for small remote sensing targets compared with YOLO-V3. Although numerous improved networks based on YOLO-V3 have been proposed, they usually detected targets in routine images. When facing complex remote sensing images, they did not do well. On the contrast, with the above measures adopted, our proposed algorithm is more suitable for remote sensing target detection than other state-of-the-art target detection algorithms. In further work, multi-receptive fields for the feature extraction of the network will be researched to boost the performance of remote sensing target detection. In addition, the latest version of YOLO: YOLO-V4 [55] has been proposed and this will be researched in further work.

Author Contributions

D.X. provided the original ideal, finished the experiment and this paper, and collected the dataset. Y.W. contributed the modifications and suggestions to the paper. All authors have read and agreed to the published version of the manuscript.

Funding

The National Nature Science Founding of China under Grant 61573183 and Open Project Program of the National Laboratory of Pattern Recognition (NLPR) under Grant 201900029 funded this research.

Acknowledgments

The authors wish to thank the editor and reviewers for their suggestions and thank Yiquan Wu for his guidance.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations:

The abbreviations in this paper are as follows:
YOLOYou Only Look Once
CVComputer Version
SVMSupport Vector Machine
HOGHistograms of Oriented Gradients
DPMDeformable Parts Model
IOUIntersection over Union
FCFull Connected Layer
FCNFull Convolutional Network
CNNConvolutional Neural Network
GTGround Truth
RPNRegion Proposal Network
FPNFeature Pyramid Network
ResNetResidual Network
DenseNetDensely Connected Network
NMSNon-Maximum Suppression
TPTrue Positive
FPFalse Positive
FNFalse Negative
APAverage Precision
mAPMean Average Precision
FPSFrames Per Second

References

  1. Shi, W.; Jiang, J.; Bao, S.; Tan, D. CISPNet: Automatic Detection of Remote Sensing Images from Google Earth in Complex Scenes Based on Context Information Scene Perception. Appl. Sci. 2019, 9, 4836. [Google Scholar] [CrossRef] [Green Version]
  2. Zhong, Y.; Weng, W.; Li, J.; Zhu, S. Collaborative Cross-Domain $k$ NN Search for Remote Sensing Image Processing. IEEE Geosci. Remote Sens. Lett. 2019, 16, 1801–1805. [Google Scholar] [CrossRef]
  3. Zhu, H.; Zhang, P.; Wang, L.; Zhang, X.; Jiao, L. A multiscale object detection approach for remote sensing images based on MSE-DenseNet and the dynamic anchor assignment. Remote Sens. Lett. 2019, 10, 959–967. [Google Scholar] [CrossRef]
  4. Zhang, Z.; Chen, J.; Liu, Z. SLIC segmentation method for full-polarised remote-sensing image. J. Eng. 2019, 2019, 6404–6407. [Google Scholar] [CrossRef]
  5. Shi, Y.; Wang, W.; Gong, Q.; Li, D. Superpixel segmentation and machine learning classification algorithm for cloud detection in remote-sensing images. J. Eng. 2019, 2019, 6675–6679. [Google Scholar] [CrossRef]
  6. Li, Y.; Xu, J.; Xia, R.; Wang, X.; Xie, W. A two-stage framework of target detection in high-resolution hyperspectral images. Signal Image Video Process. 2019, 13, 1339–1346. [Google Scholar] [CrossRef]
  7. Li, S.; Xu, Y.; Zhu, M.; Ma, S.; Tang, H. Remote Sensing Airport Detection Based on End-to-End Deep Transferable Convolutional Neural Networks. IEEE Geosci. Remote Sens. Lett. 2019, 16, 1640–1644. [Google Scholar] [CrossRef]
  8. Kujawa, S.; Mazurkiewicz, J.; Czekala, W. Using convolutional neural networks to classify the maturity of compost based on sewage sludge and rapeseed straw. J. Clean. Prod. 2020, 258, 120814. [Google Scholar] [CrossRef]
  9. Xiao, B.; Xu, Y.; Bi, X.; Zhang, J.; Ma, X. Heart sounds classification using a novel 1-D convolutional neural network with extremely low parameter consumption. Neurocomputing 2020, 392, 153–159. [Google Scholar] [CrossRef]
  10. Hashimoto, R.; Requa, J.; Dao, T.; Ninh, A.; Tran, E.; Mai, D.; Lugo, M.; El-Hage Chehade, N.; Chang, K.J.; Karnes, W.E.; et al. Artificial intelligence using convolutional neural networks for real-time detection of early esophageal neoplasia in Barrett’s esophagus (with video). Gastrointest. Endosc. 2020, 91, 1264–1271. [Google Scholar] [CrossRef]
  11. Chen, R.-C. Automatic License Plate Recognition via sliding-window darknet-YOLO deep learning. Image Vis. Comput. 2019, 87, 47–56. [Google Scholar] [CrossRef]
  12. Bilal, M.; Hanif, M.S. Benchmark Revision for HOG-SVM Pedestrian Detector Through Reinvigorated Training and Evaluation Methodologies. IEEE Trans. Intell. Transp. Syst. 2020, 21, 1277–1287. [Google Scholar] [CrossRef]
  13. Wang, L.; Wu, J.; Wu, D. Research on vehicle parts defect detection based on deep learning. J. Phys. Conf. Ser. 2020, 1437, 012004. [Google Scholar] [CrossRef]
  14. Zhang, D. Vehicle target detection methods based on color fusion deformable part model. EURASIP J. Wirel. Commun. Netw. 2018, 2018, 1–6. [Google Scholar] [CrossRef]
  15. Shen, J.; Pan, L.; Hu, X. Building Detection from High Resolution Remote Sensing Imagery Based on a Deformable Part Model. Geomat. Inf. Sci. Wuhan Univ. 2017, 42, 1285–1291. (In Chinese) [Google Scholar] [CrossRef]
  16. Chen, J.; Takiguchi, T.; Ariki, Y. Rotation-reversal invariant HOG cascade for facial expression recognition. Signal Image Video Process. 2017, 11, 1485–1492. [Google Scholar] [CrossRef]
  17. Jin, M.; Jeong, K.; Yoon, S.; Park, D.S. Real-time Pedestrian Detection based on GMM and HOG Cascade. In Sixth International Conference on Machine Vision; Verikas, A., Vuksanovic, B., Zhou, J., Eds.; SPIE: Bellingham, WA, USA, 2013; Volume 9067. [Google Scholar]
  18. Xu, Z.; Huo, Y.; Liu, K.; Liu, S. Detection of ship targets in photoelectric images based on an improved recurrent attention convolutional neural network. Int. J. Distrib. Sens. Netw. 2020, 16. [Google Scholar] [CrossRef]
  19. Liu, Z.; Zhang, G.; Zhao, J.; Yu, L.; Sheng, J.; Zhang, N.; Yuan, H. Second-Generation Sequencing with Deep Reinforcement Learning for Lung Infection Detection. J. Healthc. Eng. 2020, 2020. [Google Scholar] [CrossRef] [Green Version]
  20. Xue, D.; Sun, J.; Hu, Y.; Zheng, Y.; Zhu, Y.; Zhang, Y. Dim small target detection based on convolutinal neural network in star image. Multimed. Tools Appl. 2020, 79, 4681–4698. [Google Scholar] [CrossRef]
  21. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. IEEE. Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 24–27 June 2014; pp. 580–587. [Google Scholar] [CrossRef] [Green Version]
  22. Li, X.; Shang, M.; Qin, H.; Chen, L. Fast Accurate Fish Detection and Recognition of Underwater Images with Fast R-CNN; IEEE: Piscataway, NJ, USA, 2015; pp. 921–925. [Google Scholar]
  23. Girshick, R. IEEE. Fast R-CNN. In Proceedings of the 2015 IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar] [CrossRef]
  24. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. In Advances in Neural Information Processing Systems 28; Cortes, C., Lawrence, N.D., Lee, D.D., Sugiyama, M., Garnett, R., Eds.; IEEE Computer Society: Los Alamitos, CA, USA, 2015; Volume 28. [Google Scholar]
  25. Sun, N.; Zhu, Y.; Hu, X. Faster R-CNN Based Table Detection Combining Corner Locating; IEEE Computer Society: Los Alamitos, CA, USA, 2019; pp. 1314–1319. [Google Scholar] [CrossRef]
  26. Kaiming, H.; Gkioxari, G.; Dollar, P.; Girshick, R. Mask R-CNN. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar] [CrossRef]
  27. Huang, Z.; Zhong, Z.; Sun, L.; Huo, Q. Mask R-CNN with Pyramid Attention Network for Scene Text Detection. In Proceedings of the 2019 IEEE Winter Conference on Applications of Computer Vision (WACV), Waikoloa Village, HI, USA, 7–11 January 2019; pp. 1550–5790. [Google Scholar]
  28. Shih, K.-H.; Chiu, C.-T.; Pu, Y.-Y. IEEE. Real-Time Object Detection via Pruning and a Concatenated Multi-Feature Assisted Region Proposal Network. In Proceedings of the 2019 IEEE International Conference on Acoustics, Speech and Signal Processing, Brighton, UK, 12–17 May 2019; pp. 1398–1402. [Google Scholar]
  29. Shree, C.; Kaur, R.; Upadhyay, S.; Joshi, J. Multi-Feature Based Automated Flower Harvesting Techniques in Deep Convolutional Neural Networking. In Proceedings of the 2019 4th International Conference on Internet of Things: Smart Innovation and Usages (IoT-SIU), Ghaziabad, India, 18–19 April 2019; p. 6. [Google Scholar] [CrossRef]
  30. Yuan, J.; Xue, B.; Zhang, W.; Xu, L.; Sun, H.; Zhou, J. RPN-FCN Based Rust Detection on Power Equipment. In 2018 International Conference on Identification, Information and Knowledge in the Internet of Things; Bie, R., Sun, Y., Yu, J., Eds.; Elsevier Science Bv: Amsterdam, The Netherlands, 2019; Volume 147, pp. 349–353. [Google Scholar]
  31. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.-Y.; Berg, A.C. SSD: Single Shot MultiBox Detector. In Computer Vision—ECCV 2016, Pt I; Leibe, B., Matas, J., Sebe, N., Welling, M., Eds.; Springer International Publishing Ag: Cham, Switzerland, 2016; Volume 9905, pp. 21–37. [Google Scholar]
  32. Lin, M.; Bing, L.; Zhiyu, Z.; Aravinda, C.V.; Kamitoku, N.; Yamazaki, K. Oracle Bone Inscription Detector Based on SSD; Springer International Publishing: Cham, Switzerland, 2019; pp. 126–136. [Google Scholar] [CrossRef]
  33. Tang, J.; Yao, X.; Kang, X.; Shun, N.; Ren, F. Position-Free Hand Gesture Recognition Using Single Shot Multibox Detector Based Neural Network. In Proceedings of the 2019 IEEE International Conference on Mechatronics and Automation (ICMA), Tianjin, China, 4–7 August 2019; pp. 2251–2256. [Google Scholar] [CrossRef]
  34. Cui, L.; Ma, R.; Lv, P.; Jiang, X.; Gao, Z.; Zhou, B.; Xu, M. MDSSD: Multi-scale deconvolutional single shot detector for small objects. Sci. China Inf. Sci. 2020, 63, 120113. [Google Scholar] [CrossRef] [Green Version]
  35. Haque, M.F.; Dae-Seong, K. Multi Scale Object Detection Based on Single Shot Multibox Detector with Feature Fusion and Inception Network. J. Korean Inst. Inf. Technol. 2018, 16, 93–100. [Google Scholar] [CrossRef]
  36. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. IEEE. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar] [CrossRef] [Green Version]
  37. Zhang, X.; Qiu, Z.; Huang, P.; Hu, J.; Luo, J. IEEE. Application Research of YOLO v2 Combined with Color Identification. In Proceedings of the 2018 International Conference on Cyber-Enabled Distributed Computing and Knowledge Discovery, Zhengzhou, China, 18–20 October 2018; pp. 138–141. [Google Scholar] [CrossRef]
  38. Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement. 2018. Available online: https://pjreddie.com/media/files/papers/YOLOv3.pdf (accessed on 30 July 2020).
  39. Adarsh, P.; Rathi, P.; Kumar, M. YOLO v3-Tiny: Object Detection and Recognition Using one Stage Improved Model. In Proceedings of the 2020 6th International Conference on Advanced Computing and Communication Systems (ICACCS), Coimbatore, India, 6–7 March 2020; pp. 687–694. [Google Scholar] [CrossRef]
  40. He, W.; Huang, Z.; Wei, Z.; Li, C.; Guo, B. TF-YOLO: An Improved Incremental Network for Real-Time Object Detection. Appl. Sci. 2019, 9, 3225. [Google Scholar] [CrossRef] [Green Version]
  41. Weber, J.; Lefevre, S. A multivariate Hit-or-Miss Transform for Conjoint Spatial and Spectral Template Matching. In Image and Signal Processing; Elmoataz, A., Lezoray, O., Nouboud, F., Mammass, D., Eds.; Springer-Verlag Berlin: Berlin, Germany, 2008; Volume 5099, pp. 226–235. [Google Scholar]
  42. Feng, T.; Ma, H.; Cheng, X.; Zhang, H. Calculation of the optimal segmentation scale in object-based multiresolution segmentation based on the scene complexity of high-resolution remote sensing images. J. Appl. Remote Sens. 2018, 12, 025006. [Google Scholar] [CrossRef]
  43. Sun, H.; Sun, X.; Wang, H.; Li, Y.; Li, X. Automatic Target Detection in High-Resolution Remote Sensing Images Using Spatial Sparse Coding Bag-of-Words Model. IEEE Geosci. Remote Sens. Lett. 2012, 9, 109–113. [Google Scholar] [CrossRef]
  44. Zhang, P.; Niu, X.; Dou, Y.; Xia, F. Airport Detection on Optical Satellite Images Using Deep Convolutional Neural Networks. IEEE Geosci. Remote Sens. Lett. 2017, 14, 1183–1187. [Google Scholar] [CrossRef]
  45. Yu, Y.; Yang, X.; Xiao, S.; Lin, J. Automated Ship Detection from Optical Remote Sensing Images. In Advanced Materials in Microwaves and Optics; Wang, D., Ed.; Trans Tech Publications Ltd.: Zurich, Switzerland, 2012; Volume 500, pp. 785–791. [Google Scholar]
  46. Guo, C.; Fan, B.; Zhang, Q.; Xiang, S.; Pan, C. AugFPN: Improving Multi-scale Feature Learning for Object Detection. arXiv 2019, arXiv:1912.05384. [Google Scholar]
  47. Wong, F.; Hu, H. Adaptive learning feature pyramid for object detection. IET Comput. Vis. 2019, 13, 742–748. [Google Scholar] [CrossRef]
  48. Zeng, Y.; Ritz, C.; Zhao, J.; Lan, J. Attention-Based Residual Network with Scattering Transform Features for Hyperspectral Unmixing with Limited Training Samples. Remote Sens. 2020, 12, 400. [Google Scholar] [CrossRef] [Green Version]
  49. Li, J.; Gu, J.; Huang, Z.; Wen, J. Application Research of Improved YOLO V3 Algorithm in PCB Electronic Component Detection. Appl. Sci. 2019, 9, 3750. [Google Scholar] [CrossRef] [Green Version]
  50. Ju, M.; Luo, H.; Wang, Z.; Hui, B.; Chang, Z. The Application of Improved YOLO V3 in Multi-Scale Target Detection. Appl. Sci. 2019, 9, 3775. [Google Scholar] [CrossRef] [Green Version]
  51. Liu, G.; Nouaze, J.C.; Mbouembe, P.L.T.; Kim, J.H. YOLO-Tomato: A Robust Algorithm for Tomato Detection Based on YOLOv3. Sensors 2020, 20, 2145. [Google Scholar] [CrossRef] [Green Version]
  52. Liu, M.; Wang, X.; Zhou, A.; Fu, X.; Ma, Y.; Piao, C. UAV-YOLO: Small Object Detection on Unmanned Aerial Vehicle Perspective. Sensors 2020, 20, 2238. [Google Scholar] [CrossRef] [Green Version]
  53. Zhu, Y.; Newsam, S. IEEE. Densenet for Dense Flow. In Proceedings of the 2017 24th IEEE International Conference on Image Processing, Beijing, China, 17–20 September 2017; pp. 790–794. [Google Scholar]
  54. Huang, Z.; Wang, J. DC-SPP-YOLO: Dense connection and spatial pyramid pooling based YOLO for object detection. Inf. Sci. 2020, 522, 241–258. [Google Scholar] [CrossRef] [Green Version]
  55. Bochkovskiy, A.; Chien-Yao, W.; Liao, H.Y.M. YOLOv4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
Figure 1. The network of You Only Look Once (YOLO)-V3.
Figure 1. The network of You Only Look Once (YOLO)-V3.
Sensors 20 04276 g001
Figure 2. The structure of DenseNet.
Figure 2. The structure of DenseNet.
Sensors 20 04276 g002
Figure 3. The structure diagram of the feature extraction network.
Figure 3. The structure diagram of the feature extraction network.
Sensors 20 04276 g003
Figure 4. The Structure of the proposed network.
Figure 4. The Structure of the proposed network.
Sensors 20 04276 g004
Figure 5. The relationship between the number of clusters and average IOU by K-means clustering.
Figure 5. The relationship between the number of clusters and average IOU by K-means clustering.
Sensors 20 04276 g005
Figure 6. The final prediction.
Figure 6. The final prediction.
Sensors 20 04276 g006
Figure 7. The samples of the datasets: (a) the samples of remote sensing object detection (RSOD) dataset and (b) the samples of dataset of object detection in aerial images (UCS-AOD) dataset.
Figure 7. The samples of the datasets: (a) the samples of remote sensing object detection (RSOD) dataset and (b) the samples of dataset of object detection in aerial images (UCS-AOD) dataset.
Sensors 20 04276 g007
Figure 8. The detection results of the improved You Only Look Once (YOLO)-V3.
Figure 8. The detection results of the improved You Only Look Once (YOLO)-V3.
Sensors 20 04276 g008aSensors 20 04276 g008b
Figure 9. The comparison results of YOLO-V3 and our approach: (a1a8) The detection results of YOLO-V3; (b1b8) The detection results of Faster RCNN; (c1c8) The detection results of our approach.
Figure 9. The comparison results of YOLO-V3 and our approach: (a1a8) The detection results of YOLO-V3; (b1b8) The detection results of Faster RCNN; (c1c8) The detection results of our approach.
Sensors 20 04276 g009aSensors 20 04276 g009b
Table 1. The feature extraction network of You Only Look Once (YOLO)-V3.
Table 1. The feature extraction network of You Only Look Once (YOLO)-V3.
LayerFilterSizeOutput
Convolutional323 × 3416 × 416 × 32
Convolutional643 × 3/2208 × 208 × 64
Convolutional321 × 1
Convolutional643 × 3
Residual 208 × 208 × 64
Convolutional1283 × 3/2104 × 104 × 128
Convolutional641 × 1
Convolutional1283 × 3
Residual 104 × 104 × 128
Convolutional2563 × 3/252 × 52 × 256
Convolutional1281 × 1
Convolutional2563 × 3
Residual 52 × 52 × 256
Convolutional5123 × 3/226 × 26 × 512
Convolutional2561 × 1
Convolutional5123 × 3
Residual 26 × 26 × 512
Convolutional10243 × 3/213 × 13 × 1024
Convolutional5121 × 1
Convolutional10243 × 3
Residual 13 × 13 × 1024
Table 2. The feature extraction network of our approach.
Table 2. The feature extraction network of our approach.
LayersFilterSizeOutput
Convolutional323 × 3416 × 416 × 32
Convolutional643 × 3/2208 × 208 × 64
Convolutional321 × 1
Convolutional643 × 3
Residual 208 × 208 × 64
Convolutional1283 × 3/2104 × 104 × 128
Convolutional641 × 1
Convolutional1283 × 3
Residual 104 × 104 × 128
Convolutional2563 × 3/252 × 52 × 256
Convolutional1281 × 1
Convolutional2563 × 3
Residual 52 × 52 × 256
Convolutional321 × 1
Convolutional643 × 3
DenseNet 52 × 52 × 512
Convolutional5123 × 3/226 × 26 × 512
Convolutional2561 × 1
Convolutional5123 × 3
Residual 26 × 26 × 512
Convolutional641 × 1
Convolutional1283 × 3
DenseNet 26 × 26 × 1024
Convolutional10243 × 3/213 × 13 × 1024
Convolutional5121 × 1
Convolutional10243 × 3
Residual 13 × 13 × 1024
Table 3. The structure of residual units.
Table 3. The structure of residual units.
Convolutional5121 × 1
Convolutional10243 × 3
Residual (RES 1st) 13 × 13 × 1024
The Structure of RES 1st
Convolutional2561 × 1
Convolutional5123 × 3
Residual (RES 2nd) 26 × 26 × 512
The Structure of RES 2nd
Convolutional1281 × 1
Convolutional2563 × 3
Residual (RES 3rd) 52 × 52 × 256
The Structure of RES 3rd
Convolutional641 × 1
Convolutional1283 × 3
Residual (RES 4th) 104 × 104 × 128
The Structure of RES 4th
Table 4. The initialization parameters of training.
Table 4. The initialization parameters of training.
Input SizeBatch SizeMomentumLearning RateTraining Step
416 × 41680.90.001–0.0000150,000
Table 5. Remote sensing object detection (RSOD) dataset statistics.
Table 5. Remote sensing object detection (RSOD) dataset statistics.
DatasetClassImageInstancesTarget Amount
SmallMediumLarge
Training
Set
Aircraft44649933714833446
Oil tank1651586724713149
Overpass17618000180
Playground189 191012179
Test
Set
Aircraft1761257741359157
Oil tank6356725721397
Overpass36410041
Playground49520052
Table 6. Dataset of object detection in aerial images (UCS-AOD) dataset statistics.
Table 6. Dataset of object detection in aerial images (UCS-AOD) dataset statistics.
DatasetClassImageInstances
Training SetAircraft6003591
Car3104475
Test SetAircraft4003891
Car2002639
Table 7. Experimental comparison of accuracy and speed in the RSOD dataset.
Table 7. Experimental comparison of accuracy and speed in the RSOD dataset.
MethodBackboneMetric (%)FPS
AircraftOil TankOverpassPlaygroundmAP (IOU = 0.5)
Faster RCNNVGG-1685.8586.6788.1590.3587.766.7
SSDVGG-1669.1771.2070.2381.2672.9762.2
DSSDResNet-10172.1272.4972.1083.5675.076.1
ESSDVGG-1673.0872.9473.6184.2775.9837.3
YOLO-V2DarkNet1962.3567.7468.3878.5169.2535.6
YOLO-V3DarkNet5374.3073.8575.0885.1677.1029.7
YOLO-V3 tinyDarkNet1954.1456.2159.2864.2058.4669.8
UAV-YOLO [52]Figure 1 in [52]74.6874.2076.3285.9677.7930.1
DC-SPP-YOLO [54]Figure 5 in [54]73.1673.5274.8284.8276.5833.5
ours(Figure 3)86.4287.5789.3791.5688.7325.8
Table 8. Experimental comparison of accuracy measured by size.
Table 8. Experimental comparison of accuracy measured by size.
MethodBackboneMetric (%)Leak Detection Rate (%)
SmallMediumLarge
Faster RCNNVGG-1684.7387.8789.1811.8
SSDVGG-1670.3873.4177.5121.1
DSSDResNet-10174.4275.1877.7015.2
ESSDVGG-1675.1275.8478.1216.5
YOLO-V2DarkNet1963.2068.5369.2824.3
YOLO-V3DarkNet5374.5275.6376.1419.5
YOLO-V3 tinyDarkNet1955.2656.4760.1731.4
UAV-YOLO [52]Figure 1 in Reference [52]75.4575.1576.8517.1
DC-SPP-YOLO [54]Figure 5 in Reference [54]75.4174.6776.4115.9
ours(Figure 3)87.5187.9390.2310.2
Table 9. Experimental comparisons of accuracy and speed in the UCS-AOD dataset.
Table 9. Experimental comparisons of accuracy and speed in the UCS-AOD dataset.
MethodBackboneMetric (%)FPS
AircraftCarLeak Detection Rate (%)mAP
(IOU = 0.5)
Faster RCNNVGG-1687.3186.4813.886.906.1
SSDVGG-1670.2472.6123.771.4361.5
DSSDResNet-10173.1774.1916.173.685.2
ESSDVGG-1673.6275.0615.974.3433.2
YOLO-V2DarkNet1963.1768.4223.065.8034.3
YOLO-V3DarkNet5375.7175.6218.575.6727.6
YOLO-V3 tinyDarkNet1957.5856.3535.256.9765.3
UAV-YOLO [52]Figure 1 in Reference [52]75.1275.6016.575.3628.4
DC-SPP-YOLO [54]Figure 5 in Reference [54]76.5274.6117.475.5730.4
Ours(Figure 3)89.3188.249.388.7824.9
Table 10. Experimental comparisons of each combination in the feature extraction network.
Table 10. Experimental comparisons of each combination in the feature extraction network.
DENSE
1st
DENSE
2nd
Metric (%)FPS
AircraftOil TankOverpassPlaygroundmAP (IOU = 0.5)
1 74.3073.8575.0885.1677.1029.7
2 76.8175.3877.2185.3778.6930.9
3 77.2876.3979.6585.9279.8131.4
482.1683.5285.1286.7384.3832.3
Table 11. Experimental comparisons of each combination in detection layers.
Table 11. Experimental comparisons of each combination in detection layers.
4th ScaleRes 3Metric (%)FPS
AircraftOil TankOverpassPlaygroundmAP (IOU = 0.5)
1 77.2576.3884.3686.1281.0329.7
2 85.9785.1887.1589.6186.9824.8
3 79.3878.8585.2988.2882.9530.1
486.4287.5789.3791.5688.7325.8

Share and Cite

MDPI and ACS Style

Xu, D.; Wu, Y. Improved YOLO-V3 with DenseNet for Multi-Scale Remote Sensing Target Detection. Sensors 2020, 20, 4276. https://doi.org/10.3390/s20154276

AMA Style

Xu D, Wu Y. Improved YOLO-V3 with DenseNet for Multi-Scale Remote Sensing Target Detection. Sensors. 2020; 20(15):4276. https://doi.org/10.3390/s20154276

Chicago/Turabian Style

Xu, Danqing, and Yiquan Wu. 2020. "Improved YOLO-V3 with DenseNet for Multi-Scale Remote Sensing Target Detection" Sensors 20, no. 15: 4276. https://doi.org/10.3390/s20154276

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop