Next Article in Journal
Analysis of the Feasibility of UAS-Based EMI Sensing for Underground Utilities Detection and Mapping
Previous Article in Journal
A Novel Ground-Based Cloud Image Segmentation Method Based on a Multibranch Asymmetric Convolution Module and Attention Mechanism
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Scale Object Detection with the Pixel Attention Mechanism in a Complex Background

1
School of Electronic Information, Wuhan University, Wuhan 430064, China
2
State Key Laboratory of Information Engineering in Surveying, Mapping and Remote Sensing, Wuhan University, Wuhan 430079, China
3
Aerospace System Development Research Center, China Aerospace Science and Technology Corporation, Beijing 100094, China
4
Qian Xuesen Laboratory of Space Technology, Beijing 100094, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2022, 14(16), 3969; https://doi.org/10.3390/rs14163969
Submission received: 27 June 2022 / Revised: 3 August 2022 / Accepted: 11 August 2022 / Published: 16 August 2022

Abstract

:
The object detection task is usually affected by complex backgrounds. In this paper, a new image object detection method is proposed, which can perform multi-feature selection on multi-scale feature maps. By this method, a bidirectional multi-scale feature fusion network was designed to fuse semantic features and shallow features to improve the detection effects of small objects in complex backgrounds. When the shallow features are transferred to the top layer, a bottom-up path is added to reduce the number of network layers experienced by the feature fusion network, reducing the loss of shallow features. In addition, a multi-feature selection module based on the attention mechanism is used to minimize the interference of useless information in subsequent classification and regression, allowing the network to adaptively focus on appropriate information for classification or regression to improve detection accuracy. Because the traditional five-parameter regression method has severe boundary problems when predicting objects with large aspect ratios, the proposed network treats angle prediction as a classification task. The experimental results on the DOTA dataset, the self-made DOTA-GF dataset and the HRSC 2016 dataset show that, compared with several popular object detection algorithms, the proposed method has certain advantages in detection accuracy.

1. Introduction

Object detection in remote sensing and unmanned aerial vehicle (UAV) imagery is important in a variety of sectors, including resource monitoring, national defense, and urban planning [1,2]. Unlike typical optical images and point cloud [3,4], optical remote sensing images always have their own unique qualities, such as numerous sizes of objects, arbitrary object direction, and complex backgrounds that take up the majority of the image. Many remote sensing image object detection algorithms borrow ideas from text detection algorithms, such as RRPN [5], because the arbitrariness of the object direction in remote sensing images has a lot in common with text detection [6]. However, due to the peculiar nature of remote sensing images, directly applying text detection algorithms to remote sensing image object detection frequently yields unsatisfactory results.
For scale-differences between classes, the feature pyramid network (FPN) [7] is commonly utilized in the object detection of various remote sensing images. Shallow features in FPN, on the other hand, must ’transit’ through numerous layers to reach the top layer, resulting in significant information loss. To improve the detection effects of small objects, certain algorithms [8,9,10] optimize the structure of FPN. The traditional technique to counteract the arbitrariness of object orientation in remote sensing images is to raise the regression parameters to estimate the angles [11,12]; this technique has a severe problem of boundary discontinuities [13]. To tackle the boundary problem, the IoU constant factor is added to the smooth L 1 loss to make correct angle predictions. Because the complex background contains a lot of noise, [14,15] uses a multi-scale feature extraction method to enhance each feature map with a visual attention mechanism to lessen the impact of background noise on object detection. After using the region proposal network (RPN) to acquire regional suggestions, reference [16] uses the location-sensitive score map to anticipate the target’s local location and specifies that it can only be classified as a given category after reaching a certain local feature similarity. To some extent, this strategy can also eliminate the influence of the background.
In summary, the main issues with remote sensing image object detection are numerous scales, complex backgrounds, and poor angle prediction. This paper proposes a new remote sensing image object detection algorithm to address these issues, and the framework is shown in Figure 1.
We used a single-stage rotation detector for multi-scale objects to retain good detection accuracy and speed. The first step was to build a bidirectional multi-scale feature fusion network. To prevent information loss during the transfer of shallow features to the top layer, a bottom-up path was added to merge high-level semantic information and shallow features. Second, a multi-feature selection module based on the attention mechanism was designed to reduce the complex background’s influence on object detection. The visual attention mechanism allows the network to focus on more significant information while avoiding background noise and choosing appropriate features for classification and regression tasks. Third, to increase the accuracy of direction prediction, the proposed network treats angle prediction as a classification problem. The distribution vectors of the category labels are smoothed using the circular smooth label, which divides the angles into 180 categories. The majority of the data in open-source remote sensing image object detection datasets come from Google Earth, with only a minor amount coming from domestic satellites. Moreover, there is a lack of military targets. As a result, we gathered some GF-2 and GF-6 images and created a new dataset named DOTA-GF. On the DOTA [17] dataset and DOTA-GF dataset, the proposed method is compared to many popular remote sensing image object detection algorithms. This work makes the following contributions:
  • A bidirectional multi-scale feature fusion network was built for high-precision multi-scale object detection in remote sensing images. It is the first work that we are aware of that achieves high-precision object detection in complex backgrounds.
  • The multi-feature selection module (MFSM) based on the attention mechanism is designed to reduce the influence of useless features in feature maps in complex backgrounds with a lot of noise.
  • We propose a novel remote sensing image object detection algorithm that includes a bidirectional multi-scale feature fusion network and a multi-feature selection module. With extensive ablation experiments, we validate the effectiveness of our approach on the standard DOTA dataset and a customized dataset named DOTA-GF. Our proposed method achieves a mAP of 65.1% with the ResNet50 backbone in the DOTA dataset, and 64.1% with the ResNet50 backbone in the DOTA-GF dataset when compared to state-of-the-art methods.

2. Related Work

2.1. Object Detection Algorithms Based on Deep Learning

Object detection algorithms based on deep learning are mainly divided into two categories—one-stage algorithms and two-stage algorithms. The series of algorithms of R-CNN are typical two-stage methods, including R-CNN, Fast R-CNN, and Faster R-CNN [18]. Fast R-CNN proposed RoIpooling and used the convolution network to achieve regression and classification, while Faster R-CNN used the region proposal network (RPN) to replace the selective search and shared feature map with the subsequent classification network. The one-stage methods extract feature maps and predict the categories and locations simultaneously. SSD and YOLO are two typical one-stage methods [19]. The one-stage methods, different from the two-stage methods, are influenced by category imbalances during detection. To tackle such a problem, focal loss [20] is proposed to suppress category imbalance in the one-stage method.

2.2. Arbitrary-Oriented Object Detection

Arbitrary-oriented object detection has been widely used in remote sensing images, aerial images, natural scene texts, etc. These detectors also use rotated bounding boxes to describe the positions of objects, which are more accurate than those using horizontal bounding boxes. Recently, many detectors have been proposed. For example, RRPN [5] uses rotating anchors to improve the qualities of region proposals. R2CNN is a multi-tasking text detector that identifies both rotated and horizontal bounding boxes at the same time. However, object detection in remote sensing images is more difficult, due to multiple categories, multiple scales, and complex backgrounds. Thus, arbitrary-oriented object detection has been proposed in many remote sensing images. R3Det [12] proposed an improved one-stage rotated object detector for accurate object localization by solving the feature misalignment problem. SCRDet [13] proposed an IoU-smooth L 1 loss to solve the loss discontinuity caused by the angular periodicity. Reference [21] proposed an anchor-free oriented proposal generator (AOPG) that abandoned the horizontal box-related operations from the network architecture. The AOPG produced coarse-oriented boxes by the coarse location module in an anchor-free manner and refined them into high-quality oriented proposals. Reference [22] proposed an effective oriented object detection method, termed oriented R-CNN. Oriented R-CNN is a general two-stage oriented detector. In the first stage, the oriented region proposal network directly generates high-quality oriented proposals in a nearly cost-free manner. The second stage is the oriented R-CNN head for refining oriented regions of interest and recognizing them.

3. The Proposed Algorithm

We present an overview of our algorithm as sketched in Figure 1. It consists of four parts: The backbone, the bidirectional multi-scale feature fusion network, the multi-feature selection module based on attention mechanism, and the multi-task subnets. We used ResNet50 [23] as our backbone. The bidirectional multi-scale feature fusion network is responsible for fusing high-level semantic information and shallow feature output by the backbone. The multi-feature selection module based on the attention mechanism can select features that are appropriate for classification and regression. After feature selection, the multi-scale feature maps are sent into the classification and regression sub-networks, respectively. Only the center points, widths, and heights of the bounding boxes are predicted by the regression subnet in this case. Through the classification subnet, the categories and angles are predicted.

3.1. Bidirectional Multi-Scale Feature Fusion Network

In the early object detection algorithms, such as Faster R-CNN [18], the subsequent classification and regression are usually performed on the feature map of the last layer of the backbone, which is less computationally expensive. However, for the multi-scale object detection, the information of a single-layer feature map is not enough. In 2017, He et al. proposed FPN [20], which fuses high-level features and low-level features, and uses multi-scale fusion feature maps for subsequent detection. RetinaNet [24] also follows the idea of FPN to build a feature pyramid net, as shown in Figure 2a.
Compared with the features extracted only through the last layer of convolution, FPN can use more high-level semantic information and detailed information. The red dotted line in Figure 2a indicates that in FPN, because of the bottom-up path, shallow features need to pass through multilayer networks to reach the top layer, and the information loss is more serious. Taking ResNet50 as an example, the transfer of the C 3 layer to the C 5 layer needs to go through 27 layers of convolution operations, as shown in Figure 3. The shallow details contained in P 5 , P 6 , and P 7 are ’lacking’ to be used for subsequent detection. With the addition of the bottom-up fusion path, the detailed texture features of the C 3 layer can be transferred to P 5 , P 6 , and P 7 , with only a few layers, as indicated by the yellow dotted line in Figure 2b. Therefore, the loss of shallow features is reduced.
Therefore, we designed a new feature fusion network; a bottom-up path was added to reduce the number of network layers experienced when the shallow features were transferred to the top layer, thereby reducing the loss of shallow features. The detailed information on the network is shown in Figure 2b.
As shown in Figure 2b, 1 × 1 C o n v represents using a 1 × 1 convolution kernel to perform convolution operations and change the number of channels in the feature map. The 2   × UpSample represents the double upsampling operation of the feature map using bilinear interpolation. The 3 × 3 / 2 C o n v means using a 3 × 3 convolution kernel to perform a convolution operation with a stride of 2, reducing the size of the feature map to half of the original size. The output of the backbone is C i ( i 3 5 ) , and the feature map after feature fusion is P i ( i 3 7 ) . Using 1 × 1 convolution to reduce the dimension of C 5 to obtain P 5 , C 5 is double-downsampled to obtain P 6 , and P 6 is double downsampled to obtain P 7 . The result of the double-upsampling of P 5 is fused with C 4 to obtain P 4 . The result of the double-upsampling of P 4 is fused with C 3 to obtain P 3 . P i ( i 3 7 ) combines the information of C 3 , C 4 , and C 5 at the same time, and contains low-level detailed information and high-level semantic information. Although it has a strong characterization ability for multi-scale objects, the transmission path of shallow features to higher layers is too long, and the feature loss is severe. Therefore, we added a bottom-up path, as shown in the yellow dotted line in Figure 2b. The 3 × 3 C o n v represents a convolution operation with a stride of 1 and a 3 × 3 convolution kernel. We performed a 3 × 3 convolution operation on P 3 to obtain P 3 . The result of P 4 after the 3 × 3 convolution and the result of the double-downsampling of P 3 were fused to obtain P 4 . Then P 5 , P 6 , and P 7 were obtained in the same way.

3.2. Multi-Feature Selection Module Based on Attention Mechanism

The complex background of satellite remote sensing images occupies a large area of the whole image. The images taken by domestic satellites, such as GF-2 and GF-6, are not as clear as Google Earth images, which leads to more complex backgrounds of the images, unclear object textures, and sometimes interference from cloud and fog. Directly inputting feature maps of different scales into the subsequent classification and regression sub-networks often fails to obtain ideal results. In recent years, the attention mechanism [25] has achieved great success in computer vision tasks, such as image classification [24] and semantic segmentation [26]. Here, we designed a MFSM. MFSM uses the pixel attention mechanism to select the features suitable for classification and regression, respectively, to reduce the influence of useless information in the feature maps. Different from the spatial attention mechanism, which learns the degree of dependence on different locations in space [27], the pixel attention mechanism learns the degree of dependence on each pixel and adjusts the feature map according to the degree of dependence.
The general one-stage object detection algorithms directly input P i ( i = 3 , 4 , 5 , 6 , 7 ) into classification subnet and regression subnet. The classification subnet is to predict the category of the bounding box. The regression subnet is primarily responsible for predicting the specific position of the bounding box. The purposes of the two subnets are different. It is inappropriate to use the same feature maps to perform classification and regression tasks at the same time. Therefore, we designed the MFSM. As shown in Figure 4, the multi-scale feature maps are obtained through the feature fusion network, and then are input into two feature selection modules, respectively. Finally, the feature maps after feature selection are input into the classification subnet and regression subnet.
The network details of the feature selection module for classification and the feature selection module for regression are the same, as shown in Figure 5.
The input of the module involves the multi-scale feature map input P i ( i = 3 , 4 , 5 , 6 , 7 ) , output by the feature fusion network and the output of the module is the input series of the feature map D i ( i = 3 , 4 , 5 , 6 , 7 ) with the same dimensions as the input. The processing process for each input P i ( i = 3 , 4 , 5 , 6 , 7 ) is shown in Figure 5 and Equations (1) and (2):
A i = σ [ ϕ i ( P i ) ]
D i = A i P i + P i
where ϕ i ( P i ) means performing four layers of a 3 × 3 convolution on P i . σ is the sigmoid function that converts the value of ϕ i ( P i ) into [0–1] to obtain A i , so that it can converge faster during training. Finally, the result of multiplying the corresponding elements of P i and A i is added to P i . The multiplication operation can make the value of the functional information in P i larger and the value of the useless information smaller. The addition operation refers to the idea of the residual network [23], which can make the network converge faster. This design can make the network adaptively select features suitable for classification or regression.

3.3. Accurate Acquisition of Target Direction Based on Angle Classification

At present, most mainstream algorithms use the idea of regression for angle prediction, and the bounding box is determined by five parameters. The five-parameter regression method has a boundary discontinuity problem [13], which will make the prediction box inaccurate.
Aiming at the loss discontinuity of five-parameter regression, this paper treats the angle prediction as a classification task [28]. The angles are divided into 180 categories. We find that directly dividing the angle into 180 categories will lead to low fault tolerances of adjacent angles. Common methods include the circular smooth label (CSL) [28] method and the densely coded label (DCL) [29] method. Among them, the DCL method is an improvement of the CSL method, which solves the heavy prediction layer and the unfriendly detection of square objects. This paper directly uses the CSL method as the angle classification method. The CSL expression is as follows:
S ( x ) = f ( x ) , i f θ + τ x θ + τ , 0 , o t h e r w i s e
where γ denotes the radius of the window; θ is the angle of the current ground truth. The circular smooth label is different for each ground truth. f ( x ) is the window function, and the Gaussian function is used here, as shown in Equation (4):
G a u s s i a n ( x ) = a e ( x b ) 2 2 c 2
where a, b, and c are constants ( a > 0 ) ; in this paper, a = 1 , b = 0 , and c is equal to the radius of the window function, which is set to be 6. The CSL [28] can increase the error tolerance to the adjacent angles.
In the paper, the angles of the bounding box are divided into 180 categories. If the angle of the ground truth is 90 , the traditional label of the angle is as follows:
l a b e l = ( 1 , 0 , 0 , 0 , 0 , 0 , 0 , 0 0 , 0 , 0 , 0 )
The circular smooth label of the angle is as follows:
l a b e l c s l = ( 1 , 0.86 , 0.71 , 0.57 , 0.43 , 0.29 , 0.14 , 0 0 , 0 , 0.14 , 0.29 , 0.43 , 0.57 , 0.71 , 0.86 )
The detector has two prediction results. In the traditional method, s o f t m a x is used to calculate the probabilities of different classes. The corresponding labels are as follows:
l a b e l 1 = ( 0.03 , 0.4 , 0.03 , 0.03 , 0.03 , 0.03 , 0.03 , 0.03 0.03 , 0.03 , 0.03 ) l a b e l 2 = ( 0.03 , 0.03 , 0.03 , 0.03 , 0.03 , 0.03 , 0.4 , 0.03 0.03 , 0.03 , 0.03 )
In the proposed method, s i g m o i d is used to calculate the probabilities of different classes. The corresponding labels are as follows:
l a b e l 1 * = ( 0.1 , 0.8 , 0.1 , 0.1 , 0.1 , 0.1 , 0.1 , 0.1 , 0.1 0.1 , 0.1 , 0.1 , 0.1 , 0.1 ) l a b e l 2 * = ( 0.1 , 0.1 , 0.1 , 0.1 , 0.1 , 0.1 , 0.8 , 0.1 , 0.1 0.1 , 0.1 , 0.1 , 0.1 , 0.1 )
The predicted angle corresponding to l a b e l 1 and l a b e l 1 * is 89 , and the predicted angle corresponding to l a b e l 2 and l a b e l 2 * is 84 . Using the cross-entropy loss function as an example. In the traditional method, the losses of l a b e l 1 and l a b e l 2 to the real label are as follows:
l o s s 1 = ( 1 × l o g ( 0.03 ) + 0 × l o g ( 0.4 ) + 0 × l o g ( 0.03 ) + ) = l o g ( 0.03 ) l o s s 2 = ( 1 × l o g ( 0.03 ) + + 0 × l o g ( 0.4 ) + 0 × l o g ( 0.03 ) + ) = l o g ( 0.03 )
It is found that l o s s 1 = l o s s 2 ; that is, the l a b e l 1 and l a b e l 2 have the same loss to the ground truth. However, the predicted angle obtained by l a b e l 1 is 89 , which is only 1 different from the true angle. While the predicted angle obtained by l a b e l 2 is 84 , which is 6 different from the true angle. The first prediction result is obviously more accurate. The analysis shows that directly dividing the angle into 180 categories will lead to low fault tolerance of adjacent angles. In the proposed method, the losses of l a b e l 1 and l a b e l 2 to the real label are as follows:
l o s s 1 * = ( 1 × l o g ( 0.1 ) + 0.86 × l o g ( 0.8 ) + 0.71 × l o g ( 0.1 ) + ) 14.33 l o s s 2 * = ( 1 × l o g ( 0.1 ) + 0.86 × l o g ( 0.1 ) + 0.71 × l o g ( 0.8 ) + ) 16.12
It can be found that l o s s 1 * < l o s s 2 * ; that is, the circular smooth label makes the losses of more accurate labels smaller and increases the error tolerance to the adjacent angles.

3.4. Loss Function

The total loss function is as Equation (11):
L = 1 N n = 1 N t n j { x , y , w , h } L r e g ( v n j , v n j ) + λ 1 N n = 1 N L c l s ( p n , t n ) + λ 2 N n = 1 N L c l s θ ( θ n , θ n )
where N indicates the number of anchors, t n has two values, i.e., 0 and 1, respectively ( t n = 1 for foreground and t n = 0 for background). v i j indicates the predicted offset vector. Moreover, v i j indicates the real offset vector. t n indicates the label of the object, p n indicates the probability distribution of various classes calculated by the sigmoid function. Hyperparameters λ 1 and λ 2 are trade-off factors, which control the weights of different loss functions, and their default values are both 1. L r e g indicates Smooth L 1 Loss [18]. L c l s represents the loss of classification in the object category prediction. L c l s θ represents the loss of angle classification in the angle prediction. Both L c l s and L c l s θ use focal loss [20].

4. Experimental Results and Discussion

The GPU used in this paper was GTX1660Ti with 6G of memory. The operating system we used was Ubuntu 16.04. The deep learning framework was TensorFlow. ResNet50 was used as the backbone of the network. We conducted experiments on three datasets, and the partitioning criteria were consistent with the references. The DOTA dataset contained a total of 2806 aerial images; 1/2 of the images were selected as the training set, 1/6 as the validation set, and 1/3 as the test set. The HRSC2016 dataset contained 1061 images of ships, of which, training, validation, and testing included 436, 181, and 444 images, respectively. For the self-made DOTA-GF dataset, which contained 2994 images from Google satellites and Chinese satellites, the number of images in training, validation, and test sets were: 1541, 468, and 985.

4.1. Ablation Studies

In this section, we conducted detailed ablation on DOTA to evaluate the effectiveness of each module and illustrate the advance and generalization of the proposed method.

4.1.1. Bidirectional Multi-Scale Feature Fusion Network

To verify the effectiveness of the improved feature fusion network, we used ResNet50 as the backbone and RetinaNet as the embodiment to compare the detection result of the original FPN and the improved feature pyramid network (Improved-FPN) on the DOTA [17] dataset. We mainly considered the average precision (AP) and mean average precision (mAP) of six types of typical objects, including plane (PL), ship (SH), bridge (BG), small vehicle (SV), large vehicle (LV), and storage tank (ST). This is because among the targets in the remote sensing images, the aspect ratio of objects such as plane (PL) and storage tank (ST) was about 1:1, ship (SH), bridge (BG), small vehicle (SV), large vehicle (LV), and other targets were less than 8:1 in length and width. The experimental results are shown in Table 1.
It can be seen from Table 1 that the Improved-FPN can significantly improve the detection accuracies of typical objects in remote sensing images. Among them, the AP of the ship had the highest increase of 2.4%. That is because many ships in DOTA are small, and the shallow features have a greater impact on the detection results; the bidirectional multi-scale feature fusion network can make full use of the shallow features. The AP of the storage tank has the least increase, which is 0.6%. The mAP of the six types of objects increased by 1.4%. Experimental results show that the improved feature fusion network was more suitable for remote sensing image object detection than the original feature fusion network.

4.1.2. Multi-Feature Selection Module Based on Attention Mechanism

To further prove the effectiveness of the multi-feature selection module, the multi-feature selection module was added to RetinaNet [20] to conduct experiments on DOTA [23]. The comparative experiments of the MFSM with other attention mechanisms were supplemented too. The experimental results with MFSM, SE [30], and CBAM [27] are shown in Table 2.
Compared with RetinaNet [20], after adding the multi-feature selection module, the detection accuracies of the six types of typical objects significantly improved with AP increases of 1.2% to 1.6%. The mAP increased by 1.3%. The detection accuracy of the small vehicle had the greatest improvement, and the AP increased by 1.6%. At the same time, MFSM had a better detection performance than SE and CBAM. In SE and CBAM, an attention module was used to process the feature map, and the classification and regression subnets shared the feature map. MFSM processes feature maps for classification and regression, respectively, which can alleviate the conflicts between classification tasks and regression tasks to a certain extent. Therefore, MFSM has a simpler structure, but better performance.
Figure 6 shows a remote sensing image with cloud interference and visualization results of its feature maps. The feature map P i ( i = 3 , 4 , 5 , 6 , 7 ) was obtained by the feature fusion network. P i was input into the multi-feature selection network, and the feature map C L S i ( i = 3 , 4 , 5 , 6 , 7 ) for the classification prediction task and the feature map R E G i ( i = 3 , 4 , 5 , 6 , 7 ) for the bounding box prediction task were obtained. In Figure 6, three rows from top to bottom are P i , C L S i ( i = 3 , 4 , 5 , 6 , 7 ) , and R E G i ( i = 3 , 4 , 5 , 6 , 7 ) . Five columns from left to right are the feature maps of the 3rd, 4th, 5th, 6th, and 7th layers, respectively. For the ship in Figure 6, P 3 and P 4 in the multi-scale feature maps have greater responses. From P 3 , P 4 , C L S 3 , C L S 4 , and R E G 4 , R E G 4 , we can see that after the feature selection, the feature map has a stronger response in the object area. It shows that the multi-feature selection module based on the attention mechanism can select features suitable for classification tasks and regression tasks from multi-scale feature maps and improve the detection accuracy.

4.1.3. Accurate Acquisition of Target Direction Based on Angle Classification

To further prove that turning the angle regression problem into a classification task can improve the remote sensing image detection effect, the angle prediction in RetinaNet is regarded as a classification task with 180 categories, and CSL is used for smoothing. Comparative experiments are performed on the DOTA, and the experimental results are shown in Table 3. It can be seen from Table 3 that treating the angle prediction as a classification task can significantly improve the detection effect. Among the six types of typical targets, the APs of ships, bridges, small vehicles, and large vehicles increased by 2.7%, 2.2%, 1.9%, and 3.2%, respectively. This is because the aspect ratios of these four types of objects are relatively large, and the use of regression to predict angles has more serious loss discontinuity. For planes and storage tanks with an aspect ratio close to 1, the APs also increased by 0.8% and 0.9%. The experimental results prove that treating the angle prediction as a classification task can effectively improve the detection accuracies of objects with larger aspect ratios.
Figure 7 shows the results of the prediction angle based on the five-parameter regression method. As can be seen from the red boxes in the figure, there is a significant difference between the angles of the detected bounding boxes and the angles of the actual objects, including the large vehicles on the left and the ships on the right.
On the other hand, some visual experiment results based on the proposed classification method are shown in Figure 8. It can be seen that the results obtained by the angle prediction method based on the classification idea are more accurate when detecting the objects, while the angle prediction method based on the classification idea produces more missed detections and false detections

4.2. Results on DOTA

The DOTA [17] dataset contains 15 categories. This paper mainly analyzes six typical objects—ships, planes, bridges, small vehicles, large vehicles, and storage tanks. The evaluation indicators used are AP and mAP. CSL [28], RRPN [5], RetinaNet [20], and Xiao [11] were selected as comparative algorithms. The comparison results of different algorithms are shown in Table 4.
The data in Table 4 show that the mAP of the proposed method is better than most of the mainstream object detection algorithms. The algorithm proposed has achieved the highest AP in four types of objects: planes, ships, small vehicles, and storage tanks. Moreover, the APs of large vehicles and bridges are second only to the highest. The large vehicles in the DOTA dataset are often placed very closely, and adjacent objects have occlusion problems. This is also a problem that we will study in the future. These comparison results show that the algorithm proposed in this paper can effectively detect typical objects in remote sensing images.
The partial visual detection results of the proposed algorithm and the RetinaNet algorithm on the DOTA data set are shown in Figure 9. In order to make the comparison results clearer, some areas are enlarged.
It can be seen from the comparison results in the first column of Figure 9 that when detecting small ships, RetinaNet has a weak ability to characterize small targets, resulting in some missed detections. In the third column of comparative experimental results, RetinaNet also has missed detection when detecting cars. However, the algorithm in this paper obtains better detection results when detecting densely arranged small ships and cars, and the positioning is more accurate. The reason is that the bidirectional multi-scale feature fusion network designed by the algorithm in this paper can improve the representation ability of small targets, the angle information obtained by the idea of classification is also more accurate, and the sensitive feature selection network can further improve the network detection performance.

4.3. Results on DOTA-GF

At present, the remote sensing images in public remote sensing datasets, such as DOTA [17] and NWPU VHR-10 [31], are mainly derived from Google Earth, with only a small amount of data derived from domestic data and lack of military objects. Therefore, we collected 188 GF-2 Satellite images and GF-6 Satellite images with a resolution of 1000 × 1000 to 4000 × 4000 and labeled them using the four-point method.
The 138 domestic remote sensing images were added to the training set of DOTA as the DOTA-GF training set. The remaining 50 domestic remote sensing images were added to the DOTA testing set as the DOTA-GF testing set. Then we selected the data containing six types of objects: ships, planes, bridges, small vehicles, large vehicles, and storage tanks, and cropped them to pieces (sizes 600 × 600 ) for training. To illustrate the effectiveness of the proposed algorithm, four representative object detection algorithms, CSL [28], RRPN [5], RetinaNet [20], and R3Det [12] were selected for comparison experiments. The detection results of different algorithms are shown in Table 5.
It can be seen from Table 5 that compared with the four representative algorithms, the algorithm proposed in this paper has achieved the highest AP in four types of objects: ships, bridges, small vehicles, and storage tanks. The APs of planes and large vehicles are also much higher than the highest AP of the four types of algorithms. However, the network structure of R3Det is more complex. Both the training time and the testing time of a single image are longer than the proposed algorithm. Compared with the four comparison algorithms, the mAP of the six typical objects of the proposed algorithm is also the highest. The experimental results show that the algorithm proposed in this paper still has certain advantages on the self-made DOTA-GF dataset.
The detection results of the proposed algorithm and the RetinaNet method on high-resolution images are shown in Figure 10. In order to be more eye-catching, some areas have been partially enlarged.
From the comparison results in Figure 10, it can be seen that when there is cloud interference, RetinaNet cannot accurately capture the characteristics of ships, resulting in missed detection. The algorithm in this chapter has strong anti-cloud and fog interference ability, and can accurately detect ship targets covered by thin clouds.
From the comparison results in Figure 11, it can be seen that when the ship target is relatively small, RetinaNet cannot accurately capture the ship features, resulting in missed detection. Moreover, because of the existence of the proposed FPN module, the algorithm in this paper has a stronger ability to extract small target features and the detection results are more accurate.
From the comparison results in Figure 12, it can be seen that both the algorithm in this paper and RetinaNet have better detection results when detecting objects with large sizes and obvious features, such as aircraft and storage tubes.

4.4. Results of HRSC 2016

HRSC 2016 [32] contains lots of remote sensing ships with a large aspect ratio, scales, and arbitrary orientations. Our method achieves competitive performances on the HRSC2016 dataset. The comparison results are shown in Table 6.
From Table 6, it can be seen that compared with R2CNN [33], RRPN [5], RetinaNet [20], and the RoI transformer [34], the algorithm in this paper achieves the best detection results, with a mAP of 87.1%. The experimental results verify the effectiveness of the proposed algorithm on the HRSC 2016 dataset.

5. Conclusions

We proposed a new remote sensing image object detection algorithm aimed to target challenges, such as multi-scale objects, complex backgrounds, and boundary problems. In this algorithm, a bidirectional multi-scale feature fusion network was designed to combine the semantic features and shallow detailed features to reduce the loss of information in the process of transferring shallow features to the top layer. A multi-feature selection module based on the attention mechanism was designed to make the network focus on valuable information and select the feature maps appropriate for classification and regression tasks. To avoid the boundary discontinuity problem in the regression process, we treated angle prediction as a classification task rather than a regression task. Finally, experimental results on the DOTA dataset, the DOTA-GF dataset, and the HRSC 2016 dataset show that the proposed algorithm has certain advantages in remote sensing image object detection. However, our proposed method still has limitations in detecting dense objects. In the future, we will ’outlook’ the situation of dense object occlusion and improve our network model to better detect dense objects. The results reported in this paper can be downloaded from https://github.com/xiaojs18/Object-Detection/tree/main/Remote-Detection (accessed on 26 June 2022).

Author Contributions

Funding acquisition, J.Z.; Investigation, H.G. and S.Z.; Methodology, H.G.; Project administration, J.X. and Z.J.; Resources, Z.J.; Software, Y.Y.; Supervision, J.X.; Visualization, S.Z.; Writing—original draft, Y.Y.; Writing—review & editing, J.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (Grant No.42101448) and the Open Project Program Foundation of the Key Laboratory of Opto-Electronics Information Processing, Chinese Academy of Sciences (OEIP-O-202009). The numerical calculations in this article have been done on the supercomputing system in the Supercomputing Center of Wuhan University.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Li, K.; Wan, G.; Cheng, G.; Meng, L.; Han, J. Object detection in optical remote sensing images: A survey and a new benchmark. ISPRS J. Photogramm. Remote Sens. 2020, 159, 296–307. [Google Scholar] [CrossRef]
  2. Fatima, S.A.; Kumar, A.; Pratap, A.; Raoof, S.S. Object Recognition and Detection in Remote Sensing Images: A Comparative Study. In Proceedings of the 2020 International Conference on Artificial Intelligence and Signal Processing, AISP 2020, Amaravati, India, 10–12 January 2020. [Google Scholar]
  3. Ma, R.; Chen, C.; Yang, B.; Li, D.; Wang, H.; Cong, Y.; Hu, Z. CG-SSD: Corner guided single stage 3D object detection from LiDAR point cloud. ISPRS J. Photogramm. Remote Sens. 2022, 191, 33–48. [Google Scholar] [CrossRef]
  4. Hu, Z.; Chen, C.; Yang, B.; Wang, Z.; Ma, R.; Wu, W.; Sun, W. Geometric feature enhanced line segment extraction from large-scale point clouds with hierarchical topological optimization. Int. J. Appl. Earth Obs. Geoinf. 2022, 112, 102858. [Google Scholar] [CrossRef]
  5. Ma, J.; Shao, W.; Ye, H.; Wang, L.; Wang, H.; Zheng, Y.; Xue, X. Arbitrary-oriented scene text detection via rotation proposals. IEEE Trans. Multimed. 2018, 20, 3111–3122. [Google Scholar] [CrossRef]
  6. Liu, X.; Meng, G.; Pan, C.A. Scene text detection and recognition with advances in deep learning: A survey. Int. J. Doc. Anal. Recognit. (IJDAR) 2019, 22, 143–162. [Google Scholar] [CrossRef]
  7. Lin, T.Y.; Dollar, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Image-to-image translation with conditional adversarial networks. In Proceedings of the 30th IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  8. Liu, S.; Qi, L.; Qin, H.; Shi, J.; Jia, J. Path Aggregation Network for Instance Segmentation. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 8759–8768. [Google Scholar]
  9. Guo, C.; Fan, B.; Zhang, Q.; Xiang, S.; Pan, C. AUGFPN: Improving multi-scale feature learning for object detection. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 12592–12601. [Google Scholar]
  10. Ghiasi, G.; Lin, T.Y.; Le, Q.V. NAS-FPN: Learning scalable feature pyramid architecture for object detection. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 7029–7038. [Google Scholar]
  11. Xiao, J.; Zhang, S.; Dai, Y.; Jiang, Z.; Yi, B.; Xu, C. Multiclass Object Detection in UAV Images Based on Rotation Region Network. IEEE J. Miniaturization Air Space Syst. 2020, 1, 188–196. [Google Scholar] [CrossRef]
  12. Yang, X.; Liu, Q.; Yan, J.; Li, A. R3Det: Refined Single-Stage Detector with Feature Refinement for Rotating Object. In Proceedings of the AAAI Conference on Artificial Intelligence, Online, 2–9 February 2021. [Google Scholar]
  13. Yang, X.; Yang, J.; Yan, J.; Zhang, Y.; Zhang, T.; Guo, Z.; Sun, X.; Fu, K. SCRDet: Towards More Robust Detection for Small, Cluttered and Rotated Objects. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Korea, 27 October–2 November 2019; pp. 8231–8240. [Google Scholar]
  14. Zhang, Y.; Xiao, J.; Jinye, P.; Ding, Y.; Liu, J.; Guo, Z.; Xiaopeng, Z. Kernel Wiener Filtering Model with Low-Rank Approximation for Image Denoising. Inf. Sci. 2018, 462, 402–416. [Google Scholar] [CrossRef]
  15. Li, Q.; Mou, L.M.; Jiang, K.; Liu, Q.; Wang, Y.; Zhu, X. Hierarchical Region Based Convolution Neural Network for Multi-scale Object Detection in Remote Sensing Images. In Proceedings of the 2018 IEEE International Geoscience and Remote Sensing Symposium, Valencia, Spain, 22–27 July 2018; pp. 4355–4358. [Google Scholar]
  16. Xie, H.; Wang, T.; Qiao, M.; Zhang, M.; Shan, G.; Snoussi, H. Robust object detection for tiny and dense targets in VHR aerial images. In Proceedings of the 2017 Chinese Automation Congress, Jinan, China, 20–22 October 2017; pp. 6397–6401. [Google Scholar]
  17. Xia, G.S.; Bai, X.; Ding, J.; Zhu, Z.; Belongie, S.; Luo, J.; Datcu, M.; Pelillo, M.; Zhang, L. DOTA: A Large-scale Dataset for Object Detection in Aerial Images. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 3974–3983. [Google Scholar]
  18. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed]
  19. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A. SSD: Single shot multibox detector. In Proceedings of the 14th European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 21–37. [Google Scholar]
  20. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollar, P. Focal Loss for Dense Object Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 318–327. [Google Scholar] [CrossRef] [PubMed]
  21. Cheng, G.; Wang, J.; Li, K.; Xie, X.; Lang, C.; Yao, Y.; Han, J. Anchor-Free Oriented Proposal Generator for Object Detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–11. [Google Scholar] [CrossRef]
  22. Xie, X.; Cheng, G.; Wang, J.; Yao, X.; Han, J. Oriented R-CNN for Object Detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, BC, Canada, 11–17 October 2021; pp. 3520–3529. [Google Scholar]
  23. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  24. Wang, F.; Jiang, M.; Qian, C.; Yang, S.; Li, C.; Zhang, H.; Wang, X.; Tang, X. Residual attention network for image classification. In Proceedings of the 30th IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 6450–6458. [Google Scholar]
  25. Hang, R.; Li, Z.; Liu, Q.; Ghamisi, P.; Bhattacharyya, S.S. Hyperspectral Image Classification With Attention-Aided CNNs. IEEE Trans. Geosci. Remote Sens. 2021, 59, 2281–2293. [Google Scholar] [CrossRef]
  26. Zhong, Z.; Lin, Z.Q.; Bidart, R.; Hu, X.; Daya, I.B.; Li, Z.; Zheng, W.S.; Li, J.; Wong, A. Squeeze-and-attention networks for semantic segmentation. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 13062–13071. [Google Scholar]
  27. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. CBAM: Convolutional block attention module. In Proceedings of the Computer Vision—ECCV 2018—15th European Conference, Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  28. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollar, P. Arbitrary-Oriented Object Detection with Circular Smooth Label. Yang Xue Yan Junchi 2020, 12353, 677–694. [Google Scholar]
  29. Yang, X.; Hou, L.; Zhou, Y.; Wang, W.; Yan, J. Dense Label Encoding for Boundary Discontinuity Free Rotation Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2021, Virtual, 19–25 June 2021; pp. 15819–15829. [Google Scholar]
  30. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7132–7141. [Google Scholar]
  31. Cheng, G.; Han, J. A survey on object detection in optical remote sensing images. ISPRS J. Photogramm. Remote Sens. 2016, 117, 11–28. [Google Scholar] [CrossRef]
  32. Cheng, G.; Han, J.; Zhou, P.; Guo, L. Multi-class geospatial object detection and geographic image classification based on collection of part detectors. ISPRS J. Photogramm. Remote Sens. 2014, 98, 119–132. [Google Scholar] [CrossRef]
  33. Pang, J.; Li, C.; Shi, J.; Xu, Z.; Feng, H. R2-CNN: Fast Tiny Object Detection in Large-Scale Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2019, 57, 5512–5524. [Google Scholar] [CrossRef]
  34. Ding, J.; Xue, N.; Long, Y.; Xia, G.S.; Lu, Q. Learning roi transformer for oriented object detection in aerial images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 2849–2858. [Google Scholar]
Figure 1. The network structure of the proposed method can be divided into four parts: (a) input image, (b) feature pyramid net, (c) feature selection module, (d) multitasking subnets.
Figure 1. The network structure of the proposed method can be divided into four parts: (a) input image, (b) feature pyramid net, (c) feature selection module, (d) multitasking subnets.
Remotesensing 14 03969 g001
Figure 2. The network structure of the feature fusion network. The red dotted line—the bottom-up path of the shallow information transmitted to the high level; the yellow dotted—the new bottom-up path; 1 × 1 C o n v —convolution operation with the 1 × 1 convolution kernel; 2 × U p S a m p l e —the double upsampling operation by bilinear interpolation; 3 × 3 / 2 C o n v —convolution operation with the 3 × 3 convolution kernel and a stride of 2, 3 × 3 C o n v —convolution operation with the 3 × 3 convolution kernel and a stride of 1.
Figure 2. The network structure of the feature fusion network. The red dotted line—the bottom-up path of the shallow information transmitted to the high level; the yellow dotted—the new bottom-up path; 1 × 1 C o n v —convolution operation with the 1 × 1 convolution kernel; 2 × U p S a m p l e —the double upsampling operation by bilinear interpolation; 3 × 3 / 2 C o n v —convolution operation with the 3 × 3 convolution kernel and a stride of 2, 3 × 3 C o n v —convolution operation with the 3 × 3 convolution kernel and a stride of 1.
Remotesensing 14 03969 g002
Figure 3. ResNet50 network structure; the red arrow indicates the path from C 3 to C 5 .
Figure 3. ResNet50 network structure; the red arrow indicates the path from C 3 to C 5 .
Remotesensing 14 03969 g003
Figure 4. Multi-feature selection of multi-scale feature maps.
Figure 4. Multi-feature selection of multi-scale feature maps.
Remotesensing 14 03969 g004
Figure 5. Detailed information on the multi-feature selection module. CNNs: four layers of 3 × 3 convolution, ⊙: Hadamard product, ⊕ Matrix addition.
Figure 5. Detailed information on the multi-feature selection module. CNNs: four layers of 3 × 3 convolution, ⊙: Hadamard product, ⊕ Matrix addition.
Remotesensing 14 03969 g005
Figure 6. Visualization results of multi-scale feature maps. From top to bottom, there are the multi-scale feature maps P i ( i = 3 , 4 , 5 , 6 , 7 ) , the multi-scale feature maps C L S i ( i = 3 , 4 , 5 , 6 , 7 ) used for classification tasks, and the multi-scale feature map R E G i ( i = 3 , 4 , 5 , 6 , 7 ) used for regression tasks.
Figure 6. Visualization results of multi-scale feature maps. From top to bottom, there are the multi-scale feature maps P i ( i = 3 , 4 , 5 , 6 , 7 ) , the multi-scale feature maps C L S i ( i = 3 , 4 , 5 , 6 , 7 ) used for classification tasks, and the multi-scale feature map R E G i ( i = 3 , 4 , 5 , 6 , 7 ) used for regression tasks.
Remotesensing 14 03969 g006
Figure 7. The regression inaccuracy of the five-parameter method. RetinaNet is the base model. The cars and ships in the red box have not been accurately detected, and the angles between the prediction boxes and the ground truth are different.
Figure 7. The regression inaccuracy of the five-parameter method. RetinaNet is the base model. The cars and ships in the red box have not been accurately detected, and the angles between the prediction boxes and the ground truth are different.
Remotesensing 14 03969 g007
Figure 8. Visual detection results of some typical objects based on the proposed classification method.
Figure 8. Visual detection results of some typical objects based on the proposed classification method.
Remotesensing 14 03969 g008aRemotesensing 14 03969 g008b
Figure 9. DOTA dataset detection results (the first line is the proposed method, the second line is the RetinaNet method).
Figure 9. DOTA dataset detection results (the first line is the proposed method, the second line is the RetinaNet method).
Remotesensing 14 03969 g009
Figure 10. Detection results of ships with cloud and fog interferences on the images of the DOTA-GF dataset (left: RetinaNet method; right: our proposed method).
Figure 10. Detection results of ships with cloud and fog interferences on the images of the DOTA-GF dataset (left: RetinaNet method; right: our proposed method).
Remotesensing 14 03969 g010
Figure 11. Comparison of detection results of small ships on the DOTA-GF dataset images (left: RetinaNet method; right: our proposed method).
Figure 11. Comparison of detection results of small ships on the DOTA-GF dataset images (left: RetinaNet method; right: our proposed method).
Remotesensing 14 03969 g011
Figure 12. Comparison of large-scale target detection results in DOTA-GF dataset images (left: RetinaNet method; right: our proposed method).
Figure 12. Comparison of large-scale target detection results in DOTA-GF dataset images (left: RetinaNet method; right: our proposed method).
Remotesensing 14 03969 g012
Table 1. The experimental results of the bidirectional multi-scale feature fusion network.
Table 1. The experimental results of the bidirectional multi-scale feature fusion network.
MethodAP (%)mAP (%)
PLSHBGSVLVST
FPN83.462.232.365.748.374.961.1
our-FPN 84.5 64.6 34.0 67.2 49.2 75.5 62.5
Table 2. Experimental results of different attention mechanisms.
Table 2. Experimental results of different attention mechanisms.
MethodAP (%)mAP (%)
PLSHBGSVLVST
Baseline83.462.232.365.748.374.961.1
SE83.664.333.466.1 50.1 74.161.9
CBAM84.4 64.5 33.7 67.049.175.262.3
MFSM 84.7 63.433.6 67.3 49.5 76.1 62.4
Table 3. Experimental results of RetinaNet using classification and regression methods to predict angles.
Table 3. Experimental results of RetinaNet using classification and regression methods to predict angles.
MethodAP (%)mAP (%)
PLSHBGSVLVST
Regression83.462.232.365.748.374.961.1
Classification 84.2 64.9 34.5 67.6 51.5 75.8 63.1
Table 4. Comparison results of different algorithms on the DOTA dataset.
Table 4. Comparison results of different algorithms on the DOTA dataset.
CategoryCSLRRPNRetinaNetXiaoProposed
PL84.283.983.478 85.7
SH64.947.262.265 66.5
BG34.532.332.338 37.5
LV51.549.748.35954.2
SV67.634.765.737 69.2
ST75.848.874.950 77.3
mAP (%)63.148.061.155 65.1
Table 5. Comparison results of different algorithms on the DOTA-GF dataset.
Table 5. Comparison results of different algorithms on the DOTA-GF dataset.
CategoryCSLRRPNRetinaNetR3DetProposed
PL83.681.783.2 85.2 84.6
SH64.146.861.066.1 66.3
BG35.334.832.535.5 37.2
LV50.448.250.2 61.5 53.8
SV64.733.864.559.8 68.6
ST72.948.672.770.5 74.1
mAP(%)56.549.060.763.1 64.1
Table 6. Comparisons with different methods on the HRSC2016 dataset.
Table 6. Comparisons with different methods on the HRSC2016 dataset.
MethodsSizemAP (%)
R2CNN800 × 80073.7
RRPN800 × 80079.1
RetinaNet800 × 80081.7
RoI transformer512 × 80086.2
Proposed800 × 800 87.1
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Xiao, J.; Guo, H.; Yao, Y.; Zhang, S.; Zhou, J.; Jiang, Z. Multi-Scale Object Detection with the Pixel Attention Mechanism in a Complex Background. Remote Sens. 2022, 14, 3969. https://doi.org/10.3390/rs14163969

AMA Style

Xiao J, Guo H, Yao Y, Zhang S, Zhou J, Jiang Z. Multi-Scale Object Detection with the Pixel Attention Mechanism in a Complex Background. Remote Sensing. 2022; 14(16):3969. https://doi.org/10.3390/rs14163969

Chicago/Turabian Style

Xiao, Jinsheng, Haowen Guo, Yuntao Yao, Shuhao Zhang, Jian Zhou, and Zhijun Jiang. 2022. "Multi-Scale Object Detection with the Pixel Attention Mechanism in a Complex Background" Remote Sensing 14, no. 16: 3969. https://doi.org/10.3390/rs14163969

APA Style

Xiao, J., Guo, H., Yao, Y., Zhang, S., Zhou, J., & Jiang, Z. (2022). Multi-Scale Object Detection with the Pixel Attention Mechanism in a Complex Background. Remote Sensing, 14(16), 3969. https://doi.org/10.3390/rs14163969

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop