Next Article in Journal
Intelligent Fault Diagnosis of Industrial Bearings Using Transfer Learning and CNNs Pre-Trained for Audio Classification
Previous Article in Journal
Early Stage Forest Fire Detection from Himawari-8 AHI Images Using a Modified MOD14 Algorithm Combined with Machine Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Shared-Weight-Based Multi-Dimensional Feature Alignment Network for Oriented Object Detection in Remote Sensing Imagery

College of Information and Engineering, Shanghai Maritime University, Shanghai 201306, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(1), 207; https://doi.org/10.3390/s23010207
Submission received: 29 November 2022 / Revised: 13 December 2022 / Accepted: 14 December 2022 / Published: 25 December 2022
(This article belongs to the Section Remote Sensors)

Abstract

:
Arbitrarily Oriented Object Detection in aerial images is a highly challenging task in computer vision. The mainstream methods are based on the feature pyramid, while for remote-sensing targets, the misalignment of multi-scale features is always a thorny problem. In this article, we address the feature misalignment problem of oriented object detection from three dimensions: spatial, axial, and semantic. First, for the spatial misalignment problem, we design an intra-level alignment network based on leading features that can synchronize the location information of different pyramid features by sparse sampling. For multi-oriented aerial targets, we propose an axially aware convolution to solve the mismatch between the traditional sampling method and the orientation of instances. With the proposed collaborative optimization strategy based on shared weights, the above two modules can achieve coarse-to-fine feature alignment in spatial and axial dimensions. Last but not least, we propose a hierarchical-wise semantic alignment network to address the semantic gap between pyramid features that can cope with remote-sensing targets at varying scales by endowing the feature map with global semantic perception across pyramid levels. Extensive experiments on several challenging aerial benchmarks show state-of-the-art accuracy and appreciable inference speed. Specifically, we achieve a mean Average Precision (mAP) of 78.11% on DOTA, 90.10% on HRSC2016, and 90.29% on UCAS-AOD.

Graphical Abstract

1. Introduction

Benefiting from remote-sensing images with a broad field of view and high image quality, object detection has been widely used in various civilian and military fields, such as inshore vessel monitoring, geological exploration, and traffic diversion. With many convolutional neural networks (CNNs) and transformer-based detection algorithms proposed in the natural image domain [1,2,3,4,5,6,7], the performance of horizontal object detection has been significantly improved. However, the horizontal bounding boxes used for natural images are unsuitable for remote-sensing targets with dense arrangements and arbitrary orientations. In this case, an oriented bounding box (OBB) is a more domain-friendly solution that can perfectly distinguish between instances at high camera altitudes while reducing the redundant background inside the boxes.
Still, this annotation approach also poses several domain-specific challenges. For small remote-sensing targets with variable orientations in dense scenes, high localization accuracy and orientation consistency of detection boxes are required. For remote-sensing images with severe foreground–background imbalance and complex contexts, robust semantic representation of inter- and intra-class features is demanded. In this work, we summarize the misalignment problems that exist in oriented object detection from three aspects:
(i)
Axial misalignment: The feature extraction method of traditional convolution filters the whole image in the form of a sliding window under a fixed coordinate system, and the sampling axis of the kernel is consistent with the sliding direction. However, instance details are lost to a certain extent due to high-altitude shooting in remote-sensing imagery, which makes different instances of the same category have similar textures, while these similar targets belong to different axes in the bird’s eye view, as shown in Figure 1a. CNNs have weak axial perception, and even with the help of resampling operations such as pooling, they are still not robust enough. For a similar object in a different axis, the CNN tends to learn it as a new object instead of the rotated one [8]. Therefore, this fixed axial sampling method does not match the orientation of remote-sensing targets and OBBs. Eventually, the axial misalignment problem makes it difficult for the extracted features to represent arbitrarily oriented targets well, resulting in inaccurate regression, as shown in Figure 1b.
(ii)
Spatial misalignment: First, in the process of extracting multi-scale features from the backbone, the aliasing effect caused by the interpolation, the weak response loss caused by pooling, and the channel compression all cause different degrees of location information loss. Secondly, remote-sensing targets are small and densely arranged and vary widely in shape, rotation, and position, so this information loss significantly impacts the remote-sensing field more than other fields. The spatial misalignment problem in the detection stage leads to inaccurate regression, which further causes the predicted boxes to be incorrectly filtered by post-processing operations, as shown in Figure 1c.
(iii)
Semantic misalignment: Since the semantic information in the backbone features is distributed in several pyramid levels due to the receptive field differences, there is a huge semantic gap between layers. Although a feature pyramid network (FPN) [9] alleviates the above problem by a top-down feature fusion, its linear combination of the feature hierarchy cannot perfectly cope with the large inter- and intra-class scale variations of remote-sensing targets. Therefore, classification errors due to semantic non-alignment between pyramid levels still exist, as shown in Figure 1d.
For the above misalignment problems, many excellent algorithms have achieved considerable performance and can be divided into three directions. The first is the sampling method based on predicted boxes: [10,11,12] refine the feature points under the guidance of the key points on the rotated anchors. However, these methods rely on an additional refine head, which greatly impacts inference efficiency and parameters. Another direction is the fusion method based on feature resampling: [13,14,15,16,17,18] aggregate features from all pyramid levels and finally predict at the specified feature scale. The disadvantage of these methods is that resampling features to a single scale weakens the scale sensitivity of the network.
In this paper, we propose an intra-level alignment network based on sparse sampling and axially aware convolution for the misalignment problem in axial and spatial dimensions. The network jointly absorbs the location information and semantic knowledge from low-level and high-level features, respectively, to achieve a coarse-to-fine form of feature alignment. In addition, a collaborative optimization strategy based on shared weights is proposed to make feature alignment more stable and robust. Next, we propose a hierarchical-wise context fusion modulethat enhances the global semantics perceptivity of each level while dynamically fusing the pyramid features to achieve semantic feature alignment between different levels. Finally, our proposed method achieves remarkable performance on several public remote-sensing datasets (e.g., DOTA1.0, HRSC2016, and UCAS-AOD) and achieves accuracy comparable to state-of-the-art methods.
The key contributions of this paper are as follows:
  • First, we design a shared-weight-based multi-scale alignment network to synchronize the spatial information of different pyramid features. Following this, an axially aware convolution is proposed to align the sampling axis with remote-sensing targets in arbitrary orientations. So far, we have achieved the feature alignment of spatial and axial dimensions in a coarse-to-fine manner without the help of additional detection heads.
  • Secondly, a hierarchical-wise context fusion module is proposed to cope with the domain-specific problem of large intra-class scale variation by aligning semantic knowledge among pyramid features and fusing them dynamically.
  • Empirically, we demonstrate the effectiveness of the proposed methods through qualitative and quantitative ablation experiments. Comparing baseline and state-of-the-art methods, our OA-Det achieves comparable performance on three challenging datasets.
The rest of the paper is organized as follows. Section 2 begins with a brief introduction of the work related to this paper from three aspects. Then, the proposed feature alignment method is introduced from the three dimensions of space, axis, and semantics. Section 3 introduces the datasets and implementation details used for the experiments and then analyzes the ablation experiments of each module and the main results on the test set. Section 4 objectively discusses the limitations and future direction of the proposed method. Conclusions are given in Section 5.

2. Materials and Methods

2.1. Related Work

In this chapter, we focus on the domain-related work involved in the proposed methods. First, we introduce horizontal box-based object detection in terms of the execution stage of the algorithm and the detector type. Next, we list five representative algorithms for arbitrarily oriented object detection according to the type of method. Finally, we introduce the existing solutions related to feature alignment from both spatial and semantic perspectives, and the differences between our methods and existing methods are analyzed.

2.1.1. Horizontal Object Detection

Object detection based on horizontal bounding boxes has seen a remarkable leap in the last decade. From the traditional method based on handcrafted features [19,20], to the near-invincible deep convolutional neural network [21], to the Transformer architecture [22] that has emerged in recent years, we have been provided with the foundational support for a variety of visual detection algorithms. According to the execution stage of the algorithm, they can be divided into two-stage detectors, represented by [2,4], and single-stage detectors, represented by [6,7,23]. With its concise end-to-end network structure, the single-stage detector achieves a good trade-off in speed, parameters, and accuracy.
In recent years, anchor-free single-stage detectors have been widely proposed. CenterNet [24] predicts a center point for each object, regressing the width and height of the horizontal bounding box directly from the center point. RepPoints [25] characterize the target by adaptively learning a set of representative sampling points. FCOS [26] regresses the distances from the predicted points to the four boundaries to obtain the horizontal prediction box.
Compared with the anchor-based method, using the point as the starting state of the regression can better cope with pixel-level dense scenes. Further, the anchor-free method without anchor box hyperparameters has a more efficient inference speed, so the architecture used in this paper is a center-based anchor-free detector.

2.1.2. Arbitrarily Oriented Object Detection in Aerial Images

Remote-sensing images have more complex contextual relationships between background and foreground, and the targets to be detected have the characteristics of multi-oriented, complex backgrounds, dense arrangements, and extreme scale differences. Therefore, oriented object detection needs to address plenty of domain-specific challenges. SCRDet [27] proposes a feature fusion network based on saliency maps from the perspective of an attention mechanism to cope with small and dense remote-sensing targets. CFC-Net [28] proposes a task-specific polarization function to solve the feature incompatibility problem of classification and regression tasks; ReDet [8] extracts rotation-invariant features by reconstructing the backbone network so that the feature maps are discriminative for objects with arbitrary orientations. SASM [29] proposes a shape-sensitive label assignment method that is suitable for high aspect ratio remote-sensing targets. RBox [30] improves the performance of the Transformer architecture in rotated object detection and solves the problem of redundant feature backgrounds.
Our method is dedicated to solving feature misalignment problems for small and multi-directional remote-sensing targets and is applied to three dimensions of spatial, axial, and semantic.

2.1.3. Feature Alignment in Object Detection

The problem of feature non-alignment generally exists in feature pyramid networks. Related research on feature alignment can be divided into two main research lines:
One line of research focuses on feature alignment in the spatial dimension. R 3 Det [13] uses the best prediction box to guide feature interpolation in the refined stage. The authors of [31] predict 2D offsets for each position on the feature map in semantic segmentation tasks. These methods directly perform spatial reconstruction of feature maps in a lightweight manner. However, it is difficult to guarantee alignment stability only by relying on the predicted 2D offset map. After that, a great deal of work attempted to guide sampling locations. FaPN [32,33] combines shallow and deep features to guide the location of each sampling point; [12,34,35,36,37,38] adapt the sampling points to high-aspect-ratio remote-sensing targets through the rotated bounding box predicted by the additional head.
Another line of research focuses on feature alignment in the semantic dimension. CGNet [39] uses self-attention to enhance communication between pyramid levels. AFF-Det [14,18] maps ROIs to all levels and applies a unified supervisory signal to alleviate the semantic gap. The authors of [16,17] perform feature refinement after scaling. SCRDet++ [40] introduces instance-level denoising to decouple intra-class and inter-class features.
The above methods solve domain-related problems, but the extra detection heads increase the cost of parameters and computation. On the other hand, recursive resampling introduces cross-layer semantic interference. Our intra-level collaborative feature alignment network can act in axial and spatial dimensions without extra supervision signals. In addition, the proposed semantic alignment module addresses the semantic gap in multi-scale features.

2.2. The Proposed Method

2.2.1. Overall Pipeline

The proposed OA-Det is built based on the basic center-based anchor-free detector. The overall architecture consists of four components: the feature extraction backbone, the axially aware spatial alignment network, the hierarchical-wise semantic alignment network, and the oriented detection heads, as shown in Figure 2.
Specifically, we first feed an RGB remote-sensing image as input to the feature extraction backbone network. We adopt ResNet101 [41] pretrained on ImageNet [42] as the backbone network to generate four different scale feature maps { C 2 , C 3 , C 4 , C 5 } , where C i has 1 2 i resolution of the input image. Next, we construct a feature pyramid network to transfer the semantic information of the deep network to the shallow features layer-by-layer. We first downsample C 5 using 3×3 convolution to obtain features with smaller resolution and richer semantic information for adapting to larger-scale remote-sensing instances. Then, we fuse the pyramid features with ResNet features in a top-down manner using lateral connections and interpolation. So far, we can get the pyramid features { P 2 , P 3 , P 4 , P 5 , P 6 , P 7 } . The detection algorithm based on the above multi-scale features can better cope with large-scale variations of remote-sensing objects. It should be noted that C 2 is also fused to obtain P 2 with richer location information. However, we only feed it to the spatial alignment network instead of the final detection head for prediction. More details can be found in Section 2.2.2. Then, the intra-level collaborative alignment network performs spatial information synchronization for each pyramid feature, followed by fine feature sampling using axially aware convolution. Based on the above aligned features, we use a hierarchical-wise context fusion module to enhance the global information from the channel dimension and perform dynamic feature fusion between pyramid levels. Finally, the multi-scale features aligned in spatial, axial, and semantic dimensions are sent to task-specific detection heads. We construct the classification head and regression head by several stacked convolutional layers to predict the probability of the category and the distance vector for each position i in the feature map, respectively. The distance vector can be formulated as follows:
V i = ( l i , t i , r i , b i ) ,
where l i , t i , r i , b i are the distances to the left, top, right, and bottom boundaries of the box; centerness i [26] is predicted on another branch to suppress the low-quality bounding boxes that deviation from the center point of the ground truth. It can be calculated by the following equation:
centerness i = m i n ( l i , r i ) m a x ( l i , r i ) × m i n ( t i , b i ) m a x ( t i , b i ) .
For arbitrarily oriented object detection, the rotated bounding boxes are usually expressed by ( x c , y c , h , w , θ ) , where x c , y c are the center point of the box, and h , w , θ represent the height, width, and angle of the rotated box, respectively. Thus, we add a parallel branch to the regression head to predict θ , where angle representation is referenced to [11]. Finally, the rotated bounding box can be obtained by combining the angle θ with the above distance vector V i . Next, we present a detailed description of the proposed feature alignment module.

2.2.2. Shared-Weight-Based Multi-Scale Feature Alignment Network Using Axial-Aware Convolution

The problem of spatial misalignment is a long-standing problem for multi-scale features. Due to the recursive use of resampling operations in the process of extracting multi-scale features, this problem is gradually exacerbated, and the impact on remote-sensing object detection is even more immeasurable. On the one hand, because of the small and dense arrangement of remote-sensing targets, as long as there is a slight deviation in the deep features, the predicted boxes fall to the adjacent instance. On the other hand, there is also a domain-specific axial misalignment problem in oriented object detection that is caused by the mismatch between the sampling method of regular convolution and the orientation of remote-sensing targets.
In this section, we first design an intra-level collaborative feature alignment network to deal with the spatial misalignment problem. After that, an axially aware convolution is proposed to alleviate the axial misalignment problem of sampling points. The architecture is shown in Figure 3. First, we take the pyramid feature P n as the leading feature and perform adaptive spatially aligned resampling of features at adjacent levels; the implementation detail is as follows:
In order to obtain the prior position information of the feature maps, we predict K sets of 2D offset maps from the leading feature, where K is the number of sampling points at each position. The expression is as follows:
Δ p ( n ) = f o f f s e t ( P n ) ,
where f o f f s e t ( · ) is a convolution layer and Δ p ( n ) is the offset map predicted from the n-th level of pyramidal features. Each offset value consists of a tuple ( x k , y k ) that is used to guide each sampling point. Next, we implement spatial alignment for neighbor features:
P ^ = A P , Δ p ( n ) ,
where · represents the scaling operation and P ^ represents the feature map with pyramid level after spatial alignment. The sparse sampling method A ( · ) is implemented using deformable convolution [43]. Note that during the sampling process, we share offset maps for neighbor features to ensure the consistency of position information within the pyramid level. Next, we focus on how to calculate each point on the feature map, which can be expressed as:
x ^ ( p ) = k = 1 K w p k · x p + p k + Δ p k ( n ) ( p ) ,
where x ^ ( p ) is the value of the spatially aligned feature map at position p; w ( · ) and x ( · ) are kernel weight and input feature value, respectively; and p k is the sampling grid. When the convolution kernel is 3 and the dilation rate is 1, p k ( 1 , 1 ) , ( 1 , 0 ) , , ( 1 , 1 ) .
So far, we have obtained multiple neighbor features at each pyramid level with resolution and spatial information consistent with the leading features. Compared with the non-learnable resampling in FPN, sparse sampling based on shared offsets can reconcile the spatial information of neighbor features and leading features, thereby solving the problem of spatial misalignment.
Next, for the domain-specific axial non-alignment problem, the most commonly used method is alignment convolution [11]. The core idea is to use refined heads to make a pre-prediction before the downstream task and then use the predicted rotated box to guide the sampling position of the convolution. After our observations, this method has the following three shortcomings:
(i)
First, this anchor-aware sampling method tends to introduce foreign background interference. Since the offset is calculated from the width and height of the predicted boxes, although high-aspect-ratios targets are well adapted, there is still a lot of background inside the rotated boxes for targets with irregular shapes, as shown in Figure 4a.
(ii)
The representation of the sampling axis in alignment convolution is unique. Figure 4b shows the axial representation of different sampling methods. The alignment convolution predicts only one box at each position, and each positive sample is learned from a single ground truth. Hence, the sampling axis of each convolution kernel is in the same orientation as the rotated box, for which the representation is unique. However, the sampling axis of the convolution kernel is not affected by the angular periodicity problem [44,45]. Given the same sampling shape, an axis can be represented by four angles to achieve the same sampling effect.
(iii)
Additional supervision signals and anchor-based sub-heads increase the number of parameters and amount of computation.
Through the above analysis, we propose an axially aware convolution to deal with the domain-specific axial misalignment problem. The overall process of axial alignment is as follows: To accurately perceive the target’s orientation, we need to make full use of the neighbor features that have been aligned in spatial dimension. So we first fuse these features by element-wise addition and feed them into the spatially aware block, which can be formulated as:
X sa = P ^ n 1 + P ^ n + P ^ n + 1 X ^ sa = Proj ( F c ( F s ( X sa ) ) ) + X sa ,
where F s and F c represent the feature aggregation in spatial and channel dimensions, respectively. We use 7 × 7 depth-wise convolution combined with point-wise convolution to implement the above operation. Our idea is to extract local spatial information with a large receptive field using the group convolution of the big kernel. At the same time, we use two successive 1 × 1 convolutions to achieve multi-layer perception(MLP)-like channel projections, which can enhance the generalizability of features at a low cost.
So far, the aggregated features contain rich semantic knowledge and localization information from the upper and lower levels, respectively. We then apply the proposed axially aware convolution for refined sampling. Different from directly predicting the offset tuple, we predict a θ at each position to characterize the orientation of the sampling point set. In addition, an activation scalar is predicted to indicate the importance of each sample point after shifting [46]. The expression is as follows:
Δ θ ( n ) = f a x i s ( X ^ sa ) , Δ w ( n ) = f m a s k ( X ^ sa ) ,
where f a x i s and f m a s k are 3 × 3 convolution. We predict an axis map Δ θ ( n ) and an L scalar map Δ w ( n ) for each pyramid level. It is worth noting that we use the same weights between layers to predict the axis map, while for the scalar map, we also share weights within each level. In this way, we endow the maps with the same implicit supervision signal to achieve cross-layer axial alignment. Eventually, the finely aligned feature map can be obtained by axially aware sampling:
y ^ ( p ) = k = 1 K w p k · x ^ p + p k + Δ k ( n ) ( p ) · Δ w k ( n ) ( p ) , Δ k ( n ) = M ( Δ θ ( n ) ) · p k p k ,
where Δ k ( n ) are the axially aware offsets for each sampling point, M ( · ) represents an affine rotation matrix. In contrast to Equation (5), the first difference is that the offset of the K sampling points is not directly predicted but is calculated by an axis shared across levels. Thus, the convolutional kernel is endowed with axial perception ability. Secondly, each offset sample point is re-weighted to avoid interference from redundant backgrounds. In this work, we do not compute the loss of θ with the ground truth in an explicitly supervised manner. The reason is that we consider predicting the regression box and guiding the sampling points as different tasks. The former needs to accurately predict the angle of the rotation box, which is unique, while the axial representation can be various. Therefore, additional supervision will limit the generalization ability of axial perception. Comparative experiments of different axial alignment methods can be found in Section 3.3.3. Further, in order to ensure the stability of the predicted axis, we assign the shared weights to collaboratively optimize the axis at each position. See Section 3.3.4 for ablation experiments on weight strategies.
Correctly choosing the pyramid level that needs to be aligned is also a point that cannot be ignored. First, the location-rich P 2 is introduced to guide the upper-level features for more accurate spatial and axial alignment. However, we do not feed it into the detection head for final prediction. In this way, we can cope with dense predictions effectively. Second, since the low-resolution features contain little spatial information, the network does not act on the high-level pyramid feature P 7 . Nevertheless, our alignment networks have multiple in-degrees for each level of feature data flow during forward and backward propagation, so it can actually affect all pyramid features. Specifically, in the process of network forward inference, features can obtain rich semantic knowledge and location information from adjacent levels. Moreover, during backward propagation, the features can obtain the reverse gradient from the resampled features provided to other levels. More detailed alignment layer ablation experiments are shown in Section 3.3.2.

2.2.3. Hierarchical-wise Context Fusion Module for Semantic Alignment

So far, we have obtained neighbor features aligned in spatial and axial dimensions within each pyramid level, but there is still a problem of semantic misalignment between them. When faced with remote-sensing targets with variable scales, even if the OBBs are correctly regressed, classification errors still exist due to the semantic differences between levels. In the case of a complex contextual relationship between the background and the foreground of aerial images, the foreground is overwhelmed by a large number of object-like backgrounds. The mainstream method [9] conveys the semantic information layer-by-layer using lateral connections. Other improved works [13,14,15,16,17,18] apply semantic reconfiguration to single-scale aggregated features resampled from all pyramid levels. However, we find that these methods have certain shortcomings:
  • First, fusing all pyramid features with different receptive fields into a single scale introduces a lot of irrelevant semantic information. Figure 5a demonstrates the correlations of the backbone features and pyramid features. We can find that the values near the diagonal are significantly larger than the remote ones, whether before or after feature fusion, indicating that the features of each layer are only most correlated with neighbor features. Therefore, it is not appropriate to fuse all features after resizing them to the same scale.
  • Secondly, feature fusion based on lateral connections is equivalent to a linear combination of the feature hierarchy, which is weak in capturing highly nonlinear features. Thus, it is difficult to cope with remote-sensing targets with various shapes and scales.
  • These methods ignore feature misalignment in spatial and axial dimensions before fusing multi-scale features, resulting in semantic interference across levels.
In order to solve the above problems, we propose a hierarchical-wise global context fusion module. As shown in Figure 6, we take the spatially and axially aligned features as input and first perform global contextual semantic modeling for each pyramid level, which can be formulated as:
G = softmax ( W k · F ) F ,
where F C × H × W represents the axially aligned features of the -th layer in the current pyramid level. The learnable projections matrix W k adaptively compresses the feature channels in a query-independent manner. After that, a 2D activation map is obtained by a normalization operation. We apply it to the original feature map and calculate the weighted average of all positions.
At this point, we have obtained the global context information of each layer. However, considering remote-sensing targets’ varying scale and complex context, the extracted global information still lacks inter-layer communication. Next, we use two parallel branches to enhance this global context information from the channel and hierarchical dimensions. The importance of each layer can be extracted as follow:
𝕎 l e v e l = σ W g = 1 L G ,
where ⋃ represents channel concatenation and W g is a learnable linear transformation matrices. Here, we consider the L groups of global contextual information as a whole to predict a vector 𝕎 l e v e l , which represents the importance of each hierarchy. At the same time, we capture the inter-channel dependencies in another branch and apply the above vector to different layers, which are expressed as follows:
G = 𝕎 l e v e l = 1 L F s G ,
where ⊙ represent element-wise product. The channel-wise feature selection F s ( · ) first reduces the number of channels by using a 1 × 1 convolution followed by a layer normalization and ReLU activation function. In this way, important contextual information can be selectively retained, and useless channels are discarded so that the robustness of global contextual information is improved. Next, we use hierarchical weights 𝕎 l e v e l to re-weight channels of different layers, endowing the global contextual information with hierarchical-aware capability. Finally, we fuse feature maps from different layers with the scaled global context to obtain semantically aligned features:
Y = = 1 L G + F .
It is worth noting that the proposed alignment networks are based on neighbor features. Figure 5b visualizes the distribution of predicted quantities in the feature pyramid for different categories in the DOTA dataset. It can be found that each category concentrates on one and its adjacent level of the pyramid according to the target scale. Therefore, feature semantic alignment based on neighboring layers is the optimal level selection strategy to solve the semantic gap problem in the remote-sensing field. See Section 3.3.2 for detailed ablation experiments about selection strategies. On the one hand, our method only aligns 2× resampled neighbor features, which can significantly preserve the feature scale sensitivity. On the other hand, the weights of linear transformations (e.g., W g and W k ) in this module are shared within the hierarchy, so this co-optimization method can also alleviate the problem of semantic misalignment. More ablation experiments can be found in Section 3.3.5.

2.2.4. Loss Function

The loss function of the network consists of three parts for different tasks: the classification loss for category prediction, the regression loss for oriented box localization, and the centerness loss for anchor quality evaluation. Compared with other mainstream alignment methods in the remote-sensing field, this network does not need to calculate the extra loss of subhead, so it is easier to optimize. The definition of the loss function is as follows:
L t o t a l = μ 1 L c l s p i , p i * + μ 2 p i * L r e g v i , v i * + μ 2 p i * L c t r c i , c i * ,
where the classification loss L c l s is focal loss [6], p i represents the predicted category label at the i-th position of the feature map, μ 1 and μ 2 are the trade-off hyper-parameters and are both set to 1.0 by default, and p i * is the point sample assigned by the central portion sampling strategy and is 1 if p i * is a positive sample and 0 otherwise. The regression loss L r e g is IoU loss [47]; v i represents the rotated box decoded from the predicted distance vector and angle, while v i * is the ground truth box. The centerness loss L c t r is the cross entropy loss.

3. Results

3.1. Datasets

We conduct experiments on several remote-sensing datasets for oriented object detection, including DOTA1.0 [48], HRSC2016 [49], and UCAS-AOD [50]. Objects in all three datasets are labeled with oriented bounding boxes.
The DOTA dataset is a large-scale, publicly available image dataset for remote-sensing target detection with a total of 2806 images, ranging in size from 800 × 800 to 4000 × 4000. The dataset is annotated with 188,282 instances divided into 15 categories: bridge (BR), harbor (HA), ship (SH), plane (PL), helicopter (HC), small vehicle (SV), large vehicle (LV), baseball diamond (BD), ground track field (GTF), tennis court (TC), basketball court (BC), soccer field (SBF), roundabout (RA), swimming pool (SP), and storage tank (ST). The dataset is divided into three parts for training, validation, and testing with a ratio of 3:1:2, and the final test results are obtained through the official evaluation server. During training, we split the image into 1024 × 1024 sub-images with 200-pixel overlap.
HRSC2016 is an optical satellite dataset for oriented ship detection and includes 1061 images with size ranging from 300 × 300 to 1500 × 900. The images of this dataset were collected from six famous harbors that contain ships on the sea and close in-shore. The dataset is randomly divided into 436 images as the training set, 181 as the validation set, and 444 as the test set. In the experiments, we use the training set and validation set for training, and we evaluate the model on the test set in terms of PASCAL VOC07 metrics.
UCAS-AOD contains 1510 aerial images with two categories (car and plane) and 14,596 instances. The resolution of the images is 659 × 1280, with 510 images of cars and 1000 of airplanes. During training and inference, all images are resized to 800 × 1333 without changing the aspect ratios. For a fair comparison with the compared methods, we follow the same dataset splits setting in [51], where the ratio between the training set, validation set, and testing set is 5:2:3.

3.2. Implementation Details

We adopt the FPN-based ResNet50 [41] as our backbone network unless otherwise specified. The network is trained with a stochastic gradient descent (SGD) optimizer over a single GPU with two images per mini-batch. The weight decay and momentum are set to 10 4 and 0.9, respectively. The initial learning rate is set to 0.0025. Models are trained for 12 epochs, and the learning rate is reduced by a factor of 10 at the end of epochs 12, 16, and 19. We apply random horizontal flipping and random poly rotating with 0.5 probability for data augmentation. The inference time is calculated on a single GeForce RTX 3080 with 10GB memory. Our model is based on the MMRotate [52], an open-source toolbox for oriented object detection based on PyTorch.

3.3. Ablation Study

3.3.1. Factor-by-Factor Experiment

In this section, we first give a macroscopic overview of the performance improvement and inference efficiency of the proposed spatial alignment, axial sampling, and semantic alignment compared to the baseline, as shown in Table 1. The structure and hyperparameters of each module are the optimal configurations obtained through more-detailed ablation experiments in the following sections. All ablation experiments are tested on the validation set of the DOTA dataset.
First, we used the naive center-based anchor-free detector as our baseline, which achieved 65.25% mAP and 23.8 FPS. After introducing the intra-level multi-scale spatial alignment module, mAP is improved by 2.43%. It can be seen that the network aligns the spatial location of features in different layers under the guidance of the leading feature. Subsequently, we introduced axially aware convolution for fine alignment. We achieved 0.21% mAP improvement and moderate inference speed, which shows that the proposed axially aligned sampling can cope with the varying orientation of remote-sensing targets to a certain extent, and the extracted features can characterize the rotated targets more effectively. Next, we add a hierarchical-wise semantic alignment network based on the spatially aligned features to verify the role of the semantic alignment module. The result shows a 0.6% mAP improvement with only a slight impact on inference speed, which verifies that this module can enhance intra-level semantics and alleviate cross-level semantic gaps. Finally, our OA-Det totally improves by 4.03% mAP while maintaining considerable efficiency compared to the baseline, solving the feature misalignment problem of the remote-sensing target from three dimensions. In subsequent experiments, we go deep inside each module to verify the effectiveness of the components.

3.3.2. Evaluation of Spatial Alignment

For the proposed spatial alignment network, we tested various pyramid-level selection strategies to verify our point of view. The results are shown in Table 2.
We used the feature pyramid network as the baseline for this ablation experiment and further tested three different types of level selection strategies. The first strategy is to resample all pyramid features to the same resolution as P 4 and then perform feature refinement at all levels. This method samples up to 8x but only achieves a 0.94% improvement. Another strategy is to downsample high-resolution features to provide location information to upper features, which improves by 2.22% mAP compared to the baseline, validating the superiority of this bottom-up strategy. Next, our strategy, i.e., resampling the neighbor feature maps to the same resolution, achieves an impressive 2.3% mAP improvement. The result verifies that this strategy can provide semantic and location information to the leading level while maintaining the feature scale sensitivity. Finally, we test the effectiveness of introducing pyramid level P 2 , which further improves 0.13% mAP with a negligible impact on inference speed. In summary, the proposed pyramid level selection strategy improves 2.43% mAP compared to the baseline with considerable speed. It verifies our point of view: spatial alignment of adjacent levels is optimal.

3.3.3. Evaluation of Axially Aware Convolution

In this section, we verify the effectiveness of the proposed axially aware sampling method based on the multi-scale features extracted by the spatial alignment network. Five mainstream alignment methods are compared to test their axial alignment effects for remote-sensing targets with different orientations. Among them, regular convolution, grid sample, deformable convolution, and axially aware convolution are all sampling methods without additional detection heads. In addition, to compare with the alignment convolution, we designed an additional supervised version for the axially aware convolution, and the experimental results are shown in Table 3.
First, regular convolution achieves 66.67% mAP without feature aggregation of multi-scale features. Next, deformable convolution achieves an additional 0.34% mAP improvement over the regular grid sampling. Therefore, guiding sampling points can better alleviate the axial non-alignment problem of remote-sensing targets compared with directly predicting offset maps to interpolate features. However, the disadvantage of deformation convolution is the lack of offset restrictions, so it is easy to introduce a redundant background, as shown in Figure 7b. Then, we compare the performance of axially aware convolution and alignment convolution under sub-head guidance in the same network architecture. In this case, we aggregate features by a stack of convolution from the additional head instead of a 3 × 3 convolution. The results show that the axial convolution achieves a 0.21% mAP advantage. Although the accuracy decreases by 0.13% mAP compared with the unsupervised version of axially aware convolution, it still outperforms the alignment convolution, which shows the effectiveness of the proposed axial alignment method. After introducing the spatially aware block, the performance exceeds that of the supervised version by 0.2% mAP, proving that the large kernel can provide broader spatial information for axial prediction tasks. The visualization results are shown in Figure 7d. Axially aware convolution makes up for the shortcomings of alignment convolution, which can detect more small objects and align their orientations without introducing explicit supervision signals and detection heads.
Next, we evaluate the mean Average Orientation Error (mAOE) of the above methods, validating their axial alignment ability for different categories in DOTA. As shown in Table 4, our axially aware convolution outperforms other compared feature alignment methods and achieves minimal orientation errors on BD, GTF, LV, SP, and HC, demonstrating excellent axial alignment performance for remote-sensing targets with high aspect ratios and varying shapes.

3.3.4. Evaluation of Collaborative Optimization Strategy

Our axially aware convolution achieves better performance without explicit signal guidance due to the shared weight co-optimization strategy to improve the robustness of the predicted axis. This section conducts comparative experiments on different weight-sharing strategies of the spatial alignment network. The results are shown in Table 5.
First, we test the effect of the collaborative optimization strategy on spatial alignment. The results show that after sharing the offset weight, the performance achieved a significant improvement of 1.23%; so it can be proven that sharing offsets for adjacent layer features can help spatial alignment. Next, we focus on the axis-aware convolution of the predicted axis and sampling weight. For the sampling point weights, we find that the co-optimization strategy can achieve a slight accuracy improvement. Since most axially aligned sampling points are highly context-related compared to sparse sampling, sharing mask weight can make full use of the multi-scale features within the layer to distinguish foreground from background, thus further downweighting the sampling points still in the background. When optimizing the axial prediction task separately, the overall performance decreases. However, when axial weights are shared across levels, the network outperforms the previous stage by 0.41% mAP, proving that this strategy of sharing weights across layers can make axial prediction more robust. So far, we have verified the validity of all components inside the spatial alignment network.

3.3.5. Evaluation of Semantic Alignment

In this section, we perform ablation experiments on the proposed semantic alignment networks to verify the effectiveness of each internal component one-by-one, and the results are shown in Table 6. For a fair comparison, we first construct the network using a naive approach instead of the proposed components, where global average pooling (GAP) is performed to the extracted global context, and a 1 × 1 convolution is used to perform simple linear projection, achieving 65.13% mAP.
Next, we use the proposed query-independent global context modeling to replace GAP. We found that a 2.4% mAP improvement is achieved with almost no parameter increase. We further introduce channel-wise feature selection, which outperforms the naive method while reducing the parameter by 0.13M. Finally, the introduction of hierarchical-wise reweighting improves the accuracy by 0.71%, proving that this component empowers global context hierarchy awareness. Compared with the network of the naive method, our module brings an overall improvement of 3.15% mAP. At this point, we have verified the rationality of the internal design of our semantic alignment network, which can alleviate the semantic gap problem and strengthen the exchange of information between levels.

3.4. Main Results and Analysis

3.4.1. Results on DOTA

We show the performance of state-of-the-art methods on the DOTA dataset for two-stage, single-stage, and anchor-free detectors in Table 7. Our method achieves excellent performance of 74.82% and 78.11% mAP in single-scale and multi-scale tests, respectively, with ResNet101 as the backbone.
Note that our OA-Det achieves the best performance among the compared methods in the categories of small vehicle (SV), large vehicle (LV), and ship (SH), which shows that our method is good at dealing with small targets with different orientations in dense scenarios. Even more surprising is that we have a huge lead in the harbor (HA) and swimming pool (SP), which demonstrates that for objects with different shapes in complex backgrounds, our method can better perceive their orientations to obtain accurate regression. The visualization results of the DOTA dataset are shown in Figure 8.

3.4.2. Results on HRSC2016

We validate the effectiveness of our OA-Det on the HRSC2016 dataset. The slender ships in the dataset are challenging for anchor-free detectors, but we still achieve excellent performance with the feature alignment network.
As shown in Table 8, the results show that our OA-Det outperforms the compared anchor-free methods (e.g., RepPoints and CenterNet) and even anchor-based methods by 90.10% mAP. The visualization results of the HRSC2016 dataset are shown in Figure 9. With the proposed alignment network, we can accurately regress high-aspect-ratio objects and correctly distinguish ships from harbors.

3.4.3. Results on UCAS-AOD

The UCAS-AOD dataset contains objects with different orientations and scales. Table 9 shows that our OA-Det achieves the best performance among the compared methods. Note that Yolov3 [69] and Faster-RCNN [4] are methods based on horizontal bounding boxes, so we add angle prediction to the above two methods. The results show that our OA-Det achieves state-of-the-art performance of 90.29% mAP, outperforming other compared one-stage and two-stage detectors in both categories. The visualization of the dataset is shown in Figure 10.

4. Discussion

After analyzing the above ablation experiment and the main results from shallow-to-deep, we verified the effectiveness of each module and its internal components from several aspects. Nevertheless, we still find that the proposed method needs to be improved in some categories. For targets with extreme aspect ratios (e.g., harbors and bridges), although our alignment network can sense the target’s orientation, anchor-free detectors usually cannot accurately return the entire target. In the future, we will do further in-depth research on the OBB representation of anchor-free detectors. Further, transformer-based algorithms have achieved excellent performance in many arbitrarily oriented object detection networks. We will also try to use the Swin transformer [1] to solve the problems in the remote-sensing field.

5. Conclusions

Aiming at the problem of multi-scale feature misalignment in oriented object detection, this paper analyzes the root causes of the problem from three dimensions and gives the corresponding efficient solutions. First, we propose an intra-level collaborative feature alignment network that can absorb the semantic knowledge and localization information under different receptive fields by sparse sampling and axially aware convolution. In this way, the orientations of instances and spatial information of pyramid features can be aligned in a coarse-to-fine manner without additional supervised signals. Based on this, we further propose a hierarchical-wise global context module that alleviates the semantic gap of multi-scale features from the channel- and pyramid-level dimensions. Detailed ablation experiments and excellent performance on three testbeds demonstrate the effectiveness of the proposed method.

Author Contributions

Conceptualization, X.H.; Methodology, X.H.; Software, X.H.; Validation, X.H.; Formal analysis, X.H.; Investigation, X.H.; Resources, X.H. and C.Z.; Data curation, X.H.; Writing—original draft preparation, X.H.; Writing—review and editing, X.H. and C.Z.; Visualization, X.H.; Supervision, C.Z.; Project administration, X.H.; and Funding acquisition, C.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the National Natural Science Foundation of China (CN) under grant numbers 62276164 and 61602296 and the ‘Chenguang Program’ supported by the Shanghai Education Development Foundation and the Shanghai Municipal Education Commission under grant number 18CG54. Furthermore, this work is also sponsored by ‘Science and technology innovation action plan’ Natural Science Foundation of Shanghai under grant number 22ZR1427000. The authors would like to thank their supports.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The source code of the paper is available at https://github.com/Virusxxxxxxx/OA-Det, accessed on 20 November 2022.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 1–17 October 2021; pp. 10012–10022. [Google Scholar]
  2. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask R-CNN. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  3. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image is Worth 16 × 16 Words: Transformers for Image Recognition at Scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  4. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Cai, Z.; Vasconcelos, N. Cascade R-CNN: Delving Into High Quality Object Detection. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 6154–6162. [Google Scholar]
  6. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal Loss for Dense Object Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 318–327. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. SSD: Single Shot MultiBox Detector. In Proceedings of the Computer Vision ECCV 2016; Leibe, B., Matas, J., Sebe, N., Welling, M., Eds.; Springer International Publishing: Cham, Switzerland, 2016; pp. 21–37. [Google Scholar]
  8. Han, J.; Ding, J.; Xue, N.; Xia, G.S. ReDet: A Rotation-Equivariant Detector for Aerial Object Detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 2786–2795. [Google Scholar]
  9. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature Pyramid Networks for Object Detection. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 936–944. [Google Scholar]
  10. Cheng, G.; Wang, J.; Li, K.; Xie, X.; Lang, C.; Yao, Y.; Han, J. Anchor-Free Oriented Proposal Generator for Object Detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–11. [Google Scholar] [CrossRef]
  11. Han, J.; Ding, J.; Li, J.; Xia, G.S. Align Deep Features for Oriented Object Detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–11. [Google Scholar] [CrossRef]
  12. Yu, Y.; Yang, X.; Li, J.; Gao, X. Object Detection for Aerial Images with Feature Enhancement and Soft Label Assignment. IEEE Trans. Geosci. Remote Sens. 2022, 6, 1–11. [Google Scholar] [CrossRef]
  13. Yang, X.; Yan, J.; Feng, Z.; He, T. R3Det: Refined Single-Stage Detector with Feature Refinement for Rotating Object. Proc. Aaai Conf. Artif. Intell. 2021, 35, 3163–3171. [Google Scholar] [CrossRef]
  14. Zhen, P.; Wang, S.; Zhang, S.; Yan, X.; Wang, W.; Ji, Z.; Chen, H.B. Towards Accurate Oriented Object Detection in Aerial Images with Adaptive Multi-level Feature Fusion. ACM Trans. Multimed. Comput. Commun. Appl. 2022, 3513133. [Google Scholar] [CrossRef]
  15. Shen, Y.; Zhang, D.; Song, Z.; Jiang, X.; Ye, Q. Enhanced Neck Feature Representation for Object Detection in Aerial Images. arXiv 2022, arXiv:2204.02033. [Google Scholar]
  16. Kong, T.; Sun, F.; Huang, W.; Liu, H. Deep Feature Pyramid Reconfiguration for Object Detection. In Computer Vision—ECCV 2018; Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y., Eds.; Springer International Publishing: Cham, Switzerland, 2018; Volume 11209, pp. 172–188. [Google Scholar] [CrossRef] [Green Version]
  17. Pang, J.; Chen, K.; Shi, J.; Feng, H.; Ouyang, W.; Lin, D. Libra R-CNN: Towards Balanced Learning for Object Detection. arXiv 2019, arXiv:1904.02701. [Google Scholar]
  18. Guo, C.; Fan, B.; Zhang, Q.; Xiang, S.; Pan, C. AugFPN: Improving Multi-scale Feature Learning for Object Detection. arXiv 2019, arXiv:1912.05384. [Google Scholar]
  19. Lowe, D.G. Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
  20. Dalal, N.; Triggs, B. Histograms of oriented gradients for human detection. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), San Diego, CA, USA, 20–25 June 2005; Volume 1, pp. 886–893. [Google Scholar]
  21. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef] [Green Version]
  22. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Curran Associates Inc.: Red Hook, NY, USA, 2017; pp. 6000–6010. [Google Scholar]
  23. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  24. Zhou, X.; Wang, D.; Krähenbühl, P. Objects as Points. arXiv 2019, arXiv:1904.07850. [Google Scholar]
  25. Yang, Z.; Liu, S.; Hu, H.; Wang, L.; Lin, S. RepPoints: Point Set Representation for Object Detection. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; IEEE: Seoul, Republic of Korea, 2019; pp. 9656–9665. [Google Scholar] [CrossRef] [Green Version]
  26. Tian, Z.; Shen, C.; Chen, H.; He, T. FCOS: Fully Convolutional One-Stage Object Detection. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9626–9635. [Google Scholar]
  27. Yang, X.; Yang, J.; Yan, J.; Zhang, Y.; Zhang, T.; Guo, Z.; Sun, X.; Fu, K. SCRDet: Towards More Robust Detection for Small, Cluttered and Rotated Objects. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 8231–8240. [Google Scholar]
  28. Ming, Q.; Miao, L.; Zhou, Z.; Dong, Y. CFC-Net: A Critical Feature Capturing Network for Arbitrary-Oriented Object Detection in Remote-Sensing Images. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–14. [Google Scholar] [CrossRef]
  29. Hou, L.; Lu, K.; Xue, J.; Li, Y. Shape-Adaptive Selection and Measurement for Oriented Object Detection. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 22 February–1 March 2022. [Google Scholar]
  30. Tang, J.; Zhang, W.; Liu, H.; Yang, M.; Jiang, B.; Hu, G.; Bai, X. Few Could Be Better Than All: Feature Sampling and Grouping for Scene Text Detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 4563–4572. [Google Scholar]
  31. Huang, Z.; Wei, Y.; Wang, X.; Liu, W.; Huang, T.S.; Shi, H. Alignseg: Feature-aligned segmentation networks. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 550–557. [Google Scholar] [CrossRef]
  32. Huang, S.; Lu, Z.; Cheng, R.; He, C. FaPN: Feature-aligned pyramid network for dense image prediction. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 864–873. [Google Scholar]
  33. Yuan, Y.; Li, Z.; Ma, D. Feature-Aligned Single-Stage Rotation Object Detection With Continuous Boundary. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–11. [Google Scholar] [CrossRef]
  34. Dai, L.; Chen, H.; Li, Y.; Kong, C.; Fan, Z.; Lu, J.; Chen, X. TARDet: Two-Stage Anchor-Free Rotating Object Detector in Aerial Images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, New Orleans, LA, USA, 19–20 June 2022; pp. 4267–4275. [Google Scholar]
  35. Hou, L.; Lu, K.; Xue, J. Refined One-Stage Oriented Object Detection Method for Remote Sensing Images. IEEE Trans. Image Process. 2022, 31, 1545–1558. [Google Scholar] [CrossRef]
  36. Zou, M.; Hu, Z.; Guan, Y.; Gan, Z.; Guan, C.; Leng, S. Feature Adaption with Predicted Boxes for Oriented Object Detection in Aerial Images. In Proceedings of the PRICAI 2021: Trends in Artificial Intelligence, Hanoi, Vietnam, 8–12 November 2021; pp. 366–378. [Google Scholar]
  37. Zhang, F.; Wang, X.; Zhou, S.; Wang, Y. DARDet: A Dense Anchor-Free Rotated Object Detector in Aerial Images. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  38. Huang, Z.; Li, W.; Xia, X.G.; Wang, H.; Tao, R. Task-wise Sampling Convolutions for Arbitrary-Oriented Object Detection in Aerial Images. arXiv 2022, arXiv:2209.02200. [Google Scholar]
  39. Wei, Z.; Liang, D.; Zhang, D.; Zhang, L.; Geng, Q.; Wei, M.; Zhou, H. Learning calibrated-guidance for object detection in aerial images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 2721–2733. [Google Scholar] [CrossRef]
  40. Yang, X.; Yan, J.; Liao, W.; Yang, X.; Tang, J.; He, T. SCRDet++: Detecting Small, Cluttered and Rotated Objects via Instance-Level Feature Denoising and Rotation Loss Smoothing. IEEE Trans. Pattern Anal. Mach. Intell. 2022. [Google Scholar] [CrossRef]
  41. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27 June–30 June 2016; pp. 770–778. [Google Scholar]
  42. Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Fei-Fei, L. ImageNet: A Large-Scale Hierarchical Image Database. In Proceedings of the CVPR09, Miami, FL, USA, 20–25 June 2009. [Google Scholar]
  43. Dai, J.; Qi, H.; Xiong, Y.; Li, Y.; Zhang, G.; Hu, H.; Wei, Y. Deformable Convolutional Networks. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 764–773. [Google Scholar]
  44. Yang, X.; Yan, J.; Ming, Q.; Wang, W.; Zhang, X.; Tian, Q. Rethinking Rotated Object Detection with Gaussian Wasserstein Distance Loss. In Proceedings of the 38th International Conference on Machine Learning, Virtual, 18–24 July 2021; pp. 11830–11841. [Google Scholar]
  45. Yang, X.; Yang, X.; Yang, J.; Ming, Q.; Wang, W.; Tian, Q.; Yan, J. Learning High-Precision Bounding Box for Rotated Object Detection via Kullback-Leibler Divergence. In Proceedings of the Advances in Neural Information Processing Systems, Virtual, 6–14 December 2021; Volume 34, pp. 18381–18394. Available online: https://yangxue0827.github.io/files/neurips21_yx_poster.pdf (accessed on 28 November 2022).
  46. Zhu, X.; Hu, H.; Lin, S.; Dai, J. Deformable ConvNets v2: More Deformable, Better Results. arXiv 2018, arXiv:1811.11168. [Google Scholar]
  47. Yu, J.; Jiang, Y.; Wang, Z.; Cao, Z.; Huang, T. Unitbox: An advanced object detection network. In Proceedings of the 24th ACM international conference on Multimedia, Amsterdam, The Netherlands, 15–19 October 2016; pp. 516–520. [Google Scholar]
  48. Xia, G.S.; Bai, X.; Ding, J.; Zhu, Z.; Belongie, S.; Luo, J.; Datcu, M.; Pelillo, M.; Zhang, L. DOTA: A Large-Scale Dataset for Object Detection in Aerial Images. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3974–3983. [Google Scholar]
  49. Liu, Z.; Yuan, L.; Weng, L.; Yang, Y. A High Resolution Optical Satellite Image Dataset for Ship Recognition and Some New Baselines. In Proceedings of the 6th International Conference on Pattern Recognition Applications and Methods, SCITEPRESS—Science and Technology Publications, Porto, Portugal, 19–21 February 2017; pp. 324–331. Available online: https://pdfs.semanticscholar.org/e6a3/2b4df848fd74b43486c5232ebd362eb90416.pdf (accessed on 28 November 2022). [CrossRef]
  50. Zhu, H.; Chen, X.; Dai, W.; Fu, K.; Ye, Q.; Jiao, J. Orientation robust object detection in aerial images using deep convolutional neural network. In Proceedings of the 2015 IEEE International Conference on Image Processing (ICIP), Quebec City, QC, Canada, 27–30 September 2015; pp. 3735–3739. [Google Scholar]
  51. Li, W.; Chen, Y.; Hu, K.; Zhu, J. Oriented reppoints for aerial object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 1829–1838. [Google Scholar]
  52. Zhou, Y.; Yang, X.; Zhang, G.; Wang, J.; Liu, Y.; Hou, L.; Jiang, X.; Liu, X.; Yan, J.; Lyu, C.; et al. MMRotate: A Rotated Object Detection Benchmark using PyTorch. arXiv 2022, arXiv:2204.13317. [Google Scholar]
  53. Azimi, S.M.; Vig, E.; Bahmanyar, R.; Körner, M.; Reinartz, P. Towards Multi-class Object Detection in Unconstrained Remote Sensing Imagery. In Proceedings of the Computer Vision—ACCV 2018; Springer: Cham, Switzerland, 2019; pp. 150–165. [Google Scholar]
  54. Ding, J.; Xue, N.; Long, Y.; Xia, G.S.; Lu, Q. Learning RoI Transformer for Oriented Object Detection in Aerial Images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 2849–2858. [Google Scholar]
  55. Li, C.; Xu, C.; Cui, Z.; Wang, D.; Zhang, T.; Yang, J. Feature-Attentioned Object Detection in Remote Sensing Imagery. In Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; pp. 3886–3890. [Google Scholar]
  56. Xu, Y.; Fu, M.; Wang, Q.; Wang, Y.; Chen, K.; Xia, G.S.; Bai, X. Gliding Vertex on the Horizontal Bounding Box for Multi-Oriented Object Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 1452–1459. [Google Scholar] [CrossRef]
  57. Wang, J.; Ding, J.; Guo, H.; Cheng, W.; Pan, T.; Yang, W. Mask OBB: A Semantic Attention-Based Mask Oriented Bounding Box Representation for Multi-Category Object Detection in Aerial Images. Remote Sens. 2019, 11, 2930. [Google Scholar] [CrossRef] [Green Version]
  58. Wang, J.; Yang, W.; Li, H.C.; Zhang, H.; Xia, G.S. Learning Center Probability Map for Detecting Objects in Aerial Images. IEEE Trans. Geosci. Remote Sens. 2021, 59, 4307–4323. [Google Scholar] [CrossRef]
  59. Chen, Z.; Chen, K.; Lin, W.; See, J.; Yu, H.; Ke, Y.; Yang, C. PIoU Loss: Towards Accurate Oriented Object Detection in Complex Environments. In Proceedings of the Computer Vision—ECCV 2020; Springer: Cham, Switzerland, 2020; pp. 195–211. [Google Scholar]
  60. Xiao, Z.; Wang, K.; Wan, Q.; Tan, X.; Xu, C.; Xia, F. A2S-Det: Efficiency Anchor Matching in Aerial Image Oriented Object Detection. Remote Sens. 2021, 13, 73. [Google Scholar] [CrossRef]
  61. Ming, Q.; Zhou, Z.; Miao, L.; Zhang, H.; Li, L. Dynamic Anchor Learning for Arbitrary-Oriented Object Detection. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtually, 22 February–1 March 2021; Volume 35, pp. 2355–2363. Available online: https://arxiv.org/abs/2012.04150 (accessed on 28 November 2022).
  62. Ming, Q.; Miao, L.; Zhou, Z.; Song, J.; Yang, X. Sparse Label Assignment for Oriented Object Detection in Aerial Images. Remote Sens. 2021, 13, 2664. [Google Scholar] [CrossRef]
  63. Zhou, L.; Wei, H.; Li, H.; Zhao, W.; Zhang, Y.; Zhang, Y. Arbitrary-Oriented Object Detection in Remote Sensing Images Based on Polar Coordinates. IEEE Access 2020, 8, 223373–223384. [Google Scholar] [CrossRef]
  64. Wei, H.; Zhang, Y.; Chang, Z.; Li, H.; Wang, H.; Sun, X. Oriented objects as pairs of middle lines. ISPRS J. Photogramm. Remote Sens. 2020, 169, 268–279. [Google Scholar] [CrossRef]
  65. Yi, J.; Wu, P.; Liu, B.; Huang, Q.; Qu, H.; Metaxas, D. Oriented Object Detection in Aerial Images with Box Boundary-Aware Vectors. In Proceedings of the 2021 IEEE Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 3–8 January 2021; pp. 2149–2158. [Google Scholar]
  66. Pan, X.; Ren, Y.; Sheng, K.; Dong, W.; Yuan, H.; Guo, X.; Ma, C.; Xu, C. Dynamic Refinement Network for Oriented and Densely Packed Object Detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 11207–11216. [Google Scholar]
  67. Xiao, Z.; Qian, L.; Shao, W.; Tan, X.; Wang, K. Axis learning for orientated objects detection in aerial images. Remote Sens. 2020, 12, 908. [Google Scholar] [CrossRef] [Green Version]
  68. Yang, X.; Yan, J. On the arbitrary-oriented object detection: Classification based approaches revisited. Int. J. Comput. Vis. 2022, 130, 1340–1365. [Google Scholar] [CrossRef]
  69. Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
Figure 1. (a) The sampling axis of the regular convolution does not match the remote-sensing target with different orientations. The green box is the ground truth, the yellow points are the four vertices, and the red points are the sampling points of the convolution. (b) Axial misalignment leads to regression error in predicting rotated boxes. (c) Missed detection of densely arranged remote-sensing targets due to spatial misalignment. (d) Classification errors caused by semantic misalignment(red dashed circle). A portion of the soccer field was misdetected as a tennis court.
Figure 1. (a) The sampling axis of the regular convolution does not match the remote-sensing target with different orientations. The green box is the ground truth, the yellow points are the four vertices, and the red points are the sampling points of the convolution. (b) Axial misalignment leads to regression error in predicting rotated boxes. (c) Missed detection of densely arranged remote-sensing targets due to spatial misalignment. (d) Classification errors caused by semantic misalignment(red dashed circle). A portion of the soccer field was misdetected as a tennis court.
Sensors 23 00207 g001
Figure 2. The overall architecture of our OA-Det for oriented object detection consists of a backbone network, a feature alignment network, and an anchor-free head network. The ResNet-based feature pyramid network extracts common features from an RGB aerial image. The feature alignment network sequentially aligns pyramid features from three dimensions: (a) an intra-level collaborative alignment network in the spatial dimension, (b) an axis-aware convolution in the axial dimension, and (c) a hierarchical-wise context fusion module in the semantic dimension. Finally, aligned features are sent to the classification head and the regression head to obtain the OBB prediction result.
Figure 2. The overall architecture of our OA-Det for oriented object detection consists of a backbone network, a feature alignment network, and an anchor-free head network. The ResNet-based feature pyramid network extracts common features from an RGB aerial image. The feature alignment network sequentially aligns pyramid features from three dimensions: (a) an intra-level collaborative alignment network in the spatial dimension, (b) an axis-aware convolution in the axial dimension, and (c) a hierarchical-wise context fusion module in the semantic dimension. Finally, aligned features are sent to the classification head and the regression head to obtain the OBB prediction result.
Sensors 23 00207 g002
Figure 3. Structure of intra-level collaborative feature alignment network.
Figure 3. Structure of intra-level collaborative feature alignment network.
Sensors 23 00207 g003
Figure 4. (a) The green box is the ground truth, and the green point is its center point. The red point is the sampling position guided by the anchor on the pyramid features from P3 to P6 based on the green point. It can be seen that most of the sample points are offset to the background under the guidance of the anchor box. (b) Sampling axis representation under three sampling methods. The regular sampling method cannot align the axis of remote-sensing targets, and the anchor-aware sampling axis is represented by a unique angle. Our axially aware sampling can represent an axis in four ways.
Figure 4. (a) The green box is the ground truth, and the green point is its center point. The red point is the sampling position guided by the anchor on the pyramid features from P3 to P6 based on the green point. It can be seen that most of the sample points are offset to the background under the guidance of the anchor box. (b) Sampling axis representation under three sampling methods. The regular sampling method cannot align the axis of remote-sensing targets, and the anchor-aware sampling axis is represented by a unique angle. Our axially aware sampling can represent an axis in four ways.
Sensors 23 00207 g004
Figure 5. (a) Correlation matrix of different layers. The blue in the lower left corner and the red in the upper right corner are the correlations of the backbone features and pyramid features, respectively. (b) Pyramid-level distribution of different categories predicted by baseline. The colors of the columns represent different pyramid levels, and the categories include baseball diamond (BD), bridge (BR), tennis count (TC), soccer field (SBF), and harbors (HA).
Figure 5. (a) Correlation matrix of different layers. The blue in the lower left corner and the red in the upper right corner are the correlations of the backbone features and pyramid features, respectively. (b) Pyramid-level distribution of different categories predicted by baseline. The colors of the columns represent different pyramid levels, and the categories include baseball diamond (BD), bridge (BR), tennis count (TC), soccer field (SBF), and harbors (HA).
Sensors 23 00207 g005
Figure 6. The structure of hierarchical-wise context fusion module.
Figure 6. The structure of hierarchical-wise context fusion module.
Sensors 23 00207 g006
Figure 7. Visualization of different feature alignment methods.
Figure 7. Visualization of different feature alignment methods.
Sensors 23 00207 g007
Figure 8. Visualization of predictions on the DOTA dataset using our method OA-Det.
Figure 8. Visualization of predictions on the DOTA dataset using our method OA-Det.
Sensors 23 00207 g008
Figure 9. Visualization of predictions on the HRSC2016 dataset using our OA-Det method.
Figure 9. Visualization of predictions on the HRSC2016 dataset using our OA-Det method.
Sensors 23 00207 g009
Figure 10. Visualization of predictions on the UCAS-AOD dataset using our OA-Det.
Figure 10. Visualization of predictions on the UCAS-AOD dataset using our OA-Det.
Sensors 23 00207 g010
Table 1. Evaluation of different components: ‘✓’ indicates that the component is added on the basis of the baseline, and ‘✕’ indicates that the component is not enabled.
Table 1. Evaluation of different components: ‘✓’ indicates that the component is added on the basis of the baseline, and ‘✕’ indicates that the component is not enabled.
BaselineDifferent Settings of OA-Det
+ Spatial Alignment
+ Axial Sampling
+ Semantic Alignment
mAP(%)65.2567.6867.8968.2869.28
FPS23.821.615.019.714.3
Table 2. Multi-scale hierarchical selection strategy within different pyramid features.
Table 2. Multi-scale hierarchical selection strategy within different pyramid features.
The Strategy of Alignment LayerMax Resize FactorNumber of LayersmAP(%)FPS
baseline--65.2523.8
all level8x566.1923.2
lower level2x267.4721.9
neighbor level2x367.5520.9
neighbor level with P22x367.6820.5
Table 3. Performance comparison of different feature alignment methods.
Table 3. Performance comparison of different feature alignment methods.
Alignment MethodsAdditional Sub HeadFeature AggregationmAP(%)
Regular Conv.-66.67
Grid Sample3 × 3 conv67.00
Deformable Conv.3 × 3 conv67.34
Alignment Conv.stack of conv67.48
Axially aware Conv.stack of conv67.69
Axially aware Conv.3 × 3 conv67.56
Axially aware Conv.spatially aware block67.89
Table 4. Ablation experiment of average orientation error of mainstream feature alignment methods on DOTA1.0 dataset. All results are tested on the validation set.
Table 4. Ablation experiment of average orientation error of mainstream feature alignment methods on DOTA1.0 dataset. All results are tested on the validation set.
MethodsPLBDBRGTFSVLVSHTCBCSTSBFRAHASPHCmAOE
Regular Conv19.9014.966.202.826.454.906.062.786.856.204.189.075.3710.7220.348.45
Grid Sample17.5417.085.257.815.664.606.002.184.974.932.9910.545.688.1715.487.93
DeformConv16.8318.234.025.445.584.015.471.576.815.283.677.376.517.5718.307.78
AlignConv14.6816.833.622.484.743.004.633.023.965.865.927.884.428.2722.037.42
Axially aware Conv16.8812.954.672.335.612.984.783.934.255.903.978.475.026.8513.086.78
Table 5. Effectiveness of shared weights.
Table 5. Effectiveness of shared weights.
OffsetMaskAnglemAP(%)
SeparateSharedSeparateSharedSeparateShared
66.25
67.48
66.35
66.41
intra-level67.79
cross-level67.89
Table 6. Effectiveness of internal components of semantic alignment network.
Table 6. Effectiveness of internal components of semantic alignment network.
Global ContextChannel-WiseHierarchical-WisemAP(%)Param(M)
avg poolingsingle conv65.1342.53
single conv67.5342.53
67.5742.40
68.2844.78
Table 7. Comparisons with other state-of-the-art methods on DOTA dataset. The data in each column indicate the average precision of each category in percent. The last column indicates the mean-Average-Precision of the method: ‘†’ represents the results for multi-scale training and testing. The results with red color denote the best results in each column.
Table 7. Comparisons with other state-of-the-art methods on DOTA dataset. The data in each column indicate the average precision of each category in percent. The last column indicates the mean-Average-Precision of the method: ‘†’ represents the results for multi-scale training and testing. The results with red color denote the best results in each column.
MethodsBackbonePLBDBRGTFSVLVSHTCBCSTSBFRAHASPHCmAP
Two-stage
FR-O [48]R-10179.4277.1317.7064.0535.3038.0237.1689.4169.6459.2850.3052.9147.8947.4046.3054.13
ICN [53]R-10181.4074.3047.7070.3064.9067.8070.0090.8079.1078.2053.6062.9067.0064.2050.2068.20
RoI-Trans. [54]R-10188.6478.5243.4475.9268.8173.6883.5990.7477.2781.4658.3953.5462.8358.9347.6769.56
SCRDet [27]R-10189.9880.6552.0968.3668.3660.3272.4190.8587.9486.8665.0266.6866.2568.2465.2172.61
FADet [55]R-10190.2179.5845.4976.4173.1868.2779.5690.8383.4084.6853.4065.4274.1769.6964.8673.28
Gliding Vertex [56]R-10189.6485.0052.2677.3473.0173.1486.8290.7479.0286.8159.5570.9172.9470.8657.3275.02
Mask OBB [57]RX-10189.5685.9554.2172.9076.5274.1685.6389.8583.8186.4854.8969.6473.9469.0663.3275.33
CenterMap-O [58]R-10189.8384.4154.6070.2577.6678.3287.1990.6684.8985.2756.4669.2374.1371.5666.0676.03
Single-stage
PIoU [59]DLA-3480.969.724.160.238.364.464.890.977.270.446.537.157.161.964.060.5
A2S-Det [60]R-10189.5977.8946.3756.4775.8674.8386.0790.5881.0983.7150.2160.9465.2969.7750.9370.64
DAL [61]R-10188.6179.6946.2770.3765.8976.1078.5390.8479.9878.4158.7162.0269.2371.3260.6571.78
CFC-Net [28]R-5089.0880.4152.4170.0276.2878.1187.2190.8984.4785.6460.5161.5267.8268.0250.0973.50
R 3 Det [13]R-10188.7683.0950.9167.2776.2380.3986.7290.7884.6883.2461.9861.3566.9170.6353.9473.79
SLA [62]R-5088.3384.6748.7873.3477.4777.8286.5390.7286.9886.4358.8668.2774.1073.0969.3076.36
Anchor-free
CenterNet-O [24]H-10189.0269.7137.6263.4265.2363.7477.2890.5179.2477.9344.8354.6455.9361.1145.7165.04
P-RSDet [63]R-10189.0273.6547.3372.0370.5873.7172.7690.8280.1281.3259.4557.8760.7965.2152.5969.82
O2-DNet [64]H-10489.3182.1447.3361.2171.3274.0378.6290.7682.2381.3660.9360.1758.2166.9861.0371.04
BBAVector [65]R-10188.3579.9650.6962.1878.4378.9887.9490.8583.5884.3554.1360.2465.2264.2855.7072.32
DRN [66]H-10489.7182.3447.2264.1076.2274.4385.8490.5786.1884.8957.6561.9369.3069.6358.4873.23
AOPG [10]R-10189.1482.7451.8769.2877.6582.4288.0890.8986.2685.1360.6066.3074.0567.7658.7775.39
Ours
OA-DetR-10189.5176.3851.4768.3873.3777.0587.8590.6982.3886.2363.0965.0675.6876.2258.9274.82
OA-Det R-10188.8284.0253.4970.4281.4284.2788.6790.7183.9086.3366.7469.7977.7380.6564.6678.11
Table 8. Comparisons with high-quality detection performance on HRSC2016 dataset: mAP 07 indicates the results under VOC2007 mAP metrics.
Table 8. Comparisons with high-quality detection performance on HRSC2016 dataset: mAP 07 indicates the results under VOC2007 mAP metrics.
MethodsBackboneImage SizemAP 07 (%)
R 2 CNN [25]ResNet101800 × 80073.07
Axis Learning [67]ResNet101800 × 80078.15
RRD [25]VGG16800 × 80084.30
RepPoints [25]ResNet101512 × 80085.16
RoI-Transformer [25]ResNet101512 × 80086.20
CenterNet-O [25]DLA-34800 × 80087.89
Gliding Vertex [56]ResNet101512 × 80088.20
CSL [68]ResNet101800 × 80089.62
OA-Det(Ours)ResNet101800 × 80090.10
Table 9. Comparisons with high-quality detection performance on UCAS-AOD dataset.
Table 9. Comparisons with high-quality detection performance on UCAS-AOD dataset.
MethodsBackboneCarPlanemAP(%)
Yolov3 [69]Darknet5374.6389.5282.08
RepPoints [25]ResNet10183.0289.3486.18
Faster-RCNN [4]ResNet5086.8789.8688.36
RoI-Transformer [54]ResNet10187.9989.9088.95
CFC-Net [28]ResNet10189.2988.6989.49
DAL [61]ResNet10189.2590.4989.87
S 2 A-Net [11]ResNet5089.5690.4289.99
OA-Det(Ours)ResNet10190.0290.5690.29
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hu, X.; Zhu, C. Shared-Weight-Based Multi-Dimensional Feature Alignment Network for Oriented Object Detection in Remote Sensing Imagery. Sensors 2023, 23, 207. https://doi.org/10.3390/s23010207

AMA Style

Hu X, Zhu C. Shared-Weight-Based Multi-Dimensional Feature Alignment Network for Oriented Object Detection in Remote Sensing Imagery. Sensors. 2023; 23(1):207. https://doi.org/10.3390/s23010207

Chicago/Turabian Style

Hu, Xinxin, and Changming Zhu. 2023. "Shared-Weight-Based Multi-Dimensional Feature Alignment Network for Oriented Object Detection in Remote Sensing Imagery" Sensors 23, no. 1: 207. https://doi.org/10.3390/s23010207

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop