Next Article in Journal
Theoretical Analysis and Research on Support Reconstruction Control of Magnetic Bearing with Redundant Structure
Previous Article in Journal
DepressionMIGNN: A Multiple-Instance Learning-Based Depression Detection Model with Graph Neural Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

MCFA: Multi-Scale Cascade and Feature Adaptive Alignment Network for Cross-View Geo-Localization

College of Computer Science, Beijing Information Science and Technology University, Beijing 102206, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(14), 4519; https://doi.org/10.3390/s25144519
Submission received: 4 June 2025 / Revised: 17 July 2025 / Accepted: 18 July 2025 / Published: 21 July 2025
(This article belongs to the Section Remote Sensors)

Abstract

Cross-view geo-localization (CVGL) presents significant challenges due to the drastic variations in perspective and scene layout between unmanned aerial vehicle (UAV) and satellite images. Existing methods have made certain advancements in extracting local features from images. However, they exhibit limitations in modeling the interactions among local features and fall short in aligning cross-view representations accurately. To address these issues, we propose a Multi-Scale Cascade and Feature Adaptive Alignment (MCFA) network, which consists of a Multi-Scale Cascade Module (MSCM) and a Feature Adaptive Alignment Module (FAAM). The MSCM captures the features of the target’s adjacent regions and enhances the model’s robustness by learning key region information through association and fusion. The FAAM, with its dynamically weighted feature alignment module, adaptively adjusts feature differences across different viewpoints, achieving feature alignment between drone and satellite images. Our method achieves state-of-the-art (SOTA) performance on two public datasets, University-1652 and SUES-200. In generalization experiments, our model outperforms existing SOTA methods, with an average improvement of 1.52% in R@1 and 2.09% in AP, demonstrating its effectiveness and strong generalization in cross-view geo-localization tasks.

1. Introduction

Unmanned aerial vehicles (UAVs) are widely applied in urban planning, disaster response, and remote sensing monitoring due to their flexibility and high-resolution imaging capabilities. Traditionally, drone localization relies on the Global Navigation Satellite System (GNSS) [1]. However, in urban environments, obstacles such as high-rise buildings and dense infrastructure often cause signal blockages or multipath interference, severely degrading the accuracy and reliability of GNSS-based localization. To overcome these limitations, cross-view geo-localization (CVGL) has emerged as a promising alternative. CVGL aims to estimate the location of drone images by retrieving corresponding reference images from a geotagged satellite image database despite significant viewpoint and appearance differences [2,3]. As an image retrieval problem, CVGL eliminates the dependence on GNSS and remains robust in complex or GNSS-denied environments.
Recently, deep learning’s rapid development has propelled significant progress in CVGL. Most methods learn image feature representations through neural networks, aiming to enhance discriminative capability for precise matching of positive samples and effective differentiation of negative samples. Some works [4,5] further strengthen the model’s robustness to cross-view variations through a contrastive learning strategy, which draws positive sample pairs closer together while distancing negative pairs, thus boosting the matching accuracy. However, these methods often rely heavily on global image information, neglecting the fine-grained representation of local features, which limits their effectiveness in handling complex scenes. To address this limitation, LPN [6] employs a square-ring feature segmentation strategy to capture contextual information from target-adjacent regions. FRSA [7] segments images based on semantic information and uses implicit semantic local features for image retrieval. TransFG [8] further enhances contextual correlation by performing weighted fusion of global and local features. However, despite progress in local feature learning, these methods still struggle to model relationships among local features and fail to address cross-view feature misalignment caused by differences in target orientation (e.g., buildings viewed from different angles) or imaging platforms (e.g., drones and satellites). In real-world settings, buildings are typically embedded in complex surroundings. Although the buildings themselves may lack distinctive features, their spatial relations with nearby elements provide important information for CVGL. Consequently, capturing the associations between these local features can significantly improve retrieval accuracy. Furthermore, drone imagery is captured from varying orientations and altitudes, whereas satellite imagery is typically acquired from higher altitudes with a near-vertical perspective [9]. This substantial viewpoint difference results in cross-view feature misalignment, making it difficult for cross-view images to learn consistent representations of the same target, ultimately leading to lower image retrieval accuracy. Thus, effectively modeling local relationships and addressing cross-view misalignment are key to improving cross-view geo-localization performance.
Therefore, we propose a Multi-Scale Cascade and Feature Adaptive Alignment (MCFA) network, which addresses the challenges of target representation and viewpoint inconsistency in cross-view geo-localization. Specifically, the Multi-Scale Cascade Module (MSCM) captures informative features from target-adjacent regions and integrates them with global context to enhance feature robustness. The Feature Adaptive Alignment Module (FAAM) is designed to mitigate representation misalignment between drone and satellite images. The FAAM employs a cross-dimensional dynamic weighting strategy to jointly model spatial and channel dependencies. This mechanism adaptively reweights features based on their relevance and effectively aligns cross-view representations into a unified embedding space. As shown in Figure 1, our model achieves superior performance and generalization capability compared to current SOTA methods.
The key contributions of this paper are summarized as follows:
  • The MSCM enhances the model’s ability to capture and represent intricate spatial relationships between target-adjacent regions and global features, thereby improving robustness.
  • The FAAM effectively adapts to feature variations of targets under different orientations and platforms, achieving cross-view feature alignment between drone and satellite images.
  • Performance comparisons on benchmark datasets validate the robustness and effectiveness of our method over existing SOTA solutions. In generalization experiments, our model consistently surpasses previous SOTA approaches, achieving an average improvement of 1.74% in R@1 and 2.47% in AP, further validating its effectiveness and robustness.

2. Related Work

2.1. Cross-View Geo-Localization

Traditional works [10,11] rely on handcrafted descriptors and matching strategies, which require extensive parameter tuning and exhibiting poor generalization. Nowadays, deep learning is the mainstream approach for CVGL. Specifically, convolutional neural networks (CNNs) are employed for cross-view representation learning, demonstrating the capacity to capture semantically meaningful patterns and surpass traditional approaches [12]. Building on this, a pre-trained network was adapted through optimization of the representation discrepancy across image pairs [13], further enhancing matching accuracy. To learn discriminative feature representations, a Siamese network-based approach [14] was introduced, optimizing network parameters via contrastive loss. Additionally, the soft-margin triplet loss [15] was introduced, which encourages the reference sample to be closer to positive samples and farther from negative ones. A Siamese network incorporating NetVLAD and a weighted soft-margin ranking loss [16] was proposed to enhance training efficiency and matching accuracy. The retrieval problem was reformulated as a classification task [17] based on ResNet-50 [18]. More recently, to address the variability in drone imagery, diverse visual styles were transformed into a unified satellite image style coupled with a dynamic observation module designed to reduce surrounding noise and geographic distortions [19]. Sample4Geo [5] combined ConvNeXt [20] with contrastive learning [21,22,23] and symmetric InfoNCE loss [24], streamlining the training pipeline and enhancing generalization. MCCG [4] was based on the ConvNeXt network, capturing more discriminative information across different dimensions to obtain diverse feature representations. An Expectation-Maximization (EM)-based pseudo-label generation strategy and an information consistency module [25] were proposed to address issues such as label dependency, data association, and viewpoint differences.

2.2. Part-Based Feature Representation

Part-based feature representation has gained widespread application in deep learning. For example, in person re-identification (Re-ID) [26,27,28], extracting features from different body parts enhances recognition accuracy. Part-based Convolutional Baseline [29] segments pedestrian images into multiple parts, extracts features separately, and then combines them for matching. Pose-guided Part Attention [30] effectively mitigates pose variations, background clutter, and occlusions, leading to notable performance gains.
In cross-view geo-localization, LPN [6] captures contextual information from neighboring regions using a square-ring partitioning strategy, effectively emphasizing the context surrounding the target near the image center. Zhuang et al. [31] enhanced LPN by integrating two global blocks, aiming to mitigate the effects of scale changes and positional displacements. Chen et al. [32] further refined image partitioning by dividing images into finer-grained regions to address scenarios where targets are not located at the image center. Ge et al. [33] introduced a multi-branch joint representation learning network to capture both holistic image information and contextual relationships in adjacent regions. Ge et al. [34] further refined the approach by segmenting images into part-level and patch-level features as feedback to global features, introducing an adaptive region-erasing strategy to remove irrelevant information from global representations. Sun et al. [35] employed a self-supervised feature enhancement method to obtain building-aware masks, enabling the model to learn discriminative architectural regions and guide feature extraction. Wang et al. [36] proposed a decorrelation regularization approach with dynamically adjusted weights to suppress redundancy in spatial embeddings by minimizing correlations between feature channels. Zhao et al. [8] exploited the gradient characteristics of local features to strengthen cross-view feature representations and facilitate precise instance-level alignment under varying viewpoints. Liu et al. [37] proposed a two-stage learning framework, where the first stage focuses on extracting part representations and the second stage integrates these part representations into the global embedding to prevent all local parts from contributing equally to similarity measurement. Liu et al. [38] introduced graph convolutional networks to learn the semantic distribution and structural information of objects. Wu et al. [39] incorporated contrastive attribute mining and a location-aware partitioning strategy to align geographic features across different viewpoints and scales.
Despite their effectiveness, these methods often neglect the interactions among local features, which are crucial for robust cross-view representation learning. This paper proposes a part-based feature representation CVGL method called MCFA that captures complex relationships among local features and addresses feature misalignment caused by differences in orientations and platforms. MCFA consists of a Multi-Scale Cascaded Module (MSCM) and a Feature Adaptive Alignment (FAAM) module, where the MSCM integrates local and global features to enhance contextual understanding and the FAAM performs fine-grained alignment to achieve more precise cross-view feature correspondence.

3. Proposed Method

3.1. Problem Formulation

Consider a CVGL dataset containing K distinct geographic regions. Let the query image be denoted as x j k and the reference image as y j k , where k { 1 , 2 , , K } represents the geographic region label, and j { d , s } indicates the image perspective—drone view (d) or satellite view (s). The goal is to learn a feature embedding that aligns images from different viewpoints into a common space, ensuring that images from the same region are mapped close together while maintaining clear separation between those from different regions.

3.2. Overall Architecture

The proposed architecture comprises two parallel branches corresponding to the drone and satellite views. As illustrated in Figure 2, the framework is structured into three functional components. The first is the feature extraction module, which utilizes a pair of backbone networks with shared weights to generate feature representations from both views. The second component, the MSCM, is responsible for aggregating spatial features across multiple scales. The third component, the FAAM, refines the extracted features and aligns them to enhance cross-view correspondence.

3.3. Feature Extraction Module

The model employs ConvNeXt-Tiny [20] with shared weights to extract fine-grained features. ConvNeXt is a purely convolutional neural network that improves upon ResNet by incorporating architectural elements inspired by the Swin Transformer [40]. ConvNeXt-Tiny enlarges the convolution kernel to 7 × 7 and modifies the block stacking pattern from (3, 4, 6, 3) in ResNet to (3, 3, 9, 3). The output feature map of an input image processed by ConvNeXt-Tiny is shown as follows:
L j = F conv ( x j ) , j = d , s .
Here, x j refers to images captured from either the drone viewpoint ( j = d ) or the satellite viewpoint ( j = s ). The convolutional network F conv extracts feature representations L j R H × W × C , where the dimensions H, W, and C indicate the spatial height, the spatial width, and the number of channels in the extracted feature map, respectively.
We adopt a symmetric InfoNCE loss to improve the discriminative capability of features. Specifically, compact feature vectors are generated by applying global average pooling to the feature maps L j extracted by ConvNeXt-Tiny, serving as inputs to the loss function. This loss pulls positive pairs closer while pushing negative pairs farther apart in the feature space, enhancing feature discriminability. It is defined as
L InfoNCE ( x j k , y j k ) = log exp x j k · y j k / τ i = 1 R exp x j k · y i / τ
Here, x j k and y j k are the pooled feature vectors of a query image and its corresponding positive reference image. The set { y i } denotes reference features, including both positive and negative samples. When x j k is similar to y j k , the loss value is low; otherwise, it is high. The parameter τ is a hyperparameter that dynamically adjusts during training.

3.4. MSCM

The proposed MSCM consists of a partitioning strategy and a Hierarchical Aggregation Module (HAM). Our partitioning strategy captures richer semantic features, while the HAM integrates global and local features, significantly enhancing feature representation.
Partitioning strategy: As shown in Figure 3, the original LPN [6] enhances feature discrimination but suffers from limited semantic context due to its small central region. To address this, we expand the central area to capture richer semantics and propose an improved ring-based partitioning strategy (as shown in Figure 4) that effectively captures multi-scale local information of the target and its surrounding regions, accommodating position shifts and scale variations. Specifically, the global feature L j ( j { s , d } ) is partitioned into N local features L j n ( n = 1 , 2 , , N ), which are subsequently aggregated through average pooling:
L p j n = Pooling ( L j n )
Here, Pooling denotes average pooling, and L p j n represents the pooled feature.
Hierarchical Aggregation Module. Local features provide fine-grained information, while global features represent semantic context. To fully exploit spatial information, we propose a dynamic feature fusion method that effectively integrates global features into the fusion of local features. This method adaptively adjusts the weighting between local and global features based on task requirements, enhancing the precision of local details while maintaining the coherence of global context. The fusion process significantly enhances the model’s ability to represent features and its robustness. The fusion is defined as
L f j n = MLP ( Pooling ( L j ) ) L p j n
Here, the MLP block comprises a linear transformation, SiLU activation, a subsequent linear layer, and normalization. The symbol ⊙ represents element-wise multiplication. The feature L f j n R H × W × C , where its dimensions are the same as L j .
To fully explore the latent relationships between local features, we introduce a gated feature fusion method. This approach assigns adaptive weights to each local feature, enabling the model to focus on discriminative regions during the fusion process while suppressing or ignoring irrelevant areas. Specifically, each local feature is element-wise multiplied with its corresponding weight matrix, followed by linear weighting to integrate information from various local regions. This process ultimately generates a globally optimized feature representation. The formulation is as follows:
L + j = n = 1 N G n L f j n
where L + j R H × W × C is the globally optimized feature representation with the same dimensions as L f j n , G n is the weight matrix corresponding to the n-th local feature, and ⊙ represents element-wise multiplication.

3.5. FAAM

The proposed FAAM consists of a Cross-Dimensional Interaction Mechanism (CDIM) and Fine-Grained Alignment (FGA). The CDIM adapts to feature variations across different orientations, while FGA ensures fine-grained feature alignment.
Cross-Dimensional Interaction Mechanism. Several studies [4,41,42,43] demonstrate that incorporating attention mechanisms in both the channel and spatial dimensions can effectively enhance the model’s ability to capture critical features. Inspired by [4], we retain the first two branches of the triplet attention module. Within each branch, feature maps are transformed across spatial and channel dimensions to produce both corresponding feature maps and separate feature vectors. While both branches aim to enhance feature representation, they focus on different orientations: the first branch emphasizes the interaction between height and channels, while the second branch highlights the relationship between width and channels. However, existing methods typically assign fixed weights to these two branches, failing to account for the diverse directional arrangements that targets may exhibit under varying viewpoints or imaging conditions. Since the primary orientation of targets can be distributed at arbitrary angles (as illustrated in Figure 5), a fixed-weight strategy may limit the model’s adaptability when encountering targets with different orientations. To address this limitation, we introduce a dynamic weighting mechanism that adaptively fine-tunes the branch weights guided by the distribution of features. When the principal orientation of the target changes, this mechanism optimizes the weight allocation between the two branches, allowing the model to effectively capture critical features and adapt to targets arranged in different orientations. In this way, the dynamic weighting mechanism enhances the model’s feature representation capability, improving both the accuracy and robustness of cross-view matching. The entire procedure can be formulated as
[ D j 1 , D j 2 ] = F cross - dimension ( L + j )
M j 1 = α D j 1 · exp ( D j 1 ) exp ( D j 1 ) + exp ( D j 2 )
M j 2 = β D j 2 · exp ( D j 2 ) exp ( D j 1 ) + exp ( D j 2 )
Here, F cross - dimension ( · ) denotes the enhanced triplet attention mechanism. The tensors D j 1 , D j 2 , M j 1 , M j 2 R H × W × C represent intermediate feature maps, where D j 1 and D j 2 correspond to the outputs obtained by modeling cross-dimensional interactions between the spatial dimensions ( H , W ) and the channel dimension (C), and M j 1 and M j 2 denote the corresponding weighted features. The parameters α and β are learnable scalars used to dynamically balance the contributions of different attention branches.
Fine-Grained Alignment. To align cross-view image features, M j 1 , M j 2 are summed to generate the cross-dimension interaction feature M j . This feature is then passed through an MLP to extract alignment features f j ( j = { s , d } ), with dimensions h × w × c f , where c f = 256 . Unlike previous works [4,6], we replaced the ReLU [44] activation function with SiLU [45], whose smooth properties better handle large variations in viewpoints and scale. Next, M j and f j are concatenated in the channel dimension to form the final fused feature F j , with dimensions h × w × ( c f + C ) .
For improved cross-view feature alignment, we employ the cosine similarity loss function L Cosine as an optimization objective. This loss function evaluates the alignment of feature representations to support effective cross-view matching. It is formally defined as
L Cosine ( F d ,   F s ) = 1 N i = 1 N 1 F d i · F s i F d i 2 F s i 2
where N denotes the number of feature vectors per batch, and F d i and F s i represent the aligned feature vectors extracted from drone and satellite images, respectively. The operator · denotes the dot product, while · 2 represents the L2 norm. By minimizing the cosine distance between corresponding feature vectors, this loss function effectively enhances feature alignment across different viewpoints, improving the model’s ability to establish robust and discriminative cross-view correspondences.
Building upon previous studies [4,39], we also applied global average pooling to M j 1 , M j 2 and L j + to integrate global information. The pooled features were then encoded using a classifier module to enhance their representation. The encoded features were supervised and optimized using a cross-entropy (CE) loss, thereby improving the model’s effectiveness on classification tasks. The loss function is formulated as
L Classification ( F , y ) = log exp ( cls ( F , y ) ) i = 1 N exp ( cls ( F , y i ) )
Here, the L Classification represents the multi-class CE loss function, F is the feature representation generated by the classifier module, and y ^ is the ground-truth class label for the current input.

3.6. Loss Function

Our loss function consists of three components and is formulated as
loss = L InfoNCE + λ 1 L Classification + λ 2 L Cosine
Here, L InfoNCE is designed to handle viewpoint differences by optimizing feature representations through contrastive learning. L Classification supervises the optimization of encoded features via a CE loss, boosting the model’s accuracy on classification tasks. L Cosine focuses on distinguishing hard negative samples by dynamically mining highly similar negative pairs, which facilitates the acquisition of more discriminative feature representations. λ 1 and λ 2 are adjustable coefficients that balance the influence of individual terms.
This loss function effectively balances cross-view feature alignment, hard negative sample differentiation, and classification performance optimization, enabling robust feature alignment and precise localization under diverse scenarios and conditions.

4. Experiments

Our experiments comprise two tasks: In the drone localization task, the drone provides the query images, and its location is determined by matching the most similar satellite image. In the drone navigation task, the satellite image is used as a query to guide the drone more accurately to the target region by retrieving the most relevant drone-view images.

4.1. Datasets and Evaluation Metrics

We conducted our experiments using two authoritative public datasets: University-1652 and SUES-200. Evaluation metrics are detailed in the following section.
University-1652: The dataset [2] comprises 1652 distinct geographic landmarks collected from 72 different universities. Each landmark contains satellite-, drone-, and street-view images, but our research focuses only on satellite and drone views. The training and test sets include 701 and 951 landmarks from 33 and 39 universities, respectively. Data acquisition leveraged Google Earth, with drone images taken at varying altitudes between 121.5 m and 256 m, providing a rich variety of multi-scale and multi-view aerial images.
SUES-200: The dataset [3] comprises drone and satellite images from 200 landmarks around Shanghai University of Engineering Science, with 120 used for training and 80 for testing. Each location includes 50 drone images captured at altitudes of 150 m, 200 m, 250 m, and 300 m, and 1 satellite image from AutoNavi Map and Bing Maps.
Evaluation metrics: To evaluate retrieval performance, we employ Recall@K (R@K) and Average Precision (AP). Recall@K indicates the percentage of queries with at least one correct match among the top-K retrievals. AP measures the overall ranking quality across recall levels. These indicators comprehensively reflect both accuracy and robustness.

4.2. Implementation Details

We adopted ConvNeXt-Tiny as the backbone network, initializing weights using the Kaiming initialization method [46]. Images were resized to 384 × 384 [5] and augmented via flipping, color jittering, sharpening, rotation, and occlusion.
The AdamW optimizer [47] was used during training with a batch size of 32, consisting of 16 drone and 16 satellite images. The initial learning rate was set to 0.001 , with a warm-up period covering the first 10% of total training steps. The loss weighting factors λ 1 and λ 2 were assigned values of 0.1 and 0.4, respectively. All experiments were conducted on two NVIDIA RTX 3090 GPUs under PyTorch 2.5.0 on Ubuntu 22.04.

4.3. Comparison with SOTA Methods

4.3.1. Results on University-1652

As presented in Table 1, the proposed MCFA method achieved 94.53% R@1 and 95.40% AP on the Drone→Satellite task. For the Satellite→Drone task, the model achieved 96.43% R@1 and 93.95% AP. Importantly, our approach surpasses existing SOTA techniques. Compared with CAMP, MCFA improves R@1 by 0.07% and AP by 0.02% on the Drone→Satellite task and improves R@1 by 0.28% and AP by 1.23% on the Satellite→Drone task.

4.3.2. Results on SUES-200

Table 2 shows the results of our MCFA model across four different altitudes in the SUES200 dataset. The experimental results indicate that our model achieves superior performance compared to SOTA methods on the majority of evaluation metrics. Compared to CAMP, our method achieves improvements in R@1 and AP on the Drone→Satellite task at an altitude of 150 m, with R@1 and AP improving by 0.6% and 0.51%, respectively. At an altitude of 250 m, AP improves by 0.22%. In the Satellite→Drone task, our method improves R@1 and AP at 150 m by 1.25% and 1.07%, respectively. At 200 m, AP improves by 1.25%, and at 250 m, AP improves by 0.06%. Overall, MCFA demonstrates excellent performance and robust stability across various altitudes. On R@1 and AP, MCFA consistently matches or surpasses SOTA methods in multiple settings, highlighting its strong capability to handle complex spatial structures and viewpoint variations.

4.4. Generalization Experiments

Generalization means a model trained on one region’s data can still handle data from completely different regions. To assess the cross-domain transfer capability of our model, we trained the model on the University-1652 dataset and tested it on the SUES-200 test set. Additionally, we compared our model with current SOTA methods (MCCG, Sample4Geo, and CAMP) under the same experimental configurations. The following table summarizes the results of the generalization experiments, clearly demonstrating the excellent generalization capability of our model in the CVGL task.
Table 3 presents the evaluation results of models trained on University-1652 and tested on SUES-200. Compared to the current SOTA method CAMP, our method achieves an average improvement of 1.73% in R@1 and 2.47% in AP on the SUES-200 test set. Notably, in the 150 m height scenario, where the query and reference images exhibit significant viewpoint differences, our model outperforms CAMP by 5.75% in R@1 and 4.81% in AP for the Drone→Satellite task and by 3.75% and 7.05%, respectively, for the Satellite→Drone task. These results highlight the effectiveness of our method in handling large viewpoint variations.
However, at the 300 m altitude, our model demonstrates a slower performance improvement and slightly underperforms compared to the SOTA method. We attribute this observation to two primary factors. First, the training data employed for MCFA originates from the University-1652 dataset, which encompasses drone imagery captured at altitudes only up to 256 m. Consequently, the model was not exposed to samples from the 300 m viewpoint during training, limiting its capability to improve performance at this elevation. Second, as the acquisition altitude increases, the spatial distribution of ground-level objects such as buildings is prone to displacement toward image peripheries. This introduces heightened challenges for accurate region-level feature alignment. Despite MCFA’s incorporation of multi-scale feature fusion and adaptive alignment mechanisms designed to address such difficulties, the model’s effectiveness may still be compromised under such extreme viewpoint variations, thereby leading to a slower rate of performance enhancement.

4.5. Ablation Studies

A set of ablation studies was performed to further confirm the effectiveness of our proposed approach.

4.5.1. Effect of the Multi-Scale Cascaded Module

The MSCM enhances the model’s capacity to capture and represent spatial features by integrating contextual information related to the target. Table 4 shows that the incorporation of this module results in a significant enhancement in model performance. In the Drone→Satellite task, R@1 increased by 1.35% and AP increased by 1.18%. In the Satellite→Drone task, R@1 improved by 1.43% and AP improved by 1.28%. These results demonstrate the superiority of the MSCM in fully mining the semantic and spatial information associations.

4.5.2. Effect of the Feature Adaptive Alignment Module

The FAAM aligns and optimizes cross-view features by dynamically reweighting their significance across channels and projecting them into a unified spatial domain. Table 4 shows that after adding the Feature Adaptive alignment Module, the model achieved a 1.19% improvement in R@1 and a 0.96% improvement in AP for the Drone→Satellite task. In the Satellite→Drone task, R@1 improved by 0.85% and AP increased by 1.32%. The notable enhancement in performance demonstrates the efficiency and advantage of the FAAM in addressing the issue of cross-view feature misalignment.

4.5.3. Effect of the CDIM

To validate the effectiveness of our proposed CDIM, we conduct ablation experiments by comparing it with two widely adopted attention modules: SE [48] and CBAM [42].
As shown in Table 5, the CDIM consistently outperforms both SE and CBAM across the Drone→Satellite and Satellite→Drone tasks on the University-1652 dataset. Specifically, the CDIM achieves an R@1 of 94.53% and AP of 95.40% in the Drone→Satellite task, surpassing SE by 0.95% and 0.84% and outperforming CBAM by 1.84% and 1.59%, respectively. In the Satellite→Drone task, CDIM achieves an R@1 of 96.43% and AP of 93.95%, improving upon SE by 1.14% and 1.11% and upon CBAM by 1.42% and 1.84%.
The superior performance of our CDIM stems from its focus on modeling interactions across channels and its adaptive adjustment of attention weights along both horizontal and vertical directions based on the spatial distribution of targets. This design enables the module to effectively handle variations of objects under different orientations and learn more robust invariant features.

4.5.4. Effect of SiLU Activation in FGA

In the FGA, we adopt the SiLU activation function [45] in place of the commonly used ReLU [44]. To verify the impact of this design choice, we conducted ablation experiments comparing ReLU and SiLU within the FGA module. As shown in Table 6, the model with SiLU achieves superior performance on both CVGL tasks. In the Drone→Satellite direction, the model equipped with SiLU achieves a R@1 of 94.53% and AP of 95.40%, compared to 93.94% and 94.95% with ReLU. In the Satellite→Drone task, SiLU again shows improvements, achieving a R@1 of 96.43% and AP of 93.95%, outperforming ReLU’s 96.14% and 93.45%.
These results confirm that SiLU’s smooth activation curve helps stabilize feature propagation and enhances the discriminability of learned representations. By mitigating the gradient sparsity problem commonly encountered with ReLU, SiLU improves the model’s ability to adapt to continuous spatial transformations and appearance shifts across views.

4.5.5. Backbone Selection and Efficiency Analysis

To comprehensively assess the trade-off between computational efficiency and retrieval performance, we evaluate two variants of our MCFA framework with different backbone architectures (ConvNeXt and ViT) and compare them against two widely adopted baseline methods, LPN [6] and Sample4Geo [5]. As summarized in Table 7, we report inference latency per image, model parameter count, and retrieval metrics on both Drone→Satellite and Satellite→Drone tasks. This comparison allows us to analyze the backbone compatibility within MCFA and benchmark its efficiency and accuracy against established baseline methods.
Our MCFA with a ConvNeXt backbone achieves the best overall retrieval performance among all compared methods, reaching 94.53% R@1 and 95.40% AP on the Drone→Satellite task and 96.43% R@1 and 93.95% AP on the Satellite→Drone task. Compared to Sample4Geo, our model not only improves retrieval accuracy but also achieves comparable runtime (3.75 ms vs. 3.64 ms) despite having a slightly larger parameter count (97.57 M vs. 87.57 M). This favorable accuracy–efficiency trade-off is largely attributable to our inference pipeline, which deploys the MSCM and FAAM only during training while relying solely on the backbone during inference.
MCFA with a ViT backbone can still achieve competitive retrieval performance compared to several mainstream methods, demonstrating the strong generalization capability of our framework across different backbone architectures. However, MCFA with a ViT backbone consumes significantly more computation time (8.80 ms) due to the inherently high computational cost of the ViT architecture. Meanwhile, both LPN and Sample4Geo have been widely deployed in practical cross-view geo-localization systems, serving as reliable baselines. Our method outperforms LPN in all key metrics with faster inference (3.75 ms vs. 4.68 ms) and surpasses Sample4Geo in both retrieval accuracy and generalization while maintaining comparable runtime, making it a strong candidate for real-world applications that require both precision and efficiency.

4.5.6. Effect of Image Offset in Test Data

In practical applications, there is often a position discrepancy between the query image and the true matching image in the gallery. To evaluate the robustness of our model to positional deviations, we designed a mirror padding (Flip-Pad) experiment. Specifically, as illustrated in Figure 6, the mirror padding operation takes a region of width P from one side of the image, flips it, and attaches it back to the same side’s edge, at the same time eliminating a region of the same width from the opposite side.
Table 8 provides a comparative analysis of models designed for local region extraction, including SDPL, FRSA, LPN, and our proposed model, under different offset values ( P = 20 , 40 , 60 , 80 , 100 ), where the left edge of the image was padded by mirroring. The corresponding outcomes are illustrated in Figure 7 and Figure 8. With the increase in P, our model exhibits a noticeably slower decline in performance relative to other models and consistently maintains superior results.

4.5.7. Impact of Center Region Size on Different Cutting Methods

We investigated the effect of center region ( d 0 in Figure 4) size on the model’s performance under various partitioning strategies. We partitioned the 12 × 12 × 1024 features output by ConvNeXt-Tiny, with the center region size varying from 3 × 3 × 1024 to 7 × 7 × 1024 . As shown in Table 9, when the number of partitions is four, gradually increasing the width and height of the center region from 3 × 3 (as used in LPN) to 6 × 6 (half of the original image) leads to performance improvements in both two tasks. However, when the center region exceeds half of the original image size ( 6 × 6 ), performance declines significantly. We attribute this to the fact that a moderately sized center region allows the model to capture more complete target information, whereas an excessively large center region introduces much irrelevant information, leading to performance degradation.
To further explore this phenomenon, we increased the number of partitions to 6. As presented in Table 10, the configuration with a center region size of 4 × 4 (one-third of the original image) yields the best performance. Further increasing or decreasing the center region size results in performance drops, with a notable decline when the size exceeds half of the original image. This further confirms the significant impact of center region size on model performance.

4.5.8. Selection of the Number of Segments

The partitioning strategy of local regions directly influences the receptive field of each part, and different region selection strategies significantly impact the model’s feature learning performance. The number of local regions n is a key factor in determining model performance. Table 11 presents the experimental results of the annular segmentation method using our partitioning strategy with different values of n on the University-1652 dataset. The results indicate that our proposed region selection method achieves optimal performance when n = 6. If n is too small, the model may fail to capture sufficient local features, leading to inadequate feature representation. Conversely, if n is too large, excessive noise may be introduced, reducing generalization ability. These findings demonstrate that a well-designed region partitioning strategy can effectively enhance both feature extraction capability and model robustness.

4.5.9. Comparison of Different Loss Functions on Fine-Grained Alignment

To investigate the impact of various loss functions on Fine-Grained Alignment (FGA), we conducted experiments on the University-1652 dataset. As shown in Table 12, the superiority of cosine similarity loss lies in its direct optimization of the angular relationship between feature vectors rather than their absolute magnitudes, making it inherently scale-invariant and more effective in aligning features across different viewpoints. In contrast, MSE and triplet loss rely on Euclidean distance, which is sensitive to variations in feature magnitudes. In cross-view scenarios, perspective distortion and scale differences often lead to unstable feature magnitudes, ultimately degrading retrieval performance. InfoNCE loss tends to blur the semantic boundary between positive and negative samples for the complex features produced by FGA, increasing the risk of misclassifying hard positives as negatives and thereby weakening the model’s discriminative power and overall matching performance.

4.6. Model Visualization

Figure 9 displays the heatmaps obtained from our model and the comparison model CAMP from both drone and satellite perspectives. Compared to CAMP, our model accurately captures the key regions of the target as well as the surrounding meaningful contextual areas. This further demonstrates that during feature extraction, our approach enhances the model’s sensitivity to target-specific information, significantly improving the precision and robustness of cross-view geo-localization.
To validate the effectiveness of our model in drone localization and navigation tasks, we randomly selected three images from the query_drone and query_satellite subsets of the University-1652 test set for testing. For each query image, we retrieved the top-five matching images. Figure 10 presents the retrieval results, showing that our model exhibits excellent retrieval accuracy in drone localization and navigation tasks.

5. Limitation

5.1. Feature Separability and Failure Case Analysis

Figure 11a presents the t-SNE [50] visualization results of our model. As shown in Figure 11b, the features are well clustered, indicating successful matching. In contrast, Figure 11c illustrates a failure case, where the drone image is mismatched due to the high visual similarity between two different scenes. The model struggles to distinguish subtle differences, resulting in incorrect associations.

5.2. Robustness Analysis Under Illumination Variations

To further analyze the robustness of our model under varying illumination conditions, we visualize the attention heatmaps generated from three types of input: original images, processed high-illumination images, and processed low-illumination images. As shown in Figure 12, the model performs well under normal lighting, accurately focusing on discriminative regions such as roads and rooftops. However, under strong illumination, overexposure leads to the disappearance of key structures—roads become barely visible, and rooftops are difficult to distinguish from the surrounding environment. In low-light scenarios, buildings tend to merge with the background, resulting in suppressed attention responses and reduced feature separability. These observations suggest that extreme lighting conditions significantly degrade the model’s ability to extract informative spatial features.

5.3. Analysis at 300 m Altitude in Generalization Setting

As shown in Table 3, our MCFA model achieves superior generalization performance across most altitudes in both Drone→Satellite and Satellite→Drone directions when trained on University-1652 and evaluated on SUES-200. However, at the 300-meter altitude, CAMP slightly outperforms our model. To investigate this behavior, we visualize representative samples at 300 m in Figure 13. It can be observed that buildings exhibit noticeable spatial shifts, often appearing near the image periphery rather than near the center. This spatial displacement poses a significant challenge for region-level alignment. Although MCFA is equipped with multi-scale feature fusion and adaptive alignment modules, such extreme viewpoint variations can still restrict its ability to accurately localize and match target regions. As a result, MCFA shows limited performance improvement at 300 m compared to its strong gains at lower altitudes. Furthermore, as shown in Figure 13, our model tends to focus more on road-related structures under high-altitude conditions, whereas CAMP, benefiting from its dynamic positional encoding, adapts more effectively to building displacement. This results in CAMP achieving a slight advantage under this specific extreme setting. Nonetheless, our model remains competitive across most conditions and performs robustly in general.

6. Conclusions

To deeply explore the contextual information of the target and effectively address the issue of cross-view image feature misalignment, this paper proposes the MCFA network. This model integrates an MSCM, which captures rich contextual information by learning deep connections between neighboring regions of the target, and an FAAM, which adaptively adjusts feature importance to enhance fine-grained feature representation and achieve accurate cross-view alignment. In terms of experimental evaluation, the proposed model demonstrates leading performance on standard datasets such as University-1652 and SUES-200, significantly surpassing existing baselines in R@1 and AP metrics and outperforming SOTA methods in generalization tests, fully validating the effectiveness and superiority of our model in cross-view geo-localization tasks. We plan to further optimize the model, particularly in the fine-grained feature mining of target neighboring regions, to enhance the model’s robustness in complex environments. We hope to advance the development of cross-view geo-localization through these efforts, especially in challenging scenarios such as illumination changes and large-scale variations.

Author Contributions

Conceptualization, K.H. and Q.T.; methodology, K.H. and Q.T.; software, K.H., S.H. and N.Y.; validation, K.H. and N.Y.; formal analysis, K.H.; investigation, K.H. and N.Y.; writing—original draft preparation, K.H and Q.T.; writing—review and editing, K.H., Q.T., S.H. and X.L.; visualization, N.Y.; supervision, Q.T., S.H. and X.L.; project administration, Q.T., S.H. and X.L.; funding acquisition, Q.T. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Young Backbone Teacher Support Plan of Beijing Information Science & Technology University (YBT202426).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The public sources of the data mentioned in this study are described in the paper.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Zhu, P.; Wen, L.; Bian, X.; Ling, H.; Hu, Q. Vision meets drones: A challenge. arXiv 2018, arXiv:1804.07437. [Google Scholar]
  2. Zheng, Z.; Wei, Y.; Yang, Y. University-1652: A multi-view multi-source benchmark for drone-based geo-localization. In Proceedings of the 28th ACM international conference on Multimedia, Seattle, WA, USA, 12–16 October 2020; pp. 1395–1403. [Google Scholar]
  3. Zhu, R.; Yin, L.; Yang, M.; Wu, F.; Yang, Y.; Hu, W. SUES-200: A multi-height multi-scene cross-view image benchmark across drone and satellite. IEEE Trans. Circuits Syst. Video Technol. 2023, 33, 4825–4839. [Google Scholar] [CrossRef]
  4. Shen, T.; Wei, Y.; Kang, L.; Wan, S.; Yang, Y.H. MCCG: A ConvNeXt-based multiple-classifier method for cross-view geolocalization. IEEE Trans. Circuits Syst. Video Technol. 2023, 34, 1456–1468. [Google Scholar] [CrossRef]
  5. Deuser, F.; Habel, K.; Oswald, N. Sample4geo: Hard negative sampling for cross-view geo-localisation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 1–6 October 2023; pp. 16847–16856. [Google Scholar]
  6. Wang, T.; Zheng, Z.; Yan, C.; Zhang, J.; Sun, Y.; Zheng, B.; Yang, Y. Each part matters: Local patterns facilitate cross-view geo-localization. IEEE Trans. Circuits Syst. Video Technol. 2021, 32, 867–879. [Google Scholar] [CrossRef]
  7. Dai, M.; Hu, J.; Zhuang, J.; Zheng, E. A transformer-based feature segmentation and region alignment method for UAV-view geo-localization. IEEE Trans. Circuits Syst. Video Technol. 2021, 32, 4376–4389. [Google Scholar] [CrossRef]
  8. Zhao, H.; Ren, K.; Yue, T.; Zhang, C.; Yuan, S. TransFG: A Cross-View Geo-Localization of Satellite and UAVs Imagery Pipeline Using Transformer-Based Feature Aggregation and Gradient Guidance. IEEE Trans. Geosci. Remote Sens. 2024, 62, 4700912. [Google Scholar] [CrossRef]
  9. Chu, Y.; Tong, Q.; Liu, X.; Liu, X. ODAdapter: An Effective Method of Semi-supervised Object Detection for Aerial Images. In Proceedings of the Chinese Conference on Pattern Recognition and Computer Vision (PRCV), Urumqi, China, 18–20 October 2024; pp. 158–172. [Google Scholar]
  10. Lowe, D.G. Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
  11. Bay, H.; Ess, A.; Tuytelaars, T.; Van Gool, L. Speeded-up robust features (SURF). Comput. Vis. Image Underst. 2008, 110, 346–359. [Google Scholar] [CrossRef]
  12. Workman, S.; Jacobs, N. On the location dependence of convolutional neural network features. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Boston, MA, USA, 7–12 June 2015; pp. 70–78. [Google Scholar]
  13. Workman, S.; Souvenir, R.; Jacobs, N. Wide-area image geolocalization with aerial reference imagery. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 3961–3969. [Google Scholar]
  14. Lin, T.Y.; Cui, Y.; Belongie, S.; Hays, J. Learning deep representations for ground-to-aerial geolocalization. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 5007–5015. [Google Scholar]
  15. Vo, N.N.; Hays, J. Localizing and orienting street views using overhead imagery. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Springer: Berlin/Heidelberg, Germany, 2016; pp. 494–509. [Google Scholar]
  16. Hu, S.; Feng, M.; Nguyen, R.M.; Lee, G.H. Cvm-net: Cross-view matching network for image-based ground-to-aerial geo-localization. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7258–7267. [Google Scholar]
  17. Ding, L.; Zhou, J.; Meng, L.; Long, Z. A practical cross-view image matching method between UAV and satellite for UAV-based geo-localization. Remote Sens. 2020, 13, 47. [Google Scholar] [CrossRef]
  18. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  19. Shao, J.; Jiang, L. Style alignment-based dynamic observation method for UAV-view geo-localization. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–14. [Google Scholar] [CrossRef]
  20. Liu, Z.; Mao, H.; Wu, C.Y.; Feichtenhofer, C.; Darrell, T.; Xie, S. A convnet for the 2020s. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 11976–11986. [Google Scholar]
  21. He, K.; Fan, H.; Wu, Y.; Xie, S.; Girshick, R. Momentum contrast for unsupervised visual representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 9729–9738. [Google Scholar]
  22. Chen, T.; Kornblith, S.; Norouzi, M.; Hinton, G. A simple framework for contrastive learning of visual representations. In Proceedings of the International Conference on Machine Learning, Shenzhen, China, 15–17 February 2020; pp. 1597–1607. [Google Scholar]
  23. Grill, J.B.; Strub, F.; Altché, F.; Tallec, C.; Richemond, P.; Buchatskaya, E.; Doersch, C.; Avila Pires, B.; Guo, Z.; Gheshlaghi Azar, M.; et al. Bootstrap your own latent-a new approach to self-supervised learning. Adv. Neural Inf. Process. Syst. 2020, 33, 21271–21284. [Google Scholar]
  24. Oord, A.v.d.; Li, Y.; Vinyals, O. Representation learning with contrastive predictive coding. arXiv 2018, arXiv:1807.03748. [Google Scholar]
  25. Li, H.; Xu, C.; Yang, W.; Yu, H.; Xia, G.S. Learning cross-view visual geo-localization without ground truth. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5632017. [Google Scholar] [CrossRef]
  26. Li, D.; Chen, X.; Zhang, Z.; Huang, K. Learning deep context-aware features over body and latent parts for person re-identification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 384–393. [Google Scholar]
  27. Zheng, Z.; Yang, X.; Yu, Z.; Zheng, L.; Yang, Y.; Kautz, J. Joint discriminative and generative learning for person re-identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 2138–2147. [Google Scholar]
  28. He, S.; Luo, H.; Wang, P.; Wang, F.; Li, H.; Jiang, W. Transreid: Transformer-based object re-identification. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 15013–15022. [Google Scholar]
  29. Sun, Y.; Zheng, L.; Yang, Y.; Tian, Q.; Wang, S. Beyond part models: Person retrieval with refined part pooling (and a strong convolutional baseline). In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 480–496. [Google Scholar]
  30. Xu, J.; Zhao, R.; Zhu, F.; Wang, H.; Ouyang, W. Attention-aware compositional network for person re-identification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 2119–2128. [Google Scholar]
  31. Zhuang, J.; Dai, M.; Chen, X.; Zheng, E. A faster and more effective cross-view matching method of UAV and satellite images for UAV geolocalization. Remote Sens. 2021, 13, 3979. [Google Scholar] [CrossRef]
  32. Chen, Q.; Wang, T.; Yang, Z.; Li, H.; Lu, R.; Sun, Y.; Zheng, B.; Yan, C. SDPL: Shifting-Dense Partition Learning for UAV-View Geo-Localization. IEEE Trans. Circuits Syst. Video Technol. 2024, 34, 11810–11824. [Google Scholar] [CrossRef]
  33. Ge, F.; Zhang, Y.; Liu, Y.; Wang, G.; Coleman, S.; Kerr, D.; Wang, L. Multibranch Joint Representation Learning Based on Information Fusion Strategy for Cross-View Geo-Localization. IEEE Trans. Geosci. Remote Sens. 2024, 62, 1–16. [Google Scholar] [CrossRef]
  34. Ge, F.; Zhang, Y.; Wang, L.; Liu, W.; Liu, Y.; Coleman, S.; Kerr, D. Multi-level Feedback Joint Representation Learning Network Based on Adaptive Area Elimination for Cross-view Geo-localization. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5913915. [Google Scholar]
  35. Sun, J.; Sun, H.; Lei, L.; Ji, K.; Kuang, G. TirSA: A Three Stage Approach for UAV-Satellite Cross-View Geo-Localization Based on Self-Supervised Feature Enhancement. IEEE Trans. Circuits Syst. Video Technol. 2024, 34, 7882–7895. [Google Scholar] [CrossRef]
  36. Wang, T.; Zheng, Z.; Zhu, Z.; Sun, Y.; Yan, C.; Yang, Y. Learning cross-view geo-localization embeddings via dynamic weighted decorrelation regularization. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5647112. [Google Scholar] [CrossRef]
  37. Liu, C.; Li, S.; Du, C.; Qiu, H. Adaptive Global Embedding Learning: A Two-Stage Framework for UAV-View Geo-Localization. IEEE Signal Process. Lett. 2024, 31, 1239–1243. [Google Scholar] [CrossRef]
  38. Liu, X.; Wang, Z.; Wu, Y.; Miao, Q. SeGCN: A Semantic-Aware Graph Convolutional Network for UAV Geo-Localization. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 17, 6055–6066. [Google Scholar] [CrossRef]
  39. Wu, Q.; Wan, Y.; Zheng, Z.; Zhang, Y.; Wang, G.; Zhao, Z. Camp: A cross-view geo-localization method using contrastive attributes mining and position-aware partitioning. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5637614. [Google Scholar] [CrossRef]
  40. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
  41. Misra, D.; Nalamada, T.; Arasanipalai, A.U.; Hou, Q. Rotate to attend: Convolutional triplet attention module. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Virtual, 5–9 January 2021; pp. 3139–3148. [Google Scholar]
  42. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  43. Chen, L.; Zhang, H.; Xiao, J.; Nie, L.; Shao, J.; Liu, W.; Chua, T.S. Sca-cnn: Spatial and channel-wise attention in convolutional networks for image captioning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 5659–5667. [Google Scholar]
  44. Xu, B. Empirical evaluation of rectified activations in convolutional network. arXiv 2015, arXiv:1505.00853. [Google Scholar]
  45. Elfwing, S.; Uchibe, E.; Doya, K. Sigmoid-weighted linear units for neural network function approximation in reinforcement learning. Neural Netw. 2018, 107, 3–11. [Google Scholar] [CrossRef] [PubMed]
  46. He, K.; Zhang, X.; Ren, S.; Sun, J. Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1026–1034. [Google Scholar]
  47. Loshchilov, I. Decoupled weight decay regularization. arXiv 2017, arXiv:1711.05101. [Google Scholar]
  48. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  49. Ko, B.; Gu, G. Embedding expansion: Augmentation in embedding space for deep metric learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 7255–7264. [Google Scholar]
  50. Van der Maaten, L.; Hinton, G. Visualizing data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
Figure 1. Evaluation results of our model trained on the University-1652 dataset and tested on both the University-1652 and SUES-200 datasets.
Figure 1. Evaluation results of our model trained on the University-1652 dataset and tested on both the University-1652 and SUES-200 datasets.
Sensors 25 04519 g001
Figure 2. Architecture of the MCFA framework. The ConvNeXt-Tiny backbone extracts deep features from cross-view images. The MSCM enhances local–global interactions, while the FAAM dynamically adjusts spatial and channel importance for precise alignment. These components work together to improve feature representation and mitigate misalignment.
Figure 2. Architecture of the MCFA framework. The ConvNeXt-Tiny backbone extracts deep features from cross-view images. The MSCM enhances local–global interactions, while the FAAM dynamically adjusts spatial and channel importance for precise alignment. These components work together to improve feature representation and mitigate misalignment.
Sensors 25 04519 g002
Figure 3. The partitioning strategy influences the understanding of the content. A smaller central partition may lead to erroneous semantic information.
Figure 3. The partitioning strategy influences the understanding of the content. A smaller central partition may lead to erroneous semantic information.
Sensors 25 04519 g003
Figure 4. The partitioning method of LPN and ours. The first row represents LPN, while the second row represents ours.
Figure 4. The partitioning method of LPN and ours. The first row represents LPN, while the second row represents ours.
Sensors 25 04519 g004
Figure 5. Spatial arrangement differences of the same target captured from different orientations.
Figure 5. Spatial arrangement differences of the same target captured from different orientations.
Sensors 25 04519 g005
Figure 6. Illustration of mirror padding. The rightmost part of the image is cropped, and the leftmost region of the same width is mirrored to the left to maintain the original image size. (a) Original drone image; (b) left-side mirror padding of 20 pixels; (c) left-side mirror padding of 60 pixels; (d) left-side mirror padding of 100 pixels.
Figure 6. Illustration of mirror padding. The rightmost part of the image is cropped, and the leftmost region of the same width is mirrored to the left to maintain the original image size. (a) Original drone image; (b) left-side mirror padding of 20 pixels; (c) left-side mirror padding of 60 pixels; (d) left-side mirror padding of 100 pixels.
Sensors 25 04519 g006
Figure 7. Performance comparison of our model with SDPL, FRSA, and LPN under different mirror padding settings. Our model is represented by the blue line.
Figure 7. Performance comparison of our model with SDPL, FRSA, and LPN under different mirror padding settings. Our model is represented by the blue line.
Sensors 25 04519 g007
Figure 8. Comparison of performance degradation of our model and SDPL, FRSA, and LPN under different mirror padding settings. The blue bar in the histogram represents our model’s performance.
Figure 8. Comparison of performance degradation of our model and SDPL, FRSA, and LPN under different mirror padding settings. The blue bar in the histogram represents our model’s performance.
Sensors 25 04519 g008
Figure 9. Comparison of heatmaps between our proposed method and the existing CAMP model for both drone and satellite view images on University-1652.
Figure 9. Comparison of heatmaps between our proposed method and the existing CAMP model for both drone and satellite view images on University-1652.
Sensors 25 04519 g009
Figure 10. Top-5 retrieval results for (a) drone localization and (b) drone navigation. The yellow box indicates a correct match, while the blue box indicates an incorrect match.
Figure 10. Top-5 retrieval results for (a) drone localization and (b) drone navigation. The yellow box indicates a correct match, while the blue box indicates an incorrect match.
Sensors 25 04519 g010
Figure 11. t-SNE visualization of feature distributions. (a) shows the global view of feature embeddings. (b) highlights successful matches between drone and satellite features. (c) presents mismatches caused by high visual similarity between different scenes, where red dashed lines indicate incorrect matches.
Figure 11. t-SNE visualization of feature distributions. (a) shows the global view of feature embeddings. (b) highlights successful matches between drone and satellite features. (c) presents mismatches caused by high visual similarity between different scenes, where red dashed lines indicate incorrect matches.
Sensors 25 04519 g011
Figure 12. Heatmaps under varying illumination conditions on University-1652.
Figure 12. Heatmaps under varying illumination conditions on University-1652.
Sensors 25 04519 g012
Figure 13. Heatmaps at 300 m altitude on SUES-200.
Figure 13. Heatmaps at 300 m altitude on SUES-200.
Sensors 25 04519 g013
Table 1. Comparison with SOTA results on University-1652.
Table 1. Comparison with SOTA results on University-1652.
MethodPublicationDrone→SatelliteSatellite→Drone
R@1APR@1AP
Instance Loss [2]ACM MM’202059.6964.8073.1859.40
LPN [6]TCSVT’2275.9379.1486.4574.79
FSRA [7]TCSVT’2282.2584.8287.8781.53
TransFG [8]TGRS’2484.0186.3190.1684.61
SeGCN [38]JSTARS’2489.1890.8994.2989.65
MCCG [4]TSCVT’2389.6491.3294.3089.39
SDPL [32]TSCVT’2490.1691.6493.5889.45
Sample4Geo [5]ICCV’2392.6593.8195.1491.39
CAMP [39]TGRS’2494.4695.3896.1592.72
MCFA (ours) 94.5395.4096.4393.95
The best results are in bold.
Table 2. Comparison with SOTA results on SUES-200.
Table 2. Comparison with SOTA results on SUES-200.
Drone→Satellite
MethodPublication150 m200 m250 m300 m
R@1APR@1APR@1APR@1AP
SUES-200 Baseline [3]TCSVT’2355.6561.9266.7871.5572.0076.4374.0578.26
LPN [6]TCSVT’2261.5867.2370.8575.9680.3883.8081.4784.53
MCCG [4]TCSVT’2382.2285.4789.3891.4193.8295.0495.0796.20
SDPL [32]TCSVT’2482.9585.8292.7394.0796.0596.6997.8398.05
SeGCN [38]JSTARS’2490.8092.3291.9393.4192.5393.9093.3394.61
Sample4Geo [5]ICCV’2392.6094.0097.3897.8198.2898.6499.1899.36
CAMP [39]TGRS’2495.4096.3897.6398.1698.0598.4599.2399.46
MCFA (ours) 96.0096.8997.3897.8698.2598.6799.2899.42
Satellite→Drone
MethodPublication150 m200 m250 m300 m
R@1APR@1APR@1APR@1AP
SUES-200 Baseline [3]TCSVT’2375.0055.4685.0066.0586.2569.9488.7574.46
LPN [6]TCSVT’2283.7566.7888.7575.0192.5081.3492.5085.72
MCCG [4]TCSVT’2393.7589.7293.7592.2196.2596.1498.7596.64
SDPL [32]TCSVT’2493.7583.7596.2592.4297.5095.6596.2596.17
SeGCN [38]JSTARS’2493.7592.4595.0093.6596.2594.3997.5094.55
Sample4Geo [5]ICCV’2397.5093.6398.7596.7098.7598.2898.7598.05
CAMP [39]TGRS’2496.2593.6997.5096.7698.7598.10100.0098.85
MCFA (ours) 97.5094.7698.7596.6098.7598.1698.7598.68
The best results are in bold.
Table 3. Generalization experiment when trained on the UNIVERSITY-1652 dataset and evaluated on SUES-200.
Table 3. Generalization experiment when trained on the UNIVERSITY-1652 dataset and evaluated on SUES-200.
Drone→Satellite
Method150 m200 m250 m300 m
R@1APR@1APR@1APR@1AP
LPN [6]32.8540.1043.8050.6749.7556.5554.1060.73
MCCG [4]57.6262.8066.8371.6074.2578.3582.5585.27
Sample4Geo [5]74.7078.4781.2884.4086.8889.2889.2891.24
CAMP [39]76.5380.4787.1889.6093.7595.0496.4097.18
MCFA (ours)82.3885.2890.7092.2194.3895.2495.3096.00
Satellite→Drone
Method150 m200 m250 m300 m
R@1APR@1APR@1APR@1AP
LPN [6]32.5026.6040.0035.1046.5041.8853.5048.47
MCCG [4]61.2553.5182.5067.0681.2574.9987.5080.20
Sample4Geo [5]82.5076.2085.0082.9392.5087.7792.5088.38
CAMP [39]88.7578.1795.0088.3195.0091.8596.2593.43
MCFA (ours)92.5085.2295.0091.4695.0093.4697.5094.93
The best results are in bold.
Table 4. The results of ablation studies with MSCM and FAAM on University-1652.
Table 4. The results of ablation studies with MSCM and FAAM on University-1652.
MethodDrone→SatelliteSatellite→Drone
R@1APR@1AP
w/o both91.96 +0.0093.25 +0.0094.15 +0.0090.86 +0.00
w/MSCM93.31 +1.3594.43 +1.1895.58 +1.4392.14 +1.28
w/FAAM93.15 +1.1994.21 +0.9695.00 +0.8592.18 +1.32
MCFA (ours)94.53 +2.5795.40 +2.1596.43 +2.2893.95 +3.09
The best results are in bold.
Table 5. Comparison of different attention modules on University-1652.
Table 5. Comparison of different attention modules on University-1652.
MethodDrone→SatelliteSatellite→Drone
R@1APR@1AP
SE [48]93.5894.5695.2992.84
CBAM [42]92.6993.8195.0192.11
CDIM (ours)94.5395.4096.4393.95
The best results are in bold.
Table 6. Performance comparison of activation functions in the FGA module on University-1652.
Table 6. Performance comparison of activation functions in the FGA module on University-1652.
Activation FunctionDrone→SatelliteSatellite→Drone
R@1APR@1AP
ReLU [44]93.9494.9596.1493.45
SiLU [45]94.5395.4096.4393.95
The best results are in bold.
Table 7. Efficiency and performance comparison of different methods on University-1652.
Table 7. Efficiency and performance comparison of different methods on University-1652.
MethodRuntimeParamsDrone→SatelliteSatellite→Drone
(ms)(M)R@1APR@1AP
LPN [6]4.6858.2975.9379.1486.4574.79
Sample4Geo [5]3.6487.5792.6593.8195.1491.39
MCFA (ViT)8.8085.5188.7390.6695.0192.11
MCFA (ConvNeXt)3.7597.5794.5395.4096.4393.95
The best results are in bold.
Table 8. The effect of image offset.
Table 8. The effect of image offset.
OffsetMCFA (Ours)SDPL [32]FRSA [7]LPN [6]
R@1APR@1APR@1APR@1AP
094.53 −0.0095.40 −0.0085.25 −0.0087.48 −0.0086.41 −0.0088.34 −0.0078.08 −0.0080.94 −0.00
+2094.06 −0.4795.02 −0.3884.44 −0.8186.80 −0.6885.51 −0.9087.59 −0.7576.72 −1.3679.72 −1.22
+4092.87 −1.6694.06 −1.3482.46 −2.7985.15 −2.3382.77 −3.6485.30 −3.0473.04 −5.0476.47 −4.47
+6090.54 −3.9992.17 −3.2378.68 −6.5781.91 −5.5777.95 −8.4681.18 −7.1669.54 −8.5473.36 −7.58
+8086.49 −8.0488.87 −6.5374.48 −10.7778.33 −9.1570.90 −15.5174.99 −13.3561.13 −16.9565.80 −15.14
+10080.56 −13.9783.91 −11.4968.19 −17.0673.57 −13.9162.10 −24.3167.05 −21.2947.87 −30.2153.50 −27.44
Table 9. Impact of center region size on different cutting methods when number = 4.
Table 9. Impact of center region size on different cutting methods when number = 4.
SizeSatellite→DroneDrone→Satellite
R@1APR@1AP
3 × 3 92.72 +0.0093.88 +0.0095.29 +0.0092.40 +0.00
4 × 4 93.84 +1.1294.80 +0.9296.15 +0.8693.26 +0.86
5 × 5 93.66 +0.9494.64 +0.7696.29 +1.0093.34 +0.94
6 × 6 93.89 +1.1794.82 +0.9495.58 +0.2992.94 +0.54
7 × 7 92.34 −0.3893.51 −0.3795.15 −0.1491.99 −0.41
The best results are in bold.
Table 10. Impact of center region size on different cutting methods when number = 6.
Table 10. Impact of center region size on different cutting methods when number = 6.
SizeSatellite→DroneDrone→Satellite
R@1APR@1AP
3 × 3 93.84 +0.0094.82 +0.0095.58 +0.0093.34 +0.00
4 × 4 94.53 +0.6995.40 +0.5896.43 +0.8593.95 +0.61
5 × 5 94.03 +0.1994.97 +0.1596.15 +0.5793.59 +0.25
6 × 6 93.88 +0.0494.87 +0.0596.15 +0.5793.24 −0.10
7 × 7 92.42 −1.4293.60 −1.2295.44 −0.1491.99 −1.35
The best results are in bold.
Table 11. The results of ablation studies with different numbers of cut regions.
Table 11. The results of ablation studies with different numbers of cut regions.
NumberDrone→SatelliteSatellite→Drone
R@1APR@1AP
Ours (4)93.2694.3495.7292.73
Ours (6)94.5395.4096.4393.95
Ours (8)92.4493.5895.7292.47
Ours (10)92.4693.6395.2992.30
The best results are in bold.
Table 12. Performance of FGA with different loss functions.
Table 12. Performance of FGA with different loss functions.
Loss FunctionSatellite→DroneDrone→Satellite
R@1APR@1AP
InfoNCE91.6593.0294.7290.78
MSE loss93.0994.1595.5892.18
Triplet loss93.2194.2594.8692.09
Hard Triplet Loss [49]94.1395.0495.7293.57
Cosine Similarity Loss94.5395.4096.4393.95
The best results are in bold.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hou, K.; Tong, Q.; Yan, N.; Liu, X.; Hou, S. MCFA: Multi-Scale Cascade and Feature Adaptive Alignment Network for Cross-View Geo-Localization. Sensors 2025, 25, 4519. https://doi.org/10.3390/s25144519

AMA Style

Hou K, Tong Q, Yan N, Liu X, Hou S. MCFA: Multi-Scale Cascade and Feature Adaptive Alignment Network for Cross-View Geo-Localization. Sensors. 2025; 25(14):4519. https://doi.org/10.3390/s25144519

Chicago/Turabian Style

Hou, Kaiji, Qiang Tong, Na Yan, Xiulei Liu, and Shoulu Hou. 2025. "MCFA: Multi-Scale Cascade and Feature Adaptive Alignment Network for Cross-View Geo-Localization" Sensors 25, no. 14: 4519. https://doi.org/10.3390/s25144519

APA Style

Hou, K., Tong, Q., Yan, N., Liu, X., & Hou, S. (2025). MCFA: Multi-Scale Cascade and Feature Adaptive Alignment Network for Cross-View Geo-Localization. Sensors, 25(14), 4519. https://doi.org/10.3390/s25144519

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop