Off-Nadir Satellite Image Scene Classification: Benchmark Dataset, Angle-Aware Active Domain Adaptation, and Angular Impact Analysis
Highlights
- This paper introduces Off-nadir-Scene10, the first benchmark dataset for off-nadir satellite image scene classification, containing 5200 images across 10 categories and 26 viewing angles.
- This study proposes an angle-aware active domain adaptation method that leverages nadir imagery to improve off-nadir classification performance while reducing annotation requirements.
- This work demonstrates that training on larger off-nadir angles enhances cross-view transferability by promoting view-invariant feature learning.
- This study provides practical guidelines for dataset construction and training strategies to build robust off-nadir scene classification systems for real-world applications.
Abstract
1. Introduction
1.1. Existing Datasets and Their Limitations
1.2. Off-Nadir Imagery: Potential and Challenges
1.3. Domain Adaptation and Its Limitations
1.4. Motivation for This Work
- (1)
- Creation of Off-nadir-Scene10 benchmark dataset: We introduce Off-nadir-Scene10, the first multi-category dataset explicitly designed for off-nadir satellite image scene classification. It contains 5200 images covering 10 common remote sensing scene types captured at 26 different off-nadir angles, providing extensive angular diversity and scene complexity. This dataset addresses the critical shortage of annotated off-nadir imagery needed for advancing off-nadir scene classification research. The Off-nadir-Scene10 dataset is available at https://github.com/AIP2025-RS/Off-nadir-Scene10 (accessed on 9 November 2025).
- (2)
- Angle-aware active domain adaptation for scene classification: Leveraging the abundance of labeled nadir remote sensing images, we develop an improved active domain adaptation method that incorporates off-nadir angle awareness into sample selection and model adaptation. By integrating angle-dependent weighting into the sample selection criterion, our approach effectively transfers discriminative knowledge from nadir to off-nadir domains. This reduces the demand for labeled off-nadir samples, thereby enhancing classification accuracy.
- (3)
- Comprehensive angular impact analysis: We systematically analyze how off-nadir viewing angles affect classification performance by dividing the dataset into groups on the basis of off-nadir angle magnitude and examining intra- and cross-group training/testing accuracies. Our results reveal pronounced angular influences on generalization and provide insights into the asymmetric transferability between large- and small-angle views, which can guide future data collection and model design.
2. Off-Nadir Image Scene Classification Benchmark Dataset
2.1. Process of Dataset Building
2.1.1. Scene Category Selection
2.1.2. Data Source Acquisition
2.1.3. Sample Point Annotation
2.1.4. Image Cropping
2.2. Dataset Description
3. Methodology
3.1. Angle-Aware Active Domain Adaptation
3.2. Margin Loss Function in Model Training
3.3. Sampling Query Function
4. Experiments and Results
4.1. Experimental Setting
4.1.1. Overall Experimental Setup
4.1.2. Experiments of Representative Networks
4.1.3. Comparison Experiment of Training Modes
4.1.4. Grouping Experiment by Off-Nadir Angles
4.2. Classification Results of Representative Networks
4.3. Performance of Training Modes vs. SDM
4.4. Results of Methods for Different Training Modes
4.5. Results of Grouping Experiments by Off-Nadir Angles
5. Discussion
5.1. Contributions of the Off-Nadir-Scene10 Dataset
5.2. Role of Nadir Data and Angle-Aware Active Domain Adaptation
5.3. Angular Impact Analysis on Scene Classification
6. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Lu, S.; Guo, J.; Zimmer-Dauphinee, J.R.; Nieusma, J.M.; Wang, X.; VanValkenburgh, P.; Wernke, S.A.; Huo, Y. Vision Foundation Models in Remote Sensing: A Survey. IEEE Geosci. Remote Sens. Mag. 2025, 13, 190–215. [Google Scholar] [CrossRef]
- Wang, S.; Han, W.; Huang, X.; Zhang, X.; Wang, L.; Li, J. Trustworthy Remote Sensing Interpretation: Concepts, Technologies, and Applications. ISPRS J. Photogramm. Remote Sens. 2024, 209, 150–172. [Google Scholar] [CrossRef]
- Jia, Y.; Sun, C.; Gao, J.; Wang, Q. Few-Shot Remote Sensing Scene Classification via Parameter-Free Attention and Region Matching. ISPRS J. Photogramm. Remote Sens. 2025, 227, 265–275. [Google Scholar] [CrossRef]
- Qiu, C.; Zhang, X.; Tong, X.; Guan, N.; Yi, X.; Yang, K.; Zhu, J.; Yu, A. Few-Shot Remote Sensing Image Scene Classification: Recent Advances, New Baselines, and Future Trends. ISPRS J. Photogramm. Remote Sens. 2024, 209, 368–382. [Google Scholar] [CrossRef]
- Thapa, A.; Horanont, T.; Neupane, B.; Aryal, J. Deep Learning for Remote Sensing Image Scene Classification: A Review and Meta-Analysis. Remote Sens. 2023, 15, 4804. [Google Scholar] [CrossRef]
- Li, Z.; Chen, B.; Wu, S.; Su, M.; Chen, J.M.; Xu, B. Deep Learning for Urban Land Use Category Classification: A Review and Experimental Assessment. Remote Sens. Environ. 2024, 311, 114290. [Google Scholar] [CrossRef]
- Torres, R.N.; Fraternali, P. Learning to Identify Illegal Landfills through Scene Classification in Aerial Images. Remote Sens. 2021, 13, 4520. [Google Scholar] [CrossRef]
- Jozdani, S.; Chen, D.; Chen, W.; Leblanc, S.G.; Lovitt, J.; He, L.; Fraser, R.H.; Johnson, B.A. Evaluating Image Normalization via GANs for Environmental Mapping: A Case Study of Lichen Mapping Using High-Resolution Satellite Imagery. Remote Sens. 2021, 13, 5035. [Google Scholar] [CrossRef]
- Zhu, S.; Wu, Y.; Ma, X. Deep Learning-Based Algal Bloom Identification Method from Remote Sensing Images—Take China’s Chaohu Lake as an Example. Sustainability 2023, 15, 4545. [Google Scholar] [CrossRef]
- Wen, S.; Zhao, W.; Ji, F.; Peng, R.; Zhang, L.; Wang, Q. Recognizing Unknown Disaster Scenes with Knowledge Graph-Based Zero-Shot Learning (KG-ZSL) Model. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5621315. [Google Scholar] [CrossRef]
- Liu, L.; Xiao, Z.; Hu, J.; Han, J.; Kim, J.Y.; Sharma, R.; Li, C. FSBNet: A Classifying Framework of Disaster Scene for Volcanic Lithology through Deep-Learning Models. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2025, 18, 15101–15115. [Google Scholar] [CrossRef]
- Shi, L.; Zhang, F.; Xia, J.; Xie, J.; Zhang, Z.; Du, Z.; Liu, R. Identifying Damaged Buildings in Aerial Images Using the Object Detection Method. Remote Sens. 2021, 13, 4213. [Google Scholar] [CrossRef]
- Chen, T.; Tang, W.; Chen, S.-E.; Allan, C. SA-Encoder: A Learnt Spatial Autocorrelation Representation to Inform 3D Geospatial Object Detection. Remote Sens. 2025, 17, 3124. [Google Scholar] [CrossRef]
- Yang, Y.; Newsam, S. Bag-of-Visual-Words and Spatial Extensions for Land-Use Classification. In Proceedings of the 18th SIGSPATIAL International Conference on Advances in Geographic Information Systems, San Jose, CA, USA, 2–5 November 2010; Association for Computing Machinery: New York, NY, USA, 2010; pp. 270–279. [Google Scholar]
- Xia, G.-S.; Hu, J.; Hu, F.; Shi, B.; Bai, X.; Zhong, Y.; Zhang, L.; Lu, X. AID: A Benchmark Data Set for Performance Evaluation of Aerial Scene Classification. IEEE Trans. Geosci. Remote Sens. 2016, 55, 3965–3981. [Google Scholar] [CrossRef]
- Cheng, G.; Han, J.; Lu, X. Remote Sensing Image Scene Classification: Benchmark and State of the Art. Proc. IEEE 2017, 105, 1865–1883. [Google Scholar] [CrossRef]
- Xiao, Y.; Yuan, Q.; Jiang, K.; He, J.; Lin, C.-W.; Zhang, L. TTST: A Top-k Token Selective Transformer for Remote Sensing Image Super-Resolution. IEEE Trans. Image Process. 2024, 33, 738–752. [Google Scholar] [CrossRef] [PubMed]
- Hu, Y.; Yuan, J.; Wen, C.; Lu, X.; Liu, Y.; Li, X. RSGPT: A Remote Sensing Vision Language Model and Benchmark. ISPRS J. Photogramm. Remote Sens. 2025, 224, 272–286. [Google Scholar] [CrossRef]
- Ma, Y.; Chen, S.; Ermon, S.; Lobell, D.B. Transfer Learning in Environmental Remote Sensing. Remote Sens. Environ. 2024, 301, 113924. [Google Scholar] [CrossRef]
- McNally, S.; Nielsen, A.; Barrieau, A.; Jabari, S. Improving Off-Nadir Deep Learning-Based Change and Damage Detection through Radiometric Enhancement. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2024, 48, 33–39. [Google Scholar] [CrossRef]
- Poli, D.; Moe, K.; Legat, K.; Toschi, I.; Lago, F.; Remondino, F. Use of Vertical Aerial Images for Semi-Oblique Mapping. ISPRS—Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2017, 42, 493–498. [Google Scholar] [CrossRef]
- Zhou, H.; Kao, C.-H.; Phoo, C.P.; Mall, U.; Hariharan, B.; Bala, K. AllClear: A Comprehensive Dataset and Benchmark for Cloud Removal in Satellite Imagery. In Proceedings of the 38th International Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 10–15 December 2024; Curran Associates Inc.: Red Hook, NY, USA, 2025. [Google Scholar]
- Satriano, V.; Ciancia, E.; Pergola, N.; Tramutoli, V. A First Extension of the Robust Satellite Technique RST-FLOOD to Sentinel-2 Data for the Mapping of Flooded Areas: The Case of the Emilia Romagna (Italy) 2023 Event. Remote Sens. 2024, 16, 3450. [Google Scholar] [CrossRef]
- Dubey, A.; Ramanathan, V.; Pentland, A.; Mahajan, D. Adaptive Methods for Real-World Domain Generalization. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 14335–14344. [Google Scholar] [CrossRef]
- Pang, C.; Wu, J.; Ding, J.; Song, C.; Xia, G.-S. Detecting Building Changes with Off-Nadir Aerial Images. Sci. China Inf. Sci. 2023, 66, 140306. [Google Scholar] [CrossRef]
- Ma, Y.; He, T.; McVicar, T.R.; Liang, S.; Liu, T.; Peng, W.; Song, D.-X.; Tian, F. Quantifying How Topography Impacts Vegetation Indices at Various Spatial and Temporal Scales. Remote Sens. Environ. 2024, 312, 114311. [Google Scholar] [CrossRef]
- Wang, J.; Meng, L.; Li, W.; Yang, W.; Yu, L.; Xia, G.-S. Learning to Extract Building Footprints from Off-Nadir Aerial Images. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 1294–1301. [Google Scholar] [CrossRef]
- Zheng, J.; Zhao, Y.; Wu, W.; Chen, M.; Li, W.; Fu, H. Partial Domain Adaptation for Scene Classification from Remote Sensing Imagery. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5626911. [Google Scholar] [CrossRef]
- Wang, Q.; Liu, S.; Chanussot, J.; Li, X. Scene Classification with Recurrent Attention of VHR Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2019, 57, 1155–1167. [Google Scholar] [CrossRef]
- Chen, K.; Chen, B.; Liu, C.; Li, W.; Zou, Z.; Shi, Z. RSMamba: Remote Sensing Image Classification with State Space Model. IEEE Geosci. Remote Sens. Lett. 2024, 21, 8002605. [Google Scholar] [CrossRef]
- Reed, C.J.; Gupta, R.; Li, S.; Brockman, S.; Funk, C.; Clipp, B.; Keutzer, K.; Candido, S.; Uyttendaele, M.; Darrell, T. Scale-MAE: A Scale-Aware Masked Autoencoder for Multiscale Geospatial Representation Learning. In Proceedings of the 2023 IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 4–6 October 2023; pp. 4065–4076. [Google Scholar]
- Wang, J.; Li, W.; Zhang, M.; Tao, R.; Chanussot, J. Remote-Sensing Scene Classification via Multistage Self-Guided Separation Network. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5615312. [Google Scholar] [CrossRef]
- Weir, N.; Lindenbaum, D.; Bastidas, A.; Etten, A.V.; McPherson, S.; Shermeyer, J.; Vijay, V.; Tang, H. SpaceNet MVOI: A Multi-View Overhead Imagery Dataset. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 992–1001. [Google Scholar] [CrossRef]
- Shen, L.; Lu, Y.; Chen, H.; Wei, H.; Xie, D.; Yue, J.; Chen, R.; Zhang, Y.; Zhang, A.; Lv, S.; et al. S2Looking: A Satellite Side-Looking Dataset for Building Change Detection. Remote Sens. 2021, 13, 5094. [Google Scholar] [CrossRef]
- Machado, G.; Ferreira, E.; Nogueira, K.; Oliveira, H.; Brito, M.; Gama, P.H.T.; dos Santos, J.A. AiRound and CV-BrCT: Novel Multiview Datasets for Scene Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 488–503. [Google Scholar] [CrossRef]
- Hänsch, R.; Arndt, J.; Lunga, D.; Gibb, M.; Pedelose, T.; Boedihardjo, A.; Petrie, D.; Bacastow, T.M. Spacenet 8-the Detection of Flooded Roads and Buildings. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 1472–1480. [Google Scholar]
- Huang, X.; Wan, Y.; Zhang, Y.; Liu, X.; Zhang, B.; Wang, Y.; Guo, H.; Pei, Y.; Hu, Z. PSDA: Pyramid Spatial Deformable Aggregation for Building Segmentation in Multiview Remote Sensing Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2025, 18, 8995–9008. [Google Scholar] [CrossRef]
- Wang, Y.; Chen, S.; Zhang, R.; Xu, F.; Liang, S.; Wang, Y.; Yang, W. BuildMon: Building Extraction and Change Monitoring in Time Series Remote Sensing Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 17, 10813–10826. [Google Scholar] [CrossRef]
- Ni, L.; Huo, C.; Zhang, X.; Wang, P.; Zhang, L.; Guo, K.; Zhou, Z. NaGAN: Nadir-like Generative Adversarial Network for Off-Nadir Object Detection of Multi-View Remote Sensing Imagery. Remote Sens. 2022, 14, 975. [Google Scholar] [CrossRef]
- Saleh, T.; Weng, X.; Holail, S.; Hao, C.; Xia, G.-S. DAM-Net: Flood Detection from SAR Imagery Using Differential Attention Metric-Based Vision Transformers. ISPRS J. Photogramm. Remote Sens. 2024, 212, 440–453. [Google Scholar] [CrossRef]
- Wang, J.; Zhu, F.; Wang, Q.; Zhao, P.; Fang, Y. An Active Object-Detection Algorithm for Adaptive Attribute Adjustment of Remote-Sensing Images. Remote Sens. 2025, 17, 818. [Google Scholar] [CrossRef]
- Jin, R.; Jia, Z.; Yin, X.; Niu, Y.; Qi, Y. Domain Feature Decomposition for Efficient Object Detection in Aerial Images. Remote Sens. 2024, 16, 1626. [Google Scholar] [CrossRef]
- Toker, A.; Kondmann, L.; Weber, M.; Eisenberger, M.; Camero, A.; Hu, J.; Hoderlein, A.P.; Şenaras, Ç.; Davis, T.; Cremers, D. Dynamicearthnet: Daily Multi-Spectral Satellite Dataset for Semantic Change Segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 21158–21167. [Google Scholar]
- Han, C.; Wu, C.; Guo, H.; Hu, M.; Li, J.; Chen, H. Change Guiding Network: Incorporating Change Prior to Guide Change Detection in Remote Sensing Imagery. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 16, 8395–8407. [Google Scholar] [CrossRef]
- Li, K.; Cao, X.; Meng, D. A New Learning Paradigm for Foundation Model-Based Remote-Sensing Change Detection. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5610112. [Google Scholar] [CrossRef]
- Ding, L.; Zhu, K.; Peng, D.; Tang, H.; Yang, K.; Bruzzone, L. Adapting Segment Anything Model for Change Detection in VHR Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5611711. [Google Scholar] [CrossRef]
- Zheng, Z.; Ermon, S.; Kim, D.; Zhang, L.; Zhong, Y. Changen2: Multi-Temporal Remote Sensing Generative Change Foundation Model. IEEE Trans. Pattern Anal. Mach. Intell. 2025, 47, 725–741. [Google Scholar] [CrossRef]
- Marí, R.; Facciolo, G.; Ehret, T. Sat-NeRF: Learning Multi-View Satellite Photogrammetry with Transient Objects and Shadow Modeling Using RPC Cameras. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), New Orleans, LA, USA, 19–20 June 2022; pp. 1310–1320. [Google Scholar]
- Xie, S.; Zhang, L.; Jeon, G.; Yang, X. Remote Sensing Neural Radiance Fields for Multi-View Satellite Photogrammetry. Remote Sens. 2023, 15, 3808. [Google Scholar] [CrossRef]
- Liao, P.; Zhang, X.; Chen, G.; Wang, T.; Li, X.; Yang, H.; Zhou, W.; He, C.; Wang, Q. S2Net: A Multitask Learning Network for Semantic Stereo of Satellite Image Pairs. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5601313. [Google Scholar] [CrossRef]
- Huang, X.; Chen, H.; Gong, J. Angular Difference Feature Extraction for Urban Scene Classification Using ZY-3 Multi-Angle High-Resolution Satellite Imagery. ISPRS J. Photogramm. Remote Sens. 2018, 135, 127–141. [Google Scholar] [CrossRef]
- Huang, X.; Li, S.; Li, J.; Jia, X.; Li, J.; Zhu, X.X.; Benediktsson, J.A. A Multispectral and Multiangle 3-D Convolutional Neural Network for the Classification of ZY-3 Satellite Images over Urban Areas. IEEE Trans. Geosci. Remote Sens. 2021, 59, 10266–10285. [Google Scholar] [CrossRef]
- Zuo, R.; Huang, X.; Li, J.; Pan, X. A Cross-Angle Propagation Network for Built-Up Area Extraction by Fusing Spatial–Spectral-Angular Features From the ZY-3 Multiview Satellite Imagery: Dataset and Analysis of China’s 41 Major Cities. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5408320. [Google Scholar] [CrossRef]
- Liu, C.; Huang, X.; Zhu, Z.; Chen, H.; Tang, X.; Gong, J. Automatic Extraction of Built-up Area from ZY3 Multi-View Satellite Imagery: Analysis of 45 Global Cities. Remote Sens. Environ. 2019, 226, 51–73. [Google Scholar] [CrossRef]
- Wang, M.; Deng, W. Deep Visual Domain Adaptation: A Survey. Neurocomputing 2018, 312, 135–153. [Google Scholar] [CrossRef]
- Cui, S.; Wang, S.; Zhuo, J.; Li, L.; Huang, Q.; Tian, Q. Towards Discriminability and Diversity: Batch Nuclear-Norm Maximization under Label Insufficient Situations. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 3940–3949. [Google Scholar]
- Wang, S.; Hou, D.; Xing, H. A Self-Supervised-Driven Open-Set Unsupervised Domain Adaptation Method for Optical Remote Sensing Image Scene Classification and Retrieval. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5605515. [Google Scholar] [CrossRef]
- Zhang, J.; Liu, J.; Pan, B.; Chen, Z.; Xu, X.; Shi, Z. An Open Set Domain Adaptation Algorithm via Exploring Transferability and Discriminability for Remote Sensing Image Scene Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5609512. [Google Scholar] [CrossRef]
- Song, S.; Yu, H.; Miao, Z.; Zhang, Q.; Lin, Y.; Wang, S. Domain Adaptation for Convolutional Neural Networks-Based Remote Sensing Scene Classification. IEEE Geosci. Remote Sens. Lett. 2019, 16, 1324–1328. [Google Scholar] [CrossRef]
- Xu, Q.; Shi, Y.; Yuan, X.; Zhu, X.X. Universal Domain Adaptation for Remote Sensing Image Scene Classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 4700515. [Google Scholar] [CrossRef]
- Xie, M.; Li, Y.; Wang, Y.; Luo, Z.; Gan, Z.; Sun, Z.; Chi, M.; Wang, C.; Wang, P. Learning Distinctive Margin toward Active Domain Adaptation. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 7983–7992. [Google Scholar]
- Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
- Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the Inception Architecture for Computer Vision. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
- Iandola, F.N.; Moskewicz, M.W.; Ashraf, K.; Han, S.; Dally, W.J.; Keutzer, K. SqueezeNet: AlexNet-Level Accuracy with 50x Fewer Parameters and <1MB Model Size. arXiv 2016, arXiv:1602.07360. [Google Scholar]
- Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely Connected Convolutional Networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 2261–2269. [Google Scholar]
- Ma, N.; Zhang, X.; Zheng, H.-T.; Sun, J. ShuffleNet V2: Practical Guidelines for Efficient CNN Architecture Design. In Proceedings of the Computer Vision—ECCV 2018, Munich, Germany, 14–18 September 2018; Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y., Eds.; Springer International Publishing: Cham, Switzerland, 2018; pp. 122–138. [Google Scholar]
- Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.-C. MobileNetV2: Inverted Residuals and Linear Bottlenecks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4510–4520. [Google Scholar]
- Xie, S.; Girshick, R.; Dollár, P.; Tu, Z.; He, K. Aggregated Residual Transformations for Deep Neural Networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 5987–5995. [Google Scholar]
- Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer Using Shifted Windows. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 9992–10002. [Google Scholar] [CrossRef]
- Woo, S.; Debnath, S.; Hu, R.; Chen, X.; Liu, Z.; Kweon, I.S.; Xie, S. ConvNeXt V2: Co-Designing and Scaling ConvNets with Masked Autoencoders. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023; pp. 16133–16142. [Google Scholar]
- Zhang, H.; Wu, C.; Zhang, Z.; Zhu, Y.; Lin, H.; Zhang, Z.; Sun, Y.; He, T.; Mueller, J.; Manmatha, R.; et al. ResNeSt: Split-Attention Networks. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), New Orleans, LA, USA, 19–20 June 2022; pp. 2735–2745. [Google Scholar]
- Wang, W.; Xie, E.; Li, X.; Fan, D.-P.; Song, K.; Liang, D.; Lu, T.; Luo, P.; Shao, L. PVT v2: Improved Baselines with Pyramid Vision Transformer. Comput. Vis. Media 2022, 8, 415–424. [Google Scholar] [CrossRef]
- Yu, W.; Luo, M.; Zhou, P.; Si, C.; Zhou, Y.; Wang, X.; Feng, J.; Yan, S. MetaFormer Is Actually What You Need for Vision. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 10809–10819. [Google Scholar]
- Ding, M.; Xiao, B.; Codella, N.; Luo, P.; Wang, J.; Yuan, L. DaViT: Dual Attention Vision Transformers. In Proceedings of the Computer Vision—ECCV 2022, Tel Aviv, Israel, 23–27 October 2022; Avidan, S., Brostow, G., Cissé, M., Farinella, G.M., Hassner, T., Eds.; Springer Nature: Cham, Switzerland, 2022; pp. 74–92. [Google Scholar]
- Maaz, M.; Shaker, A.; Cholakkal, H.; Khan, S.; Zamir, S.W.; Anwer, R.M.; Shahbaz Khan, F. EdgeNeXt: Efficiently Amalgamated CNN-Transformer Architecture for Mobile Vision Applications. In Proceedings of the Computer Vision—ECCV 2022 Workshops, Tel Aviv, Israel, 23–27 October 2022; Karlinsky, L., Michaeli, T., Nishino, K., Eds.; Springer Nature: Cham, Switzerland, 2023; pp. 3–20. [Google Scholar]
- Wu, K.; Zhang, J.; Peng, H.; Liu, M.; Xiao, B.; Fu, J.; Yuan, L. TinyViT: Fast Pretraining Distillation for Small Vision Transformers. In Proceedings of the Computer Vision—ECCV 2022, Tel Aviv, Israel, 23–27 October 2022; Avidan, S., Brostow, G., Cissé, M., Farinella, G.M., Hassner, T., Eds.; Springer Nature: Cham, Switzerland, 2022; pp. 68–85. [Google Scholar]
- Tatsunami, Y.; Taki, M. Sequencer: Deep LSTM for Image Classification. arXiv 2022, arXiv:2205.01972. [Google Scholar]
- Yu, W.; Zhou, P.; Yan, S.; Wang, X. InceptionNeXt: When Inception Meets ConvNeXt. In Proceedings of the 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 16–22 June 2024; pp. 5672–5683. [Google Scholar]
- Vasu, P.K.A.; Gabriel, J.; Zhu, J.; Tuzel, O.; Ranjan, A. FastViT: A Fast Hybrid Vision Transformer Using Structural Reparameterization. In Proceedings of the 2023 IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 1–6 October 2023; pp. 5762–5772. [Google Scholar]
- Chen, C.; Guo, Z.; Zeng, H.; Xiong, P.; Dong, J. RepGhost: A Hardware-Efficient Ghost Module via Re-Parameterization. arXiv 2022, arXiv:2211.06088. [Google Scholar]
- Wang, A.; Chen, H.; Lin, Z.; Han, J.; Ding, G. Rep ViT: Revisiting Mobile CNN From ViT Perspective. In Proceedings of the 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 16–22 June 2024; pp. 15909–15920. [Google Scholar]
- Cai, H.; Li, J.; Hu, M.; Gan, C.; Han, S. EfficientViT: Lightweight Multi-Scale Attention for High-Resolution Dense Prediction. In Proceedings of the 2023 IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 1–6 October 2023; pp. 17256–17267. [Google Scholar]






| Off-Nadir ID | Size (Pixel) | GSD (m) | In-Track View Angle (°) | Cross-Track View Angle (°) | Off-Nadir Angle (°) |
|---|---|---|---|---|---|
| OA8 | 902 | 0.474 | 3.4 | −7.4 | 8.1 |
| OA9 | 899 | 0.475 | −4.4 | −7.5 | 8.7 |
| OA11 | 889 | 0.481 | −7.1 | −7.9 | 10.6 |
| OA14 | 869 | 0.492 | 12.4 | −6.9 | 14.2 |
| OA15 | 862 | 0.496 | −12.3 | −8.2 | 14.8 |
| OA18 | 847 | 0.505 | 17.0 | −6.5 | 18.1 |
| OA19 | 826 | 0.517 | −17.5 | −8.3 | 19.3 |
| OA22 | 807 | 0.530 | 20.6 | −6.1 | 21.5 |
| OA24 | 783 | 0.546 | −22.1 | −8.6 | 23.6 |
| OA26 | 763 | 0.56 | 25.9 | −5.9 | 26.5 |
| OA27 | 735 | 0.581 | −25.8 | −9.1 | 27.3 |
| OA30 | 717 | 0.596 | 29.8 | −5.6 | 30.2 |
| OA31 | 687 | 0.623 | −30.0 | −9 | 31.2 |
| OA32 | 670 | 0.638 | 31.8 | −5.1 | 32.1 |
| OA33 | 644 | 0.664 | −32 | −9.4 | 33.2 |
| OA36 | 596 | 0.717 | −35.1 | −9.6 | 36.2 |
| OA40 | 551 | 0.776 | −39.0 | −9.5 | 40 |
| OA42 | 507 | 0.842 | −41.0 | −9.8 | 41.9 |
| OA44 | 467 | 0.916 | −42.7 | −10.2 | 43.7 |
| OA46 | 429 | 0.997 | −44.7 | −10.3 | 45.6 |
| OA48 | 391 | 1.093 | −47.4 | −10.1 | 48.2 |
| OA49 | 358 | 1.194 | −48.1 | −10.6 | 49 |
| OA51 | 327 | 1.307 | −50.4 | −10.3 | 51.2 |
| OA52 | 297 | 1.441 | −51.7 | −10.4 | 52.4 |
| OA53 | 270 | 1.582 | −52.4 | −10.8 | 53.1 |
| OA54 | 256 | 1.670 | −53.0 | −11.1 | 53.8 |
| Type | Number | 20% | 50% | 80% |
|---|---|---|---|---|
| training | locations per class | 4 | 10 | 16 |
| locations | 104 | 260 | 416 | |
| images | 1040 | 2600 | 4160 | |
| testing | locations per class | 16 | 10 | 4 |
| locations | 416 | 260 | 104 | |
| images | 4160 | 2600 | 1040 |
| Group Name | Off-Nadir ID | Range of Off-Nadir Angles (°) | Medium Off-Nadir Angle (°) | Mean Off-Nadir Angle (°) |
|---|---|---|---|---|
| A | OA8, OA9, OA11, OA14, OA15, OA18, OA19, OA22, OA24, OA26, OA27, OA30, OA31 | 8.1~31.2 | 19.3 | 19.5 |
| B | OA32, OA33, OA36, OA40, OA42, OA44, OA46, OA48, OA49, OA51, OA52, OA53, OA54 | 32.1~53.8 | 45.6 | 44.6 |
| Networks | Training Ratio | ||
|---|---|---|---|
| 20% | 50% | 80% | |
| VGG16 | 80.24 | 86.50 | 94.62 |
| Inception-v3 | 76.49 | 85.92 | 95.38 |
| ResNet-50 | 80.12 | 84.69 | 91.06 |
| SqueezeNet-10 | 79.90 | 84.27 | 94.13 |
| DenseNet-121 | 80.29 | 87.42 | 94.13 |
| ShuffleNetV2 | 85.77 | 88.23 | 94.52 |
| MobileNetV2 | 79.74 | 86.92 | 92.69 |
| ResNeXt-50 | 79.64 | 86.58 | 92.31 |
| Swin-T | 83.32 | 87.27 | 93.94 |
| ConvNeXt V2 Tiny | 79.74 | 86.65 | 93.46 |
| ResNeSt-50 | 80.67 | 87.69 | 93.65 |
| PVT v2-B2 | 81.61 | 87.65 | 95.87 |
| PoolFormer-S24 | 82.04 | 87.35 | 95.77 |
| DaViT-Tiny | 84.90 | 88.92 | 95.87 |
| EdgeNeXt-S | 80.87 | 89.35 | 95.38 |
| TinyViT-21M | 85.75 | 89.88 | 96.83 |
| Sequencer2D-S | 81.35 | 87.81 | 94.52 |
| InceptionNeXt-T | 77.84 | 86.58 | 93.65 |
| FastViT-SA24 | 77.79 | 87.88 | 93.46 |
| RepGhostNet | 77.16 | 85.58 | 93.27 |
| RepViT-M1.1 | 76.85 | 84.96 | 92.50 |
| EfficientViT-B1 | 79.33 | 85.35 | 93.85 |
| Training Ratio | Finetune- ImageNet | Finetune- NWPU10 | AADA | AADA-600 | SDM |
|---|---|---|---|---|---|
| 20% | 0.8012 | 0.8159 | 0.8548 | 0.8560 | 0.8387 |
| 50% | 0.8469 | 0.8562 | 0.8985 | 0.8996 | 0.8801 |
| 80% | 0.9106 | 0.9154 | 0.9212 | 0.9225 | 0.9024 |
| Training Ratio | Finetune- ImageNet | Finetune- NWPU10 | AADA | AADA-600 | SDM |
|---|---|---|---|---|---|
| 20% | 3.8 | 4.2 | 4.4 | 8.3 | 4.3 |
| 50% | 4.2 | 4.6 | 4.8 | 9.2 | 4.7 |
| 80% | 4.7 | 5.1 | 5.3 | 10.1 | 5.2 |
| Training Ratio | A_A | A_B | B_A | B_B |
|---|---|---|---|---|
| 20% | 0.8452 | 0.8240 | 0.8279 | 0.8659 |
| 50% | 0.8977 | 0.8769 | 0.8792 | 0.8877 |
| 80% | 0.9135 | 0.8923 | 0.9077 | 0.9096 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Peng, F.; Guo, M.; Hu, H.; Yan, T.; Jiang, L. Off-Nadir Satellite Image Scene Classification: Benchmark Dataset, Angle-Aware Active Domain Adaptation, and Angular Impact Analysis. Remote Sens. 2025, 17, 3697. https://doi.org/10.3390/rs17223697
Peng F, Guo M, Hu H, Yan T, Jiang L. Off-Nadir Satellite Image Scene Classification: Benchmark Dataset, Angle-Aware Active Domain Adaptation, and Angular Impact Analysis. Remote Sensing. 2025; 17(22):3697. https://doi.org/10.3390/rs17223697
Chicago/Turabian StylePeng, Feifei, Mengchu Guo, Haoqing Hu, Tongtong Yan, and Liangcun Jiang. 2025. "Off-Nadir Satellite Image Scene Classification: Benchmark Dataset, Angle-Aware Active Domain Adaptation, and Angular Impact Analysis" Remote Sensing 17, no. 22: 3697. https://doi.org/10.3390/rs17223697
APA StylePeng, F., Guo, M., Hu, H., Yan, T., & Jiang, L. (2025). Off-Nadir Satellite Image Scene Classification: Benchmark Dataset, Angle-Aware Active Domain Adaptation, and Angular Impact Analysis. Remote Sensing, 17(22), 3697. https://doi.org/10.3390/rs17223697

