RSWFormer: A Multi-Scale Fusion Network from Local to Global with Multiple Stages for Regional Geological Mapping
Abstract
:1. Introduction
- High interclass similarity and intraclass variability in GERS elements. For example, diorite and granite often show similar texture and tone characteristics in remote sensing images, while arenaceous rock shows diverse spatial distribution patterns after experiencing long geological processes and changes.
- The boundaries of GERS elements such as rocks and soil are gradational. Interactions between different geological features make it difficult to accurately delineate boundaries.
- Distribution complexity and noise interference. As shown in Figure 1, the distribution space of different GERS elements is complex, and the non-uniform distribution varies greatly. In addition, disturbances in atmospheric conditions such as clouds and fog also significantly reduce the accuracy and reliability of remote sensing image data.
- To the best of our knowledge, this study is the first to propose a meter-level high-resolution GERS multi-source dataset (Multi-GL9) which is used in scene classification tasks to fill the lack of datasets of complex regional GERS elements, such as rocks, soils, and surface water. It can provide rich regional and large-scale GERS elements information and further use for the interpretation of complex geological environments.
- In order to meet the needs of regional GERS interpretation, a novel RSWFormer is proposed to capture local to global semantic information, extract high-dimensional feature information of GERS elements, and effectively fuse semantic features of different scales.
- Experiments were conducted on the Multi-GL9 dataset. Compared with existing scene classification models, RSWFormer achieved an overall accuracy improvement of 0.61% and 1.7% on the Gaofen-2 and Landsat-8 datasets, respectively.
2. Related Work
2.1. Datasets in the Field of GERS Interpretation
2.2. Image Classification Based on Deep Learning (CNNs and Transformer)
3. Methodology
3.1. Overall Architecture and Processes
- The MGHS module aims to extract high-dimensional geological features, in which the LTG block effectively captures and analyzes geological data at each stage to achieve the extraction of feature information from local to global. More details are presented in Section 3.2.
- The MGCE module enhances the understanding of contextual semantics by fusing geological semantic information at different scales. Detailed information is introduced in Section 3.3.
- The image size is and is first input into a Patchify stem (Inchannel = 3, Outchannel = embed_dim, Kernel_size = Stride = Patch_size, called a Patchify stem). It uses a convolution kernel (stride 4) to extract high-resolution features and outputs features with a channel count of 96 followed by a LayerNorm [54] (LN).
- In four consecutive stages (), the depth of the feature map is gradually deepened and its spatial dimension is reduced at the same time, aiming to capture more abstract and advanced geological remote sensing image features. In order to maintain a high spatial resolution and effectively fuse features and retain key feature information, downsampling is not performed in the . The output feature maps are labeled .
- performs dimension transformation through a size convolution kernel to unify them to the dimension . Subsequently, the feature maps are upsampled by 2 times one by one and added together to obtain . Parameter information at each stage is shown in Table 1.
- By upsampling to the same resolution as through bilinear interpolation and splicing in the channel dimension, finally, a feature with rich GERS details is formed to be used in the classification task.
3.2. Multi-Stage Geosemantic Hierarchical Sampling Module (MGHS)
- As shown in Figure 3a, in W-MSA, the feature map is divided into different windows according to (M = 7), and multi-head self-attention is used within each window. Using this method, the computational complexity is , which is smaller than the computational complexity of when using the MSA method.
- As shown in Figure 3b, in SW-MSA, the window position in W-MSA is offset by in the x and y directions (for example, Window_size = , then offset by 2 pixels). This is combined with the masked MSA method, and each offset window is calculated independently to improve the ability to capture global information.
- When i takes the value of 1, the initial feature () first passes through in (one characteristic of a Transformer is that it does not change the resolution and dimension of the input data).
- Secondly, when passing through the layer, it halves the H and W of the feature map and uses the linear transformation weight to reduce the merged patch channel dimension from the original four times to two times.
- Then, it generates feature , completing the conversion of doubling the quantity of channels and halving the spatial size.
- Finally, is used as new input data to further generate and then is used to generate .
3.3. Multi-Scale Geological Context Enhancement Module (MGCE)
- Firstly, considering the idea of feature fusion of different scales, we use (kernel size is ) to adjust the channel to for .
- Then, we obtain the by twice up-sampling operation, and the computational procedure is summarized as follows:
- Next, are dimensioned to the same width and height as using the bilinear interpolation method .
- Finally, the feature maps are connected in the channel dimension using the concatenation function .
- Afterwards, the feature maps are normalized by the LN layer. The calculations are as follows:
4. Experiments and Analysis
4.1. Study Area and Dataset
- Preprocessing of Figure 4a (stretching, sampling, and denoising).
- The uniqueness, diversity, and serious homogenization of GERS elements have led to a complex classification system and a lack of unified standards. Based on this, we first built on the theoretical foundation of constructing 9 and 13 types of GERS elements in [6,7] and then conducted in-depth research on a large number of geological survey reports (covering a wide area) based on field investigations and on-site data collection. Next, with the help of texture features captured by remote sensing satellites and the interaction between GERS elements, we analyzed the differences in tonal and texture features of the data in Figure 4a. Therefore, we finally divided the GERS elements into nine categories: slate, granite, gabbro, diorite, silt, surface water, soil, arenaceous rock, and carbonate rock.
- As shown in Figure 5a, crop GERS data in a top-to-bottom, left-to-right, non-overlapping window sliding manner, and the cropping size was set to in consideration of the high-spatial resolution of Gaofen-2, while the cropping size of Landsat-8 data was set to .
- When performing manual annotation, we specify the label category according to the principle of the largest area in the image block. For mixed transition zones, we divide them into a certain type of elements based on prior knowledge identification and the proportion of each geological element component; for the finely distributed and smaller areas, they are divided into existing categories or the category names are redefined based on geological theoretical knowledge.
4.2. Experimental Environment and Evaluation Metrics
4.3. Analysis of Results
4.4. Comparative Experiment of Multi-GL9 Dataset in the Same Area
4.4.1. Comparative Experiments and Analysis of Gaofen-2 Dataset
4.4.2. Comparative Experiments and Analysis of Landsat-8 Dataset
5. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Price, N.J.; Cosgrove, J.W. Analysis of Geological Structures; Cambridge University Press: Cambridge, UK, 1990. [Google Scholar]
- Sabins, F.F. Remote sensing for mineral exploration. Ore Geol. 1999, 14, 157–183. [Google Scholar] [CrossRef]
- Gutiérrez, F.; Parise, M.; Waele, J.D.; Jourde, H. A review on natural and human-induced geohazards and impacts in karst. Earth-Sci. Rev. 2014, 138, 61–88. [Google Scholar] [CrossRef]
- Goetz, A.F.; Rowan, L.C. Geologic remote sensing. Science 1981, 211, 781–791. [Google Scholar] [CrossRef]
- Han, W.; Zhang, X.; Wang, Y.; Wang, L.; Huang, X.; Li, J.; Wang, S.; Chen, W.; Li, X.; Feng, R.; et al. A survey of machine learning and deep learning in remote sensing of geological environment: Challenges, advances, and opportunities. Isprs J. Photogramm. Remote Sens. 2023, 202, 87–113. [Google Scholar] [CrossRef]
- Han, W.; Li, J.; Wang, S.; Zhang, X.; Dong, Y.; Fan, R.; Zhang, X.; Wang, L. Geological remote sensing interpretation using deep learning feature and an adaptive multisource data fusion network. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–14. [Google Scholar] [CrossRef]
- Wang, S.; Huang, X.; Han, W.; Li, J.; Zhang, X.; Wang, L. Lithological mapping of geological remote sensing via adversarial semi-supervised segmentation network. Int. J. Appl. Earth Obs. Geoinf. 2023, 125, 103536. [Google Scholar] [CrossRef]
- Hearst, M.; Dumais, S.; Osuna, E.; Platt, J.; Scholkopf, B. Support vector machines. IEEE Intell. Syst. Their Appl. 1998, 13, 18–28. [Google Scholar] [CrossRef]
- Peterson, L.E. K-nearest neighbor. Scholarpedia 2009, 4, 1883, revision #137311. [Google Scholar] [CrossRef]
- Shlens, J. A tutorial on principal component analysis. arXiv 2014, arXiv:1404.1100. [Google Scholar]
- Fels, A.E.A.E.; Ghorfi, M.E. Using remote sensing data for geological mapping in semi-arid environment: A machine learning approach. Earth Sci. Inform. 2022, 15, 485–496. [Google Scholar] [CrossRef]
- Cracknell, M.J.; Reading, A.M. Geological mapping using remote sensing data: A comparison of five machine learning algorithms, their response to variations in the spatial distribution of training data and the use of explicit spatial information. Comput. Geosci. 2014, 63, 22–33. [Google Scholar] [CrossRef]
- Rigatti, S.J. Random forest. J. Insur. Med. 2017, 47, 31–39. [Google Scholar] [CrossRef] [PubMed]
- Zuo, W.; Zeng, X.; Gao, X.; Zhang, Z.; Liu, D.; Li, C. Machine learning fusion multi-source data features for classification prediction of lunar surface geological units. Remote Sens. 2022, 14, 5075. [Google Scholar] [CrossRef]
- Rezaei, A.; Hassani, H.; Moarefvand, P.; Golmohammadi, A. Lithological mapping in sangan region in northeast iran using aster satellite data and image processing methods. Geol. Ecol. Landsc. 2020, 4, 59–70. [Google Scholar] [CrossRef]
- Kruse, F.A.; Lefkoff, A.B.; Boardman, J.W.; Heidebrecht, K.B.; Shapiro, A.; Barloon, P.J.; Goetz, A.F.H. The Spectral Image Processing System (Sips): Software for Integrated Analysis of Aviris Data. 1992. Available online: https://api.semanticscholar.org/CorpusID:60199064 (accessed on 24 June 2024).
- Yuan, Q.; Shen, H.; Li, T.; Li, Z.; Li, S.; Jiang, Y.; Xu, H.; Tan, W.; Yang, Q.; Wang, J.; et al. Deep learning in environmental remote sensing: Achievements and challenges. Remote Sens. Environ. 2020, 241, 111716. [Google Scholar] [CrossRef]
- Han, W.; Chen, J.; Wang, L.; Feng, R.; Li, F.; Wu, L.; Tian, T.; Yan, J. Methods for small, weak object detection in optical high-resolution remote sensing images: A survey of advances and challenges. IEEE Geosci. Remote Sens. Mag. 2021, 9, 8–34. [Google Scholar] [CrossRef]
- Han, W.; Feng, R.; Wang, L.; Cheng, Y. A semi-supervised generative framework with deep learning features for high-resolution remote sensing image scene classification. Isprs J. Photogramm. Remote Sens. 2018, 145, 23–43. [Google Scholar] [CrossRef]
- Zhang, W.; Tang, P.; Zhao, L. Remote sensing image scene classification using cnn-capsnet. Remote Sens. 2019, 11, 494. [Google Scholar] [CrossRef]
- Wang, Z.; Zuo, R.; Liu, H. Lithological mapping based on fully convolutional network and multi-source geological data. Remote Sens. 2021, 13, 4860. [Google Scholar] [CrossRef]
- Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
- Brandmeier, M.; Chen, Y. Lithological classification using multi-sensor data and convolutional neural networks. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2019, 42, 55–59. [Google Scholar] [CrossRef]
- Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015: 18th International Conference, Part III 18, Munich, Germany, 5–9 October 2015; Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
- Hajaj, S.; Harti, A.E.; Jellouli, A.; Pour, A.B.; Himyari, S.M.; Hamzaoui, A.; Hashim, M. Evaluating the performance of machine learning and deep learning techniques to hymap imagery for lithological mapping in a semi-arid region: Case study from western anti-atlas, morocco. Minerals 2023, 13, 766. [Google Scholar] [CrossRef]
- der Meer, F.D.V.; Werff, H.M.V.D.; Ruitenbeek, F.J.V.; Hecker, C.A.; Bakker, W.H.; Noomen, M.F.; Meijde, M.V.D.; Carranza, E.J.M.; Smeth, J.B.; Woldai, T. Multi-and hyperspectral geologic remote sensing: A review. Int. J. Appl. Earth Obs. Geoinf. 2012, 14, 112–128. [Google Scholar] [CrossRef]
- Cheng, G.; Han, J.; Lu, X. Remote sensing image scene classification: Benchmark and state of the art. Proc. IEEE 2017, 105, 1865–1883. [Google Scholar] [CrossRef]
- Zhang, Y.; Li, M.; Han, S. Automatic identification and classification in lithology based on deep learning in rock images. Yanshi Xuebao/Acta Petrol. Sin. 2018, 34, 333–342. [Google Scholar]
- Kumar, C.; Chatterjee, S.; Oommen, T.; Guha, A.; Mukherjee, A. Multi-sensor datasets-based optimal integration of spectral, textural, and morphological characteristics of rocks for lithological classification using machine learning models. Geocarto Int. 2022, 37, 6004–6032. [Google Scholar] [CrossRef]
- Li, D.; Zhao, J.; Liu, Z. A novel method of multitype hybrid rock lithology classification based on convolutional neural networks. Sensors 2022, 22, 1574. [Google Scholar] [CrossRef] [PubMed]
- Shi, H.; Xu, Z.; Lin, P.; Ma, W. Refined lithology identification: Methodology, challenges and prospects. Geoenergy Sci. Eng. 2023, 231, 212382. [Google Scholar] [CrossRef]
- Wang, S.; Han, W.; Zhang, X.; Li, J.; Wang, L. Geospatial remote sensing interpretation: From perception to cognition. Innov. Geosci. 2024, 2, 100056. [Google Scholar] [CrossRef]
- Wang, S.; Han, W.; Huang, X.; Zhang, X.; Wang, L.; Li, J. Trustworthy remote sensing interpretation: Concepts, technologies, and applications. ISPRS J. Photogramm. Remote Sens. 2024, 209, 150–172. [Google Scholar] [CrossRef]
- Han, W.; Li, J.; Wang, S.; Wang, Y.; Yan, J.; Fan, R.; Zhang, X.; Wang, L. A context-scale-aware detector and a new benchmark for remote sensing small weak object detection in unmanned aerial vehicle images. Int. J. Appl. Earth Obs. Geoinf. 2022, 112, 102966. [Google Scholar] [CrossRef]
- Zhang, X.; Wang, L.; Li, J.; Han, W.; Fan, R.; Wang, S. Satellite-derived sediment distribution mapping using icesat-2 and superdove. Isprs J. Photogramm. Remote. Sens. 2023, 202, 545–564. [Google Scholar] [CrossRef]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar]
- Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is all you need. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Ser. NIPS’17. Curran Associates Inc.: Red Hook, NY, USA, 2017; pp. 6000–6010. [Google Scholar]
- Latifovic, R.; Pouliot, D.; Campbell, J. Assessment of convolution neural networks for surficial geology mapping in the south rae geological region, northwest territories, canada. Remote Sens. 2018, 10, 307. [Google Scholar] [CrossRef]
- Sang, X.; Xue, L.; Ran, X.; Li, X.; Liu, J.; Liu, Z. Intelligent high-resolution geological mapping based on slic-cnn. Isprs Int. J. Geo-Inf. 2020, 9, 99. [Google Scholar] [CrossRef]
- Shirmard, H.; Farahbakhsh, E.; Heidari, E.; Pour, A.B.; Pradhan, B.; Müller, D.; Chandra, R. A comparative study of convolutional neural networks and conventional machine learning models for lithological mapping using remote sensing data. Remote Sens. 2022, 14, 819. [Google Scholar] [CrossRef]
- Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
- Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
- Huang, G.; Liu, Z.; Maaten, L.V.D.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
- Pascual, A.D.P.; Shu, L.; Szoke-Sieswerda, J.; McIsaac, K.; Osinski, G. Towards natural scene rock image classification with convolutional neural networks. In Proceedings of the 2019 IEEE Canadian Conference of Electrical and Computer Engineering (CCECE), Edmonton, AB, Canada, 5–8 May 2019; pp. 1–4. [Google Scholar]
- Lynda, N.O. Systematic survey of convolutional neural network in satellite image classification for geological mapping. In Proceedings of the 2019 15th International Conference on Electronics, Computer and Computation (ICECCO), Abuja, Nigeria, 10–12 December 2019; pp. 1–6. [Google Scholar]
- Fu, D.; Su, C.; Wang, W.; Yuan, R. Deep learning based lithology classification of drill core images. PLoS ONE 2022, 17, e0270826. [Google Scholar] [CrossRef]
- Huang, Z.; Su, L.; Wu, J.; Chen, Y. Rock image classification based on efficientnet and triplet attention mechanism. Appl. Sci. 2023, 13, 3180. [Google Scholar] [CrossRef]
- Dong, Y.; Yang, Z.; Liu, Q.; Zuo, R.; Wang, Z. Fusion of gaofen-5 and sentinel-2b data for lithological mapping using vision transformer dynamic graph convolutional network. Int. J. Appl. Earth Obs. Geoinf. 2024, 129, 103780. [Google Scholar] [CrossRef]
- Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
- Koeshidayatullah, A.; Al-Azani, S.; Baraboshkin, E.E.; Alfarraj, M. Faciesvit: Vision transformer for an improved core lithofacies prediction. Front. Earth Sci. 2022, 10, 992442. [Google Scholar] [CrossRef]
- Jacinto, M.V.G.; Silva, M.A.; de Oliveira, L.H.L.; Medeiros, D.R.; de Medeiros, G.C.; Rodrigues, T.C.; de Montalvão, L.C.; de Almeida, R.V. Lithostratigraphy modeling with transformer-based deep learning and natural language processing techniques. In Proceedings of the Abu Dhabi International Petroleum Exhibition and Conference, Abu Dhabi, United Arab Emirates, 2–5 October 2023; p. D031S110R003. [Google Scholar]
- Cheng, L.; Keyan, X.; Li, S.; Rui, T.; Xuchao, D.; Baocheng, Q.; Dahong, X. Cnn-transformers for mineral prospectivity mapping in the maodeng—Baiyinchagan area, southern great xing’an range. Ore Geol. Rev. 2024, 167, 106007. [Google Scholar]
- Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Virtual, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
- Ba, J.L.; Kiros, J.R.; Hinton, G.E. Layer normalization. arXiv 2016, arXiv:1607.06450. [Google Scholar]
- Lin, T.-Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
- Liu, S.; Qi, L.; Qin, H.; Shi, J.; Jia, J. Path aggregation network for instance segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, June 18–22 June 2018; pp. 8759–8768. [Google Scholar]
- Tan, M.; Le, Q.V. Efficientnet: Rethinking model scaling for convolutional neural networks. In Proceedings of the 36th International Conference on Machine Learning, ICML 2019, Long Beach, CA, USA, 9–15 June 2019; Chaudhuri, K., Salakhutdinov, R., Eds.; Volume 97, pp. 6105–6114. [Google Scholar]
- Radosavovic, I.; Kosaraju, R.P.; Girshick, R.; He, K.; Dollár, P. Designing network design spaces. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–18 June 2020; pp. 10428–10436. [Google Scholar]
- Yuan, K.; Guo, S.; Liu, Z.; Zhou, A.; Yu, F.; Wu, W. Incorporating convolution designs into visual transformers. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Virtual, 11–17 October 2021; pp. 579–588. [Google Scholar]
- Touvron, H.; Cord, M.; Douze, M.; Massa, F.; Sablayrolles, A.; Jégou, H. Training data-efficient image transformers & distillation through attention. In Proceedings of the 38th International Conference on Machine Learning, ICML 2021, Virtual Event, 18–24 July 2021; Meila, M., Zhang, T., Eds.; Volume 139, pp. 10347–10357. [Google Scholar]
- Wang, W.; Xie, E.; Li, X.; Fan, D.-P.; Song, K.; Liang, D.; Lu, T.; Luo, P.; Shao, L. Pyramid vision transformer: A versatile backbone for dense prediction without convolutions. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Virtual, 11–17 October 2021; pp. 568–578. [Google Scholar]
/ | Outchannel | Iterations |
---|---|---|
Stage1 | 192 | 2 |
Stage2 | 384 | 2 |
Stage3 | 768 | 6 |
Stage4 | 768 | 2 |
Serial No. | Name | Gaofen-2 | Landsat-8 |
---|---|---|---|
1 | Slate | 207 | 103 |
2 | Granite | 207 | 101 |
3 | Diorite | 175 | 82 |
4 | Gabbro | 148 | 102 |
5 | Arenaceous Rock | 267 | 102 |
6 | Soil | 247 | 144 |
7 | Surface Water | 167 | 65 |
8 | Silt | 105 | 100 |
9 | Carbonate Rock | 145 | 100 |
Total number of datasets | 1668 | 899 |
Experimental Platform | Parameter Name |
---|---|
GPU | RTX 4090 |
CPU | Intel (R) Xeon (R) Gold 6430 |
Memory | 32 GB |
DL Framework | Pytorch 1.10.0 |
Operation System | Windows 10 |
Programming language | Python 3.8 |
Dataset | SL | GR | DI | GA | AR | SO | SW | SI | CR | OA |
---|---|---|---|---|---|---|---|---|---|---|
Gaofen-2 | 90.24 | 90.24 | 94.29 | 100.0 | 77.36 | 97.96 | 100.0 | 95.24 | 93.10 | 92.15 |
Landsat-8 | 95.00 | 85.00 | 56.25 | 75.00 | 40.00 | 82.14 | 100.0 | 100.0 | 90.00 | 80.23 |
Model | SL | GR | DI | GA | AR | SO | SW | SI | CR | OA |
---|---|---|---|---|---|---|---|---|---|---|
EfficientNet | 70.73 | 87.80 | 40.00 | 58.62 | 39.62 | 83.67 | 96.97 | 95.24 | 82.67 | 70.69 |
RegNet | 73.17 | 87.80 | 74.29 | 58.62 | 45.28 | 85.71 | 100.0 | 90.48 | 82.76 | 75.83 |
CeiT | 82.93 | 82.93 | 80.00 | 86.21 | 64.15 | 93.88 | 96.97 | 90.48 | 72.41 | 82.48 |
DeiT | 85.37 | 90.24 | 68.57 | 96.55 | 73.58 | 93.88 | 100.0 | 100.0 | 72.41 | 85.80 |
ViT | 92.68 | 97.65 | 88.57 | 89.66 | 86.79 | 95.92 | 96.97 | 100.0 | 72.41 | 91.24 |
PVT | 87.80 | 95.12 | 85.71 | 93.10 | 88.68 | 95.92 | 100.0 | 100.0 | 79.31 | 91.54 |
RSWFormer | 90.24 | 90.24 | 94.29 | 100.0 | 77.36 | 97.96 | 100.0 | 95.24 | 93.10 | 92.15 |
Model | SL | GR | DI | GA | AR | SO | SW | SI | CR | OA |
---|---|---|---|---|---|---|---|---|---|---|
EfficientNet | 70.00 | 45.00 | 31.25 | 30.00 | 50.00 | 39.29 | 76.92 | 80.00 | 65.00 | 53.11 |
RegNet | 80.00 | 75.00 | 50.00 | 50.00 | 60.00 | 75.00 | 92.31 | 60.00 | 65.00 | 67.23 |
CeiT | 70.00 | 100.0 | 50.00 | 45.00 | 15.00 | 71.43 | 92.31 | 70.00 | 55.00 | 62.71 |
DeiT | 75.00 | 95.00 | 65.50 | 70.00 | 95.00 | 75.00 | 100.0 | 85.00 | 95.00 | 76.27 |
ViT | 95.00 | 95.00 | 68.75 | 60.00 | 30.00 | 75.00 | 100.0 | 95.00 | 95.00 | 78.53 |
PVT | 90.00 | 95.00 | 50.00 | 70.00 | 45.00 | 75.00 | 100.0 | 90.00 | 95.00 | 78.53 |
RSWFormer | 95.00 | 85.00 | 56.25 | 75.00 | 40.00 | 82.14 | 100.0 | 100.0 | 90.00 | 80.23 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Han, S.; Wan, Z.; Deng, J.; Zhang, C.; Liu, X.; Zhu, T.; Zhao, J. RSWFormer: A Multi-Scale Fusion Network from Local to Global with Multiple Stages for Regional Geological Mapping. Remote Sens. 2024, 16, 2548. https://doi.org/10.3390/rs16142548
Han S, Wan Z, Deng J, Zhang C, Liu X, Zhu T, Zhao J. RSWFormer: A Multi-Scale Fusion Network from Local to Global with Multiple Stages for Regional Geological Mapping. Remote Sensing. 2024; 16(14):2548. https://doi.org/10.3390/rs16142548
Chicago/Turabian StyleHan, Sipeng, Zhipeng Wan, Junfeng Deng, Congyuan Zhang, Xingwu Liu, Tong Zhu, and Junli Zhao. 2024. "RSWFormer: A Multi-Scale Fusion Network from Local to Global with Multiple Stages for Regional Geological Mapping" Remote Sensing 16, no. 14: 2548. https://doi.org/10.3390/rs16142548
APA StyleHan, S., Wan, Z., Deng, J., Zhang, C., Liu, X., Zhu, T., & Zhao, J. (2024). RSWFormer: A Multi-Scale Fusion Network from Local to Global with Multiple Stages for Regional Geological Mapping. Remote Sensing, 16(14), 2548. https://doi.org/10.3390/rs16142548