Next Article in Journal
An Image Quality Improvement Method in Side-Scan Sonar Based on Deconvolution
Previous Article in Journal
PM2.5 Estimation in Day/Night-Time from Himawari-8 Infrared Bands via a Deep Learning Neural Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Technical Note

A Seabed Terrain Feature Extraction Transformer for the Super-Resolution of the Digital Bathymetric Model

1
The First Institute of Oceanography, Ministry of Natural Resources, Qingdao 266061, China
2
The Key Laboratory of Ocean Geomatics, Ministry of Natural Resources, Qingdao 266590, China
3
State Key Laboratory of Information Engineering in Surveying, Mapping and Remote Sensing, Wuhan University, Wuhan 430079, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(20), 4906; https://doi.org/10.3390/rs15204906
Submission received: 27 July 2023 / Revised: 25 September 2023 / Accepted: 26 September 2023 / Published: 11 October 2023

Abstract

:
The acquisition of high-resolution (HR) digital bathymetric models (DBMs) is crucial for oceanic research activities. However, obtaining HR DBM data is challenging, which has led to the use of super-resolution (SR) methods to improve the DBM’s resolution, as, unfortunately, existing interpolation methods for DBMs suffer from low precision, which limits their practicality. To address this issue, we propose a seabed terrain feature extraction transform model that combines the seabed terrain feature extraction module with the efficient transform module, focusing on the terrain characteristics of DBMs. By taking advantage of these two modules, we improved the efficient extraction of seabed terrain features both locally and globally, and as a result, we obtained a highly accurate SR reconstruction of DBM data within the study area, including the Mariana Trench in the Pacific Ocean and the adjacent sea. A comparative analysis with bicubic interpolation, SRCNN, SRGAN, and SRResNet shows that the proposed method decreases the root mean square error (RMSE) by 16%, 10%, 13%, and 12%, respectively. These experimental results confirm the high accuracy of the proposed method in terms of reconstructing HR DBMs.

1. Introduction

Bathymetry provides fundamental information for all marine activities, and numerous marine scientific research endeavors rely on digital bathymetry model (DBM) data. High-resolution (HR) DBMs are commonly used for charting purposes [1]; at the same time, they play a crucial role in establishing hydrodynamic fluid models [2], constructing marine biological habitats and ecosystems [3], and other significant studies. In addition, HR DBMs are valuable for maritime search and rescue operations. For example, the search and rescue operations for Malaysia Airlines flight MH370 were severely hampered by the lack of accurate bathymetric data in the corresponding area [4]. Therefore, the acquisition of HR DBMs is of immense importance not only for scientific research but also for various social activities. However, the scarcity of HR DBMs severely limits their application and development.
HR DBMs are typically derived from ship-based acoustic surveys. On the one hand, these surveys face challenges and inefficiencies, resulting in the limited coverage of HR bathymetric data. According to statistics [5], ship-borne sonar-based HR data (≤800 m) cover only approximately 6.2% of the global seafloor. Consequently, the availability of HR bathymetric data to construct DBMs is severely limited for most of the global ocean. On the other hand, many institutions have obtained global low-resolution (LR) DBM data, including EPOTO and GEBCO. To address the shortage of HR DBMs, researchers have employed various super-resolution (SR) methods to improve the resolution of DBMs; DBM SR aims to recover detailed topographic information from LR DBMs to generate HR DBMs. Currently, SR methods for DBMs mainly rely on traditional interpolation techniques, such as inverse distance-weighted interpolation [6], bilinear interpolation [7], bicubic spline interpolation [8], continuous curvature tensor spline interpolation [9], natural neighborhood interpolation [10], and kriging interpolation [11]. Although these interpolation methods are widely used due to their simplicity and speed, they can present certain problems, particularly when generating large-scale DBMs. One of their main drawbacks is the potential to be over-smooth, as these methods do not adequately exploit local terrain features during the interpolation process [12]. In addition to traditional interpolation methods, many researchers have employed deep learning-based methods to enhance the resolution of terrestrial digital elevation models (DEMs). These methods have the advantage of not relying on additional multisource data and instead generate HR DEMs solely through the reconstruction of LR DEMs [13]. Similarly, in situations in which data are limited, the resolution of DBMs can theoretically be improved using deep learning-based SR methods. This approach not only addresses the issue of insufficient data but also serves as a valuable source of HR data for DBM data fusion.
Current deep learning-based SR models are mainly based on convolutional neural networks (CNNs) and other networks based on CNNs, such as residual networks (ResNets), generative adversarial networks (GANs), and transformers. The pioneering CNN model for image SR is SRCNN, which consists of three convolutional layers that perform feature extraction, non-linear mapping, and image reconstruction to enhance image resolution [14]. The advancement of neural networks has highlighted their crucial role in feature extraction; however, the stacking of network layers can reduce the effectiveness of this network and lead to degradation problems [15]. To address this problem, He et al. [16] employed ResNet to extract shallow features and transmit mapping to the deep network. This identification mapping does not require any additional parameters or calculations, ensuring that the characteristics of this network do not deteriorate as the number of layers increases. In the context of DEM SR, Chen et al. [17] were the first to apply SRCNN to DEM data for SR reconstruction. Subsequently, other researchers have demonstrated the strong feature extraction capabilities and applicability of ResNet to DEM SR [18,19]. Peak signal-to-noise ratio (PSNR) evaluations of SR reconstruction accuracy are frequently performed in the CV domain. However, studies have revealed a mixed relationship between PSNR and visual quality, suggesting that a greater PSNR does not always translate into superior visual perception [20]. In order to enhance perception, Ledig et al. [20] integrated a perceptual loss function with a generative adversarial network for SR. Given the distinct characteristics of DEM data compared to natural images, Zhang and Yu [21] applied SRGAN, ESRGAN, and CEDGAN to DEM SR, and the experimental results showed that SRGAN was the most effective in terms of extracting terrain features, outperforming other GANs. This result suggests that neural networks with superior performance in the CV domain might not be equally effective when extracting topographic features. The current focus of DEM SR research is centered on fully and effectively extracting terrain features. Zhou et al. [22] introduced dual filters into ResNet to accomplish this goal, while Chen et al. [12] incorporated a spatial attention mechanism module into CNN to improve its feature extraction efficiency. Zhang et al. [23] proposed the addition of slope loss to the loss function, which significantly improved feature extraction effectiveness. Zhou et al. [24] integrated vector terrain features before outputting HR DEM, which significantly improved the accuracy of the model. To solve practical problems, researchers are now placing their research objects in hard-to-access topographic data. Jiang et al. [25] used high mountain data for the study data and constructed a new loss function in ResNet, which combined the terrain parameters of slope and curvature.
However, the development of global high-precision HR DBMs has been slow due to the challenges experienced when obtaining bathymetric data, particularly the lack of HR data. The resolution of HR DBMs is significantly lower than that of terrestrial DEMs, and the application of SR research based on deep learning in DBMs is limited. Yutani et al. [26] utilized sparse coding and dictionary learning techniques for SR in DBMs, resulting in improved results compared to bicubic interpolation, and Hidaka et al. [27] employed deep learning for SR in DBMs for the first time and compared the SR reconstruction effects of various deep learning-based methods, including SRCNN, FSRCNN, ESPCN, SRGAN, and ESRGAN. The results were superior to bicubic interpolation but did not contribute to building a deep learning-based model for DBM terrain characteristics. Additionally, Zhang et al. [28] enhanced the resolution of the global GEBCO_2021 dataset from 15 arc seconds to 3 arc seconds using ResNet and migration learning. Although many deep learning-based SR methods utilize ResNet for DEMs, there are some general issues with this approach. First, the use of regular convolution kernels in ResNet might not be optimal for recovering different topographic features due to the irregular terrain. Second, the convolution operation does not adequately consider the influence of distant pixels. The transformer, with its self-attention mechanism, possesses strong capabilities in terms of capturing global information and long-range interactions among similar features, making it successful in various visual tasks. Some researchers [29] have employed this transformer in the computer vision domain to enhance SR accuracy. However, this transformer is currently underutilized in SR for terrain models. Zheng et al. [30] designed a transformer model for SR in DEMs via leveraging terrain self-similarity. The reconstruction effect was superior to SRCNN and SRGAN, but improvements specific to local features were not made. Additionally, SR is a low-level vision task, and using a complete transformer structure can reduce computational efficiency. Therefore, modifications to the transformer structure are necessary for its application in DBM SR.
In this paper, considering the complex and extensive nature of DBM seabed terrain features, a seabed terrain feature extraction transformer (STFET) is proposed to achieve DBM SR, combining ResNet with deformable convolutional layers and an efficient transformer. The following are this paper’s key contributions:
  • This study is an early attempt to restore HR DBMs from LR DBMs using a transformer. In addition, we utilize the proposed transformer-based model, combined with ResNet and deformable convolutional layers, which ensures that STFET can capture both local and global seafloor topographic features.
  • Given the characteristics of large changes and rapid fluctuations in the DBM terrain, the traditional convolutional layers in ResNet are replaced by deformable convolutional layers, which have the ability to flexibly modify the sampling position of the convolution kernel to align with irregular features, improving the extraction of local seabed terrain features.
  • The transformer’s self-attention mechanism allows long-term dependencies to be established between similar regions in the DBM, improving global terrain feature reconstruction. In addition, the matrix is broken down into smaller matrices for parameter manipulation, increasing the speed of model operations.

2. Methodology

The StfeT method includes the following: (1) data-processing; (2) training in the proposed model; and (3) testing and evaluation. The workflow is summarized in Figure 1. More detail is provided in other sections.

2.1. Network Architecture

In view of the complex terrain characteristics of DBMs, in this paper, we propose a seabed terrain feature extraction transformer model for DBM SR. This model mainly consists of three parts: shallow terrain feature extraction, deep feature extraction, including the seabed terrain feature extraction module and the efficient transformer, and sub-pixel convolution reconstruction, as shown in Figure 2. The utilization of convolutional layers offers a straightforward and efficient approach to incorporating feature maps, demonstrating a consistent performance across various computer vision tasks.
To extract the shallow features, we employed a regular convolutional layer f N (channel = 64, kernel size = 9, stride = 1) as the first STFET module to extract the DBM feature F s h a l l o w :
F s h a l l o w = f N I L R
where I L R represents LR DBM. After extracting the shallow features to extract deep features, we employed the seabed terrain feature extraction module and the effective transformer module. Seabed terrain features were extracted adaptively using ResNet with deformable convolutional layers, and the intermediate features are represented by A 1 , A 2 ,…, A n :
A i = R e s A i 1 + A i 1 , i = 1,2 , , n
A 1 = R e s F s h a l l o w + F s h a l l o w
F d e e p = E T A n
where R e s represents a residual block with deformable convolutional layers. Through residual learning, seabed terrain features could be effectively extracted, and the deformable convolutional layer could focus on high-frequency information, thereby enhancing the extraction of local features. E T represents the efficient transformer, and the feature maps after residual learning were learned using E T . Through the self-attention mechanism, the correlation between similar seabed terrain features was modeled, allowing the model to focus on other positions with similar seabed terrain features to extract global seabed terrain features. The deep feature F d e e p was obtained through these two modules.
After obtaining the deep seabed terrain features, the conventional convolutional layers were used to extract and transform the high-frequency features in the residual information. B N is batch normalization, which is then used to hasten model convergence and improve the model’s robustness. Furthermore, a skip connection is used to connect the shallow topographic features with deep topographic features to obtain all topographic features F a l l .
F a l l = B N f N F d e e p + F s h a l l o w
Finally, the LR feature maps were transformed into HR feature maps using sub-pixel convolution which were reconstructed to generate I S R , which is SR DBM. During sub-pixel convolution, the HR feature maps were obtained using convolution and multichannel recombination, avoiding a large number of zero-filled areas in general deconvolution upsampling. Moreover, we used the dropout to prevent model overfitting before reconstruction.
I S R = f P F d e e p
where f P is a sub-pixel convolutional layer. STFET uses a large number of convolutional layers to extract and transform features. The local seabed terrain features and global seabed terrain features were extracted mainly through ResNet with deformable convolutional layers and the efficient transformer. Finally, a sub-pixel convolutional layer was used to reconstruct the SR DBM.

2.2. Seabed Terrain Feature Extraction Module

DBM often has clear topographic undulations in local areas, generating different topographies, and the depth distribution of DBM is closely related to the topographic features of local areas. The effective extraction of seafloor topographic features can provide more accurate feature maps, thereby improving the quality of SR DBM. Very deep convolutional networks (VDSRs) [31] were the first to use ResNet for the SR domain of images, as they can utilize both deeper neural networks and mitigate the problems of disappearing or exploding gradients and network degradation. Seabed terrain features have different shapes, and the convolution kernel used in the traditional convolution layer is usually in the shape of an n × n regular grid, which makes it difficult to extract complicated seabed terrain features. To extract complex-shaped objects or terrain features, researchers have replaced conventional convolutional layers with deformable convolutional layers [32,33]. Deformable convolution adds an offset to the regular convolutional layer, which is implemented using an additional convolutional layer, as shown in Figure 3. In order to always be able to cover complicated terrain features, the deformable convolution mechanism makes sure that offsets can be learned during the learning process. Therefore, the deformable convolution layer can improve its ability to extract terrain features. In this paper, we replaced regular convolutional layers in the residual network with deformable convolutional layers, and the feature extraction process of the conventional convolutional layer and the deformable convolutional layer in the DBM is shown in Figure 4. When the conventional convolutional layer extracts terrain features because the convolution kernel is regular, the edge terrain features are easily overlooked, whereas the deformable convolutional layer can adaptively extract seabed terrain features by learning during training.
The seabed terrain feature extraction module consists of N residual blocks. In previous studies, the higher the number of residual blocks, the higher the accuracy, but at the same time, it increases the computation time. Therefore, it is necessary to balance the reconstruction accuracy and computation time to choose an appropriate number of residual blocks. Each residual block has five layers: a ReLU activation function, two deformable convolutional layers (channel = 64, kernel size = 3, stride = 1), and two batch normalization layers. The deformable convolutional layer is used to extract high-frequency details of the DBM to adaptively extract seabed terrain features. DBMs are different from natural images because the difference between the maximum value and the minimum value is often on the kilometer scale, and batch normalization can reduce the impact of extreme water depth. Furthermore, the ReLU activation function can enhance the non-linear relationship between layers of the neural network, which can improve the model’s fitting ability. Therefore, the seabed terrain feature extraction module can effectively extract topographic features from the seafloor.

2.3. Efficient Transformer Module

According to the first law of geography, the closer things are, the more related they are. Therefore, in close regions, DBMs generally have a similar topography globally. The texture details of the current image block can be recovered by referring to other image blocks in single-image super-resolution (SISR), which allows similar image blocks in an image to be used as references for one another. The transformer self-attention mechanism can realize this function very well, so the transformer has a theoretical basis to recover the current area by referring to other areas to extract similar terrain features.
The structure of the efficient transformer used in this paper is shown in Figure 5. Similar to the Vision Transformer (ViT) [34], only the encoding part in the standard transformer is used, and it is composed of Multi-Layer Perceptron (MLP) and efficient multi-head self-attention (EMSA). At the same time, we used layer normalization (LN) before each module to accelerate the convergence speed of the model. Then, after each module, the residual connection was achieved on an element-by-element basis. This process can be represented using the following formula:
E = E M S A L N U n A n + U n A n
F d e e p = F o l d M L P L N E + E
where U n denotes the unfold function, whose purpose is to generate a local sliding block from the feature map extracted by the seabed terrain feature extraction module. E represents the feature extracted using E M S A , and F o l d denotes the inverse operation of U n , which converts the sliding block into a feature map.
The EMSA structure is shown in Figure 5. First, features were embedded in query Q, key K, and value V using a linear layer. As in the transformer, we projected Q, K, and V m times, where m is the multi-head number, and the multi-head self-attention (MSA) directly computes Q, K, and V. The computations typically use a large matrix, which consumes a large amount of GPU memory and computation. Unlike MSA, the DBM matrix we used for training to extract terrain features is relatively large due to the LR of DBM and the large scale of the terrain features. Therefore, we used feature splitting to divide Q, K, and V into s parts to obtain Qs, Ks, and Vs, where s is the splitting factor. Then, we use Scaled Dot Product Attention (SDPA) to perform attention to generate multiple outputs O1...Os. This process is shown on the right side of Figure 5. Finally, all outputs were concatenated to produce the overall output feature O.

3. Experimental Setup

3.1. Network Hyperparameters

In the proposed STFET, to ensure accuracy and computational efficiency, the number of residual blocks N was set to 16. The multi-head number in the efficient transformer module was set to 8, and the hidden layer size of the MLP was m l p d i m = 85, and was obtained using the following formula:
m l p d i m = i n c h a n n e l + o u t c h a n n l × 2 / 3
where i n c h a n n e l and o u t c h a n n e l are the numbers of MLP input channels and output channels, respectively, which are both 64. To train the proposed STFET, we used adaptive moment estimation (Adam) with a 0.0002 learning rate, a random seed of 42, and a dropout rate of 0.2. The loss in STFET was the L2 loss MSE. For CNN-based algorithms, all the models were coded using the PyTorch framework. In addition, we used 8 GB of RAM and trained the models on a deep-learning server with an Nvidia RTX 2070S GPU. Moreover, the hyperparameters were adjusted to retrain the comparison models in the optimal way possible.

3.2. Study Area and Data

Unlike natural images, DBM is a grayscale image with only one channel. DBMs tend to have topographic features. In this study, the Mariana Trench in the Pacific Ocean and the surrounding waters to the east were selected as the experimental area, constituting a range of 0°−45°N, 135°–180°E. This area is large and rich in seafloor topography and essentially includes most of the typical seafloor topographic features, such as ridges, trenches, seamounts, sea basins, and other representative seafloor topographic features. The values of natural images are pixel values, ranging from 0 to 255. However, the elevation of our study area was between −10,923 and 3659 m. The study area is shown in Figure 6.
The DBM data used in this paper were gebco_2022_sub_ice, sourced from https://www.gebco.net/data_and_products/gridded_bathymetry_data/gebco_2022 (accessed on 10 September 2023), which has a resolution of 15 arc seconds. Since original DBM data are large, it was necessary to preprocess them to train the data. We cropped the original DBM data into 128 × 128 patches and removed most of the land area, and a total of 6386 DBM images were obtained as HR DBMs. To obtain paired HR DBMs and LR DBMs as datasets for training, many previous DEM SR studies were scaled 4 times for verification experiments [12,22,24]. In this study, we also used bicubic interpolation to reduce the data size 4-fold and obtain 128 × 128 patch DBM images with resolutions of 60 arc seconds as LR DBMs. The ratio of the training set data, verification set data, and test set data was 8:1:1 through random allocation. Before training, due to the large difference in the DBM water depth values DBMs needed to be standardized, and therefore, each DBM water depth value was standardized to [−1,1]. The normalization formula is expressed as follows:
D B M i = 2 × ( D B M i H m i n ) / ( H m a x H m i n ) 1
where H m i n is the minimum value of DBM and H m a x is the maximum value of DBM.
Since the format of DBMs is usually .tif, there was only one channel, while natural images had three channels. Therefore, before inputting data into the model, we needed to set the number of input channels in the convolutional layer to 1 for training. Similarly, it was necessary to set the number of output channels of the convolutional layer to 1 before reconstruction.

3.3. Evaluation Methods

To test the effectiveness of STFET, we selected bicubic interpolation [35], SRCNN, SRGAN, and SRResNet (SRGAN generator) as comparison methods. Bicubic interpolation is one of the simplest and most effective interpolation methods and is often used as a baseline method for SR [21]. The first deep learning-based SR method was SRCNN. This network is simple and effective, and some scholars have proven its practicality in DEM SR. SRGAN is the first model to employ the generation confrontation network for the SR task, and it emphasizes that high accuracy does not always equate to superior visual quality using the perceptual loss function to enhance the picture. This finding has been supported by later research [36,37], and the focus of several studies has shifted from improving image accuracy to improving image data quality. Therefore, in this paper, we chose SRGAN to verify whether the perceptual loss function could acquire better visual quality in DBMs. SRResNet is the generator component of SRGAN, and it improves SRGAN’s accuracy while using only the mean squared error (MSE) loss function to avoid imbalance during the training process.
DBM water depth values generally vary widely, reaching several thousand meters underwater or even tens of thousands of meters underwater. Therefore, to quantitatively compare the performance of various DBM SR methods, we used common DBM evaluation indices, the root mean squared error (RMSE) and mean absolute error (MAE), as evaluation indices and used the maximum bathymetric error (Emax) as an evaluation index to reflect the stability of the SR methods. These evaluation indices can be calculated as follows:
R M S E = 1 N i = 1 N ( h i h ^ i ) 2
M A E = 1 N i N | h i h ^ i |
E m a x = m a x ( a b s ( h i h ^ i ) )
where N is the total number of pixels in a DBM image. h i and h ^ i represent the elevation values of the HR DBM and the corresponding SR DBM, respectively.

4. Results and Discussions

4.1. Results on DBMs

Table 1 shows the quantitative results of 638 DBMs in the test set in terms of the RMSE, MAE, and Emax indices. Among these results, those in bold highlight the best performance of each indicator, and the underlined values are the second-best results. From Table 1, we can see that, compared to other methods, STFET had the best effect on all evaluation indices. Moreover, the accuracy of all deep learning-based methods was better than bicubic interpolation, which shows that deep learning methods have an application value in DBM SR. Similar to the results for natural images [20], SRGAN achieves a lower accuracy than SRResNet based on the RMSE accuracy index.
From Table 1, it can also be seen that SRCNN performs better than SRGAN and SRResNet; these results are different from those obtained in the CV domain. STFET uses the seabed terrain feature extraction module and effective transformer module, and SR’s accuracy is significantly improved. Compared with bicubic interpolation, SRCNN, SRGAN, and SRResNet, the performance results in terms of RMSE decreased by 16%, 10%, 13%, and 12%, respectively. The test set includes complex-area and flat-area DBMs, and the STFET results are optimal, proving that this method is universal and reliable.

4.2. Quantitative Evaluation Results on Complex Regions

To evaluate the performance of DBM in different terrain regions, we selected five regions with more significant topographic features from the test set to conduct an accuracy comparison. These five regions contained highly undulating and complex terrain, including trenches and seamounts. The locations are shown in Figure 6. Among them, the extremum between area 2 and area 5 was small, and the terrain undulation was relatively moderate, while the topographic features of areas 1, 3, and 4 were more obvious. The original DBM for the five regions and the SR DBM reconstructed using different SR methods were combined for comparison, and the quantitative results are shown in Table 2.
From Table 2, it can be seen that the STFET results were still optimal for complex terrain. In area 2 and area 5, all methods obtained quite accurate results, and STFET exhibited the greatest improvement. However, in area 1, area 3, and area 4, the terrain features were complex. For example, area 1 comprised a large area containing trenches with topographic features. The reconstruction effect of bicubic interpolation and SRCNN was poor, while the reconstruction effect of SRGAN and SRResNet was much better in area 1. This observation illustrates the ability of residual networks to extract features significantly. STFET has a seabed terrain feature extraction module and an efficient transformer module and, therefore, it can enhance the extraction of local terrain features and similar global terrain features so that the reconstruction results are significantly improved. Compared with bicubic interpolation, SRCNN, SRGAN and SRResNet, the accuracy in terms of the RMSE decreased by 63%, 62%, 21%, and 44%, respectively, in area 1. These results demonstrate the effectiveness of STFET in terms of extracting seafloor topographic features.

4.3. Visual Evaluation of Different Methods

To compare the visual quality of reconstructions using different methods, we combined HR DBMs with SR DBMs generated using different methods for the corresponding regions, and these results are shown in Figure 7. We can see that, on the whole, the SR DBM reconstructed by all methods was relatively close to the HR DBM. However, in some areas with large terrain fluctuations and obvious topographic features, subtle differences in terrain could still be observed. As shown in Figure 8, we zoomed in on the area with obvious seabed terrain features in Figure 7; the reconstruction effect of bicubic interpolation and SRCNN in areas with large terrain fluctuations was too smooth, and the visual quality was poor, which is consistent with the quantitative results in Table 2. At the same time, the proposed method had a better reconstruction effect compared to other methods, and the result is closer to an HR DBM. Therefore, the results confirm the effectiveness of STFET in terms of both visual quality and reconstruction accuracy.

4.4. Residual Evaluation of Different Methods

To visually compare the differences between the DBM and HR DBM generated using each SR method, we plotted the water depth difference between the SR DBM and HR DBM in the five districts as an error map, and the results are shown in Figure 9. The main difference in the reconstructed DBM lies in special terrain areas, such as the trench in area 1 and the seamount in area 4 in the figure, where the error is relatively large. From the error map results, we can clearly see that the error area generated using STFET was much smaller overall. The error in the local area was also greatly reduced closer to the HR DBM, which indicates that the effectiveness of the seabed terrain extraction module and the efficient transformer module was further validated. Meanwhile, similar to natural images, SRGAN outperformed SRResNet visually. However, the performance of bicubic interpolation and SRCNN was poor, and their error range was large; therefore, they were not suitable for DBM SR in complex terrain areas. As a result, the main challenge of DBM SR was to restore these areas. At the same time, it was necessary to ensure the reconstruction effect of the overall terrain.
To quantitatively analyze the error accuracy of various methods, we took 0–25 m, 25–100 m, 100–200 m, and more than 200 m as the statistical conditions and counted the number of error points generated using different methods of DBMs in Figure 9. The results are shown in Table 3. From the results, we can clearly see that the DBM reconstruction errors generated by StfeT were more concentrated at 0–25 m, and this reconstruction accuracy was better than other methods.

4.5. Ablation Study

In this section, we verify the effectiveness of the seabed terrain feature extraction module and the efficient transformer module in DBM SR. We added the seabed terrain feature extraction module and the efficient transformer module into other models, with SRResNet as the baseline. SRResNet-dconv represents SRResNet with the seabed terrain feature extraction module, and SRResNet-ET represents SRResNet with the efficient transformer module. The comparison results are shown in Table 4.
Common convolutional layers were used in conventional residual networks, and their convolution kernels were usually regular; however, the DBM seabed terrain features were irregular. By using deformable convolutional layers, features that more closely resembled the actual terrain could be obtained to some extent, thereby improving the accuracy of super-resolution. In addition, the efficient transformer could use a self-attention mechanism that allowed each pixel to interact with all pixels in the entire image, and this nonlocal operation could better capture long-range correlations in images, helping to recover missing details more accurately. The results in Table 3 show that the seabed terrain feature extraction module and the efficient transformer module have the potential to improve the accuracy of different deep learning-based SR methods. The effectiveness of the seabed terrain extraction module and the efficient transformer is demonstrated. However, when the seabed terrain feature extraction module and the efficient transformer module were used alone, this improvement effect was not very satisfactory. This observation shows that when using deep learning-based SR methods to extract features from DBMs, it is necessary to both extract local features and strengthen the extraction of similar global features. Only in this way can high-accuracy HR DBM be obtained more effectively.

5. Conclusions

In most previous studies, scholars have often used interpolation methods to improve DBM resolution, but the accuracy of generating HR DBM was poor. In this paper, we propose a seabed terrain feature extraction transformer for DBM SR, which mainly consists of the seabed terrain feature extraction module and the efficient transformer module. First, the seabed terrain feature extraction module has an important function, and it is able to adaptively extract DBM seabed terrain features. This adaptive ability enables the model to perceive DBM seabed terrain features at different scales and strengthen the extraction of local DBM seabed terrain features. Second, the efficient transformer module finds similar seabed terrain features through the self-attention mechanism, which enhances the extraction effect of similar global seabed terrain features. The effectiveness of these two modules was verified through ablation experiments, which demonstrated that, compared with bicubic interpolation and SRCNN, SRGAN, and SRResNet, STFET decreased the RMSE by 16%, 10%, 13%, and 12%, respectively. Furthermore, especially in areas with complex seabed terrain features, the accuracy of SR DBM generated using STFET reconstruction was significantly improved.
The method proposed in this paper considers global and local seabed terrain features for DBM SR but is not limited to DBM SR. While making up for the lack of interpolation accuracy, it can also be combined with the interpolation method for data fusion. Our next step considers that the proposed STFET can be used to transfer the trained model to the sea domain with only LR DBM, and we plan to rebuild and generate HR DBMs to compensate for the deficiency of LR DBMs and meet the needs of practical applications.

Author Contributions

W.C. collected and processed the data, performed analysis, and wrote the Paper; Y.L. proposed the main idea and made suggestions on the experiments; Y.C. and Z.D. helped to write and edit the article; H.Y. and N.L. contributed to the validation. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Shandong Natural Science Foundation (grant number ZR2023QD113), the Shandong Postdoctoral Innovation Project (grant number SDCX-ZG-202202041), and the Qingdao Natural Science Foundation (grant number 23-2-1-73-zyyd-jch).

Data Availability Statement

The digital bathymetric model data used in this paper can be downloaded at: https://www.gebco.net/data_and_products/gridded_bathymetry_data/gebco_2022/ (accessed on 10 September 2023).

Conflicts of Interest

The authors declare that they have no known competing financial interest or personal relationships that could have appeared to influence the work reported in this paper.

References

  1. Wu, Z.; Li, J.; Jin, X.; Shang, J.; Li, S.; Jin, X. Distribution, features, and influence factors of the submarine topographic boundaries of the Okinawa Trough. Sci. China Earth Sci. 2014, 57, 1885–1896. [Google Scholar] [CrossRef]
  2. Sandwell, D.T.; Smith, W.H. Marine gravity anomaly from Geosat and ERS 1 satellite altimetry. J. Geophys. Res. Solid Earth 1997, 102, 10039–10054. [Google Scholar] [CrossRef]
  3. Wang, M.; Wu, Z.; Yang, F.; Ma, Y.; Wang, X.H.; Zhao, D. Multifeature extraction and seafloor classification combining LiDAR and MBES data around Yuanzhi Island in the South China Sea. Sensors 2018, 18, 3828. [Google Scholar] [CrossRef]
  4. Picard, K.; Brooke, B.P.; Harris, P.T.; Siwabessy, P.J.; Coffin, M.F.; Tran, M.; Spinoccia, M.; Weales, J.; Macmillan-Lawler, M.; Sullivan, J. Malaysia Airlines flight MH370 search data reveal geomorphology and seafloor processes in the remote southeast Indian Ocean. Mar. Geol. 2018, 395, 301–319. [Google Scholar] [CrossRef]
  5. Mayer, L.; Jakobsson, M.; Allen, G.; Dorschel, B.; Falconer, R.; Ferrini, V.; Lamarche, G.; Snaith, H.; Weatherall, P. The Nippon Foundation—GEBCO seabed 2030 project: The quest to see the world’s oceans completely mapped by 2030. Geosciences 2018, 8, 63. [Google Scholar] [CrossRef]
  6. Schaffer, J.; Timmermann, R.; Arndt, J.E.; Kristensen, S.S.; Mayer, C.; Morlighem, M.; Steinhage, D. A global, high-resolution data set of ice sheet topography, cavity geometry, and ocean bathymetry. Earth Syst. Sci. Data 2016, 8, 543–557. [Google Scholar] [CrossRef]
  7. Ramillien, G.; Cazenave, A. Global bathymetry derived from altimeter data of the ERS-1 geodetic mission. J. Geodyn. 1997, 23, 129–149. [Google Scholar] [CrossRef]
  8. Briggs, I.C. Machine contouring using minimum curvature. Geophysics 1974, 39, 39–48. [Google Scholar] [CrossRef]
  9. Smith, W.H.; Wessel, P. Gridding with continuous curvature splines in tension. Geophysics 1990, 55, 293–305. [Google Scholar] [CrossRef]
  10. Glenn, J.; Tonina, D.; Morehead, M.D.; Fiedler, F.; Benjankar, R. Effect of transect location, transect spacing and interpolation methods on river bathymetry accuracy. Earth Surf. Process. Landforms 2016, 41, 1185–1198. [Google Scholar] [CrossRef]
  11. Merwade, V. Effect of spatial trends on interpolation of river bathymetry. J. Hydrol. 2009, 371, 169–181. [Google Scholar] [CrossRef]
  12. Chen, G.; Chen, Y.; Wilson, J.P.; Zhou, A.; Chen, Y.; Su, H. An Enhanced Residual Feature Fusion Network Integrated with a Terrain Weight Module for Digital Elevation Model Super-Resolution. Remote Sens. 2023, 15, 1038. [Google Scholar] [CrossRef]
  13. Habib, M. Evaluation of DEM interpolation techniques for characterizing terrain roughness. Catena 2021, 198, 105072. [Google Scholar] [CrossRef]
  14. Dong, C.; Loy, C.C.; He, K.; Tang, X. Learning a Deep Convolutional Network for Image Super-Resolution. In Proceedings of the ECCV: 13th European Conference, Zurich, Switzerland, 6–12 September 2014. [Google Scholar]
  15. Yang, W.; Zhang, X.; Tian, Y.; Wang, W.; Xue, J.H.; Liao, Q. Deep learning for single image super-resolution: A brief review. IEEE Trans. Multimed. 2019, 21, 3106–3121. [Google Scholar] [CrossRef]
  16. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef]
  17. Chen, Z.; Wang, X.; Xu, Z.; Hou, W. Convolutional neural network based dem super resolution. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. ISPRS Arch. 2016, 41, 247–250. [Google Scholar] [CrossRef]
  18. Wu, Z.; Shen, C.; Van Den Hengel, A. Wider or deeper: Revisiting the resnet model for visual recognition. Pattern Recognit. 2019, 90, 119–133. [Google Scholar] [CrossRef]
  19. Xu, Z.; Chen, Z.; Yi, W.; Gui, Q.; Hou, W.; Ding, M. Deep gradient prior network for DEM super-resolution: Transfer learning from image to DEM. ISPRS J. Photogramm. Remote Sens. 2019, 150, 80–90. [Google Scholar] [CrossRef]
  20. Ledig, C.; Theis, L.; Husz’ar, F.; Caballero, J.; Cunningham, A.; Acosta, A.; Aitken, A.; Tejani, A.; Totz, J.; Wang, Z.; et al. Photo-Realistic Single Image Super-Resolution Using a Generative Adversarial Network. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 105–114. [Google Scholar] [CrossRef]
  21. Zhang, Y.; Yu, W. Comparison of DEM Super-Resolution Methods Based on Interpolation and Neural Networks. Sensors 2022, 22, 745. [Google Scholar] [CrossRef]
  22. Zhou, A.; Chen, Y.; Wilson, J.P.; Su, H.; Xiong, Z.; Cheng, Q. An Enhanced double-filter deep residual neural network for generating super resolution DEMs. Remote Sens. 2021, 13, 3089. [Google Scholar] [CrossRef]
  23. Zhang, B.; Xiong, W.; Ma, M.; Wang, M.; Wang, D.; Huang, X.; Yu, L.; Zhang, Q.; Lu, H.; Hong, D. Super-resolution reconstruction of a 3 arc-second global DEM dataset. Sci. Bull. 2022, 67, 2526–2530. [Google Scholar] [CrossRef]
  24. Zhou, A.; Chen, Y.; Wilson, J.P.; Chen, G.; Min, W.; Xu, R. A multi-terrain feature-based deep convolutional neural network for constructing super-resolution DEMs. Int. J. Appl. Earth Obs. Geoinf. 2023, 120, 103338. [Google Scholar] [CrossRef]
  25. Jiang, Y.; Xiong, L.; Huang, X.; Li, S.; Shen, W. Super-resolution for terrain modeling using deep learning in high mountain Asia. Int. J. Appl. Earth Obs. Geoinf. 2023, 118, 103296. [Google Scholar] [CrossRef]
  26. Yutani, T.; Yono, O.; Kuwatani, T.; Matsuoka, D.; Kaneko, J.; Hidaka, M.; Kasaya, T.; Kido, Y.; Ishikawa, Y.; Ueki, T. Super-Resolution and Feature Extraction for Ocean Bathymetric Maps Using Sparse Coding. Sensors 2022, 22, 3198. [Google Scholar] [CrossRef] [PubMed]
  27. Hidaka, M.; Matsuoka, D.; Kuwatani, T.; Kaneko, J.; Kasaya, T.; Kido, Y.; Ishikawa, Y.; Kikawa, E. Super-resolution for Ocean Bathymetric Maps Using Deep Learning Approaches: A Comparison and Validation. Geoinformatics 2021, 32, 3–13. (In Japanese) [Google Scholar] [CrossRef]
  28. Zhang, Y.; Yu, W.; Zhu, D. Terrain feature-aware deep learning network for digital elevation model superresolution. ISPRS J. Photogramm. Remote Sens. 2022, 189, 143–162. [Google Scholar] [CrossRef]
  29. Lu, Z.; Li, J.; Liu, H.; Huang, C.; Zhang, L.; Zeng, T. Transformer for single image super-resolution. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 19–20 June 2022; pp. 457–466. [Google Scholar] [CrossRef]
  30. Zheng, X.; Bao, Z.; Yin, Q. Terrain Self-Similarity-Based Transformer for Generating Super Resolution DEMs. Remote Sens. 2023, 15, 1954. [Google Scholar] [CrossRef]
  31. Kim, J.; Lee, J.K.; Lee, K.M. Accurate Image Super-Resolution Using Very DeeConvolutional Networks. In Proceedings of the 2016 IEEE Conference on Computer Vision and PatternRecognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 1646–1654. [Google Scholar] [CrossRef]
  32. Dai, J.; Qi, H.; Xiong, Y.; Li, Y.; Zhang, G.; Hu, H.; Wei, Y. Deformable convolutional networks. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 764–773. [Google Scholar] [CrossRef]
  33. Zhu, X.; Hu, H.; Lin, S.; Dai, J. Deformable convnets v2: More deformable, better results. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 9308–9316. [Google Scholar] [CrossRef]
  34. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S. An image is worth 16 × 16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  35. Han, D. Comparison of Commonly Used Image Interpolation Methods. In Proceedings of the 2nd International Conference on Computer Science and Electronics Engineering (ICCSEE 2013), Hangzhou, China, 22–23 March 2013; pp. 1556–1559. [Google Scholar]
  36. Wang, X.; Yu, K.; Wu, S.; Gu, J.; Liu, Y.; Dong, C.; Qiao, Y.; Change Loy, C. ESRGAN: Enhanced Super-Resolution Generative Adversarial Networks. In Proceedings of the European Conference on Computer Vision (ECCV) Workshops, Munich, Germany, 8–14 September 2018; pp. 63–79. [Google Scholar]
  37. Zhang, W.; Liu, Y.; Dong, C.; Qiao, Y. Ranksrgan: Generative adversarial networks with ranker for image super-resolution. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 3096–3105. [Google Scholar] [CrossRef]
Figure 1. The StfeT workflow.
Figure 1. The StfeT workflow.
Remotesensing 15 04906 g001
Figure 2. Framework of the proposed STFET.
Figure 2. Framework of the proposed STFET.
Remotesensing 15 04906 g002
Figure 3. Schematic of the deformable convolution principle.
Figure 3. Schematic of the deformable convolution principle.
Remotesensing 15 04906 g003
Figure 4. Deformable convolution and regular convolution for feature extraction on DBM.
Figure 4. Deformable convolution and regular convolution for feature extraction on DBM.
Remotesensing 15 04906 g004
Figure 5. Structure of the efficient transformer and EMSA.
Figure 5. Structure of the efficient transformer and EMSA.
Remotesensing 15 04906 g005
Figure 6. Study area.
Figure 6. Study area.
Remotesensing 15 04906 g006
Figure 7. Comparison of results using different SR methods in various topographic areas.
Figure 7. Comparison of results using different SR methods in various topographic areas.
Remotesensing 15 04906 g007
Figure 8. Comparison of the results after zooming in to Figure 7.
Figure 8. Comparison of the results after zooming in to Figure 7.
Remotesensing 15 04906 g008
Figure 9. Comparison of SR DBM and HR DBM error maps for different methods in various regions.
Figure 9. Comparison of SR DBM and HR DBM error maps for different methods in various regions.
Remotesensing 15 04906 g009
Table 1. Quantitative results on the test set. Results in bold are best, and those underlined are second best.
Table 1. Quantitative results on the test set. Results in bold are best, and those underlined are second best.
MethodRMSE (m)MAE (m)Emax (m)
Bicubic15.858.08295.80
SRCNN14.787.26293.48
SRGAN15.277.87306.98
SRResNet15.137.89300.07
STFET13.306.88262.05
Table 2. Quantitative results for complex regions. Results in bold are the best, and those underlined are the second best.
Table 2. Quantitative results for complex regions. Results in bold are the best, and those underlined are the second best.
AreaMethodRMSE (m)MAE (m)Emax (m)
1Bicubic32.2624.90157.34
SRCNN31.6924.72201.30
SRGAN15.1111.03175.39
SRResNet21.4016.25140.28
STFET11.988.41198.65
2Bicubic17.589.78494.95
SRCNN17.269.73483.92
SRGAN14.497.44484.46
SRResNet14.697.87492.02
STFET12.576.12491.90
3Bicubic30.0817.66189.57
SRCNN29.7718.04196.98
SRGAN12.449.27136.85
SRResNet15.2210.97126.56
STFET10.928.18125.52
4Bicubic27.1014.30314.91
SRCNN26.4214.63295.98
SRGAN13.997.94357.04
SRResNet16.699.50310.11
STFET12.846.99360.42
5Bicubic16.636.42268.70
SRCNN15.635.05279.80
SRGAN14.165.45259.50
SRResNet12.935.27207.15
STFET11.294.58194.97
Table 3. The number of errors generated using different methods.
Table 3. The number of errors generated using different methods.
MethodNumber (Error < 25 m)Number (25 m < Error < 100 m)Number (100 m < Error < 200 mNumber (Error > 200 m
Bicubic61,71119,39079326
SRCNN61,37819,87664224
SRGAN77,49443446022
SRResNet71,02910,70416324
StfeT78,89329495622
Table 4. Ablation study results for the proposed STFET. Results in bold are best.
Table 4. Ablation study results for the proposed STFET. Results in bold are best.
MethodRMSE (m)MAE (m)Emax (m)
SRResNet15.137.89300.07
SRResNet-dconv15.107.70322.98
SRResNet-ET15.017.70319.30
STFET13.306.88262.05
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Cai, W.; Liu, Y.; Chen, Y.; Dong, Z.; Yuan, H.; Li, N. A Seabed Terrain Feature Extraction Transformer for the Super-Resolution of the Digital Bathymetric Model. Remote Sens. 2023, 15, 4906. https://doi.org/10.3390/rs15204906

AMA Style

Cai W, Liu Y, Chen Y, Dong Z, Yuan H, Li N. A Seabed Terrain Feature Extraction Transformer for the Super-Resolution of the Digital Bathymetric Model. Remote Sensing. 2023; 15(20):4906. https://doi.org/10.3390/rs15204906

Chicago/Turabian Style

Cai, Wuxu, Yanxiong Liu, Yilan Chen, Zhipeng Dong, Hanxiao Yuan, and Ningning Li. 2023. "A Seabed Terrain Feature Extraction Transformer for the Super-Resolution of the Digital Bathymetric Model" Remote Sensing 15, no. 20: 4906. https://doi.org/10.3390/rs15204906

APA Style

Cai, W., Liu, Y., Chen, Y., Dong, Z., Yuan, H., & Li, N. (2023). A Seabed Terrain Feature Extraction Transformer for the Super-Resolution of the Digital Bathymetric Model. Remote Sensing, 15(20), 4906. https://doi.org/10.3390/rs15204906

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop