Next Article in Journal
Advancements in Remote Photoplethysmography
Previous Article in Journal
Enhancing Domain-Specific Knowledge Graph Reasoning via Metapath-Based Large Model Prompt Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Soil Moisture Monitoring Based on Deformable Convolution Unit Net Algorithm Combined with Water Area Changes

1
School of Electronics and Electrical Engineering, Ningxia University, Yinchuan 750021, China
2
School of Information Engineering, North China University of Water Resources and Electric Power, Zhengzhou 450046, China
*
Author to whom correspondence should be addressed.
Electronics 2025, 14(5), 1011; https://doi.org/10.3390/electronics14051011
Submission received: 10 February 2025 / Revised: 22 February 2025 / Accepted: 27 February 2025 / Published: 3 March 2025

Abstract

:
In response to the issue that existing soil moisture monitoring methods are significantly affected by surface roughness and the complex environment around water bodies, leading to a need for improvement in the accuracy of soil moisture inversion, a soil moisture detection algorithm based on a DCU-Net (Deformable Conv Unit-Net) water body extraction model is proposed, using the Ningxia region as the study area. The algorithm introduces the DCU (Deformable Conv Unit) module, which addresses the problem of extracting small water bodies at large scales with low resolution; reduces the probability of misjudgment during water body extraction caused by shadows from mountains, buildings, and other objects; and enhances the robustness and adaptability of the water body extraction algorithm. The method first creates a water body extraction dataset based on multi-year remote sensing images from Ningxia Province and trains the proposed DCU-Net model; then, it selects remote sensing images from certain areas for water body extraction; finally, it conducts regression analysis between the water body areas of Ningxia Province at different times and the corresponding measured soil moisture data to establish the intrinsic relationship between water body areas and soil moisture in the study area, achieving real-time regional soil moisture monitoring. The water body extraction performance of DCU-Net is verified based on extraction accuracy, with U-Net selected as the baseline network. The experimental results show that DCU-Net leads to improvements of 2.98%, 1.37%, 0.36%, and 1.49% in terms of IoU, Precision, Recall, and F 1 , respectively. The algorithm is more sensitive to water body feature information, can more accurately identify water bodies, and extracts water body contours more accurately. Additionally, a soil moisture inversion method based on a cubic polynomial is constructed. These results indicate that DCU-Net can precisely extract water body contours and accurately invert regional soil moisture, thereby providing support for the monitoring of large-scale soil moisture.

1. Introduction

Surface soil moisture (SSM) is a key terrestrial environmental variable that connects agriculture, ecology, and hydrology, and is also a critical parameter in hydrological, meteorological, and agricultural applications [1,2]. The detection of SSM can be used for planning irrigation, monitoring water quality, managing water resources, and estimating crop yields. Understanding the temporal and spatial distribution of and dynamic changes in SSM can allow us to better address natural disasters such as droughts and floods, and more effectively guide agricultural production. Therefore, obtaining surface soil moisture data in a timely and accurate manner is of significant importance for studying climate issues, establishing hydrological models, and monitoring droughts [3,4].
Traditional SSM monitoring typically employs methods such as gravimetric, resistance, and probe techniques [5]. Although these methods offer quite good accuracy and are simple to operate, they require a significant workforce and lots of material resources. Additionally, they are susceptible to environmental and human factors. Moreover, due to the limited number of sampling points, they can only reflect the local conditions of the area being measured, making it difficult to obtain a large amount of SSM information in a short time, and making it hard to meet the needs of soil moisture monitoring over a wide area.
The development of remote sensing technology has made it possible to monitor soil moisture dynamics over large areas and over long periods. Currently, there are two main types of soil moisture monitoring based on remote sensing technology: the first is microwave remote sensing-based soil moisture monitoring, which mainly utilizes the nonlinear relationship between radar backscatter coefficients and other parameters with soil moisture, employing active and passive microwave remote sensing and their synergistic effects to invert soil moisture. However, this method is significantly affected by surface roughness and vegetation cover, and its accuracy needs improvement; the second type is soil moisture monitoring based on optical remote sensing images, which involves extracting water surface areas to establish related inversion models. This method is greatly affected by weather conditions, and its inversion accuracy is closely related to its imaging quality [6]. A comparison of soil moisture monitoring is shown in Table 1. Therefore, the direct inversion of soil moisture based on remote sensing technology is significantly affected by environmental factors such as weather conditions and surface roughness, which limits the improvement of its inversion accuracy.
Remote sensing technology, as an important means of obtaining surface information, provides a rich data source for the extraction of water area. However, traditional remote sensing image processing methods are mostly based on manual operations and simple image analysis algorithms, which have problems such as low efficiency and a low degree of automation. With the continuous development of machine learning and artificial intelligence technologies, new solutions and methods have been provided for the extraction of water areas [7]. Machine learning is an algorithm driven by data that can automatically learn from large amounts of data and extract useful features and information. In the extraction of water area, machine learning algorithms can utilize various features, such as spectral, texture, and shape features, from remote sensing images to achieve the automatic identification and extraction of water area [8]. However, this task has several problems: (1) Large-scale water bodies are mainly of medium to low resolution, and the spatial details of the water bodies are not obvious, making it difficult to effectively extract small water bodies, resulting in low overall accuracy. (2) The spectral characteristics of water bodies are mainly related to the interaction of factors such as the bands reflected by the surface, sediment concentration, and transmission depth. Changes in these factors lead to different reflection and absorption characteristics of water bodies at different bands, increasing the difficulty of water area extraction. (3) The texture and shape of water areas in remote sensing images are influenced by various factors, such as terrain and landforms, wind waves, and vegetation cover, requiring algorithms to have high robustness and adaptability. (4) The shadows of buildings, mountains, and other objects in remote sensing images may be similar to the spectral characteristics of water bodies, leading to misjudgments during water body extraction.
Convolutional neural network-based algorithms have made significant contributions to the intelligent interpretation of water area extraction in remote sensing, achieving inputs of images of arbitrary size and obtaining pixel-level attribute information. To address the issues in the automatic identification and extraction of water areas mentioned above, it is common to increase the size of the convolutional kernel. This is because a larger convolutional kernel means a larger receptive field, which can capture more contextual information. The network can extract richer feature information, thereby improving the accuracy of water area extraction. However, large kernel convolutions have more parameters, leading to a more complex model, increasing the risk of overfitting, and significantly increasing the computational overhead and memory consumption of the model [9].
In this study, we designed a CNN-based fundamental model with a larger receptive field, which preserves long-range dependencies within images and extracts richer feature information. This model can obtain a large amount of surface soil moisture information in a short time, meeting soil moisture monitoring needs within a wide area, and providing more accurate data for soil moisture inversion. Overall, the contributions of this study include the following:
(1)
Monitoring soil moisture based on remote sensing images by establishing a related inversion model through the extraction of water area.
(2)
Proposing a new water area extraction model, DCU-Net, which improves the accuracy of water area extraction, thereby enabling more accurate monitoring of soil moisture.

2. Related Work

With the rapid development of large-scale data and computing resources, convolutional neural networks have also seen richer extensions. Based on AlexNet [10], many deeper and more efficient convolutional neural network architectures have been proposed, such as VGG [11], GoogleNet [12], ResNet [13,14], EfficientNet [15,16], and others. In addition to architectural design, more complex convolutional operations have been developed [17], such as deep convolution and deformable convolution [18,19], which have a larger receptive field, retain more global information, and can improve the problem of losing target features in deep network layers.
With the remarkable achievements of Transformers in large-scale language models [20,21,22], Vision Transformers (ViTs) [23,24,25] have also rapidly emerged in the field of computer vision and become the preferred architecture for research on and the practice of large-scale vision models. Some research has demonstrated that ViTs can outperform CNNs and significantly enhance performance in various computer vision tasks. Through in-depth analysis, it can be seen that CNNs and ViTs can be compared in the following two dimensions:
(1)
At the Core Operator Level: The multi-head self-attention mechanism (MHSA) of ViTs has the ability to model long-range dependencies and the characteristic of adaptive spatial aggregation, enabling it to learn more powerful robustness from massive data than CNNs.
(2)
From an Architectural Perspective: ViTs integrate a series of advanced components that are not available in conventional CNNs, such as Layer Normalization and Feed-Forward Networks (FFNs). However, the MHSA in ViTs has high computational costs and memory complexities, which, to some extent, restricts its wide application in downstream tasks.
The algorithm for water extraction based on convolutional neural Networks (CNNs) applied to soil moisture monitoring has gained the favor of many researchers due to its speed and efficiency. Ronneberger et al. proposed the U-Net network [26] on the basis of the FCN network [27], constructing a shallow convolutional neural network based on the Encoder–Decoder concept and combining high-level semantic information with low-level semantic information through skip connections, enabling the precise pixel-level extraction of targets under conditions of small sample sizes. Given that the water body contours and color information in high-resolution remote sensing images are significantly different from the background, Liu Jiadian et al. proposed an improved U-Net model for water body extraction from remote sensing images [28]. By reducing the number of layers in the U-Net network, they achieved rapid and accurate extraction of water areas. However, the lightweight model failed to balance the relationship between computational speed and extraction accuracy, and the extraction accuracy in complex images needs improvement compared to deep convolutional neural networks. On the other hand, the extraction speed of deep convolutional neural networks cannot easily meet the needs of large-scale water body extraction tasks. To balance computational speed and accuracy, Wang et al. proposed the InternImage feature extraction network [29], which uses Deformable convolution to replace conventional convolution kernels, thus introducing dynamic weights with long-range dependencies, obtaining a larger receptive field while controlling the amount of computation, and extracting more global information.

3. Materials and Methods

3.1. Overview of the Study Area

The Ningxia Hui Autonomous Region (35.44′ N–39.47′ N, 105.45′ E–111.30′ E), located in the northwest of China, borders the Yellow River to the north and is part of the Yellow River basin. The terrain of Ningxia can be mainly divided into three parts: the western mountainous area; the Loess Plateau; and the Hetao Plain. The climate of Ningxia is typically a temperate continental climate with sparse precipitation, with annual rainfall ranging roughly from 100 mm to 400 mm, mostly concentrated in the summer. Water resources in Ningxia are relatively scarce, relying heavily on the Yellow River, which flows through the central part of Ningxia, providing significant support for agricultural irrigation in the region. Therefore, soil moisture monitoring based on water body changes in Ningxia is of great importance for the growth of crops and the evolution of the ecological pattern in the area.

3.2. Research Data

The study period was from 2017 to 2022, and we used remote sensing images of the Yellow River basin in different periods within Ningxia Hui Autonomous Region as the data source, all of which were remote sensing data with cloud cover of less than 10%.

3.3. Water Extraction Model (Deformable Conv Unit-Net)

Typically, images of large-scale water bodies are dominated by medium to low resolutions, making it difficult to effectively extract small water bodies and affecting overall accuracy; water bodies are significantly influenced by various factors, such as terrain, which places high adaptability requirements on water extraction algorithms; the spectral characteristics of various objects are similar to those of water bodies, and the extraction algorithm needs to retain as much detail information from the image as possible. To better address these issues, the water body extraction algorithm is required to have a larger receptive field while ensuring adequate extraction speed.
To increase the effective receptive field of the model and its ability to mine detailed features, large kernel convolutions and increased network depth are commonly used. However, this significantly increases computational costs and affects extraction speed. Additionally, standard convolutions have a high degree of inductive bias and lack adaptive spatial aggregation capabilities, which limits their ability to learn more general patterns from large-scale data.
Therefore, this paper proposes the DCU-Net model, whose network structure is shown in Figure 1. The DCU module based on deformable convolution is used in the network layers where feature information is rich. The following are true in this module: (1) The flexible receptive field adapts to low-resolution images, where water area details are sparse, making it difficult for traditional convolutions with fixed receptive fields to effectively capture key features. The DCU introduces learnable offsets, allowing the sampling points of the convolutional kernel to dynamically adjust based on the input content. (2) Local feature enhancement: The DCU module can focus on key regions of the target in low-resolution images, enhancing its ability to extract local features. (3) Diversity and generalization capability in large-scale water body data: In large-scale datasets, water bodies exhibit high diversity in scale, shape, and posture. The DCU, by dynamically adjusting sampling positions, can better adapt to this data diversity. (4) Information recovery capability in low-resolution images: in low-resolution images, information loss is severe, and traditional convolutions may overlook key features. DCU, through offset learning, can recover crucial water body features from sparse information. The DCU module has a lightweight design while acquiring richer gradient information flow, accelerating convergence speed, and enhancing training efficiency. This helps the network better learn the features of the input data, thereby improving the model’s accuracy in extracting water information.

3.4. Deformable Convolution Unit

Under normal circumstances, water remote sensing images are characterized by their large scale, low resolution, and complex surrounding environments. Conventional semantic segmentation models extract limited image detail information, resulting in their inability to accurately extract all water information, which affects their accuracy in soil moisture monitoring.
To address these issues, in the task of water body extraction from remote sensing images, we propose the Deformable Convolution Unit, as shown in Figure 2. In this module, we use two layers of regular convolution operations to retain more semantic information. In the lower layers, where semantic information is richer, we apply Deformable convolution, whose sampling positions better match the shape and size of water boundaries. It is capable of learning the detailed features of objects, better excluding the interference of complex background environments, and obtaining a more complete water image.
In deformable convolution, the following process is carried out: (1) We detach the original convolution weights w k into depth-wise and point-wise parts, where the depth-wise part is responsible for the original location-aware modulation scalar m k , and the point-wise part represents the shared projection weights w among sampling points. (2) Introducing the multi-group mechanism, we split the spatial aggregation process into G groups, each of which has individual sampling offsets p g k and modulation scales m g k , and thus, different groups on a single convolution layer can have different spatial aggregation patterns, resulting in stronger features for downstream tasks. (3) To alleviate the instability issues, the sum of the modulation scalars is constrained to 1, which makes the training process of models at different scales more stable [29].
The calculation process is shown in Equation (1).
y p 0 = g = 1 G k = 1 K w g m g k X g p 0 + p k + p g k
where G denotes the total number of aggregation groups. For the g-th group, w g R C × C denotes the location-irrelevant projection weights of the group, where C = C / G represents the group dimension. m g k R denotes the modulation scalar of the k-th sampling point in the g-th group, normalized by the sofmax function along the dimension K . X g R C × H × W represents the sliced input-feature map. p g k is the offset corresponding to the grid sampling location pk in the g-th group.

4. Experiments

4.1. Implementation Details

The DCUNet water extraction algorithm was built using the Pytorch 10.2 deep learning framework and Python language. The model’s training, optimization, and testing were performed on an Ubuntu 18.04 operating system, with a memory capacity of 128 GB, utilizing two high-performance NVIDIA RTX 3090 graphics cards (24 GB each). During the experiment, the initial learning rate was set to 0.01, and the batch_size was set to 16. It took approximately 5 min to train the entire network in each round. During the whole training process, the training model with the best performance was saved.

4.2. Dataset

The data used in this study contained remote sensing images of the Yellow River basin in Ningxia Province from different periods between 2017 and 2022, with an original size of 10,980 pixels × 10,980 pixels. To improve the speed of model training, this study cropped the images around the water area and resized the cropped images to 512 pixels × 512 pixels, as shown in Table 2. The data annotation software labelme was used for manual annotation of the water area’s edge lines to create the dataset. A total of 12,600 samples were created, including 9600 positive samples and 3000 negative samples without water labels. The dataset was divided into training, validation, and test sets in an 8:1:1 ratio.
To verify the effect of the model before and after the improvement in the water body extraction tasks, I O U , P r e c i s i o n , R e c a l l , and F 1 were chosen as evaluation metrics, with their calculation methods shown in Equations (2)–(5).
I O U = T P T P + F P + F N
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
F 1 = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l
where T P represents the number of pixels correctly extracted as water bodies, F P represents the number of pixels incorrectly extracted as water bodies, and F N represents the number of water body pixels incorrectly classified as background.

4.3. Ablation Study

To improve the integrity of water body extraction, this study constructed a water body extraction algorithm named DCU-Net, which mainly includes the DCU module that introduces deformable convolution. To verify the effectiveness of the DCU module in water body extraction tasks, ablation experiments were conducted using the U-Net model as the baseline network. The DCU module was added under the same conditions to compare the performance of different models. The experimental results are shown in Table 3.
In Table 3, we can see that in the water extraction task, the baseline network achieved an IOU of 89.31%, while the DCU-Net network with the DCU module achieved an IOU of 92.29%, an improvement of 2.98% over the baseline network. This indicates that the water areas extracted by our proposed algorithm are closer to the true value labels. The baseline network achieved a Precision of 94.76%, while the DCU-Net achieved a Precision of 96.13%, an improvement of 1.37%. This shows that DCU-Net can more accurately identify water areas, reducing the impact of complex environments on water extraction effectiveness. The baseline network’s Recall reached 91.78%, and DCU-Net’s Recall reached 92.14%, an increase of 0.36%. This indicates that DCU-Net is more sensitive to water feature information in images, filtering out invalid information that does not belong to water features. The F 1 score of DCU-Net reached 94.32%, an improvement of 1.49% over the baseline network. This demonstrates that DCU-Net has superior performance in the water extraction task and can better cope with complex environments.

4.4. Comparison Study

To demonstrate the effectiveness of DCU-Net in water body extraction tasks, we conducted a series of comparative experiments. Several mainstream image segmentation algorithms were implemented, including U-Net, DenseUNet, UNet++, and Deeplabv3. The same experimental parameters were used in the experiments, and the results are shown in Table 4.
From the comparative experiments above, it can be seen that compared with mainstream image segmentation algorithms, DCU-Net achieves the highest scores in all four metrics (IOU, Precision, Recall, and F 1 ). Compared to the second-best network, Deeplabv3, DCU-Net leads to improvements of 0.95%, 0.26%, 0.71%, and 0.47%, respectively, in IOU, Precision, Recall, and F 1 . DCU-Net exhibits the lowest latency performance, with the processing speed of a single image reaching 33.6 ms. It possesses a more rapid inference speed and thus has a significant advantage in real-time performance. These results indicate that DCU-Net can more effectively extract the spatial detail information of water bodies in medium- and low-resolution images, cope with the challenges brought by complex environmental changes, and improve the accuracy of water body extraction.
To verify the extraction effect of DCU-Net, some images were selected from the dataset we created as detection samples and compared with the baseline network. The detection results are shown in Table 5. In the images in Table 5, the white boxes represent areas where DCU-Net extracted water bodies more accurately compared to the baseline network, while the yellow boxes indicate areas that the baseline network failed to extract. From these results, we can see that DCU-Net has higher accuracy in areas where the water environment is complex and the water feature information is not obvious. The water boundaries are more accurate, meaning this model can provide more precise data support for soil moisture monitoring and has good applicability.

5. Discussion

The variation in soil moisture within a region is a result of the combined effects of regional climate, surface water infiltration capacity, and human activities. During the transformation process from atmospheric precipitation to surface water and then to soil water, the replenishment sources for soil water and surface water are the same, and soil water mainly comes from the infiltration of surface water. Therefore, this study determines the relationship between the two based on the surface water area data in Ningxia Province and the measured soil moisture data at different depths during the same period, achieving soil moisture inversion within a wide area based on the surface water area.
The goodness of fit ( R 2 ) is commonly used to measure the correlation between two variables. In this study, the effectiveness of surface water area on soil moisture inversion is estimated by calculating the correlation coefficient. When the correlation coefficient between surface water area and soil moisture is positive, it indicates a positive correlation between the two, and vice versa for a negative correlation. Furthermore, the larger the R 2 value, the stronger the correlation between the two. Regression analysis was performed using the surface water area data in Ningxia Province extracted by DCU-Net and the measured soil moisture data at a depth of 20 cm below the surface during the same period, with the results shown in Figure 3. The goodness of fit between the surface water area and soil moisture in the Ningxia region is 0.5962, indicating a strong correlation.

6. Conclusions

This study created a large-scale, multi-scenario, pixel-level water body extraction benchmark dataset based on multi-year remote sensing images from Ningxia Province. The dataset was constructed considering various factors, such as different light intensities, types of water bodies, and levels of complexity of the background. Utilizing this dataset, the DCU-Net network was applied for the pixel-level extraction of water body areas.
To address issues encountered during water body extraction in the region, while controlling the computational load to ensure extraction speed, the DCU module was proposed. This module was applied in the lower layers of the network, which are rich in semantic information, to build the DCU-Net network. Our extensive experiments verify that DCU-Net has excellent performance. Compared to the baseline network, our proposed algorithm shows improvements in four metrics: IoU, Precision, Recall, and F 1 score. This indicates that the algorithm performs better in water body extraction tasks, is better equipped to handle complex environments, and provides more accurate data support for soil moisture monitoring.
Based on the water body area data from different periods in Ningxia Province and the measured soil moisture data from the same time, this study determined the relationship between soil moisture and water body area using a cubic fitting polynomial method, achieving high-precision inversion of soil moisture. Our results indicate a strong correlation between soil moisture and water body area. Therefore, the accuracy of water body extraction by DCU-Net can provide robust data support for soil moisture monitoring in Ningxia Province.

Author Contributions

Methodology, Z.N. and Y.Z.; Software, Z.N.; Resources, Z.G.; Data curation, Z.G.; Writing—original draft, Z.N.; Writing—review & editing, Y.Z.; Supervision, Z.G.; Project administration, Z.N.; Funding acquisition, Z.G. All authors have read and agreed to the published version of the manuscript.

Funding

The research presented herein was funded by the National Natural Science Foundation of China (Grant No. 62365016). The research presented herein was funded by the Central Government supported the Local Special Fund Project of China (Grant No. 2023FRD05034).

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kerr, Y.H.; Waldteufel, P.; Wigneron, J.P.; Delwart, S.; Cabot, F.; Boutin, J.; Escorihuela, M.J.; Font, J.; Reul, N.; Gruhier, C.; et al. The SMOS mission: New tool for monitoring key elements of the global water cycle. Proc. IEEE 2010, 98, 666–687. [Google Scholar] [CrossRef]
  2. Durre, I.; Wallace, J.M.; Lettenmaier, D.P. Dependence of extreme daily maximum temperatures on antecedent soil moisture in the contiguous United States during summer. J. Clim. 1998, 13, 2641–2651. [Google Scholar] [CrossRef]
  3. Filintas, A. Soil moisture depletion modelling using a TDR multi-sensor system, GIS, soil analyzes, precision agriculture and remote sensing on maize for improved irrigation-fertilization decisions. Eng. Proc. 2021, 9, 36. [Google Scholar] [CrossRef]
  4. Babaeian, E.; Sadeghi, M.; Jones, S.B.; Montzka, C.; Vereecken, H.; Tuller, M. Ground, proximal, and satellite remote sensing of soil moisture. Rev. Geophys. 2019, 57, 530–616. [Google Scholar] [CrossRef]
  5. Pei, Y.; Han, X. Overview of GNSS-R detection of soil moisture. Geod. Geodyn. 2021, 41, 140–144. [Google Scholar]
  6. Xiao, B.; Sha, J. Overview of remote sensing inversion methods for soil moisture. Remote Sens. Inf. 2007, 94–98. [Google Scholar]
  7. Lievens, H.; Martens, B.; Verhoest, N.E.C.; Hahn, S.; Reichle, R.H.; Miralles, D.G. Assimilation of global radar backscatter and radiometer brightness temperature observations to improve soil moisture and land evaporation estimates. Remote Sens. Environ. 2017, 189, 194–210. [Google Scholar] [CrossRef]
  8. Zhao, T.; Zhang, L.; Jiang, L.; Chen, Q.; Zhang, Z.; Zhang, Y. Joint inversion of soil moisture using active and passive microwave data. Prog. Earth Sci. 2009, 24, 769–775. [Google Scholar]
  9. Liu, Z.; Hu, H.; Lin, Y.; Yao, Z.; Xie, Z.; Wei, Y.; Ning, J.; Cao, Y.; Zhang, Z.; Dong, L.; et al. Swin transformer v2: Scaling up capacity and resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 12009–12019. [Google Scholar]
  10. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classifcation with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  11. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  12. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  13. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  14. Xie, S.; Girshick, R.; Dollar, P.; Tu, Z.; He, K. Aggregated residual transformations for deep neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1492–1500. [Google Scholar]
  15. Tan, M.; Le, Q. Efcientnet: Rethinking model scaling for convolutional neural networks. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 10–15 June 2019; pp. 6105–6114. [Google Scholar]
  16. Tan, M.; Le, Q. Efcientnetv2: Smaller models and faster training. In Proceedings of the International Conference on Machine Learning, Online, 18–24 July 2021; pp. 10096–10106. [Google Scholar]
  17. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, M. Mobilenets: Efcient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
  18. Dai, J.; Qi, H.; Xiong, Y.; Li, Y.; Zhang, G.; Hu, H.; Wei, Y. Deformable convolutional networks. In Proceedings of the International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 764–773. [Google Scholar]
  19. Zhu, X.; Hu, H.; Lin, S.; Dai, J. Deformable convnets v2: More deformable, better results. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 9308–9316. [Google Scholar]
  20. Raffel, C.; Shazeer, N.; Roberts, A.; Lee, K.; Narang, S.; Matena, M.; Zhou, Y.; Li, W.; Liu, P.J. Exploring the limits of transfer learning with a unified text-to-text transformer. J. Mach. Learn. Res. 2020, 21, 1–67. [Google Scholar]
  21. Radford, A.; Wu, J.; Child, R.; Luan, D.; Amodei, D.; Sutskever, I. Language models are unsupervised multitask learners. OpenAI Blog 2019, 1, 9. [Google Scholar]
  22. Shoeybi, M.; Patwary, M.; Puri, R.; LeGresley, P.; Casper, J.; Catanzaro, B. Megatron-LM: Training Multi-Billion Parameter Language Models Using GPU Model Parallelism. arXiv 2019, arXiv:1909.08053. [Google Scholar]
  23. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  24. Wang, W.; Xie, E.; Li, X.; Fan, D.P.; Song, K.; Liang, D.; Lu, T.; Luo, P.; Shao, L. Pyramid vision transformer: A versatile backbone for dense prediction without convolutions. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 568–578. [Google Scholar]
  25. Wang, W.; Xie, E.; Li, X.; Fan, D.P.; Song, K.; Liang, D.; Lu, T.; Luo, P.; Shao, L. Pvt v2: Improved baselines with pyramid vision transformer. Comput. Vis. Media 2022, 8, 415–424. [Google Scholar] [CrossRef]
  26. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional networks for biomedical image segmentation. In Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015, Proceedings of the 18th International Conference, Munich, Germany, 5–9 October 2015; Springer: Berlin/Heidelberg, Germany, 2015. [Google Scholar]
  27. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networksfor semantic segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 39, 640–651. [Google Scholar]
  28. Liu, J.; Song, W.; Zhu, H.; Zhang, B.; Dai, J. Research on Remote Sensing Image Water Extraction with Improved U-Net Model. Surv. Sci. 2022, 47, 230–239. [Google Scholar]
  29. Wang, W.; Dai, J.; Chen, Z.; Huang, Z.; Li, Z.; Zhu, X.; Hu, X.; Lu, T.; Lu, L.; Li, H.; et al. Internimage: Exploring large-scale vision foundation models with deformable convolutions. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 14408–14419. [Google Scholar]
  30. Yahyatabar, M.; Jouvet, P.; Cheriet, F. Dense-Unet: A light model for lung fields segmentation in Chest X-Ray images. In Proceedings of the 42nd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Montreal, QC, Canada, 20–24 July 2020; IEEE: New York, NY, USA, 2020; pp. 1242–1245. [Google Scholar]
  31. Zhou, Z.; Rahman Siddiquee, M.M.; Tajbakhsh, N.; Liang, J. Unet++: A nested u-net architecture for medical image segmentation. In Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support, Proceedings of the 4th International Workshop, DLMIA 2018, and 8th International Workshop, ML-CDS 2018, held in Conjunction with MICCAI 2018, Granada, Spain, 20 September 2018, Proceedings 4; Springer International Publishing: Berlin/Heidelberg, Germany, 2018; pp. 3–11. [Google Scholar]
  32. Yurtkulu, S.C.; Şahin, Y.H.; Unal, G. Semantic segmentation with extended DeepLabv3 architecture. In Proceedings of the 27th Signal Processing and Communications Applications Conference (SIU), Sivas, Turkey, 24–26 April 2019; IEEE: New York, NY, USA, 2019; pp. 1–4. [Google Scholar]
Figure 1. DCU-Net network architecture diagram.
Figure 1. DCU-Net network architecture diagram.
Electronics 14 01011 g001
Figure 2. DCU module structure diagram.
Figure 2. DCU module structure diagram.
Electronics 14 01011 g002
Figure 3. Relationship between water area and soil moisture.
Figure 3. Relationship between water area and soil moisture.
Electronics 14 01011 g003
Table 1. Comparison of soil moisture remote sensing detection.
Table 1. Comparison of soil moisture remote sensing detection.
Remote Sensing Monitoring TechnologySpectral RangeAdvantagesDisadvantages
Optical remote sensingVisible lightHigh spatial resolution, wide range of data sources, relatively mature technology.Weak penetration capability, easily affected by weather, unable to conduct all-weather observations.
Thermal infrared
Microwave remote sensingActive microwaveLittle atmospheric interference, strong penetration, not affected by weather.Low spatial resolution, easily affected by surface roughness and vegetation.
Passive microwave
Passive and active microwave
Table 2. Remote sensing images of water bodies in the Yellow River basin, Ningxia Province.
Table 2. Remote sensing images of water bodies in the Yellow River basin, Ningxia Province.
CategoryImage
Example 1Electronics 14 01011 i001
Example 2Electronics 14 01011 i002
Example 3Electronics 14 01011 i003
Table 3. Ablation study.
Table 3. Ablation study.
Model.DCU ModuleIoU/%Precision/%Recall/% F 1 /%Flops/G
U-Net89.3194.7691.7892.8390.2
DCU-Net (ours)92.2996.1392.1494.3292.1
Table 4. The water extraction results using different algorithms on a custom dataset.
Table 4. The water extraction results using different algorithms on a custom dataset.
ModelIoU/%Precision/%Recall/% F 1 /%Latency
U-Net89.3194.7691.7892.8352.9 ms
DenseUNet [30]90.6293.8291.8792.9049.6 ms
UNet++ [31]91.1194.8692.0893.2440.1 ms
Deeplabv3 [32]91.3495.8791.4393.8535.1 ms
DCU-Net (ours)92.2996.1392.1494.3233.6 ms
Table 5. Comparison of water extraction effects using DCU-Net.
Table 5. Comparison of water extraction effects using DCU-Net.
ExampleInitial ImageDCU-NetU-Net
example 1Electronics 14 01011 i004Electronics 14 01011 i005Electronics 14 01011 i006
example 2Electronics 14 01011 i007Electronics 14 01011 i008Electronics 14 01011 i009
example 3Electronics 14 01011 i010Electronics 14 01011 i011Electronics 14 01011 i012
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Na, Z.; Guo, Z.; Zhu, Y. Soil Moisture Monitoring Based on Deformable Convolution Unit Net Algorithm Combined with Water Area Changes. Electronics 2025, 14, 1011. https://doi.org/10.3390/electronics14051011

AMA Style

Na Z, Guo Z, Zhu Y. Soil Moisture Monitoring Based on Deformable Convolution Unit Net Algorithm Combined with Water Area Changes. Electronics. 2025; 14(5):1011. https://doi.org/10.3390/electronics14051011

Chicago/Turabian Style

Na, Zihao, Zhonghua Guo, and Yang Zhu. 2025. "Soil Moisture Monitoring Based on Deformable Convolution Unit Net Algorithm Combined with Water Area Changes" Electronics 14, no. 5: 1011. https://doi.org/10.3390/electronics14051011

APA Style

Na, Z., Guo, Z., & Zhu, Y. (2025). Soil Moisture Monitoring Based on Deformable Convolution Unit Net Algorithm Combined with Water Area Changes. Electronics, 14(5), 1011. https://doi.org/10.3390/electronics14051011

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop