Next Article in Journal
A Novel Technique for Solving the Nonlinear Fractional-Order Smoking Model
Next Article in Special Issue
Fractional-Order Modeling and Identification for an SCR Denitrification Process
Previous Article in Journal
Screen-Printed Metamaterial Absorber Using Fractal Metal Mesh for Optical Transparency and Flexibility
Previous Article in Special Issue
Adaptive Fractional-Order Multi-Scale Optimization TV-L1 Optical Flow Algorithm
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Crop and Weed Segmentation and Fractal Dimension Estimation Using Small Training Data in Heterogeneous Data Environment

Division of Electronics and Electrical Engineering, Dongguk University, 30 Pildong-ro, 1-gil, Jung-gu, Seoul 04620, Republic of Korea
*
Author to whom correspondence should be addressed.
Fractal Fract. 2024, 8(5), 285; https://doi.org/10.3390/fractalfract8050285
Submission received: 20 March 2024 / Revised: 8 May 2024 / Accepted: 8 May 2024 / Published: 10 May 2024

Abstract

:
The segmentation of crops and weeds from camera-captured images is a demanding research area for advancing agricultural and smart farming systems. Previously, the segmentation of crops and weeds was conducted within a homogeneous data environment where training and testing data were from the same database. However, in the real-world application of advancing agricultural and smart farming systems, it is often the case of a heterogeneous data environment where a system trained with one database should be used for testing with a different database without additional training. This study pioneers the use of heterogeneous data for crop and weed segmentation, addressing the issue of degraded accuracy. Through adjusting the mean and standard deviation, we minimize the variability in pixel value and contrast, enhancing segmentation robustness. Unlike previous methods relying on extensive training data, our approach achieves real-world applicability with just one training sample for deep learning-based semantic segmentation. Moreover, we seamlessly integrated a method for estimating fractal dimensions into our system, incorporating it as an end-to-end task to provide important information on the distributional characteristics of crops and weeds. We evaluated our framework using the BoniRob dataset and the CWFID. When trained with the BoniRob dataset and tested with the CWFID, we obtained a mean intersection of union (mIoU) of 62% and an F1-score of 75.2%. Furthermore, when trained with the CWFID and tested with the BoniRob dataset, we obtained an mIoU of 63.7% and an F1-score of 74.3%. We confirmed that these values are higher than those obtained by state-of-the-art methods.

1. Introduction

Expanding crop efficiency is becoming increasingly important as food security concerns are increasing globally. However, it faces many challenges, such as a lack of manpower, uncertain environmental conditions, soil factors, and scarcity of water. To achieve high yields and address such problems, previous research used plant phenotyping methods and monitoring systems to strengthen crop productivity in precision agriculture [1,2]. Over time, traditional farming has shifted toward modern automated farming to increase yields and minimize labor costs, individual effort, and time.
Image processing has been widely adopted in various fields [3,4,5,6,7,8,9,10,11]. In addition, recently, deep learning methods have offered various solutions, and the use of computer vision has grown significantly in various applications including building monitoring, image enhancement, medical image processing, biomedical engineering, and underwater computer vision, where some research has adopted fractal-related perspectives, also in [9,10,11,12,13,14,15,16]. Although the studies adopt similar concepts of fractal dimension (FD) estimations [12,13,15], their applications are different in terms of building monitoring and medical image processing. Semantic segmentation also exhibits a fundamental role in the accurate recognition of crops and weeds [17,18]. Two mainstream methods exist for crop and weed detection. The first is box-based detection [19,20]; however, it has the drawback of overlooking specific regions in weeds and crops. The second type is semantic segmentation or pixel-based detection [21,22,23], which detects precise regions of weeds or crops at the pixel level. Therefore, correctly identifying the crop and weed segments is essential. Crop and weed segments typically have irregular shapes, and training with these irregularly shaped segments comes with an imbalance in the data. The irregular shapes of “sugar beets” (BoniRob) [24] and crop/weed field image datasets (CWFIDs) [25] are shown in Figure 1. A considerable amount of data is usually required during training, particularly in segmentation cases. To acquire more data, it may be necessary for experts to create a large number of annotations, which would require considerable time and effort. In some scenarios, if a considerable amount of training data are unavailable, the testing performance decreases. To avoid such a drop in performance owing to insufficient training data, a method using small amounts of training data was proposed by Nguyen et al. [26], which also achieved good performance. Homogenous data usually show satisfactory results; however, when applied to heterogeneous environments, where the testing and training datasets are completely different, the overall results decrease significantly. Abdalla et al. [27] showed that the efficacy of these algorithms is compromised in complex environments due to their heavy reliance on various factors, including lighting conditions and weed density, for feature extraction. Thus, it is essential to propose an effective and reliable framework that segments crops and weeds accurately, even in complex heterogeneous environments.
To solve these issues, we propose an approach for weed and crop segmentation and fractal dimension estimation using a small amount of training data in a heterogeneous data environment. The contributions of this study include the following:
  • This is the first study that considers the segmentation of crops and weeds within a heterogeneous environmental setup utilizing one training data sample. We rigorously investigate the factors that cause performance degradation in heterogeneous datasets, including variations in illumination and contrast. To address this problem, we propose a method that applies the Reinhard (RH) transformation, leveraging the mean and standard deviation (std) adjustments.
  • We address the issue of high data availability for real-world applications. For this purpose, we improved the performance using a small amount of training data. The small amount of additional training data significantly improves the segmentation performance while requiring fewer computational resources and less training time.
  • We introduce the FD estimation approach in our framework, which is seamlessly combined as an end-to-end task to provide important information on the distributional features of crops and weeds.
  • It is noteworthy that our proposed framework [28] is publicly accessible for a fair comparison with other studies.
The structure of the remaining sections of this study is as follows: Multiple related studies are described in Section 2. In Section 3, we outline the proposed approach. Section 4 describes a comparison between the proposed framework and the state-of-the-art (SOTA) methods in terms of their performance. Section 5 presents a discussion, while Section 6 presents the conclusion and outlines future work.

2. Related Work

We classified previous studies on crop and weed segmentation into homogenous data-based methods and heterogeneous data-based methods as follows.

2.1. Homogenous Data-Based Methods

Homogeneous data-based methods mostly exhibit high accuracy because their training and testing data distributions originate from the same dataset. Many related studies using homogenous data have been conducted to date, and they have been highly effective in multiple domains, not only in agriculture. In general, the term “a large amount of training data” is used extensively in the literature. Typically, a considerable amount of training data is needed to efficiently train the model and achieve higher accuracy. Many prior studies have been conducted using a large amount of training data. Furthermore, learning-based methods are grouped into two main groups: the handcrafted feature-based and the deep learning-based methods.

2.1.1. Handcrafted Feature-Based Methods

Before the significant advancements in deep learning, features were often manually engineered, referred to as manual or handcrafted features, as they were developed progressively. In [29], a random forest classifier (RFC) is used to handle the overlap of together-grown different crops and weed plants. A Markov random field was also applied to smooth the sparse pixels. Another study by Lottes et al. [30] used the same RFC for vegetation detection using local and object-based features. Lottes et al. [31] used unmanned aerial vehicles (UAVs) and various robots to monitor weeds and crops. They implemented and evaluated plant-tailored feature extraction. Many systems rely on these techniques, primarily because they require fewer computations and have shorter execution times.

2.1.2. Deep Feature-Based Methods

Deep learning techniques utilizing deep features are advancing to automate precision agriculture [32], particularly by making intelligent decisions in the semantic segmentation of crops and weeds. Pixelwise classification networks play a crucial role in detecting objects and properly delimiting their boundaries so that automated robotic weeders can perform precision spraying and weeding operations. Commonly used base networks in semantic segmentation studies include DeepLab [33], fully convolutional networks [34], U-Net [35], and SegNet [36]. These networks, along with certain blocks proposed in various studies, employ an encoder–decoder architecture for crop and weed segmentation. The encoder architecture transforms input data into a compressed representation capturing their key features, while the decoder module upsamples and restores the spatial features of areas where the edges of objects are absent. The base U-Net encoder–decoder network has undergone modifications into multiple architectures, as seen in the work by Zou et al. [37]. They achieved this by reducing feature extraction in the encoder and adding a skip connection at the output layer to recover object details, thereby enhancing model accuracy. They conducted two-stage training to accurately segment weeds and showcased greater applicability in the field. However, the robustness of these models has not been tested using heterogeneous datasets.
Milioto et al. [38] designed an end-to-end model identical to the previously used encoder–decoder format. This network is narrow and fast; however, the dataset used here covers a very small portion of crops and weeds. The model was developed by modifying Enet [39] and SegNet [36] through the replacement of convolutional (Conv) layers with residual blocks. Fathipoor et al. [40], based on an encoder–decoder in U-Net and U-Net++ [41] architectures, demonstrated promising overall results for weed segmentation in the early stages. In a prior study [18], a two-stage approach named MTS-CNN was proposed to segment crops and weeds utilizing U-Net with a visual geometry group (VGG)-16 [42]. The model separates object segmentation from crop and weed segmentation in two stages to enhance accuracy, also creating a loss function to address the class imbalance problem of the crop and weed dataset. However, errors made by the first model can impact overall model performance, and training takes a considerable amount of time. Another study [43] relied on images captured by different cameras mounted on an unmanned aerial vehicle (UAV) for crop and weed segmentation. The authors used a modified VGG-16 encoder and modified U-Net decoder architecture, concatenating images of different formats into channel directions to improve segmentation accuracy. Alongside recent developments in deep convolutional neural networks (CNNs), several new networks have been designed to enhance crop and weed segmentation. Dilated convolution [44] and atrous convolution [33] were integrated into the network alongside a universal function approximation block (UFAB) [45] to improve segmentation. However, these networks require inputs of near-infrared (NIR) light and red, green, and blue (RGB) channels that are unavailable in real time. Wang et al. [46] devised a dual attention network (DA-Net) bridging the gap between low- and high-level featured data using branch and spatial attention. The employed self-attention is computationally demanding due to the size of the spatial features. Siddiqui et al. [47] explored data augmentation (DA) using CNN methods to distinguish weeds from crops. In another study, Khan et al. [48] introduced a new cascaded encoder–decoder network (CED-Net) modifying the base network U-Net into four stages to distinguish between weeds and crops. The inclusion of stages in the network enhanced crop and weed segmentation accuracy. From the above discussion, we can conclude that these deep learning-based methods offer greater accuracy than handcrafted feature-based methods. However, all prior studies were conducted in a homogeneous data environment, where training and testing were performed using the same dataset.

2.2. Heterogeneous Data-Based Methods

Previous studies on crop and weed segmentation have not explored the use of heterogeneous data, where training and testing are conducted using different datasets. However, a model trained with the first dataset is often applied to the second dataset without intensive training using the second dataset. Additionally, sufficient training data cannot often be acquired for real-world applications. However, no previous studies have considered insufficient training data for crop or weed segmentation. Therefore, we propose a framework for weed and crop segmentation and FD estimation utilizing limited training data in a heterogeneous data environment. The strengths and weaknesses of the proposed framework for crop and weed segmentation relative to other techniques are listed in Table 1.

3. Proposed Method

3.1. Overview of Proposed Method

Figure 2 depicts an overview of the flow of our framework. During training, we train the conventional semantic segmentation model with Dataset A. Next, we preprocess the images of Dataset B using the RH transformation, which is based on the mean and std of Datasets A and B. This transformation adjusts the visual properties of the images, such as intensity, illumination, and contrast, to make them similar to the reference image in a heterogeneous environment. Then, we select one training data from the preprocessed Dataset B and perform DA on it to augment the training data. Afterward, we perform fine-tuning and train the model from scratch with Dataset A using the augmented data, and we perform the semantic segmentation of weeds and crops utilizing the testing data from Dataset B.

3.2. Preprocessing

3.2.1. RH Transformation

Many factors such as variations in intensity, illumination, and contrast cause performance degradation, particularly in heterogeneous environments. Transferring colors from one image (reference) to another (target) is a significant problem, particularly when the color information in the reference image does not match the newly generated image, causing quality and performance degradation. To address performance degradation using heterogeneous data, we adopted the RH transformation [49] with average mean and std adjustments to enhance the visual attributes of the images. Many prior transformation methods and color spaces enhance the visual characteristics of an image. In the Ɩαβ color space (LCS) [50], “Ɩ” indicates a brightness channel that captures the brightness independently of color attributes, the “α” channel encapsulates yellow and blue hues, and the “β” channel encloses the interplay between red and green shades. In the RH transformation, the LCS is used. The LCS minimizes the relative significance of each weight, with a matrix proposed for converting vectors from RGB to LCS. The RH transformation changes the mean and std values of the color channels based on the LCS that consistently represents the pixel colors in an image. As shown in Equation (1), it uses the ƒ mapping function with δ parameters to transform the d a t a B into preprocessed d a t a B .
d a t a B = ƒ ( d a t a B ,   δ )
For the transformation, the RGB images are manipulated in the LCS and then transformed into the long, medium, short (LMS) cone space, and the logarithmic of the LMS space is obtained to reduce skewness [49]. The mean and std for all the axes in the LCS are separately calculated to make images more synthetic. Moreover, the color space is normalized by subtracting the mean of the data points from the original data point value, as follows:
Ɩ ^ = Ɩ Ɩ m ,
α ^ = α α m ,
β ^ = β β m
where Ɩ m ,   α m ,   β m denote the average mean data point values of the Ɩ , α ,   a n d   β data points, and Ɩ ^ ,     α ^ ,   a n d   β ^ show the normalized space data points. Upon normalizing the space, data points are scaled by std σ j Ɩ ,   σ j α ,   σ j β for the reference images and σ i Ɩ ,   σ i α ,   σ i β for the target images. Finally, Ɩ ° , α ° ,   β ° are the scaled points for the transformed images, as follows:
Ɩ ° = σ i Ɩ σ j Ɩ Ɩ ^ ,
α ° = σ i α σ j α α ^ ,
β ° = σ i β σ j β β ^
The pseudo-code of the RH transformation is provided in Algorithm 1.
Algorithm 1: RH transformation with pseudo-code
Input: {dataB} tn; The total of t data samples, dataB: input dataset B images
dataAm: training data from dataset A.
Output: (dataB’) the preprocessed sample
1:  Compute the std of the dataset B (input images) by
   dataB _std = std (dataB (:, :))
2:  Compute the mean of the dataB
   dataB _mean = mean (dataB (:, :))
3:  Compute the average std of the training data from dataset A using
   dataA_std = avg (std (dataAm (:, :))
4:  Compute the average mean of the training data from dataset A using
   dataA_mean = avg (mean (dataAm (:, :))
5:  Apply the RH transformation
   for m = 1: x
    for n = 1: y
     dataB’ (m, n) = [(dataB (m, n) − dataB_mean) ×
     (dataA_std/dataB_std)] + dataA_mean
    end
   end
   return dataB’
Histograms of the RGB color channel distribution for the reference, target, and newly transformed images are shown in Figure 3a–c, respectively. The reference image is from Dataset A, the target image is from Dataset B, and the RH-transformed image is a new image generated by RH transformation. In Figure 3b, the histograms show only two channels; apparently, the red and blue channels overlap because they are replicas of each other [25]. The histogram shows the pixel values on the x-axis and the probability distribution of the pixels on the y-axis. As shown in Figure 3c, the histogram of the target image is more akin to that of the reference image in Figure 3a by RH transformation than to that of the target image in Figure 3b before RH transformation.

3.2.2. RH Transformation with Additional Adjustments

However, despite the RH transformation, there is still a difference in the relative probability distribution of each channel in Figure 3a,c. Therefore, we introduce an additional adjustment to obtain the relative probability distribution of each channel of the RH-transformed image that is more similar to that of the reference image as follows:
m i n s e l e c t p i o p t   R t R r p i 2 ,
m i n s e l e c t q i o p t   G t G r q i 2 ,
m i n s e l e c t r i o p t   B t B r r i 2
where R t ,   G t , a n d   B t represent the pixels in the target image, and R r ,   G r ,   a n d   B r represent the pixels in the reference image. Moreover, p i o p t ,   q i o p t ,   a n d   r i o p t are the optimal values generated by varying the p i ,   q i , a n d r i values, respectively, to minimize the distance between the corresponding red, green, and blue channel distributions. To obtain this channel distribution, the following additional adjustments were made:
Ɩ a d j = Ɩ m p i o p t ,
α a d j = α m q i o p t ,
β a d j = β m r i o p t
where Ɩ m ,   α m ,   a n d   β m represent the average mean data point values, and p i o p t ,   q i o p t ,   a n d   r i o p t represent the values that make the channel distribution look more similar among the reference and transformed images. These p i o p t ,   q i o p t ,   r i o p t values are subtracted from the average mean values to adjust the Ɩ a d j ,   α a d j ,   a n d   β a d j data points. Furthermore, the transformation with additional adjustments is as follows:
Ɩ = Ɩ Ɩ a d j ,
α = α α a d j ,
β = β β a d j
where Ɩ a d j ,   α a d j ,   a n d   β a d j are the different data point values for additional adjustments, which are experimentally chosen to make the channel distribution similar to the reference image. These values are subtracted from the Ɩ , α ,   β data points, and the result is Ɩ ,     α ,   a n d   β , which shows the normalized space data points. The newly generated histograms are visually presented in Figure 3d. As demonstrated in Figure 3d, the channel distributions resulting from the RH transformation with additional adjustments closely resemble those of the reference image in Figure 3a, in contrast to the RH without additional adjustment in Figure 3c.
Figure 4 illustrates the sample images of the referenced image, targeted image, RH-transformed image, and RH-transformed image with additional adjustments from the BoniRob dataset and CWFID.

3.3. Data Augmentation of One Training Data

Data augmentation (DA) involves employing various transformations on existing data to expand and diversify a training dataset artificially. In image datasets, DA can involve procedures such as in-plane rotation, flipping, zooming, and modifying the contrast and brightness levels of the images within the training dataset [51]. The purpose behind DA is to enhance the capacity of a model to generalize and perform effectively on novel and unseen data by exposing it to a broader spectrum of variations that may be encountered in real-world contexts. We randomly selected a single image from the training data and augmented it for training. By adding augmentation to this small training dataset, we improved the segmentation performance while reducing the training time. A simple 180-degree in-plane rotation was used for DA, as depicted in Figure 5.

3.4. Semantic Segmentation Networks

For the semantic segmentation of weeds and crops, the following conventional semantic segmentation models were adopted:

3.4.1. U-Net

U-Net [35] features a fully CNN with a U-shaped encoder–decoder framework. The input image of the encoder and the output image of the decoder have the same image size. The decoder is more or less symmetric than the encoder, and the encoder has a large feature channel for propagating contextual information to the high-resolution layer. The architecture includes two 3 × 3 Conv layers, each accompanied by a max-pooling (MP) layer of 2 × 2 kernel size, with stride 2, and a ReLU [52]. In the decoder, the tensor map was upsampled using a 2 × 2 up-convolution and concatenated with the encoder features. A final layer of 1 × 1 convolution was incorporated into the U-Net network. Figure 6 illustrates its architecture. Furthermore, the first stages of the encoder (including and before the first max-pooling layer as E n c o d e r 1 ) and decoder (first up Conv layer with subsequent layers having the same spatial dimensions as D e c o d e r 1 ) of the U-Net shown in Figure 6 are mathematically represented as follows:
E n c o d e r 1 = P 1 C 1   w h e r e   C 1 = C o n v C o n v X , W 1 , P 1 = M a x p o o l C 1
D e c o d e r 1 = C R 1 C c o n c a t 1 C c r o p 1 U 1 ,
w h e r e   U 1 = U p c o n v B ,   W u p 1 , C c r o p 1 = C r o p R , s i z e U 1 , C c o n c a t 1 = C o n c a t e n a t i o n U 1 , C c r o p 1 , C R 1 = C o n v C o n v C c o n c a t 1 , W c o n v
In E n c o d e r 1 , X represents the input feature, and W 1 indicates the weight tensor for convolution operations. In every convolution operation for both the encoder and decoder, ReLU activation functions and batch normalization are applied. In D e c o d e r 1 , B represents the tensor of the previous layer, and W u p 1 in U p c o n v ( B , W u p 1 ) shows the weight tensor with an up-convolution operation. R in C r o p R , s i z e ( U 1 ) indicates the input image size after convolutions from the encoder side with the same size as U 1 , which is a result of the previous up-convolution. C c o n c a t 1 represents the concatenation of U 1 and C c r o p 1 which are the final calculations for the previous layers. W C o n v represents weight tensors with convolution operations. Finally, after the convolution in the decoder, it is fed to C R 1 , and the subsequent levels of decoder operations continue in a similar manner.

3.4.2. Modified U-Net

Modified U-Net [37] is an updated variant of the U-Net, featuring an encoder–decoder framework. Like U-Net, the modified version includes two 3 × 3 Conv layers, each accompanied by an MP layer of 2 × 2 kernel size of stride 2, and an activation function named the exponential linear unit [53]. In the decoder, the tensor map is upsampled using a 2 × 2 up-convolution and concatenated with the encoder features. To overcome the overfitting issue, a dropout layer was placed between the Conv layers. A last layer with 1 × 1 Conv is used in the modified U-Net architecture. Moreover, the stochastic gradient descent optimization is replaced with the Adadelta algorithm. Figure 7 shows the structure of the modified U-Net. Furthermore, the first stages of the encoder (before and including the first pooling layer as E n c o d e r 1 ) and decoder (first up Conv layer with the subsequent layer including convolutions as D e c o d e r 1 ) of Figure 7 are mathematically expressed as follows:
E n c o d e r 1 = P 1 C 1 D 1 B 1   w h e r e   B 1 = C o n v X , W b 1 , D 1 = D r o p o u t B 1 ,   d r o p o u t v a l u e , C 1 = C o n v D 1 , W c 1 , P 1 = M a x p o o l C 1
D e c o d e r 1 = C R 1 R 1 C B 1 U 1 C C 1 ,
w h e r e   C C 1 = C o n c a t e n a t e B , D n , U 1 = U p c o n v C C 1 , W u 1 , C B 1 = C o n v U 1 , W c b 1 , R 1 = D r o p o u t C B 1 ,   d r o p o u t v a l u e , C R 1 = C o n v ( R 1 , W c r 1 )
In E n c o d e r 1 , X represents the input image, and W b 1 represents the weight tensor for convolution operations. Following every C o n v operation in both the encoder and decoder batch normalization, ELU is applied. In E L U ( C o n v ( D 1 , W c 1 ) , D 1 represents the dropout layer, which is further processed by convolution and E L U resulting in C 1 . C 1 is downsampled using the MP layer. In D e c o d e r 1 , B represents the tensor map of the prior layer, and D n represents the dropout feature map from the same level connection. W u 1 in U p c o n v ( C C 1 , W u 1 ) represents the weight tensor with an upconvolution operation. Subsequently, a convolution is applied to the W c b 1 weight tensor. R 1 has a dropout value on which convolution is applied with the W c r 1 weight tensor. After the activation function, the final results are concluded in C R 1 with the completion of the first decoder-level operations.

3.4.3. CED-Net

Although many other architectures for segmentation networks are deep, with a large number of parameters, CED-Net [48] is a simple cascading semantic segmentation model with a smaller number of parameters. The complete network comprises two training levels. Model 1 is trained at level 1 using the following steps. Model 1 is trained for predicting weeds, while Model 3 is trained for the prediction of crops. The results of Models 1 and 3 are upsampled and stitched to scale for each input image and then utilized as inputs for Models 2 and 4, respectively. The models are trained separately at each level. At Level 1, the images with their segmentation ground truths are reduced in size to spatial dimensions of 448 × 448. To train the Level 2 models, the 448 × 448 spatial dimension is upsampled to 896 × 896 using bilinear interpolation. The only difference from the U-Net structure is the highest feature map of 256 sizes in the bottleneck layer. All models and levels use the same encoder–decoder network architecture for CED-Net, as shown in Figure 8. Moreover, the first stages of CED-Net at each level in the model are akin to those of the U-Net encoder–decoder, other than the network depth and the number of convolutions. The encoder (including and before the first MP layer as E n c o d e r 1 ) and decoder (first up Conv layer with more layers having convolutions as D e c o d e r 1 ) of CED-Net, shown in Figure 8, are mathematically represented as follows:
E n c o d e r 1 = P 1 D 1   w h e r e   D 1 = C o n v C o n v Y , W 1 , P 1 = M a x p o o l ( C 1 )
D e c o d e r 1 = F R 1 C c o n c a t 1 C c r o p 1 U 1 ,
w h e r e   U 1 = U p c o n v B ,   W u p 1 , C c r o p 1 = C r o p R , s i z e U 1 , C c o n c a t 1 = C o n c a t e n a t i o n U 1 , C c r o p 1 , F R 1 = C o n v C o n v C o n v C c o n c a t 1 , W c o n v
In E n c o d e r 1 , Y represents the input feature, and W 1 represents the weight tensor for the convolution operations. ReLU and batch normalization are employed in each Conv operation for both the encoder and decoder. In D e c o d e r 1 , B depicts the feature map of the prior layer, and W u p 1 in U p c o n v ( B , W u p 1 ) shows the weight tensor with an upconvolution operation. In C r o p R , s i z e ( U 1 ) , R represents the input image size after convolutions on the encoder side, converted to the same size as U 1 , which is the result of a previous upconvolution. C c o n c a t 1 represents the concatenation of U 1 and C c r o p 1 which are the final calculations for the previous layers. W C o n v represents weight tensors with convolution operations. Finally, after convolutions in the decoder, it is fed to F R 1 , and the next stages of the decoder computations continue in a similar manner. At each level of CED-Net, a similar encoder–decoder architecture is used for network operations.
Figure 8. Encoder and decoder architecture of CED-Net.
Figure 8. Encoder and decoder architecture of CED-Net.
Fractalfract 08 00285 g008

4. Experimental Results

4.1. Experimental Dataset and Setup

We used two open datasets: the BoniRob dataset [24] as Dataset A and the CWFID [25] as Dataset B. The BoniRob dataset contains 496 images with pixel-level annotated masks, while the CWFID contains 60 images with the same pixel-level annotated masks. Both dataset images were captured by an autonomous robot in the field at a resolution of 1296 × 966 pixels. For small training datasets, we use only a single image, and augmentation produces an additional image. We utilized two images (one augmented and one small training data) for training. During training, we reduced the image size to 512 × 512 pixels. We experimented on a Windows-based desktop system with an Intel Core i5-2320 CPU @ 3.00 GHz processor [54], a GPU of NVIDIA GeForce GTX 1070 [55] with 8 gigabytes of memory, and 16 gigabytes of RAM. For development, we utilized the PyTorch [56] platform in Python version 3.8 [57].

4.2. Training Setup

For Experiment 1, Dataset A was first used for training with 70% of the data. Dataset B, transformed using the proposed method, was subsequently divided into two equal portions: one serving as testing data and the other as small training data. A single image of the small training data was then augmented and used for training, whereas the testing data remained unchanged. In Experiment 2, 70% of the data in Dataset B were used for training. Next, the data were preprocessed, and Dataset A was subsequently divided into two equal portions: one serving as testing data and the other as small training data. We augmented a single image of the small training data and further utilized it for training, while the test data remained unchanged. During training, the images underwent resizing to a resolution of 512 × 512 pixels, and the network was trained using a batch size of two for 150 epochs. We employed an Adam optimizer [58] with an initial learning rate (LR) of 1 × 10−5 and utilized a cosine annealing strategy [59] to steadily reduce the LR during training. The training loss was calculated using dice loss [60]. We trained the proposed framework using the U-Net, modified U-Net, and CED-Net architectures. Figure 9 depicts the loss and training graphs for the BoniRob dataset and the CWFID using U-Net. Furthermore, the convergence of the loss and accuracy curves of the validation and training data demonstrates that the network is adequately trained and avoids overfitting.

4.3. Evaluation Metrics

The experiment aimed to assess the semantic segmentation performance across three classes (background, crop, and weed) using evaluation metrics including precision, the mean intersection of union (mIoU), recall, and weighted harmonic mean of precision and recall (F1-score), as outlined in Equations (23)–(27). These values were used to calculate the overall segmentation performance of the proposed framework. The number of classes was set to three. The evaluation metrics used to determine the accuracy of segmentation included true negative (TN), true positive (TP), false negative (FN), and false positive (FP) values. When the true and false labels match the prediction, the scenario is usually referred to as a TN or TP. FP and FN are terms used to describe scenarios in which an incorrect label is mistakenly anticipated as true and a valid label is mistakenly anticipated as a false label.
I o U = T P T P + F N + F P ,
m I o U = j = 1 C l s I O U j C l s ,
R e c a l l = T P T P + F N ,
P r e c i s i o n = T P T P + F P ,
F 1 - score = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l

4.4. Testing for the Proposed Framework

Below is the explanation of the experimental results of the proposed framework for semantic segmentation in Experiments 1 and 2.

4.4.1. Testing with CWFID after Training with BoniRob Dataset

Ablation Study

In the ablation studies, we conducted several experiments across various scenarios, encompassing six different cases outlined in Table 2.
We repeated these six cases in Experiments 1 and 2. Table 3, Table 4 and Table 5 present the results of Experiment 1 for semantic segmentation using segmentation networks including U-Net, modified U-Net, and CED-Net. Across all experiments, Case I showed the lowest performance, and Case VI (the proposed method) showed the highest performance. This confirms that the proposed schemes of using RH transformation, one training dataset, and DA can enhance segmentation accuracy in all the segmentation models. In addition, the semantic segmentation performance of Case VI and U-Net is the best.
An ablation study is presented in Table 6 to show the significance of the selection of the mean and std values for the RH transformation in the proposed framework. In this ablation study, we used the same training data as in the ablation study experiments with U-Net of Case VI (Table 3) and performed testing to validate the performance difference between the RH transformation and RH transformation with additional adjustment values that deviate from the selected values. We set the RH transformation with additional adjustment values to make the distribution of channels akin to that of the reference image in the histograms, as shown in Figure 3 and explained in Section 3.2.2. The performance results (F1-score) with additional adjustment values are 13.1% lower than those with the RH transformation without additional adjustments, as listed in Table 6. Although the RH transformation with additional adjustments makes the relative channel distributions similar between the target and reference images, as shown in Figure 3, it results in a greater decrease in the absolute numbers of red and green channels. Additionally, the mean of all channels decreases and moves toward zero, which reduces contrast and illumination, as shown in Figure 3. Moreover, the variance also increases, and all factors collectively cause the performance degradation of the RH transformation with additional adjustments.
Visual examples of the semantic segmentation results using U-Net, modified U-Net, and CED-Net are illustrated in Figure 10. In this illustration, red pixels represent the TP of crops, black pixels represent the TP of the background, and blue pixels signify the TP of the weed. Yellow pixels represent errors where crops were mistakenly identified as background or weeds, while orange pixels represent errors where weeds were mistakenly identified as background or crops. Gray pixels indicate errors where the background was mistakenly identified as weeds or crops. As depicted in the figure, the utilization of U-Net in the proposed framework demonstrates superior semantic segmentation accuracy.

Performance Comparisons of the Proposed Framework with SOTA Transformations

We analyzed the SOTA transformations with those of the proposed method employing U-Net, modified U-Net, and CED-Net, as outlined in Table 7. Our findings confirm that the proposed framework yields the highest outcomes across all segmentation networks. Additionally, Table 7 reveals that U-Net within the proposed framework attains the highest segmentation performance for both crops and weeds.
Visual examples of SOTA transformations and the proposed method using U-Net are shown in Figure 11. In this illustration, red pixels represent the TP of crops, black pixels represent the TP of the background, and blue pixels signify the TP of the weed. Yellow pixels represent errors where crops were mistakenly identified as background or weeds, while orange pixels represent errors where weeds were mistakenly identified as background or crops. Gray pixels indicate errors where the background was mistakenly identified as weeds or crops. As depicted in the figure, the proposed framework demonstrates the highest semantic segmentation accuracy.

4.4.2. Testing with BoniRob Dataset after Training with CWFID

Ablation Study

For the ablation studies, six cases were considered in Experiment 2, as described in the Ablation Study of Section 4.4.1. Table 8, Table 9 and Table 10 present the results of Experiment 2 for semantic segmentation using segmentation networks. From these tables, we can observe that the accuracy is lower for Cases I, III, and IV without RH transformation for the test data, but the performance improved when we used RH transformation in Cases II, IV, and VI. In Case VI, the proposed framework shows the highest accuracy among all segmentation networks and the best accuracy with U-Net.
We can see the visual examples of semantic segmentation outputs using U-Net, modified U-Net, and CED-Net in Figure 12. In this illustration, red pixels represent the TP of crops, black pixels represent the TP of the background, and blue pixels signify the TP of the weed. Yellow pixels represent errors where crops were mistakenly identified as background or weeds, while orange pixels represent errors where weeds were mistakenly identified as background or crops. Gray pixels indicate errors where the background was mistakenly identified as weeds or crops. As depicted in the figure, the proposed framework with U-Net demonstrates the highest semantic segmentation accuracy.

Performance Comparisons of the Proposed Framework with SOTA Transformations

In this subsection, we compared the SOTA transformations with those of the proposed framework using U-Net, modified U-Net, and CED-Net, as listed in Table 11. We verify that the proposed framework shows superior performance over all segmentation networks. Moreover, Table 11 highlights that U-Net, when utilized within the proposed framework, obtained the highest semantic segmentation accuracy for crops and weeds.
Visual examples of SOTA transformations and the proposed method using U-Net are illustrated in Figure 13. In this illustration, red pixels represent the TP of crops, black pixels represent the TP of the background, and blue pixels signify the TP of the weed. Yellow pixels represent errors where crops were mistakenly identified as background or weeds, while orange pixels represent errors where weeds were mistakenly identified as background or crops. Gray pixels indicate errors where the background was mistakenly identified as weeds or crops. As illustrated in the figure, the proposed framework demonstrates the highest semantic segmentation accuracy.

4.5. Fractal Dimension Estimation

The FD is used as a mathematical metric to characterize the complexity of geometric structures, particularly fractal shapes, which exhibit self-similarity across different scales. Fractal shapes possess similar patterns or structures across different scales, and the complexity of these shapes can be quantified using a numerical value of FD that generally ranges between one and two [65]. A higher FD value indicates greater complexity. A common method for computing the FD is the box-counting algorithm [66]. We refer to the box-counting algorithm [12] to compute the FD, which was implemented by the PyTorch [56] platform in Python version 3.8 [57]. Algorithm 2 provides the pseudo-code to measure the FD.
Algorithm 2: Pseudo-code for measuring FD
Input: image (path to the input image)
Output: Fractal dimension (FD) value
1:  Read the input image and further convert it into grayscale
2:   Set the maximum box-size with the power of 2 and ensure the dimensions
     s = 2^(log(max(size(image))/log2)]
    Add the padding if required to match the dimensions
3:   Compute the number of boxes N(s) till minimum pixels
4:   Reduce box size by 2 and recalculate N(s) iteratively
    while s > 1
5:   Compute log(N(s)) and log(1/s) for each s
6:   Draw a fitted line to the points (log(N(s)) and log(1/s))
7:  FD value is the slope of the fitted line
   Return Fractal Dimension Value
The algorithm enables the calculation of the FD for both precise and approximate self-affine patterns and has wide applications across various natural and manmade systems [13,67]. The formula employed to estimate the FD utilizing the box-counting algorithm [12] is as follows:
FD = lim s 0 log ( N ( s ) ) log ( 1 / s )
where N ( s ) represents the sum of boxes of size s , and FD signifies the fractal dimension defining the curve being analyzed. We employed a box-counting technique to calculate the approximate FD of various shapes. The method was evaluated using two datasets, BoniRob and CWFID. The experimental results are presented in Table 12, demonstrating the FD values and distribution of crops and weeds across different sections of the field. The 1st–3rd row FD values in Table 12 were computed from the 1st~3rd row images in Figure 14. In addition, the 4th~6th row FD values in Table 12 were computed from the 4th~6th row images in Figure 15. Higher FD values for crops and weeds represent the high complexities of crops and weeds, suggesting that farming experts or robots should pay more attention to discriminating between crops and weeds. This estimation technique can also automate farming systems by targeting and eliminating weeds through precise spraying in areas with high weed complexity, ultimately increasing the crop yield.

4.6. Comparisons of Processing Time

In this section, we compare the average processing times per image obtained using the proposed method with those obtained using the SOTA method. The unit ms denotes milliseconds. Table 13 illustrates that the method by Xiao et al. [61] exhibits the highest processing time, whereas the method by Nguyen et al. [64] demonstrates the lowest processing time. Interestingly, our proposed framework falls at the second lowest in processing time (as indicated in Table 13). Although the proposed framework has a higher processing time than Nguyen et al. [64], our goal is to achieve high accuracies of crop and weed segmentation. The proposed framework with U-Net yields better results than the SOTA methods, as evidenced in Table 7 and Table 11 and Figure 11 and Figure 13.

5. Discussion

We performed statistical analysis using the Student’s t-test [68] and calculated the Cohen’s d-value [69]. For this purpose, we calculated the mean and std of the mIoU using our method with U-Net, as shown in Table 7 and Table 11, respectively. Additionally, we calculated the mean and std of the mIoU using the second-best method (Pitie et al. [62]) with U-Net, as presented in Table 7 and Table 11. The measured p-value is 0.041, showing a confidence level of 95% with a significant difference, as depicted in Figure 16, and Cohen’s d-value is 0.936, representing a large effect size. This confirms that the proposed framework statistically surpasses the second-best method and achieves higher segmentation accuracy.
We also analyzed the semantic segmentation performance using the proposed method with U-Net based on gradient-weighted class activation mapping (Grad-CAM) [70], as depicted in Figure 17. Grad-CAM typically depicts important features as reddish and yellowish colors, whereas unimportant features are shown in bluish colors as explainable artificial intelligence. Figure 17 shows the Grad-CAM image, obtained after the fourth convolution layer of U-Net, illustrating the accurate extraction of crucial features for crop and weed segmentation, confirming that our method can generate correct images in heterogeneous datasets for the accurate segmentation of crops and weeds.
Figure 18 shows examples of incorrect segmentation using the proposed method with U-Net. In this illustration, red pixels represent the TP of crops, black pixels represent the TP of the background, and blue pixels signify the TP of the weed. Yellow pixels represent errors where crops were mistakenly identified as background or weeds, while orange pixels represent errors where weeds were mistakenly identified as background or crops. Gray pixels indicate errors where the background was mistakenly identified as weeds or crops. The reason for the incorrect segmentation is that crops and weeds have similar shapes and colors, making them hard to distinguish, especially in cases where objects possess thin regions.
As automation in agriculture is a subject of intensive research, it has led to many reformations in agriculture, such as saving time, reducing manpower, increasing yields through proper monitoring systems, and precise crop and weed detection. In agricultural precision for heterogeneous data, our proposed framework exhibits great performance, segregating crops and weeds accurately, which can increase yield and drive the transition to modern agriculture.

6. Conclusions

In this study, we introduce an approach for segmenting crops and weeds and estimating the FD using small amounts of training data in a heterogeneous data environment. This is the first study on the segmentation of crops and weeds within a heterogeneous environmental setup utilizing one training data sample. We rigorously investigated the factors that cause performance degradation in heterogeneous datasets, including variations in illumination and contrast. To solve this challenge, we proposed a framework that leverages mean and standard deviation adjustments using the RH transformation. Furthermore, we improved the performance using a small amount of training data. This additional small amount of training data significantly improves the segmentation performance with reduced computational power and training time. In addition, we introduced an FD estimation approach in our system, which was smoothly combined as an end-to-end task to furnish crucial insights into the distributional characteristics of crops and weeds. Through experiments using two open databases, we proved that our approach outperforms the SOTA method. Additionally, we confirmed that our approach showed statistically superior results than the second-best method in terms of the t-test and Cohen’s d-value. Furthermore, using Grad-CAM images, we validated the capability of the proposed method to extract essential features necessary for the accurate segmentation of crops and weeds. Nonetheless, instances of inaccurate segmentation were observed in scenarios where crops and weeds exhibit comparable colors, shapes, and thin regions, as depicted in Figure 18.
To address this issue, we would research the generative adversarial network-based transformation method for the correct segmentation of small-sized and thin crops and weeds having similar colors and shapes in heterogeneous data environments. Moreover, we applied our method to other tasks, such as box-based detection or classification in heterogeneous data environments. In addition, we would check the possibility of applying our method to other application areas including crack detection for building monitoring, medical image detection, and underwater computer vision.

Author Contributions

Methodology, R.A.; conceptualization, J.S.H.; data curation, S.G.K.; formal analysis, H.S.; resources, M.U.; visualization, H.A.H.G.; software, M.H.T.; validation, N.U.; supervision, K.R.P.; writing—original draft, R.A.; writing—review and editing, K.R.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported in part by the National Research Foundation of Korea (NRF) funded by the Ministry of Science and ICT (MSIT) through the Basic Science Research Program (NRF-2022R1F1A1064291) and in part by MSIT, Korea, under the Information Technology Research Center (ITRC) support program (IITP-2024-2020-0-01789) supervised by the Institute for Information & Communications Technology Planning & Evaluation (IITP).

Data Availability Statement

The datasets are available in [24,25], and the proposed framework is available in [28].

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Jiang, Y.; Li, C. Convolutional Neural Networks for Image-Based High-Throughput Plant Phenotyping: A Review. Plant Phenomics 2020, 2020, 4152816. [Google Scholar] [CrossRef] [PubMed]
  2. Fathipoor, H.; Arefi, H.; Shah-Hosseini, R.; Moghadam, H. Corn Forage Yield Prediction Using Unmanned Aerial Vehicle Images at Mid-Season Growth Stage. J. Appl. Remote Sens 2019, 13, 034503. [Google Scholar] [CrossRef]
  3. Yang, Q.; Wang, Y.; Liu, L.; Zhang, X. Adaptive Fractional-Order Multi-Scale Optimization TV-L1 Optical Flow Algorithm. Fractal Fract. 2024, 8, 8040179. [Google Scholar] [CrossRef]
  4. Huang, T.; Wang, X.; Xie, D.; Wang, C.; Liu, X. Depth Image Enhancement Algorithm Based on Fractional Differentiation. Fractal Fract. 2023, 7, 7050394. [Google Scholar] [CrossRef]
  5. Bai, X.; Zhang, D.; Shi, S.; Yao, W.; Guo, Z.; Sun, J. A Fractional-Order Telegraph Diffusion Model for Restoring Texture Images with Multiplicative Noise. Fractal Fract. 2023, 7, 7010064. [Google Scholar] [CrossRef]
  6. AlSheikh, M.H.; Al-Saidi, N.M.G.; Ibrahim, R.W. Dental X-ray Identification System Based on Association Rules Extracted by k-Symbol Fractional Haar Functions. Fractal Fract. 2022, 6, 6110669. [Google Scholar] [CrossRef]
  7. Zhang, Y.; Yang, L.; Li, Y. A Novel Adaptive Fractional Differential Active Contour Image Segmentation Method. Fractal Fract. 2022, 6, 6100579. [Google Scholar] [CrossRef]
  8. Zhang, Y.; Liu, T.; Yang, F.; Yang, Q. A Study of Adaptive Fractional-Order Total Variational Medical Image Denoising. Fractal Fract. 2022, 6, 6090508. [Google Scholar] [CrossRef]
  9. Jiao, Q.; Liu, M.; Ning, B.; Zhao, F.; Dong, L.; Kong, L.; Hui, M.; Zhao, Y. Image Dehazing Based on Local and Non-Local Features. Fractal Fract. 2022, 6, 6050262. [Google Scholar] [CrossRef]
  10. Zhang, X.; Dai, L. Image Enhancement Based on Rough Set and Fractional Order Differentiator. Fractal Fract. 2022, 6, 6040214. [Google Scholar] [CrossRef]
  11. Zhang, X.; Liu, R.; Ren, J.; Gui, Q. Adaptive Fractional Image Enhancement Algorithm Based on Rough Set and Particle Swarm Optimization. Fractal Fract. 2022, 6, 6020100. [Google Scholar] [CrossRef]
  12. Cheng, J.; Chen, Q.; Huang, X. An Algorithm for Crack Detection, Segmentation, and Fractal Dimension Estimation in Low-Light Environments by Fusing FFT and Convolutional Neural Network. Fractal Fract. 2023, 7, 7110820. [Google Scholar] [CrossRef]
  13. An, Q.; Chen, X.; Wang, H.; Yang, H.; Yang, Y.; Huang, W.; Wang, L. Segmentation of Concrete Cracks by Using Fractal Dimension and UHK-Net. Fractal Fract. 2022, 6, 6020095. [Google Scholar] [CrossRef]
  14. Sultan, H.; Owais, M.; Park, C.; Mahmood, T.; Haider, A.; Park, K.R. Artificial Intelligence-Based Recognition of Different Types of Shoulder Implants in X-Ray Scans Based on Dense Residual Ensemble-Network for Personalized Medicine. J. Pers. Med. 2021, 11, 11060482. [Google Scholar] [CrossRef] [PubMed]
  15. Arsalan, M.; Haider, A.; Hong, J.S.; Kim, J.S.; Park, K.R. Deep Learning-Based Detection of Human Blastocyst Compartments with Fractal Dimension Estimation. Fractal Fract. 2024, 8, 8050267. [Google Scholar] [CrossRef]
  16. González-Sabbagh, S.P.; Robles-Kelly, A. A Survey on Underwater Computer Vision. ACM Comput. Surv. 2023, 55, 1–39. [Google Scholar] [CrossRef]
  17. Madokoro, H.; Takahashi, K.; Yamamoto, S.; Nix, S.; Chiyonobu, S.; Saruta, K.; Saito, T.K.; Nishimura, Y.; Sato, K. Semantic Segmentation of Agricultural Images Based on Style Transfer Using Conditional and Unconditional Generative Adversarial Networks. Appl. Sci. 2022, 12, 12157785. [Google Scholar] [CrossRef]
  18. Kim, Y.; Park, K.R. MTS-CNN: MTS-CNN: Multi-Task Semantic Segmentation-Convolutional Neural Network for Detecting Crops and Weeds. Comput. Electron. Agric. 2022, 199, 107146. [Google Scholar] [CrossRef]
  19. Wang, R.; Jiao, L.; Xie, C.; Chen, P.; Du, J.; Li, R. S-RPN: Sampling-Balanced Region Proposal Network for Small Crop Pest Detection. Comput. Electron. Agric. 2021, 187, 106290. [Google Scholar] [CrossRef]
  20. Huang, S.; Wu, S.; Sun, C.; Ma, X.; Jiang, Y.; Qi, L. Deep Localization Model for Intra-Row Crop Detection in Paddy Field. Comput. Electron. Agric. 2020, 169, 105203. [Google Scholar] [CrossRef]
  21. Kang, J.; Liu, L.; Zhang, F.; Shen, C.; Wang, N.; Shao, L. Semantic Segmentation Model of Cotton Roots In-Situ Image Based on Attention Mechanism. Comput. Electron. Agric. 2021, 189, 106370. [Google Scholar] [CrossRef]
  22. Le Louëdec, J.; Cielniak, G. 3D Shape Sensing and Deep Learning-Based Segmentation of Strawberries. Comput. Electron. Agric. 2021, 190, 106374. [Google Scholar] [CrossRef]
  23. Brilhador, A.; Gutoski, M.; Hattori, L.T.; de Souza Inácio, A.; Lazzaretti, A.E.; Lopes, H.S. Classification of Weeds and Crops at the Pixel-Level Using Convolutional Neural Networks and Data Augmentation. In Proceedings of the IEEE Latin American Conference on Computational Intelligence, Guayaquil, Ecuador, 11–15 November 2019; pp. 1–6. [Google Scholar] [CrossRef]
  24. Chebrolu, N.; Lottes, P.; Schaefer, A.; Winterhalter, W.; Burgard, W.; Stachniss, C. Agricultural Robot Dataset for Plant Classification, Localization and Mapping on Sugar Beet Fields. Int. J. Robot. Res. 2017, 36, 1045–1052. [Google Scholar] [CrossRef]
  25. Haug, S.; Ostermann, J. A Crop/Weed Field Image Dataset for the Evaluation of Computer Vision Based Precision Agriculture Tasks. In Proceedings of the Computer Vision—ECCV 2014 Workshops, Zurich, Switzerland, 6–7 and 12 September 2014; pp. 105–116. [Google Scholar] [CrossRef]
  26. Nguyen, D.T.; Nam, S.H.; Batchuluun, G.; Owais, M.; Park, K.R. An Ensemble Classification Method for Brain Tumor Images Using Small Training Data. Mathematics 2022, 10, 10234566. [Google Scholar] [CrossRef]
  27. Abdalla, A.; Cen, H.; Wan, L.; Rashid, R.; Weng, H.; Zhou, W.; He, Y. Fine-Tuning Convolutional Neural Network with Transfer Learning for Semantic Segmentation of Ground-Level Oilseed Rape Images in a Field with High Weed Pressure. Comput. Electron. Agric. 2019, 167, 105091. [Google Scholar] [CrossRef]
  28. Crops and Weeds Segmentation Method in Heterogeneous Environment. Available online: https://github.com/iamrehanch/crops_and_weeds_semantic_segmentation (accessed on 9 March 2023).
  29. Haug, S.; Michaels, A.; Biber, P.; Ostermann, J. Plant Classification System for Crop/Weed Discrimination without Segmentation. In Proceedings of the IEEE Winter Conference on Applications of Computer Vision, Steamboat Springs, CO, USA, 24–26 March 2014; pp. 1142–1149. [Google Scholar] [CrossRef]
  30. Lottes, P.; Hörferlin, M.; Sander, S.; Stachniss, C. Effective Vision-Based Classification for Separating Sugar Beets and Weeds for Precision Farming: Effective Vision-Based Classification. J. Field Robot. 2017, 34, 1160–1178. [Google Scholar] [CrossRef]
  31. Lottes, P.; Khanna, R.; Pfeifer, J.; Siegwart, R.; Stachniss, C. UAV-Based Crop and Weed Classification for Smart Farming. In Proceedings of the IEEE International Conference on Robotics and Automation, Singapore, Singapore, 29 May–3 June 2017; pp. 3024–3031. [Google Scholar] [CrossRef]
  32. Yang, B.; Xu, Y. Applications of Deep-Learning Approaches in Horticultural Research: A Review. Hortic. Res. 2021, 8, 123. [Google Scholar] [CrossRef]
  33. Chen, L.C.; Papandreou, G.; Schroff, F.; Adam, H. Rethinking Atrous Convolution for Semantic Image Segmentation. arXiv 2015, arXiv:1706.05587v3. [Google Scholar]
  34. Long, J.; Shelhamer, E.; Darrell, T. Fully Convolutional Networks for Semantic Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar] [CrossRef]
  35. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar] [CrossRef]
  36. Badrinarayanan, V.; Kendall, A.; Cipolla, R. SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef]
  37. Zou, K.; Chen, X.; Wang, Y.; Zhang, C.; Zhang, F. A Modified U-Net with a Specific Data Argumentation Method for Semantic Segmentation of Weed Images in the Field. Comput. Electron. Agric. 2021, 187, 106242. [Google Scholar] [CrossRef]
  38. Milioto, A.; Lottes, P.; Stachniss, C. Real-Time Semantic Segmentation of Crop and Weed for Precision Agriculture Robots Leveraging Background Knowledge in CNNs. In Proceedings of the IEEE International Conference on Robotics and Automation, Brisbane, Australia, 21–25 May 2018; pp. 2235–2299. [Google Scholar] [CrossRef]
  39. Paszke, A.; Chaurasia, A.; Kim, S.; Culurciello, E. ENet: A Deep Neural Network Architecture for Real-Time Semantic Segmentation. arXiv 2016, arXiv:1606.02147v1. [Google Scholar]
  40. Fathipoor, H.; Shah-Hosseini, R.; Arefi, H. Crop and Weed Segmentation on Ground_Based Images using Deep Convolutional Neural Network. In Proceedings of the ISPRS Annals of the Photogrammetry Remote Sensing and Spatial Information Sciences, Tehran, Iran, 13 January 2023; pp. 195–200. [Google Scholar] [CrossRef]
  41. Zhou, Z.; Siddiquee, M.M.R.; Tajbakhsh, N.; Liang, J. UNet++: A Nested U-Net Architecture for Medical Image Segmentation. In Proceedings of the Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support, Granada, Spain, 20 September 2018; pp. 3–11. [Google Scholar] [CrossRef]
  42. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2014, arXiv:1409.1556v6. [Google Scholar]
  43. Fawakherji, M.; Potena, C.; Bloisi, D.D.; Imperoli, M.; Pretto, A.; Nardi, D. UAV Image Based Crop and Weed Distribution Estimation on Embedded GPU Boards. In Proceedings of the Computer Analysis of Images and Patterns, Salerno, Italy, 6 September 2019; pp. 100–108. [Google Scholar] [CrossRef]
  44. Chakraborty, R.; Zhen, X.; Vogt, N.; Bendlin, B.; Singh, V. Dilated Convolutional Neural Networks for Sequential Manifold-Valued Data. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, South Korea, 27 October—2 November 2019; pp. 10620–10630. [Google Scholar] [CrossRef]
  45. You, J.; Liu, W.; Lee, J. A DNN-Based Semantic Segmentation for Detecting Weed and Crop. Comput. Electron. Agric. 2020, 178, 10570. [Google Scholar] [CrossRef]
  46. Wang, H.; Song, H.; Wu, H.; Zhang, Z.; Deng, S.; Feng, X.; Chen, Y. Multilayer Feature Fusion and Attention-Based Network for Crops and Weeds Segmentation. J. Plant Dis. Prot. 2022, 129, 1475–1489. [Google Scholar] [CrossRef]
  47. Siddiqui, S.A.; Fatima, N.; Ahmad, A. Neural Network Based Smart Weed Detection System. In Proceedings of the International Conference on Communication, Control and Information Sciences, Idukki, India, 16–18 June 2021; pp. 1–5. [Google Scholar] [CrossRef]
  48. Khan, A.; Ilyas, T.; Umraiz, M.; Mannan, Z.I.; Kim, H. CED-Net: Crops and Weeds Segmentation for Smart Farming Using a Small Cascaded Encoder-Decoder Architecture. Electronics 2020, 9, 9101602. [Google Scholar] [CrossRef]
  49. Reinhard, E.; Adhikhmin, M.; Gooch, B.; Shirley, P. Color Transfer between Images. IEEE Comput. Graph. Appl. 2001, 21, 34–41. [Google Scholar] [CrossRef]
  50. Ruderman, D.L.; Cronin, T.W.; Chiao, C.C. Statistics of Cone Responses to Natural Images: Implications for Visual Coding. J. Opt. Soc. Am. A-Opt. Image Sci. Vis. 1998, 15, 2036–2045. [Google Scholar] [CrossRef]
  51. Mikołajczyk, A.; Grochowski, M. Data Augmentation for Improving Deep Learning in Image Classification Problem. In Proceedings of the International Interdisciplinary PhD Workshop, Świnouście, Poland, 9–12 May 2018; pp. 117–122. [Google Scholar] [CrossRef]
  52. Agarap, A.F. Deep Learning Using Rectified Linear Units (ReLU). arXiv 2019, arXiv:1803.08375v2. [Google Scholar]
  53. Clevert, D.A.; Unterthiner, T.; Hochreiter, S. Fast and Accurate Deep Network Learning by Exponential Linear Units (ELUs). arXiv 2015, arXiv:1511.07289v5. [Google Scholar]
  54. Intel Core i5-2320. Available online: https://www.intel.com/content/www/us/en/products/sku/53446/intel-core-i52320-processor-6m-cache-up_to-3-30-ghz/specifications.html (accessed on 5 October 2023).
  55. NVIDIA GeForce GTX 1070. Available online: https://www.nvidia.com/en-gb/geforce/10-series/ (accessed on 5 October 2023).
  56. PyTorch. Available online: https://pytorch.org/ (accessed on 5 October 2023).
  57. Python 3.8. Available online: https://www.python.org/downloads/release/python-380/ (accessed on 5 October 2023).
  58. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. arXiv 2014, arXiv:1412.6980v9. [Google Scholar]
  59. Loshchilov, I.; Hutter, F. SGDR: Stochastic Gradient Descent with Warm Restarts. arXiv 2016, arXiv:1608.03983v5. Available online: https://arxiv.org/abs/1608.03983 (accessed on 5 October 2023).
  60. Sudre, C.H.; Li, W.; Vercauteren, T.; Ourselin, S.; Cardoso, M.J. Generalised Dice Overlap as a Deep Learning Loss Function for Highly Unbalanced Segmentations. In Proceedings of the International Conference on Communication, Control and Information Sciences, Québec City, Canada, 9 September 2017; pp. 240–248. [Google Scholar] [CrossRef]
  61. Xiao, X.; Ma, L. Color Transfer in Correlated Color Space. In Proceedings of the ACM International Conference on Virtual Reality Continuum and Its Applications, Hong Kong, China, 14 June 2006; pp. 305–309. [Google Scholar] [CrossRef]
  62. Pitié, F.; Kokaram, A.C.; Dahyot, R. Automated Colour Grading Using Colour Distribution Transfer. Comput. Vis. Image Underst. 2007, 107, 123–137. [Google Scholar] [CrossRef]
  63. Gatys, L.A.; Ecker, A.S.; Bethge, M. Image Style Transfer Using Convolutional Neural Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 23–27 June 2016; pp. 2414–2423. [Google Scholar] [CrossRef]
  64. Nguyen, R.M.H.; Kim, S.J.; Brown, M.S. Illuminant Aware Gamut-Based Color Transfer. Comput. Graph. Forum 2014, 33, 319–328. [Google Scholar] [CrossRef]
  65. Rezaie, A.; Mauron, A.J.; Beyer, K. Sensitivity Analysis of Fractal Dimensions of Crack Maps on Concrete and Masonry Walls. Autom.Constr. 2020, 117, 103258. [Google Scholar] [CrossRef]
  66. Wu, J.; Jin, X.; Mi, S.; Tang, J. An Effective Method to Compute the Box-counting Dimension Based on the Mathematical Definition and Intervals. Results Eng. 2020, 6, 100106. [Google Scholar] [CrossRef]
  67. Xie, Y. The Application of Fractal Theory in Real-life. In Proceedings of the International Conference on Computing Innovation and Applied Physics, Qingdao, Shandong, China, 30 November 2023; pp. 132–136. [Google Scholar] [CrossRef]
  68. Mishra, P.; Singh, U.; Pandey, C.M.; Mishra, P.; Pandey, G. Application of Student’s t-test, Analysis of Variance, and Covariance. Ann. Card. Anaesth. 2019, 22, 407–411. [Google Scholar] [CrossRef]
  69. Cohen, J. A Power Primer. Psychol. Bull. 1992, 112, 155–159. [Google Scholar] [CrossRef]
  70. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-CAM: Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar] [CrossRef]
Figure 1. Sample images (left) and ground-truth masks (right) for crops (white pixels) and weeds (gray pixels) in BoniRob dataset (upper) and CWFID (lower).
Figure 1. Sample images (left) and ground-truth masks (right) for crops (white pixels) and weeds (gray pixels) in BoniRob dataset (upper) and CWFID (lower).
Fractalfract 08 00285 g001
Figure 2. Overview of the proposed framework flow.
Figure 2. Overview of the proposed framework flow.
Fractalfract 08 00285 g002
Figure 3. Sample histograms of (a) referenced image, (b) targeted image, (c) RH-transformed image, and (d) RH-transformed image with additional adjustment.
Figure 3. Sample histograms of (a) referenced image, (b) targeted image, (c) RH-transformed image, and (d) RH-transformed image with additional adjustment.
Fractalfract 08 00285 g003
Figure 4. Sample images of the BoniRob dataset (upper images) and CWFID (lower images). (a) Referenced image, (b) targeted image, (c) RH-transformed image, and (d) RH-transformed image with additional adjustments.
Figure 4. Sample images of the BoniRob dataset (upper images) and CWFID (lower images). (a) Referenced image, (b) targeted image, (c) RH-transformed image, and (d) RH-transformed image with additional adjustments.
Fractalfract 08 00285 g004
Figure 5. Sample training image (upper) and augmented image (lower) for training. (a) BoniRob dataset (left images) and (b) CWFID (right images).
Figure 5. Sample training image (upper) and augmented image (lower) for training. (a) BoniRob dataset (left images) and (b) CWFID (right images).
Fractalfract 08 00285 g005
Figure 6. U-Net architecture.
Figure 6. U-Net architecture.
Fractalfract 08 00285 g006
Figure 7. Modified U-Net architecture.
Figure 7. Modified U-Net architecture.
Fractalfract 08 00285 g007
Figure 9. Accuracy and loss graphs of training and validation data from (a) BoniRob dataset and (b) CWFID.
Figure 9. Accuracy and loss graphs of training and validation data from (a) BoniRob dataset and (b) CWFID.
Fractalfract 08 00285 g009
Figure 10. Visual comparisons of various semantic segmentation outputs with proposed framework (Experiment 1): (a) original image; (b) RH-transformed image; (c) ground-truth mask; semantic segmentation results with (d) CED-Net; (e) modified U-Net, and (f) U-Net.
Figure 10. Visual comparisons of various semantic segmentation outputs with proposed framework (Experiment 1): (a) original image; (b) RH-transformed image; (c) ground-truth mask; semantic segmentation results with (d) CED-Net; (e) modified U-Net, and (f) U-Net.
Fractalfract 08 00285 g010
Figure 11. Visual comparisons of various SOTA transformations with the proposed framework using U-Net (Experiment 1): (a) original image; (b) ground-truth mask; semantic segmentation results with (c) Xiao et al.; (d) Pitie et al.; (e) Gatys et al.; (f) Nguyen et al.; and (g) proposed method.
Figure 11. Visual comparisons of various SOTA transformations with the proposed framework using U-Net (Experiment 1): (a) original image; (b) ground-truth mask; semantic segmentation results with (c) Xiao et al.; (d) Pitie et al.; (e) Gatys et al.; (f) Nguyen et al.; and (g) proposed method.
Fractalfract 08 00285 g011
Figure 12. Visual comparisons of various semantic segmentation outputs with proposed framework (Experiment 1): (a) original image; (b) RH-transformed image; (c) ground-truth mask; semantic segmentation results with (d) CED-Net; (e) modified U-Net, and (f) U-Net.
Figure 12. Visual comparisons of various semantic segmentation outputs with proposed framework (Experiment 1): (a) original image; (b) RH-transformed image; (c) ground-truth mask; semantic segmentation results with (d) CED-Net; (e) modified U-Net, and (f) U-Net.
Fractalfract 08 00285 g012
Figure 13. Visual comparisons of various SOTA transformations with proposed framework using U-Net (Experiment 1): (a) original image; (b) ground-truth mask; semantic segmentation results with (c) Xiao et al.; (d) Pitie et al.; (e) Gatys et al.; (f) Nguyen et al.; and (g) proposed method.
Figure 13. Visual comparisons of various SOTA transformations with proposed framework using U-Net (Experiment 1): (a) original image; (b) ground-truth mask; semantic segmentation results with (c) Xiao et al.; (d) Pitie et al.; (e) Gatys et al.; (f) Nguyen et al.; and (g) proposed method.
Fractalfract 08 00285 g013
Figure 14. Visual representation of crops and weeds with samples from the CWFID for estimating FD values: (a) whole segmentation results; (b) weed segmentation result; and (c) crop segmentation result.
Figure 14. Visual representation of crops and weeds with samples from the CWFID for estimating FD values: (a) whole segmentation results; (b) weed segmentation result; and (c) crop segmentation result.
Fractalfract 08 00285 g014
Figure 15. Visual representation of crops and weeds with samples from the BoniRob dataset for estimating FD values: (a) whole segmentation results; (b) weed segmentation result; and (c) crop segmentation result.
Figure 15. Visual representation of crops and weeds with samples from the BoniRob dataset for estimating FD values: (a) whole segmentation results; (b) weed segmentation result; and (c) crop segmentation result.
Fractalfract 08 00285 g015
Figure 16. T-test results.
Figure 16. T-test results.
Fractalfract 08 00285 g016
Figure 17. Sample images of the CWFID with Grad-CAM visualization after the 4th convolution layer. (a) Input image, (b) ground-truth mask (gray, white, and black pixels mean weeds, crops, and background, respectively), (c) Grad-CAM of class 1 (weed), (d) Grad-CAM of class 2 (crop), and (e) segmented results.
Figure 17. Sample images of the CWFID with Grad-CAM visualization after the 4th convolution layer. (a) Input image, (b) ground-truth mask (gray, white, and black pixels mean weeds, crops, and background, respectively), (c) Grad-CAM of class 1 (weed), (d) Grad-CAM of class 2 (crop), and (e) segmented results.
Fractalfract 08 00285 g017
Figure 18. Samples of incorrect segmentation generated by proposed method with U-Net: (a) original image; (b) RH-transformed image; (c) ground-truth mask; and (d) semantic segmentation results.
Figure 18. Samples of incorrect segmentation generated by proposed method with U-Net: (a) original image; (b) RH-transformed image; (c) ground-truth mask; and (d) semantic segmentation results.
Fractalfract 08 00285 g018
Table 1. Comparisons of proposed method with previous ones on crop and weed segmentation.
Table 1. Comparisons of proposed method with previous ones on crop and weed segmentation.
DataTypeMethodStrength/MotivationWeakness
Homogeneous data-basedHandcrafted
feature-based
RFC [29]Handling overlapping of crops and weedsOverlapping of multiple plants with the same class cannot be split
RFC + vegetation detection [30]Detection of local and object-based featuresSmoothing as post-processing on only local features
Plant-tailored feature extraction [31]UAV intra-row-space-based weed detection in challenging conditionsThe number of weeds is much smaller in the datasets used
Deep feature-basedMTS-CNN [18]Separate object segmentation to avoid background-biased learningDependency of first stage network on second stage network
Modified U-Net [37]Effective two-stage training method with large applicabilityOnly weed-targeted segmentation
SegNet + Enet [38]Fast and more accurate pixelwise predictionsImages contain very small portions of crops and weeds
U-Net and U-Net++ [40]Detecting weeds in the early stages of growthUses a very small dataset and has no suitable real-time application
Modified U-Net + modified VGG-16 [43]Effective result for distribution estimation problem with graphics processing unit (GPU)-based embedded boardNot focusing on the exact location of weeds in the images
UFAB [45]Reducing redundancy by strengthening the model diversityUnavailability of RGB and NIR input
DA-Net [46]Expanding receptive field without affecting the computational costHard and time-consuming mechanism to parallelize the system using attention modules
4-layered CNN + data augmentation [47]Good for the early detection of weeds, improving production, and is easy to deploy because of the cheap cost Minimizing accuracy if weeds are not detected at the early stages
CED-Net [48]Using a light model and achieving efficient resultsError at any level among the four levels affects the overall performance
Heterogeneous data-basedProposed framework (proposed)Use of small training images in a heterogeneous environmentPreprocessing steps are included
Table 2. Representation of various cases.
Table 2. Representation of various cases.
CasesRH TransformationUsing One Training DataData Augmentation
Case I
Case II
Case III
Case IV
Case V
Case VI
Table 3. Comparisons of different cases using U-Net (Experiment 1) (Cr means crop, Wd means weed, Bg means background, Re means recall, and Pre means precision).
Table 3. Comparisons of different cases using U-Net (Experiment 1) (Cr means crop, Wd means weed, Bg means background, Re means recall, and Pre means precision).
ModelCasesmIoUIoU (Cr)IoU (Wd)IoU (Bg)RePreF1-Score
U-NetCase I0.3840.0010.1950.3840.6430.4030.495
Case II0.4230.0980.2060.9650.6380.5670.594
Case III0.4930.3220.1750.9820.6210.6050.611
Case IV0.5890.4720.3090.9850.7320.7210.724
Case V0.4990.2940.2210.9820.6520.6270.639
Case VI
(proposed)
0.6200.5240.3490.9860.7620.7490.752
Table 4. Comparisons of different cases using modified U-Net (Experiment 1) (Cr means crop, Wd means weed, Bg means background, Re means recall, and Pre means precision).
Table 4. Comparisons of different cases using modified U-Net (Experiment 1) (Cr means crop, Wd means weed, Bg means background, Re means recall, and Pre means precision).
ModelCasesmIoUIoU (Cr)IoU (Wd)IoU (Bg)RePreF1-Score
Modified U-NetCase I0.3800.0060.1950.9410.6450.4060.497
Case II0.4270.1040.2090.9680.6570.5630.601
Case III0.4830.2440.2210.9830.6460.6080.625
Case IV0.5230.3680.2170.9840.6640.6350.648
Case V0.4710.1950.2340.9830.6560.6150.634
Case VI
(proposed)
0.5390.3820.2510.9840.6860.6650.674
Table 5. Comparisons of different cases using CED-Net (Experiment 1) (Cr means crop, Wd means weed, Bg means background, Re means recall, and Pre means precision).
Table 5. Comparisons of different cases using CED-Net (Experiment 1) (Cr means crop, Wd means weed, Bg means background, Re means recall, and Pre means precision).
ModelCasesmIoUIoU (Cr)IoU (Wd)IoU (Bg)RePreF1-Score
CED-NetCase I0.3870.0090.1960.9560.6340.4410.518
Case II0.3960.0180.2090.9620.6160.4650.528
Case III0.4610.2340.1750.9740.6030.5590.579
Case IV0.5160.4090.1680.9710.6400.5920.614
Case V0.4660.2350.1880.9740.6170.5640.589
Case VI
(proposed)
0.5210.4720.1200.9720.6370.6130.624
Table 6. Comparison of results between RH transformation and RH transformation with additional adjustments (Cr means crop, Wd means weed, Bg means background, Re means recall, and Pre means precision).
Table 6. Comparison of results between RH transformation and RH transformation with additional adjustments (Cr means crop, Wd means weed, Bg means background, Re means recall, and Pre means precision).
ExperimentmIoUIoU (Cr)IoU (Wd)IoU (Bg)RePreF1-Score
RH transformation0.6200.5240.3490.9860.7620.7490.752
RH transformation with additional adjustments0.4680.1990.2290.9750.6310.6150.621
Table 7. Performance comparisons of proposed method and SOTA transformations (Cr means crop, Wd means weed, Bg means background, Re means recall, and Pre means precision).
Table 7. Performance comparisons of proposed method and SOTA transformations (Cr means crop, Wd means weed, Bg means background, Re means recall, and Pre means precision).
Segmentation ModelTransformationmIoUIoU (Cr)IoU (Wd)IoU (Bg)RePreF1-Score
U-NetXiao et al. [61]0.4960.2570.2740.9580.7170.5830.640
Pitie et al. [62]0.5480.3780.3120.9530.7760.6010.675
Gatys et al. [63]0.4570.3130.1010.9580.5580.5750.563
Nguyen et al. [64]0.4870.4860.0100.9640.5630.5800.569
Proposed0.6200.5240.3490.9860.7620.7490.752
Modified U-NetXiao et al. [61]0.3870.0660.1610.9340.6510.4940.558
Pitie et al. [62]0.4620.2000.2210.9660.7160.5690.630
Gatys et al. [63]0.3960.1570.0830.9480.4520.5440.490
Nguyen et al. [64]0.4270.1360.1850.9590.5180.5800.545
Proposed0.5390.3820.2510.9840.6860.6650.674
CED-NetXiao et al. [61]0.3900.0360.2040.9280.6180.4470.518
Pitie et al. [62]0.3940.1090.1800.8940.6730.4720.553
Gatys et al. [63]0.3600.0000.1430.9380.6050.3780.465
Nguyen et al. [64]0.3100.0000.0650.8650.4790.3490.402
Proposed0.5210.4720.1200.9720.6370.6130.624
Table 8. Comparison of different cases using U-Net (Experiment 2) (Cr means crop, Wd means weed, Bg means background, Re means recall, and Pre means precision).
Table 8. Comparison of different cases using U-Net (Experiment 2) (Cr means crop, Wd means weed, Bg means background, Re means recall, and Pre means precision).
ModelCasesmIoUIoU (Cr)IoU (Wd)IoU (Bg)RePreF1-Score
U-NetCase I0.3160.0000.0000.9480.3330.3290.330
Case II0.5700.2300.5080.9710.7040.6790.688
Case III0.4940.2270.2860.9690.6300.6730.649
Case IV0.6210.2720.6210.9690.7790.6890.730
Case V0.5070.2320.3220.9680.6450.6790.659
Case VI
(proposed)
0.6370.2920.6470.9710.7870.7080.743
Table 9. Comparison of different cases using modified U-Net (Experiment 2) (Cr means crop, Wd means weed, Bg means background, Re means recall, and Pre means precision).
Table 9. Comparison of different cases using modified U-Net (Experiment 2) (Cr means crop, Wd means weed, Bg means background, Re means recall, and Pre means precision).
ModelCasesmIoUIoU (Cr)IoU (Wd)IoU (Bg)RePreF1-Score
Modified U-NetCase I0.3160.0000.0000.9480.3330.3160.324
Case II0.5290.2350.3830.9700.6820.6650.668
Case III0.4480.2120.1630.9680.5880.5960.589
Case IV0.6050.2660.5930.9550.8330.6510.728
Case V0.4770.2320.2320.9660.6330.6110.619
Case VI
(proposed)
0.6220.2940.6100.9620.8370.6670.739
Table 10. Comparison of different cases using CED-Net (Experiment 2) (Cr means crop, Wd means weed, Bg means background, Re means recall, and Pre means precision).
Table 10. Comparison of different cases using CED-Net (Experiment 2) (Cr means crop, Wd means weed, Bg means background, Re means recall, and Pre means precision).
ModelCasesmIoUIoU (Cr)IoU (Wd)IoU (Bg)RePreF1-Score
CED-NetCase I0.3150.0000.0000.9460.3320.3180.322
Case II0.4870.1760.3360.9510.6990.5640.621
Case III0.4880.0130.5720.8790.6240.5370.575
Case IV0.5520.2180.5040.9350.8380.5920.691
Case V0.4850.01430.5810.8600.6230.5390.576
Case VI
(proposed)
0.5700.2440.5190.9460.8360.6110.703
Table 11. Comparisons of the performance of the proposed method and SOTA transformations (Cr means crop, Wd means weed, Bg means background, Re means recall, and Pre means precision).
Table 11. Comparisons of the performance of the proposed method and SOTA transformations (Cr means crop, Wd means weed, Bg means background, Re means recall, and Pre means precision).
ModelTransformationmIoUIoU (Cr)IoU (Wd)IoU (Bg)RePreF1-Score
U-NetXiao et al. [61]0.5300.5050.1240.9620.6960.5890.635
Pitie et al. [62]0.5430.4750.1930.9590.7290.6050.657
Gatys et al. [63]0.3160.0000.0000.9480.3330.3810.352
Nguyen et al. [64]0.3320.0490.0000.9460.3520.4170.379
Proposed0.6370.6470.2920.9710.7870.7080.743
Modified U-NetXiao et al. [61]0.5260.5210.1450.9110.7870.5730.661
Pitie et al. [62]0.5160.5220.1410.8860.8170.5670.667
Gatys et al. [63]0.3160.0000.0000.9480.3330.3200.326
Nguyen et al. [64]0.1870.0140.0310.5160.3870.3680.375
Proposed0.6220.6100.2940.9620.8370.6670.739
CED-NetXiao et al. [61]0.4230.3390.0670.8620.7090.4750.566
Pitie et al. [62]0.4660.4420.1070.8500.8070.5240.632
Gatys et al. [63]0.3180.0050.0180.9320.3380.3400.339
Nguyen et al. [64]0.3160.0000.0010.9480.3330.3750.349
Proposed0.5700.5190.2440.9460.8360.6110.703
Table 12. FD values of images from CWFID and BoniRob dataset. The 1st~3rd row FD values are computed from the 1st~3rd row images of Figure 14. The 4th~6th row FD values are computed from the 4th~6th row images of Figure 15.
Table 12. FD values of images from CWFID and BoniRob dataset. The 1st~3rd row FD values are computed from the 1st~3rd row images of Figure 14. The 4th~6th row FD values are computed from the 4th~6th row images of Figure 15.
DatasetWeed FDCrop FD
CWFID 1.611.26
0.761.43
1.531.21
BoniRob dataset1.270.91
1.321.31
0.971.54
Table 13. Comparisons of average processing time by proposed and SOTA methods (unit: ms).
Table 13. Comparisons of average processing time by proposed and SOTA methods (unit: ms).
MethodsProcessing Time
Xiao et al. [61]1270
Pitie et al. [62]2920
Gatys et al. [63]2210
Nguyen et al. [64]1030
Proposed1080
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Akram, R.; Hong, J.S.; Kim, S.G.; Sultan, H.; Usman, M.; Gondal, H.A.H.; Tariq, M.H.; Ullah, N.; Park, K.R. Crop and Weed Segmentation and Fractal Dimension Estimation Using Small Training Data in Heterogeneous Data Environment. Fractal Fract. 2024, 8, 285. https://doi.org/10.3390/fractalfract8050285

AMA Style

Akram R, Hong JS, Kim SG, Sultan H, Usman M, Gondal HAH, Tariq MH, Ullah N, Park KR. Crop and Weed Segmentation and Fractal Dimension Estimation Using Small Training Data in Heterogeneous Data Environment. Fractal and Fractional. 2024; 8(5):285. https://doi.org/10.3390/fractalfract8050285

Chicago/Turabian Style

Akram, Rehan, Jin Seong Hong, Seung Gu Kim, Haseeb Sultan, Muhammad Usman, Hafiz Ali Hamza Gondal, Muhammad Hamza Tariq, Nadeem Ullah, and Kang Ryoung Park. 2024. "Crop and Weed Segmentation and Fractal Dimension Estimation Using Small Training Data in Heterogeneous Data Environment" Fractal and Fractional 8, no. 5: 285. https://doi.org/10.3390/fractalfract8050285

APA Style

Akram, R., Hong, J. S., Kim, S. G., Sultan, H., Usman, M., Gondal, H. A. H., Tariq, M. H., Ullah, N., & Park, K. R. (2024). Crop and Weed Segmentation and Fractal Dimension Estimation Using Small Training Data in Heterogeneous Data Environment. Fractal and Fractional, 8(5), 285. https://doi.org/10.3390/fractalfract8050285

Article Metrics

Back to TopTop