You are currently viewing a new version of our website. To view the old version click .
Fractal and Fractional
  • Article
  • Open Access

10 May 2024

Crop and Weed Segmentation and Fractal Dimension Estimation Using Small Training Data in Heterogeneous Data Environment

,
,
,
,
,
,
,
and
Division of Electronics and Electrical Engineering, Dongguk University, 30 Pildong-ro, 1-gil, Jung-gu, Seoul 04620, Republic of Korea
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Fractional Order Complex Systems: Advanced Control, Intelligent Estimation and Reinforcement Learning Image Processing​ Algorithms

Abstract

The segmentation of crops and weeds from camera-captured images is a demanding research area for advancing agricultural and smart farming systems. Previously, the segmentation of crops and weeds was conducted within a homogeneous data environment where training and testing data were from the same database. However, in the real-world application of advancing agricultural and smart farming systems, it is often the case of a heterogeneous data environment where a system trained with one database should be used for testing with a different database without additional training. This study pioneers the use of heterogeneous data for crop and weed segmentation, addressing the issue of degraded accuracy. Through adjusting the mean and standard deviation, we minimize the variability in pixel value and contrast, enhancing segmentation robustness. Unlike previous methods relying on extensive training data, our approach achieves real-world applicability with just one training sample for deep learning-based semantic segmentation. Moreover, we seamlessly integrated a method for estimating fractal dimensions into our system, incorporating it as an end-to-end task to provide important information on the distributional characteristics of crops and weeds. We evaluated our framework using the BoniRob dataset and the CWFID. When trained with the BoniRob dataset and tested with the CWFID, we obtained a mean intersection of union (mIoU) of 62% and an F1-score of 75.2%. Furthermore, when trained with the CWFID and tested with the BoniRob dataset, we obtained an mIoU of 63.7% and an F1-score of 74.3%. We confirmed that these values are higher than those obtained by state-of-the-art methods.

1. Introduction

Expanding crop efficiency is becoming increasingly important as food security concerns are increasing globally. However, it faces many challenges, such as a lack of manpower, uncertain environmental conditions, soil factors, and scarcity of water. To achieve high yields and address such problems, previous research used plant phenotyping methods and monitoring systems to strengthen crop productivity in precision agriculture [,]. Over time, traditional farming has shifted toward modern automated farming to increase yields and minimize labor costs, individual effort, and time.
Image processing has been widely adopted in various fields [,,,,,,,,]. In addition, recently, deep learning methods have offered various solutions, and the use of computer vision has grown significantly in various applications including building monitoring, image enhancement, medical image processing, biomedical engineering, and underwater computer vision, where some research has adopted fractal-related perspectives, also in [,,,,,,,]. Although the studies adopt similar concepts of fractal dimension (FD) estimations [,,], their applications are different in terms of building monitoring and medical image processing. Semantic segmentation also exhibits a fundamental role in the accurate recognition of crops and weeds [,]. Two mainstream methods exist for crop and weed detection. The first is box-based detection [,]; however, it has the drawback of overlooking specific regions in weeds and crops. The second type is semantic segmentation or pixel-based detection [,,], which detects precise regions of weeds or crops at the pixel level. Therefore, correctly identifying the crop and weed segments is essential. Crop and weed segments typically have irregular shapes, and training with these irregularly shaped segments comes with an imbalance in the data. The irregular shapes of “sugar beets” (BoniRob) [] and crop/weed field image datasets (CWFIDs) [] are shown in Figure 1. A considerable amount of data is usually required during training, particularly in segmentation cases. To acquire more data, it may be necessary for experts to create a large number of annotations, which would require considerable time and effort. In some scenarios, if a considerable amount of training data are unavailable, the testing performance decreases. To avoid such a drop in performance owing to insufficient training data, a method using small amounts of training data was proposed by Nguyen et al. [], which also achieved good performance. Homogenous data usually show satisfactory results; however, when applied to heterogeneous environments, where the testing and training datasets are completely different, the overall results decrease significantly. Abdalla et al. [] showed that the efficacy of these algorithms is compromised in complex environments due to their heavy reliance on various factors, including lighting conditions and weed density, for feature extraction. Thus, it is essential to propose an effective and reliable framework that segments crops and weeds accurately, even in complex heterogeneous environments.
Figure 1. Sample images (left) and ground-truth masks (right) for crops (white pixels) and weeds (gray pixels) in BoniRob dataset (upper) and CWFID (lower).
To solve these issues, we propose an approach for weed and crop segmentation and fractal dimension estimation using a small amount of training data in a heterogeneous data environment. The contributions of this study include the following:
  • This is the first study that considers the segmentation of crops and weeds within a heterogeneous environmental setup utilizing one training data sample. We rigorously investigate the factors that cause performance degradation in heterogeneous datasets, including variations in illumination and contrast. To address this problem, we propose a method that applies the Reinhard (RH) transformation, leveraging the mean and standard deviation (std) adjustments.
  • We address the issue of high data availability for real-world applications. For this purpose, we improved the performance using a small amount of training data. The small amount of additional training data significantly improves the segmentation performance while requiring fewer computational resources and less training time.
  • We introduce the FD estimation approach in our framework, which is seamlessly combined as an end-to-end task to provide important information on the distributional features of crops and weeds.
  • It is noteworthy that our proposed framework [] is publicly accessible for a fair comparison with other studies.
The structure of the remaining sections of this study is as follows: Multiple related studies are described in Section 2. In Section 3, we outline the proposed approach. Section 4 describes a comparison between the proposed framework and the state-of-the-art (SOTA) methods in terms of their performance. Section 5 presents a discussion, while Section 6 presents the conclusion and outlines future work.

3. Proposed Method

3.1. Overview of Proposed Method

Figure 2 depicts an overview of the flow of our framework. During training, we train the conventional semantic segmentation model with Dataset A. Next, we preprocess the images of Dataset B using the RH transformation, which is based on the mean and std of Datasets A and B. This transformation adjusts the visual properties of the images, such as intensity, illumination, and contrast, to make them similar to the reference image in a heterogeneous environment. Then, we select one training data from the preprocessed Dataset B and perform DA on it to augment the training data. Afterward, we perform fine-tuning and train the model from scratch with Dataset A using the augmented data, and we perform the semantic segmentation of weeds and crops utilizing the testing data from Dataset B.
Figure 2. Overview of the proposed framework flow.

3.2. Preprocessing

3.2.1. RH Transformation

Many factors such as variations in intensity, illumination, and contrast cause performance degradation, particularly in heterogeneous environments. Transferring colors from one image (reference) to another (target) is a significant problem, particularly when the color information in the reference image does not match the newly generated image, causing quality and performance degradation. To address performance degradation using heterogeneous data, we adopted the RH transformation [] with average mean and std adjustments to enhance the visual attributes of the images. Many prior transformation methods and color spaces enhance the visual characteristics of an image. In the Ɩαβ color space (LCS) [], “Ɩ” indicates a brightness channel that captures the brightness independently of color attributes, the “α” channel encapsulates yellow and blue hues, and the “β” channel encloses the interplay between red and green shades. In the RH transformation, the LCS is used. The LCS minimizes the relative significance of each weight, with a matrix proposed for converting vectors from RGB to LCS. The RH transformation changes the mean and std values of the color channels based on the LCS that consistently represents the pixel colors in an image. As shown in Equation (1), it uses the ƒ mapping function with δ parameters to transform the d a t a B into preprocessed d a t a B .
d a t a B = ƒ ( d a t a B ,   δ )
For the transformation, the RGB images are manipulated in the LCS and then transformed into the long, medium, short (LMS) cone space, and the logarithmic of the LMS space is obtained to reduce skewness []. The mean and std for all the axes in the LCS are separately calculated to make images more synthetic. Moreover, the color space is normalized by subtracting the mean of the data points from the original data point value, as follows:
Ɩ ^ = Ɩ Ɩ m ,
α ^ = α α m ,
β ^ = β β m
where Ɩ m ,   α m ,   β m denote the average mean data point values of the Ɩ , α ,   a n d   β data points, and Ɩ ^ ,     α ^ ,   a n d   β ^ show the normalized space data points. Upon normalizing the space, data points are scaled by std σ j Ɩ ,   σ j α ,   σ j β for the reference images and σ i Ɩ ,   σ i α ,   σ i β for the target images. Finally, Ɩ ° , α ° ,   β ° are the scaled points for the transformed images, as follows:
Ɩ ° = σ i Ɩ σ j Ɩ Ɩ ^ ,
α ° = σ i α σ j α α ^ ,
β ° = σ i β σ j β β ^
The pseudo-code of the RH transformation is provided in Algorithm 1.
Algorithm 1: RH transformation with pseudo-code
Input: {dataB} tn; The total of t data samples, dataB: input dataset B images
dataAm: training data from dataset A.
Output: (dataB’) the preprocessed sample
1:  Compute the std of the dataset B (input images) by
   dataB _std = std (dataB (:, :))
2:  Compute the mean of the dataB
   dataB _mean = mean (dataB (:, :))
3:  Compute the average std of the training data from dataset A using
   dataA_std = avg (std (dataAm (:, :))
4:  Compute the average mean of the training data from dataset A using
   dataA_mean = avg (mean (dataAm (:, :))
5:  Apply the RH transformation
   for m = 1: x
    for n = 1: y
     dataB’ (m, n) = [(dataB (m, n) − dataB_mean) ×
     (dataA_std/dataB_std)] + dataA_mean
    end
   end
   return dataB’
Histograms of the RGB color channel distribution for the reference, target, and newly transformed images are shown in Figure 3a–c, respectively. The reference image is from Dataset A, the target image is from Dataset B, and the RH-transformed image is a new image generated by RH transformation. In Figure 3b, the histograms show only two channels; apparently, the red and blue channels overlap because they are replicas of each other []. The histogram shows the pixel values on the x-axis and the probability distribution of the pixels on the y-axis. As shown in Figure 3c, the histogram of the target image is more akin to that of the reference image in Figure 3a by RH transformation than to that of the target image in Figure 3b before RH transformation.
Figure 3. Sample histograms of (a) referenced image, (b) targeted image, (c) RH-transformed image, and (d) RH-transformed image with additional adjustment.

3.2.2. RH Transformation with Additional Adjustments

However, despite the RH transformation, there is still a difference in the relative probability distribution of each channel in Figure 3a,c. Therefore, we introduce an additional adjustment to obtain the relative probability distribution of each channel of the RH-transformed image that is more similar to that of the reference image as follows:
m i n s e l e c t p i o p t   R t R r p i 2 ,
m i n s e l e c t q i o p t   G t G r q i 2 ,
m i n s e l e c t r i o p t   B t B r r i 2
where R t ,   G t , a n d   B t represent the pixels in the target image, and R r ,   G r ,   a n d   B r represent the pixels in the reference image. Moreover, p i o p t ,   q i o p t ,   a n d   r i o p t are the optimal values generated by varying the p i ,   q i , a n d r i values, respectively, to minimize the distance between the corresponding red, green, and blue channel distributions. To obtain this channel distribution, the following additional adjustments were made:
Ɩ a d j = Ɩ m p i o p t ,
α a d j = α m q i o p t ,
β a d j = β m r i o p t
where Ɩ m ,   α m ,   a n d   β m represent the average mean data point values, and p i o p t ,   q i o p t ,   a n d   r i o p t represent the values that make the channel distribution look more similar among the reference and transformed images. These p i o p t ,   q i o p t ,   r i o p t values are subtracted from the average mean values to adjust the Ɩ a d j ,   α a d j ,   a n d   β a d j data points. Furthermore, the transformation with additional adjustments is as follows:
Ɩ = Ɩ Ɩ a d j ,
α = α α a d j ,
β = β β a d j
where Ɩ a d j ,   α a d j ,   a n d   β a d j are the different data point values for additional adjustments, which are experimentally chosen to make the channel distribution similar to the reference image. These values are subtracted from the Ɩ , α ,   β data points, and the result is Ɩ ,     α ,   a n d   β , which shows the normalized space data points. The newly generated histograms are visually presented in Figure 3d. As demonstrated in Figure 3d, the channel distributions resulting from the RH transformation with additional adjustments closely resemble those of the reference image in Figure 3a, in contrast to the RH without additional adjustment in Figure 3c.
Figure 4 illustrates the sample images of the referenced image, targeted image, RH-transformed image, and RH-transformed image with additional adjustments from the BoniRob dataset and CWFID.
Figure 4. Sample images of the BoniRob dataset (upper images) and CWFID (lower images). (a) Referenced image, (b) targeted image, (c) RH-transformed image, and (d) RH-transformed image with additional adjustments.

3.3. Data Augmentation of One Training Data

Data augmentation (DA) involves employing various transformations on existing data to expand and diversify a training dataset artificially. In image datasets, DA can involve procedures such as in-plane rotation, flipping, zooming, and modifying the contrast and brightness levels of the images within the training dataset []. The purpose behind DA is to enhance the capacity of a model to generalize and perform effectively on novel and unseen data by exposing it to a broader spectrum of variations that may be encountered in real-world contexts. We randomly selected a single image from the training data and augmented it for training. By adding augmentation to this small training dataset, we improved the segmentation performance while reducing the training time. A simple 180-degree in-plane rotation was used for DA, as depicted in Figure 5.
Figure 5. Sample training image (upper) and augmented image (lower) for training. (a) BoniRob dataset (left images) and (b) CWFID (right images).

3.4. Semantic Segmentation Networks

For the semantic segmentation of weeds and crops, the following conventional semantic segmentation models were adopted:

3.4.1. U-Net

U-Net [] features a fully CNN with a U-shaped encoder–decoder framework. The input image of the encoder and the output image of the decoder have the same image size. The decoder is more or less symmetric than the encoder, and the encoder has a large feature channel for propagating contextual information to the high-resolution layer. The architecture includes two 3 × 3 Conv layers, each accompanied by a max-pooling (MP) layer of 2 × 2 kernel size, with stride 2, and a ReLU []. In the decoder, the tensor map was upsampled using a 2 × 2 up-convolution and concatenated with the encoder features. A final layer of 1 × 1 convolution was incorporated into the U-Net network. Figure 6 illustrates its architecture. Furthermore, the first stages of the encoder (including and before the first max-pooling layer as E n c o d e r 1 ) and decoder (first up Conv layer with subsequent layers having the same spatial dimensions as D e c o d e r 1 ) of the U-Net shown in Figure 6 are mathematically represented as follows:
E n c o d e r 1 = P 1 C 1   w h e r e   C 1 = C o n v C o n v X , W 1 , P 1 = M a x p o o l C 1
D e c o d e r 1 = C R 1 C c o n c a t 1 C c r o p 1 U 1 ,
w h e r e   U 1 = U p c o n v B ,   W u p 1 , C c r o p 1 = C r o p R , s i z e U 1 , C c o n c a t 1 = C o n c a t e n a t i o n U 1 , C c r o p 1 , C R 1 = C o n v C o n v C c o n c a t 1 , W c o n v
In E n c o d e r 1 , X represents the input feature, and W 1 indicates the weight tensor for convolution operations. In every convolution operation for both the encoder and decoder, ReLU activation functions and batch normalization are applied. In D e c o d e r 1 , B represents the tensor of the previous layer, and W u p 1 in U p c o n v ( B , W u p 1 ) shows the weight tensor with an up-convolution operation. R in C r o p R , s i z e ( U 1 ) indicates the input image size after convolutions from the encoder side with the same size as U 1 , which is a result of the previous up-convolution. C c o n c a t 1 represents the concatenation of U 1 and C c r o p 1 which are the final calculations for the previous layers. W C o n v represents weight tensors with convolution operations. Finally, after the convolution in the decoder, it is fed to C R 1 , and the subsequent levels of decoder operations continue in a similar manner.
Figure 6. U-Net architecture.

3.4.2. Modified U-Net

Modified U-Net [] is an updated variant of the U-Net, featuring an encoder–decoder framework. Like U-Net, the modified version includes two 3 × 3 Conv layers, each accompanied by an MP layer of 2 × 2 kernel size of stride 2, and an activation function named the exponential linear unit []. In the decoder, the tensor map is upsampled using a 2 × 2 up-convolution and concatenated with the encoder features. To overcome the overfitting issue, a dropout layer was placed between the Conv layers. A last layer with 1 × 1 Conv is used in the modified U-Net architecture. Moreover, the stochastic gradient descent optimization is replaced with the Adadelta algorithm. Figure 7 shows the structure of the modified U-Net. Furthermore, the first stages of the encoder (before and including the first pooling layer as E n c o d e r 1 ) and decoder (first up Conv layer with the subsequent layer including convolutions as D e c o d e r 1 ) of Figure 7 are mathematically expressed as follows:
E n c o d e r 1 = P 1 C 1 D 1 B 1   w h e r e   B 1 = C o n v X , W b 1 , D 1 = D r o p o u t B 1 ,   d r o p o u t v a l u e , C 1 = C o n v D 1 , W c 1 , P 1 = M a x p o o l C 1
D e c o d e r 1 = C R 1 R 1 C B 1 U 1 C C 1 ,
w h e r e   C C 1 = C o n c a t e n a t e B , D n , U 1 = U p c o n v C C 1 , W u 1 , C B 1 = C o n v U 1 , W c b 1 , R 1 = D r o p o u t C B 1 ,   d r o p o u t v a l u e , C R 1 = C o n v ( R 1 , W c r 1 )
In E n c o d e r 1 , X represents the input image, and W b 1 represents the weight tensor for convolution operations. Following every C o n v operation in both the encoder and decoder batch normalization, ELU is applied. In E L U ( C o n v ( D 1 , W c 1 ) , D 1 represents the dropout layer, which is further processed by convolution and E L U resulting in C 1 . C 1 is downsampled using the MP layer. In D e c o d e r 1 , B represents the tensor map of the prior layer, and D n represents the dropout feature map from the same level connection. W u 1 in U p c o n v ( C C 1 , W u 1 ) represents the weight tensor with an upconvolution operation. Subsequently, a convolution is applied to the W c b 1 weight tensor. R 1 has a dropout value on which convolution is applied with the W c r 1 weight tensor. After the activation function, the final results are concluded in C R 1 with the completion of the first decoder-level operations.
Figure 7. Modified U-Net architecture.

3.4.3. CED-Net

Although many other architectures for segmentation networks are deep, with a large number of parameters, CED-Net [] is a simple cascading semantic segmentation model with a smaller number of parameters. The complete network comprises two training levels. Model 1 is trained at level 1 using the following steps. Model 1 is trained for predicting weeds, while Model 3 is trained for the prediction of crops. The results of Models 1 and 3 are upsampled and stitched to scale for each input image and then utilized as inputs for Models 2 and 4, respectively. The models are trained separately at each level. At Level 1, the images with their segmentation ground truths are reduced in size to spatial dimensions of 448 × 448. To train the Level 2 models, the 448 × 448 spatial dimension is upsampled to 896 × 896 using bilinear interpolation. The only difference from the U-Net structure is the highest feature map of 256 sizes in the bottleneck layer. All models and levels use the same encoder–decoder network architecture for CED-Net, as shown in Figure 8. Moreover, the first stages of CED-Net at each level in the model are akin to those of the U-Net encoder–decoder, other than the network depth and the number of convolutions. The encoder (including and before the first MP layer as E n c o d e r 1 ) and decoder (first up Conv layer with more layers having convolutions as D e c o d e r 1 ) of CED-Net, shown in Figure 8, are mathematically represented as follows:
E n c o d e r 1 = P 1 D 1   w h e r e   D 1 = C o n v C o n v Y , W 1 , P 1 = M a x p o o l ( C 1 )
D e c o d e r 1 = F R 1 C c o n c a t 1 C c r o p 1 U 1 ,
w h e r e   U 1 = U p c o n v B ,   W u p 1 , C c r o p 1 = C r o p R , s i z e U 1 , C c o n c a t 1 = C o n c a t e n a t i o n U 1 , C c r o p 1 , F R 1 = C o n v C o n v C o n v C c o n c a t 1 , W c o n v
In E n c o d e r 1 , Y represents the input feature, and W 1 represents the weight tensor for the convolution operations. ReLU and batch normalization are employed in each Conv operation for both the encoder and decoder. In D e c o d e r 1 , B depicts the feature map of the prior layer, and W u p 1 in U p c o n v ( B , W u p 1 ) shows the weight tensor with an upconvolution operation. In C r o p R , s i z e ( U 1 ) , R represents the input image size after convolutions on the encoder side, converted to the same size as U 1 , which is the result of a previous upconvolution. C c o n c a t 1 represents the concatenation of U 1 and C c r o p 1 which are the final calculations for the previous layers. W C o n v represents weight tensors with convolution operations. Finally, after convolutions in the decoder, it is fed to F R 1 , and the next stages of the decoder computations continue in a similar manner. At each level of CED-Net, a similar encoder–decoder architecture is used for network operations.
Figure 8. Encoder and decoder architecture of CED-Net.

4. Experimental Results

4.1. Experimental Dataset and Setup

We used two open datasets: the BoniRob dataset [] as Dataset A and the CWFID [] as Dataset B. The BoniRob dataset contains 496 images with pixel-level annotated masks, while the CWFID contains 60 images with the same pixel-level annotated masks. Both dataset images were captured by an autonomous robot in the field at a resolution of 1296 × 966 pixels. For small training datasets, we use only a single image, and augmentation produces an additional image. We utilized two images (one augmented and one small training data) for training. During training, we reduced the image size to 512 × 512 pixels. We experimented on a Windows-based desktop system with an Intel Core i5-2320 CPU @ 3.00 GHz processor [], a GPU of NVIDIA GeForce GTX 1070 [] with 8 gigabytes of memory, and 16 gigabytes of RAM. For development, we utilized the PyTorch [] platform in Python version 3.8 [].

4.2. Training Setup

For Experiment 1, Dataset A was first used for training with 70% of the data. Dataset B, transformed using the proposed method, was subsequently divided into two equal portions: one serving as testing data and the other as small training data. A single image of the small training data was then augmented and used for training, whereas the testing data remained unchanged. In Experiment 2, 70% of the data in Dataset B were used for training. Next, the data were preprocessed, and Dataset A was subsequently divided into two equal portions: one serving as testing data and the other as small training data. We augmented a single image of the small training data and further utilized it for training, while the test data remained unchanged. During training, the images underwent resizing to a resolution of 512 × 512 pixels, and the network was trained using a batch size of two for 150 epochs. We employed an Adam optimizer [] with an initial learning rate (LR) of 1 × 10−5 and utilized a cosine annealing strategy [] to steadily reduce the LR during training. The training loss was calculated using dice loss []. We trained the proposed framework using the U-Net, modified U-Net, and CED-Net architectures. Figure 9 depicts the loss and training graphs for the BoniRob dataset and the CWFID using U-Net. Furthermore, the convergence of the loss and accuracy curves of the validation and training data demonstrates that the network is adequately trained and avoids overfitting.
Figure 9. Accuracy and loss graphs of training and validation data from (a) BoniRob dataset and (b) CWFID.

4.3. Evaluation Metrics

The experiment aimed to assess the semantic segmentation performance across three classes (background, crop, and weed) using evaluation metrics including precision, the mean intersection of union (mIoU), recall, and weighted harmonic mean of precision and recall (F1-score), as outlined in Equations (23)–(27). These values were used to calculate the overall segmentation performance of the proposed framework. The number of classes was set to three. The evaluation metrics used to determine the accuracy of segmentation included true negative (TN), true positive (TP), false negative (FN), and false positive (FP) values. When the true and false labels match the prediction, the scenario is usually referred to as a TN or TP. FP and FN are terms used to describe scenarios in which an incorrect label is mistakenly anticipated as true and a valid label is mistakenly anticipated as a false label.
I o U = T P T P + F N + F P ,
m I o U = j = 1 C l s I O U j C l s ,
R e c a l l = T P T P + F N ,
P r e c i s i o n = T P T P + F P ,
F 1 - score = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l

4.4. Testing for the Proposed Framework

Below is the explanation of the experimental results of the proposed framework for semantic segmentation in Experiments 1 and 2.

4.4.1. Testing with CWFID after Training with BoniRob Dataset

Ablation Study
In the ablation studies, we conducted several experiments across various scenarios, encompassing six different cases outlined in Table 2.
Table 2. Representation of various cases.
We repeated these six cases in Experiments 1 and 2. Table 3, Table 4 and Table 5 present the results of Experiment 1 for semantic segmentation using segmentation networks including U-Net, modified U-Net, and CED-Net. Across all experiments, Case I showed the lowest performance, and Case VI (the proposed method) showed the highest performance. This confirms that the proposed schemes of using RH transformation, one training dataset, and DA can enhance segmentation accuracy in all the segmentation models. In addition, the semantic segmentation performance of Case VI and U-Net is the best.
Table 3. Comparisons of different cases using U-Net (Experiment 1) (Cr means crop, Wd means weed, Bg means background, Re means recall, and Pre means precision).
Table 4. Comparisons of different cases using modified U-Net (Experiment 1) (Cr means crop, Wd means weed, Bg means background, Re means recall, and Pre means precision).
Table 5. Comparisons of different cases using CED-Net (Experiment 1) (Cr means crop, Wd means weed, Bg means background, Re means recall, and Pre means precision).
An ablation study is presented in Table 6 to show the significance of the selection of the mean and std values for the RH transformation in the proposed framework. In this ablation study, we used the same training data as in the ablation study experiments with U-Net of Case VI (Table 3) and performed testing to validate the performance difference between the RH transformation and RH transformation with additional adjustment values that deviate from the selected values. We set the RH transformation with additional adjustment values to make the distribution of channels akin to that of the reference image in the histograms, as shown in Figure 3 and explained in Section 3.2.2. The performance results (F1-score) with additional adjustment values are 13.1% lower than those with the RH transformation without additional adjustments, as listed in Table 6. Although the RH transformation with additional adjustments makes the relative channel distributions similar between the target and reference images, as shown in Figure 3, it results in a greater decrease in the absolute numbers of red and green channels. Additionally, the mean of all channels decreases and moves toward zero, which reduces contrast and illumination, as shown in Figure 3. Moreover, the variance also increases, and all factors collectively cause the performance degradation of the RH transformation with additional adjustments.
Table 6. Comparison of results between RH transformation and RH transformation with additional adjustments (Cr means crop, Wd means weed, Bg means background, Re means recall, and Pre means precision).
Visual examples of the semantic segmentation results using U-Net, modified U-Net, and CED-Net are illustrated in Figure 10. In this illustration, red pixels represent the TP of crops, black pixels represent the TP of the background, and blue pixels signify the TP of the weed. Yellow pixels represent errors where crops were mistakenly identified as background or weeds, while orange pixels represent errors where weeds were mistakenly identified as background or crops. Gray pixels indicate errors where the background was mistakenly identified as weeds or crops. As depicted in the figure, the utilization of U-Net in the proposed framework demonstrates superior semantic segmentation accuracy.
Figure 10. Visual comparisons of various semantic segmentation outputs with proposed framework (Experiment 1): (a) original image; (b) RH-transformed image; (c) ground-truth mask; semantic segmentation results with (d) CED-Net; (e) modified U-Net, and (f) U-Net.
Performance Comparisons of the Proposed Framework with SOTA Transformations
We analyzed the SOTA transformations with those of the proposed method employing U-Net, modified U-Net, and CED-Net, as outlined in Table 7. Our findings confirm that the proposed framework yields the highest outcomes across all segmentation networks. Additionally, Table 7 reveals that U-Net within the proposed framework attains the highest segmentation performance for both crops and weeds.
Table 7. Performance comparisons of proposed method and SOTA transformations (Cr means crop, Wd means weed, Bg means background, Re means recall, and Pre means precision).
Visual examples of SOTA transformations and the proposed method using U-Net are shown in Figure 11. In this illustration, red pixels represent the TP of crops, black pixels represent the TP of the background, and blue pixels signify the TP of the weed. Yellow pixels represent errors where crops were mistakenly identified as background or weeds, while orange pixels represent errors where weeds were mistakenly identified as background or crops. Gray pixels indicate errors where the background was mistakenly identified as weeds or crops. As depicted in the figure, the proposed framework demonstrates the highest semantic segmentation accuracy.
Figure 11. Visual comparisons of various SOTA transformations with the proposed framework using U-Net (Experiment 1): (a) original image; (b) ground-truth mask; semantic segmentation results with (c) Xiao et al.; (d) Pitie et al.; (e) Gatys et al.; (f) Nguyen et al.; and (g) proposed method.

4.4.2. Testing with BoniRob Dataset after Training with CWFID

Ablation Study
For the ablation studies, six cases were considered in Experiment 2, as described in the Ablation Study of Section 4.4.1. Table 8, Table 9 and Table 10 present the results of Experiment 2 for semantic segmentation using segmentation networks. From these tables, we can observe that the accuracy is lower for Cases I, III, and IV without RH transformation for the test data, but the performance improved when we used RH transformation in Cases II, IV, and VI. In Case VI, the proposed framework shows the highest accuracy among all segmentation networks and the best accuracy with U-Net.
Table 8. Comparison of different cases using U-Net (Experiment 2) (Cr means crop, Wd means weed, Bg means background, Re means recall, and Pre means precision).
Table 9. Comparison of different cases using modified U-Net (Experiment 2) (Cr means crop, Wd means weed, Bg means background, Re means recall, and Pre means precision).
Table 10. Comparison of different cases using CED-Net (Experiment 2) (Cr means crop, Wd means weed, Bg means background, Re means recall, and Pre means precision).
We can see the visual examples of semantic segmentation outputs using U-Net, modified U-Net, and CED-Net in Figure 12. In this illustration, red pixels represent the TP of crops, black pixels represent the TP of the background, and blue pixels signify the TP of the weed. Yellow pixels represent errors where crops were mistakenly identified as background or weeds, while orange pixels represent errors where weeds were mistakenly identified as background or crops. Gray pixels indicate errors where the background was mistakenly identified as weeds or crops. As depicted in the figure, the proposed framework with U-Net demonstrates the highest semantic segmentation accuracy.
Figure 12. Visual comparisons of various semantic segmentation outputs with proposed framework (Experiment 1): (a) original image; (b) RH-transformed image; (c) ground-truth mask; semantic segmentation results with (d) CED-Net; (e) modified U-Net, and (f) U-Net.

Performance Comparisons of the Proposed Framework with SOTA Transformations

In this subsection, we compared the SOTA transformations with those of the proposed framework using U-Net, modified U-Net, and CED-Net, as listed in Table 11. We verify that the proposed framework shows superior performance over all segmentation networks. Moreover, Table 11 highlights that U-Net, when utilized within the proposed framework, obtained the highest semantic segmentation accuracy for crops and weeds.
Table 11. Comparisons of the performance of the proposed method and SOTA transformations (Cr means crop, Wd means weed, Bg means background, Re means recall, and Pre means precision).
Visual examples of SOTA transformations and the proposed method using U-Net are illustrated in Figure 13. In this illustration, red pixels represent the TP of crops, black pixels represent the TP of the background, and blue pixels signify the TP of the weed. Yellow pixels represent errors where crops were mistakenly identified as background or weeds, while orange pixels represent errors where weeds were mistakenly identified as background or crops. Gray pixels indicate errors where the background was mistakenly identified as weeds or crops. As illustrated in the figure, the proposed framework demonstrates the highest semantic segmentation accuracy.
Figure 13. Visual comparisons of various SOTA transformations with proposed framework using U-Net (Experiment 1): (a) original image; (b) ground-truth mask; semantic segmentation results with (c) Xiao et al.; (d) Pitie et al.; (e) Gatys et al.; (f) Nguyen et al.; and (g) proposed method.

4.5. Fractal Dimension Estimation

The FD is used as a mathematical metric to characterize the complexity of geometric structures, particularly fractal shapes, which exhibit self-similarity across different scales. Fractal shapes possess similar patterns or structures across different scales, and the complexity of these shapes can be quantified using a numerical value of FD that generally ranges between one and two []. A higher FD value indicates greater complexity. A common method for computing the FD is the box-counting algorithm []. We refer to the box-counting algorithm [] to compute the FD, which was implemented by the PyTorch [] platform in Python version 3.8 []. Algorithm 2 provides the pseudo-code to measure the FD.
Algorithm 2: Pseudo-code for measuring FD
Input: image (path to the input image)
Output: Fractal dimension (FD) value
1:  Read the input image and further convert it into grayscale
2:   Set the maximum box-size with the power of 2 and ensure the dimensions
     s = 2^(log(max(size(image))/log2)]
    Add the padding if required to match the dimensions
3:   Compute the number of boxes N(s) till minimum pixels
4:   Reduce box size by 2 and recalculate N(s) iteratively
    while s > 1
5:   Compute log(N(s)) and log(1/s) for each s
6:   Draw a fitted line to the points (log(N(s)) and log(1/s))
7:  FD value is the slope of the fitted line
   Return Fractal Dimension Value
The algorithm enables the calculation of the FD for both precise and approximate self-affine patterns and has wide applications across various natural and manmade systems [,]. The formula employed to estimate the FD utilizing the box-counting algorithm [] is as follows:
FD = lim s 0 log ( N ( s ) ) log ( 1 / s )
where N ( s ) represents the sum of boxes of size s , and FD signifies the fractal dimension defining the curve being analyzed. We employed a box-counting technique to calculate the approximate FD of various shapes. The method was evaluated using two datasets, BoniRob and CWFID. The experimental results are presented in Table 12, demonstrating the FD values and distribution of crops and weeds across different sections of the field. The 1st–3rd row FD values in Table 12 were computed from the 1st~3rd row images in Figure 14. In addition, the 4th~6th row FD values in Table 12 were computed from the 4th~6th row images in Figure 15. Higher FD values for crops and weeds represent the high complexities of crops and weeds, suggesting that farming experts or robots should pay more attention to discriminating between crops and weeds. This estimation technique can also automate farming systems by targeting and eliminating weeds through precise spraying in areas with high weed complexity, ultimately increasing the crop yield.
Table 12. FD values of images from CWFID and BoniRob dataset. The 1st~3rd row FD values are computed from the 1st~3rd row images of Figure 14. The 4th~6th row FD values are computed from the 4th~6th row images of Figure 15.
Figure 14. Visual representation of crops and weeds with samples from the CWFID for estimating FD values: (a) whole segmentation results; (b) weed segmentation result; and (c) crop segmentation result.
Figure 15. Visual representation of crops and weeds with samples from the BoniRob dataset for estimating FD values: (a) whole segmentation results; (b) weed segmentation result; and (c) crop segmentation result.

4.6. Comparisons of Processing Time

In this section, we compare the average processing times per image obtained using the proposed method with those obtained using the SOTA method. The unit ms denotes milliseconds. Table 13 illustrates that the method by Xiao et al. [] exhibits the highest processing time, whereas the method by Nguyen et al. [] demonstrates the lowest processing time. Interestingly, our proposed framework falls at the second lowest in processing time (as indicated in Table 13). Although the proposed framework has a higher processing time than Nguyen et al. [], our goal is to achieve high accuracies of crop and weed segmentation. The proposed framework with U-Net yields better results than the SOTA methods, as evidenced in Table 7 and Table 11 and Figure 11 and Figure 13.
Table 13. Comparisons of average processing time by proposed and SOTA methods (unit: ms).

5. Discussion

We performed statistical analysis using the Student’s t-test [] and calculated the Cohen’s d-value []. For this purpose, we calculated the mean and std of the mIoU using our method with U-Net, as shown in Table 7 and Table 11, respectively. Additionally, we calculated the mean and std of the mIoU using the second-best method (Pitie et al. []) with U-Net, as presented in Table 7 and Table 11. The measured p-value is 0.041, showing a confidence level of 95% with a significant difference, as depicted in Figure 16, and Cohen’s d-value is 0.936, representing a large effect size. This confirms that the proposed framework statistically surpasses the second-best method and achieves higher segmentation accuracy.
Figure 16. T-test results.
We also analyzed the semantic segmentation performance using the proposed method with U-Net based on gradient-weighted class activation mapping (Grad-CAM) [], as depicted in Figure 17. Grad-CAM typically depicts important features as reddish and yellowish colors, whereas unimportant features are shown in bluish colors as explainable artificial intelligence. Figure 17 shows the Grad-CAM image, obtained after the fourth convolution layer of U-Net, illustrating the accurate extraction of crucial features for crop and weed segmentation, confirming that our method can generate correct images in heterogeneous datasets for the accurate segmentation of crops and weeds.
Figure 17. Sample images of the CWFID with Grad-CAM visualization after the 4th convolution layer. (a) Input image, (b) ground-truth mask (gray, white, and black pixels mean weeds, crops, and background, respectively), (c) Grad-CAM of class 1 (weed), (d) Grad-CAM of class 2 (crop), and (e) segmented results.
Figure 18 shows examples of incorrect segmentation using the proposed method with U-Net. In this illustration, red pixels represent the TP of crops, black pixels represent the TP of the background, and blue pixels signify the TP of the weed. Yellow pixels represent errors where crops were mistakenly identified as background or weeds, while orange pixels represent errors where weeds were mistakenly identified as background or crops. Gray pixels indicate errors where the background was mistakenly identified as weeds or crops. The reason for the incorrect segmentation is that crops and weeds have similar shapes and colors, making them hard to distinguish, especially in cases where objects possess thin regions.
Figure 18. Samples of incorrect segmentation generated by proposed method with U-Net: (a) original image; (b) RH-transformed image; (c) ground-truth mask; and (d) semantic segmentation results.
As automation in agriculture is a subject of intensive research, it has led to many reformations in agriculture, such as saving time, reducing manpower, increasing yields through proper monitoring systems, and precise crop and weed detection. In agricultural precision for heterogeneous data, our proposed framework exhibits great performance, segregating crops and weeds accurately, which can increase yield and drive the transition to modern agriculture.

6. Conclusions

In this study, we introduce an approach for segmenting crops and weeds and estimating the FD using small amounts of training data in a heterogeneous data environment. This is the first study on the segmentation of crops and weeds within a heterogeneous environmental setup utilizing one training data sample. We rigorously investigated the factors that cause performance degradation in heterogeneous datasets, including variations in illumination and contrast. To solve this challenge, we proposed a framework that leverages mean and standard deviation adjustments using the RH transformation. Furthermore, we improved the performance using a small amount of training data. This additional small amount of training data significantly improves the segmentation performance with reduced computational power and training time. In addition, we introduced an FD estimation approach in our system, which was smoothly combined as an end-to-end task to furnish crucial insights into the distributional characteristics of crops and weeds. Through experiments using two open databases, we proved that our approach outperforms the SOTA method. Additionally, we confirmed that our approach showed statistically superior results than the second-best method in terms of the t-test and Cohen’s d-value. Furthermore, using Grad-CAM images, we validated the capability of the proposed method to extract essential features necessary for the accurate segmentation of crops and weeds. Nonetheless, instances of inaccurate segmentation were observed in scenarios where crops and weeds exhibit comparable colors, shapes, and thin regions, as depicted in Figure 18.
To address this issue, we would research the generative adversarial network-based transformation method for the correct segmentation of small-sized and thin crops and weeds having similar colors and shapes in heterogeneous data environments. Moreover, we applied our method to other tasks, such as box-based detection or classification in heterogeneous data environments. In addition, we would check the possibility of applying our method to other application areas including crack detection for building monitoring, medical image detection, and underwater computer vision.

Author Contributions

Methodology, R.A.; conceptualization, J.S.H.; data curation, S.G.K.; formal analysis, H.S.; resources, M.U.; visualization, H.A.H.G.; software, M.H.T.; validation, N.U.; supervision, K.R.P.; writing—original draft, R.A.; writing—review and editing, K.R.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported in part by the National Research Foundation of Korea (NRF) funded by the Ministry of Science and ICT (MSIT) through the Basic Science Research Program (NRF-2022R1F1A1064291) and in part by MSIT, Korea, under the Information Technology Research Center (ITRC) support program (IITP-2024-2020-0-01789) supervised by the Institute for Information & Communications Technology Planning & Evaluation (IITP).

Data Availability Statement

The datasets are available in [,], and the proposed framework is available in [].

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Jiang, Y.; Li, C. Convolutional Neural Networks for Image-Based High-Throughput Plant Phenotyping: A Review. Plant Phenomics 2020, 2020, 4152816. [Google Scholar] [CrossRef] [PubMed]
  2. Fathipoor, H.; Arefi, H.; Shah-Hosseini, R.; Moghadam, H. Corn Forage Yield Prediction Using Unmanned Aerial Vehicle Images at Mid-Season Growth Stage. J. Appl. Remote Sens 2019, 13, 034503. [Google Scholar] [CrossRef]
  3. Yang, Q.; Wang, Y.; Liu, L.; Zhang, X. Adaptive Fractional-Order Multi-Scale Optimization TV-L1 Optical Flow Algorithm. Fractal Fract. 2024, 8, 8040179. [Google Scholar] [CrossRef]
  4. Huang, T.; Wang, X.; Xie, D.; Wang, C.; Liu, X. Depth Image Enhancement Algorithm Based on Fractional Differentiation. Fractal Fract. 2023, 7, 7050394. [Google Scholar] [CrossRef]
  5. Bai, X.; Zhang, D.; Shi, S.; Yao, W.; Guo, Z.; Sun, J. A Fractional-Order Telegraph Diffusion Model for Restoring Texture Images with Multiplicative Noise. Fractal Fract. 2023, 7, 7010064. [Google Scholar] [CrossRef]
  6. AlSheikh, M.H.; Al-Saidi, N.M.G.; Ibrahim, R.W. Dental X-ray Identification System Based on Association Rules Extracted by k-Symbol Fractional Haar Functions. Fractal Fract. 2022, 6, 6110669. [Google Scholar] [CrossRef]
  7. Zhang, Y.; Yang, L.; Li, Y. A Novel Adaptive Fractional Differential Active Contour Image Segmentation Method. Fractal Fract. 2022, 6, 6100579. [Google Scholar] [CrossRef]
  8. Zhang, Y.; Liu, T.; Yang, F.; Yang, Q. A Study of Adaptive Fractional-Order Total Variational Medical Image Denoising. Fractal Fract. 2022, 6, 6090508. [Google Scholar] [CrossRef]
  9. Jiao, Q.; Liu, M.; Ning, B.; Zhao, F.; Dong, L.; Kong, L.; Hui, M.; Zhao, Y. Image Dehazing Based on Local and Non-Local Features. Fractal Fract. 2022, 6, 6050262. [Google Scholar] [CrossRef]
  10. Zhang, X.; Dai, L. Image Enhancement Based on Rough Set and Fractional Order Differentiator. Fractal Fract. 2022, 6, 6040214. [Google Scholar] [CrossRef]
  11. Zhang, X.; Liu, R.; Ren, J.; Gui, Q. Adaptive Fractional Image Enhancement Algorithm Based on Rough Set and Particle Swarm Optimization. Fractal Fract. 2022, 6, 6020100. [Google Scholar] [CrossRef]
  12. Cheng, J.; Chen, Q.; Huang, X. An Algorithm for Crack Detection, Segmentation, and Fractal Dimension Estimation in Low-Light Environments by Fusing FFT and Convolutional Neural Network. Fractal Fract. 2023, 7, 7110820. [Google Scholar] [CrossRef]
  13. An, Q.; Chen, X.; Wang, H.; Yang, H.; Yang, Y.; Huang, W.; Wang, L. Segmentation of Concrete Cracks by Using Fractal Dimension and UHK-Net. Fractal Fract. 2022, 6, 6020095. [Google Scholar] [CrossRef]
  14. Sultan, H.; Owais, M.; Park, C.; Mahmood, T.; Haider, A.; Park, K.R. Artificial Intelligence-Based Recognition of Different Types of Shoulder Implants in X-Ray Scans Based on Dense Residual Ensemble-Network for Personalized Medicine. J. Pers. Med. 2021, 11, 11060482. [Google Scholar] [CrossRef] [PubMed]
  15. Arsalan, M.; Haider, A.; Hong, J.S.; Kim, J.S.; Park, K.R. Deep Learning-Based Detection of Human Blastocyst Compartments with Fractal Dimension Estimation. Fractal Fract. 2024, 8, 8050267. [Google Scholar] [CrossRef]
  16. González-Sabbagh, S.P.; Robles-Kelly, A. A Survey on Underwater Computer Vision. ACM Comput. Surv. 2023, 55, 1–39. [Google Scholar] [CrossRef]
  17. Madokoro, H.; Takahashi, K.; Yamamoto, S.; Nix, S.; Chiyonobu, S.; Saruta, K.; Saito, T.K.; Nishimura, Y.; Sato, K. Semantic Segmentation of Agricultural Images Based on Style Transfer Using Conditional and Unconditional Generative Adversarial Networks. Appl. Sci. 2022, 12, 12157785. [Google Scholar] [CrossRef]
  18. Kim, Y.; Park, K.R. MTS-CNN: MTS-CNN: Multi-Task Semantic Segmentation-Convolutional Neural Network for Detecting Crops and Weeds. Comput. Electron. Agric. 2022, 199, 107146. [Google Scholar] [CrossRef]
  19. Wang, R.; Jiao, L.; Xie, C.; Chen, P.; Du, J.; Li, R. S-RPN: Sampling-Balanced Region Proposal Network for Small Crop Pest Detection. Comput. Electron. Agric. 2021, 187, 106290. [Google Scholar] [CrossRef]
  20. Huang, S.; Wu, S.; Sun, C.; Ma, X.; Jiang, Y.; Qi, L. Deep Localization Model for Intra-Row Crop Detection in Paddy Field. Comput. Electron. Agric. 2020, 169, 105203. [Google Scholar] [CrossRef]
  21. Kang, J.; Liu, L.; Zhang, F.; Shen, C.; Wang, N.; Shao, L. Semantic Segmentation Model of Cotton Roots In-Situ Image Based on Attention Mechanism. Comput. Electron. Agric. 2021, 189, 106370. [Google Scholar] [CrossRef]
  22. Le Louëdec, J.; Cielniak, G. 3D Shape Sensing and Deep Learning-Based Segmentation of Strawberries. Comput. Electron. Agric. 2021, 190, 106374. [Google Scholar] [CrossRef]
  23. Brilhador, A.; Gutoski, M.; Hattori, L.T.; de Souza Inácio, A.; Lazzaretti, A.E.; Lopes, H.S. Classification of Weeds and Crops at the Pixel-Level Using Convolutional Neural Networks and Data Augmentation. In Proceedings of the IEEE Latin American Conference on Computational Intelligence, Guayaquil, Ecuador, 11–15 November 2019; pp. 1–6. [Google Scholar] [CrossRef]
  24. Chebrolu, N.; Lottes, P.; Schaefer, A.; Winterhalter, W.; Burgard, W.; Stachniss, C. Agricultural Robot Dataset for Plant Classification, Localization and Mapping on Sugar Beet Fields. Int. J. Robot. Res. 2017, 36, 1045–1052. [Google Scholar] [CrossRef]
  25. Haug, S.; Ostermann, J. A Crop/Weed Field Image Dataset for the Evaluation of Computer Vision Based Precision Agriculture Tasks. In Proceedings of the Computer Vision—ECCV 2014 Workshops, Zurich, Switzerland, 6–7 and 12 September 2014; pp. 105–116. [Google Scholar] [CrossRef]
  26. Nguyen, D.T.; Nam, S.H.; Batchuluun, G.; Owais, M.; Park, K.R. An Ensemble Classification Method for Brain Tumor Images Using Small Training Data. Mathematics 2022, 10, 10234566. [Google Scholar] [CrossRef]
  27. Abdalla, A.; Cen, H.; Wan, L.; Rashid, R.; Weng, H.; Zhou, W.; He, Y. Fine-Tuning Convolutional Neural Network with Transfer Learning for Semantic Segmentation of Ground-Level Oilseed Rape Images in a Field with High Weed Pressure. Comput. Electron. Agric. 2019, 167, 105091. [Google Scholar] [CrossRef]
  28. Crops and Weeds Segmentation Method in Heterogeneous Environment. Available online: https://github.com/iamrehanch/crops_and_weeds_semantic_segmentation (accessed on 9 March 2023).
  29. Haug, S.; Michaels, A.; Biber, P.; Ostermann, J. Plant Classification System for Crop/Weed Discrimination without Segmentation. In Proceedings of the IEEE Winter Conference on Applications of Computer Vision, Steamboat Springs, CO, USA, 24–26 March 2014; pp. 1142–1149. [Google Scholar] [CrossRef]
  30. Lottes, P.; Hörferlin, M.; Sander, S.; Stachniss, C. Effective Vision-Based Classification for Separating Sugar Beets and Weeds for Precision Farming: Effective Vision-Based Classification. J. Field Robot. 2017, 34, 1160–1178. [Google Scholar] [CrossRef]
  31. Lottes, P.; Khanna, R.; Pfeifer, J.; Siegwart, R.; Stachniss, C. UAV-Based Crop and Weed Classification for Smart Farming. In Proceedings of the IEEE International Conference on Robotics and Automation, Singapore, Singapore, 29 May–3 June 2017; pp. 3024–3031. [Google Scholar] [CrossRef]
  32. Yang, B.; Xu, Y. Applications of Deep-Learning Approaches in Horticultural Research: A Review. Hortic. Res. 2021, 8, 123. [Google Scholar] [CrossRef]
  33. Chen, L.C.; Papandreou, G.; Schroff, F.; Adam, H. Rethinking Atrous Convolution for Semantic Image Segmentation. arXiv 2015, arXiv:1706.05587v3. [Google Scholar]
  34. Long, J.; Shelhamer, E.; Darrell, T. Fully Convolutional Networks for Semantic Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar] [CrossRef]
  35. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar] [CrossRef]
  36. Badrinarayanan, V.; Kendall, A.; Cipolla, R. SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef]
  37. Zou, K.; Chen, X.; Wang, Y.; Zhang, C.; Zhang, F. A Modified U-Net with a Specific Data Argumentation Method for Semantic Segmentation of Weed Images in the Field. Comput. Electron. Agric. 2021, 187, 106242. [Google Scholar] [CrossRef]
  38. Milioto, A.; Lottes, P.; Stachniss, C. Real-Time Semantic Segmentation of Crop and Weed for Precision Agriculture Robots Leveraging Background Knowledge in CNNs. In Proceedings of the IEEE International Conference on Robotics and Automation, Brisbane, Australia, 21–25 May 2018; pp. 2235–2299. [Google Scholar] [CrossRef]
  39. Paszke, A.; Chaurasia, A.; Kim, S.; Culurciello, E. ENet: A Deep Neural Network Architecture for Real-Time Semantic Segmentation. arXiv 2016, arXiv:1606.02147v1. [Google Scholar]
  40. Fathipoor, H.; Shah-Hosseini, R.; Arefi, H. Crop and Weed Segmentation on Ground_Based Images using Deep Convolutional Neural Network. In Proceedings of the ISPRS Annals of the Photogrammetry Remote Sensing and Spatial Information Sciences, Tehran, Iran, 13 January 2023; pp. 195–200. [Google Scholar] [CrossRef]
  41. Zhou, Z.; Siddiquee, M.M.R.; Tajbakhsh, N.; Liang, J. UNet++: A Nested U-Net Architecture for Medical Image Segmentation. In Proceedings of the Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support, Granada, Spain, 20 September 2018; pp. 3–11. [Google Scholar] [CrossRef]
  42. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2014, arXiv:1409.1556v6. [Google Scholar]
  43. Fawakherji, M.; Potena, C.; Bloisi, D.D.; Imperoli, M.; Pretto, A.; Nardi, D. UAV Image Based Crop and Weed Distribution Estimation on Embedded GPU Boards. In Proceedings of the Computer Analysis of Images and Patterns, Salerno, Italy, 6 September 2019; pp. 100–108. [Google Scholar] [CrossRef]
  44. Chakraborty, R.; Zhen, X.; Vogt, N.; Bendlin, B.; Singh, V. Dilated Convolutional Neural Networks for Sequential Manifold-Valued Data. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, South Korea, 27 October—2 November 2019; pp. 10620–10630. [Google Scholar] [CrossRef]
  45. You, J.; Liu, W.; Lee, J. A DNN-Based Semantic Segmentation for Detecting Weed and Crop. Comput. Electron. Agric. 2020, 178, 10570. [Google Scholar] [CrossRef]
  46. Wang, H.; Song, H.; Wu, H.; Zhang, Z.; Deng, S.; Feng, X.; Chen, Y. Multilayer Feature Fusion and Attention-Based Network for Crops and Weeds Segmentation. J. Plant Dis. Prot. 2022, 129, 1475–1489. [Google Scholar] [CrossRef]
  47. Siddiqui, S.A.; Fatima, N.; Ahmad, A. Neural Network Based Smart Weed Detection System. In Proceedings of the International Conference on Communication, Control and Information Sciences, Idukki, India, 16–18 June 2021; pp. 1–5. [Google Scholar] [CrossRef]
  48. Khan, A.; Ilyas, T.; Umraiz, M.; Mannan, Z.I.; Kim, H. CED-Net: Crops and Weeds Segmentation for Smart Farming Using a Small Cascaded Encoder-Decoder Architecture. Electronics 2020, 9, 9101602. [Google Scholar] [CrossRef]
  49. Reinhard, E.; Adhikhmin, M.; Gooch, B.; Shirley, P. Color Transfer between Images. IEEE Comput. Graph. Appl. 2001, 21, 34–41. [Google Scholar] [CrossRef]
  50. Ruderman, D.L.; Cronin, T.W.; Chiao, C.C. Statistics of Cone Responses to Natural Images: Implications for Visual Coding. J. Opt. Soc. Am. A-Opt. Image Sci. Vis. 1998, 15, 2036–2045. [Google Scholar] [CrossRef]
  51. Mikołajczyk, A.; Grochowski, M. Data Augmentation for Improving Deep Learning in Image Classification Problem. In Proceedings of the International Interdisciplinary PhD Workshop, Świnouście, Poland, 9–12 May 2018; pp. 117–122. [Google Scholar] [CrossRef]
  52. Agarap, A.F. Deep Learning Using Rectified Linear Units (ReLU). arXiv 2019, arXiv:1803.08375v2. [Google Scholar]
  53. Clevert, D.A.; Unterthiner, T.; Hochreiter, S. Fast and Accurate Deep Network Learning by Exponential Linear Units (ELUs). arXiv 2015, arXiv:1511.07289v5. [Google Scholar]
  54. Intel Core i5-2320. Available online: https://www.intel.com/content/www/us/en/products/sku/53446/intel-core-i52320-processor-6m-cache-up_to-3-30-ghz/specifications.html (accessed on 5 October 2023).
  55. NVIDIA GeForce GTX 1070. Available online: https://www.nvidia.com/en-gb/geforce/10-series/ (accessed on 5 October 2023).
  56. PyTorch. Available online: https://pytorch.org/ (accessed on 5 October 2023).
  57. Python 3.8. Available online: https://www.python.org/downloads/release/python-380/ (accessed on 5 October 2023).
  58. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. arXiv 2014, arXiv:1412.6980v9. [Google Scholar]
  59. Loshchilov, I.; Hutter, F. SGDR: Stochastic Gradient Descent with Warm Restarts. arXiv 2016, arXiv:1608.03983v5. Available online: https://arxiv.org/abs/1608.03983 (accessed on 5 October 2023).
  60. Sudre, C.H.; Li, W.; Vercauteren, T.; Ourselin, S.; Cardoso, M.J. Generalised Dice Overlap as a Deep Learning Loss Function for Highly Unbalanced Segmentations. In Proceedings of the International Conference on Communication, Control and Information Sciences, Québec City, Canada, 9 September 2017; pp. 240–248. [Google Scholar] [CrossRef]
  61. Xiao, X.; Ma, L. Color Transfer in Correlated Color Space. In Proceedings of the ACM International Conference on Virtual Reality Continuum and Its Applications, Hong Kong, China, 14 June 2006; pp. 305–309. [Google Scholar] [CrossRef]
  62. Pitié, F.; Kokaram, A.C.; Dahyot, R. Automated Colour Grading Using Colour Distribution Transfer. Comput. Vis. Image Underst. 2007, 107, 123–137. [Google Scholar] [CrossRef]
  63. Gatys, L.A.; Ecker, A.S.; Bethge, M. Image Style Transfer Using Convolutional Neural Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 23–27 June 2016; pp. 2414–2423. [Google Scholar] [CrossRef]
  64. Nguyen, R.M.H.; Kim, S.J.; Brown, M.S. Illuminant Aware Gamut-Based Color Transfer. Comput. Graph. Forum 2014, 33, 319–328. [Google Scholar] [CrossRef]
  65. Rezaie, A.; Mauron, A.J.; Beyer, K. Sensitivity Analysis of Fractal Dimensions of Crack Maps on Concrete and Masonry Walls. Autom.Constr. 2020, 117, 103258. [Google Scholar] [CrossRef]
  66. Wu, J.; Jin, X.; Mi, S.; Tang, J. An Effective Method to Compute the Box-counting Dimension Based on the Mathematical Definition and Intervals. Results Eng. 2020, 6, 100106. [Google Scholar] [CrossRef]
  67. Xie, Y. The Application of Fractal Theory in Real-life. In Proceedings of the International Conference on Computing Innovation and Applied Physics, Qingdao, Shandong, China, 30 November 2023; pp. 132–136. [Google Scholar] [CrossRef]
  68. Mishra, P.; Singh, U.; Pandey, C.M.; Mishra, P.; Pandey, G. Application of Student’s t-test, Analysis of Variance, and Covariance. Ann. Card. Anaesth. 2019, 22, 407–411. [Google Scholar] [CrossRef]
  69. Cohen, J. A Power Primer. Psychol. Bull. 1992, 112, 155–159. [Google Scholar] [CrossRef]
  70. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-CAM: Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.