Deep Learning-Based Weed–Crop Recognition for Smart Agricultural Equipment: A Review

: Weeds and crops engage in a relentless battle for the same resources, leading to potential reductions in crop yields and increased agricultural costs. Traditional methods of weed control, such as heavy herbicide use, come with the drawback of promoting weed resistance and environmental pollution. As the demand for pollution-free and organic agricultural products rises, there is a pressing need for innovative solutions. The emergence of smart agricultural equipment, including intelligent robots, unmanned aerial vehicles and satellite technology, proves to be pivotal in addressing weed-related challenges. The effectiveness of smart agricultural equipment, however, hinges on accurate detection, a task influenced by various factors, like growth stages, environmental conditions and shading. To achieve precise crop identification, it is essential to employ suitable sensors and optimized algorithms. Deep learning plays a crucial role in enhancing weed recognition accuracy. This advancement enables targeted actions such as minimal pesticide spraying or precise laser excision of weeds, effectively reducing the overall cost of agricultural production. This paper provides a thorough overview of the application of deep learning for crop and weed recognition in smart agricultural equipment. Starting with an overview of intelligent agricultural tools, sensors and identification algorithms, the discussion delves into instructive examples, show-casing the technology’s prowess in distinguishing between weeds and crops. The narrative highlights recent breakthroughs in automated technologies for precision plant identification while ac-knowledging existing challenges and proposing prospects. By marrying cutting-edge technology with sustainable agricultural practices, the adoption of intelligent equipment presents a promising path toward efficient and eco-friendly weed management in modern agriculture.


Introduction
Weeds are a big threat in agriculture as they occur in all parts of the field and compete with crop plants for resources.The result of competition for resources is reduced crop yields.Yield losses depend on factors, such as weed species, population density and relative time of emergence and distribution, as well as on the soil type, soil moisture levels, pH and fertility [1,2].For decades, researchers and farmers have struggled to control weeds to overcome the thorny challenges they pose.Weeds in the field compete with crops for water, nutrients and sunlight.If not controlled properly, weeds can adversely affect crop yield and quality.In addition, research has shown that there is a significant link between reduced crop yields and weed competition [2].For example, the annual cost of weeds in Australia within grain production systems is USD 3.3 billion, comprising USD 2.6 billion in costs for weed control and USD 0.7 billion in lost yield [3].
In today's agricultural sector, accurately identifying crops and weeds is crucial for improving agricultural productivity, reducing production costs and achieving sustainable agricultural development.The fast development of deep learning techniques for wide application in computer vision provides new opportunities for crop and weed Academic Editor: Sung-Cheol Koh recognition.The high automation and learning capabilities of deep learning models enable them to learn from large datasets and gradually improve their performance, bringing unprecedented breakthroughs to precision agriculture.Recently, the main methods of weed control in agricultural fields have included hand weeding, mechanical weeding, laser weeding and chemical weeding.Chemical weeding provides the advantage of low cost, and it is unaffected by terrain factors.It is widely used all over the world [4].The heavy use of herbicides increases weed resistance and increases the cost of agricultural inputs.Reducing the use of herbicides is also a critical step towards sustainable agriculture.Site-specific weed control can save up to 90% of herbicide expenditures.In addition, annual sales of pesticides worldwide amount to about USD 100 billion.If this idea can become reality, it will significantly reduce agricultural expenditure [5].Spraying pesticides over large areas can also pollute the environment.For example, indiscriminate broadcast spraying throughout tobacco fields, especially during the early growth phase, can lead to unnecessarily spraying bare soil off target between any two contiguous tobacco plants, causing environmental pollution and pesticide seepage into the ground [6,7].Pesticide use also has an impact on human health.The WHO has estimated that 1 million adverse reactions have been reported when hand-sprayed insecticides are used in crop fields [8].In order to better control the use of herbicides, due to the massive increase in over-reliance on herbicides and herbicide-resistant weeds, the EU's agricultural system has become more fragile and unsustainable.The EU Green Deal has a goal of cutting the use and risks of chemical fertilizers by 50 percent by 2030 [9].The European Food Safety Authority (EFSA) has announced that 98.9% of food products contain agrochemical residues (of which 1.5% exceed legal limits).In addition, plants resistance to agrochemicals (e.g., herbicides) is becoming a huge threat to crop yields in many countries [10].
Manual weeding is not only a heavy workload but also cannot easily detect weeds in a timely manner.The only solution to the problem is to increase manpower, but this will inevitably increase agricultural costs.Mechanical weed control is especially suitable for weed control in organic farmland and can also be useful in traditional farmland.On the other hand, the utilization of machines may also have a downside effect by damaging and eroding crops and the environment [11].Currently, weed removal in crop rows still relies on manual removal in many cases, but manual weeding is less efficient.With the development of deep learning algorithms, weed management has achieved successful results.Agricultural robotics research has increased over the past few years due to the potential applications of robots and industry efforts in robot development.The role of robots in many agricultural tasks has been studied, focusing mainly on improving the automation of traditional agricultural machinery and weeding processes [12,13].It can accurately recognize weeds and accurately deal with them, which greatly saves the use of herbicides, avoids environmental pollution and reduces agricultural costs.In smart agriculture, using sensors installed on satellites, unmanned aerial vehicles or ground tractors to separate them between weeds and crops is becoming an effective method of weed management.Remote sensing technology allows for quickly charting the distribution of weeds and crops over large areas [14].An SVM-based system for a crop/weed detection system for tractor boom sprayers to spot spray tobacco crops in the field was constructed.Its classification accuracy is 96% [6].In the last decade or so, Earth observation satellites have provided higher-resolution free remote sensing data, making the detection of agriculture by high-resolution satellites possible.Google Street View images were tested using a convolutional neural network (CNN), with an overall accuracy of 83.3 percent [15].Laser weed control also offers a new possibility for weed removal.A YOLOX convolutional neural network-based weeding robot utilizes a blue laser to weed with a weed recognition rate of 88.94% [16].Drones are considered to be more efficient than robotic or satellite acquisition because they can rapidly collect field data at very high spatial resolution and at low cost [17][18][19].The most widely used application is the use of drones, which are utilized to capture RGB images and tested on a test set using SVM, KNN, AdaBoost and CNN, whose accuracies for recognizing rice weeds are 89.75%,85.58%, 90.25% and 92.41%, respectively [20].
This paper reviews the current state of research on applying deep learning to crop and weed recognition for smart agricultural equipment.There are many previous review articles related to this topic.For example, Imran Zualkernan et al. [21] focused on new deep learning models and architectures for research using drone image data since 2018.Jiayou Shi et al. [22] presented a thorough review of the methods and applications related to crop row inspection in agricultural machinery navigation.They paid special attention to sensors and systems used for crop row detection in order to validate their sensing and detection capabilities and, thus, improve their sensing and inspection capabilities.Ana I. de Castro et al. [23] reviewed the sensor types, configurations and image processing algorithms of UAVs for agriculture and forestry applications.WenHao Su [14] discussed RGB, hyperspectral and spot spectroscopy in sensors for crop and weed identification.However, they did not provide a comprehensive introduction to intelligent agricultural equipment.We briefly describe the need for intelligent weed management and then present aspects of weed control.Section 2 focuses on the image recognition steps for smart devices, including image collection, image preprocessing and feature extraction.Section 3 describes the application of using deep learning algorithmic models for recognizing weeds in smart agricultural equipment.They mainly utilize convolutional neural networks (CNNs) and their variants, such as Faster RCNN [24], MTS-CNN [25], FHGSO-based Deep CNN [26] and DRCNN [27].In addition to this, support vector machine (SVM) is also heavily used, mostly in agricultural equipment, such as tractors, drones, etc. [4,6,24].Most notably, a Transformer Neural Network and its variants, for example, vit [28], Swin-DeepLab+ [29] and Deformable DETR, are used [30].Moreover, the vit model is a relatively new proposed model, which outperforms some advanced models such as EfficientNet and ResNet, so this model has great potential [28].With the rapid transformation of agricultural landscapes, driven by technological innovations, this review aims to synthesize the current state of the art in the application of deep learning-based smart agricultural equipment for weed and crop differentiation.By elucidating state-of-the-art technologies, identifying research gaps and suggesting potential directions for future research, this study aims to contribute to the development of intelligent and autonomous systems that empower farmers with the tools to address weed management challenges, leading to sustainable and efficient agricultural management.

Weed Detection Using Remote Sensing Technique
The workflow of image recognition of crops and weeds can generally be divided into four steps: image data acquisition, preprocessing, feature extraction and classification of weeds and crops [31].The specific details are shown in Figure 1.

Image Data Collection
DL-based weed inspection and classification techniques require a sufficient quantity of labeled data.Data can be gathered using various types of sensors mounted on various smart agricultural devices.The main sensors commonly used are as follows: RGB sensors, multispectral sensors, hyperspectral sensors and LiDAR sensors.Table 1 shows the images collected by different sensors.
Visible Light Sensors are most commonly used by UAVs in precision agriculture and related smart agriculture applications.RGB imaging or color imaging has gained popularity due to its clear color-revealing principle, simple hardware structure and proven production process.The costs of the RGB cameras are comparatively inexpensive, light weight and perform well in drawing orthophoto maps that capture images and aerial videos of the entire field in a single instance.UAVs equipped with RGB cameras have the benefits of small size, low cost, great productivity and mobility [32].Meanwhile, RGB imaging has many benefits, providing only limited data at a limited number of wavelengths [33].
Recently, technologies such as hyperspectral imaging (HSI) systems have provided a chance to quickly categorize plant species, both in the laboratory and the field.The advantage of HSI is to provide an integrated analysis of spectroscopy and the relationship between various chemical components and absorption in the spectrum.The principle of HSI spectroscopy is based on the vibration of muscles in the infrared region.Therefore, absorbance at specific wavelengths, which might be related to specific chemical bands, can be used for different materials' classification and quality determination.Weed identification techniques based on RGB imaging are based on shape, size and color discrimination, while the use of HSI increases the value of such techniques [34].However, hyperspectral images typically contain a great number of superfluous information, which may mask the real information of ground objects and adversely affect spectral data recognition.In addition, high-dimensional spectral data not only increase temporal and spatial complexity but also tend to cause a dimensional disaster.To address the above problems, Zhihua Diao et al. [35] proposed a lightweight three-dimensional convolution neural network model.An image enhancement method was used to improve the training results to address the problem of sparse training samples in hyperspectral images.A lightweight unit module was introduced on this basis to reduce the number of parameters in the network.Meanwhile, Zhaoxia Lou proposed a 3D-CNN model for predicting the CCI of a competition index.There are two key aspects of hyperspectral band selection, the effective preservation of information and the elimination of redundancy.Many hyperspectral studies use VIP for band selection because it performs better in terms of information preservation.However, this method has the limitation of retaining an excessive number of bands in the band selection process, which may identify irrelevant bands as significant.Therefore, the use of the VIP method in band selection may require further research [36,37].
Compared to hyperspectral cameras, a multispectral camera is lightweight, low-cost and has high spatial resolution, making it suitable for large areas [38].In contrast to RGB cameras, multispectral cameras have additional spectral bands and are capable of sensing radiation in both the invisible (red-edge and near-infrared) and visible segments of the spectrum, typically spanning four to six bands.The inclusion of a reflectance calibration panel makes multispectral cameras less susceptible to environmental variation [39,40].A multispectral image is essentially a collection of grayscale images, with each image corresponding to a specific wavelength or band of wavelengths in the electromagnetic spectrum.Multispectral imaging (MSI) involves capturing images from various spectral bands to gather both spatial and spectral information.MSI technology enables the creation of wavelength channels in the near-UV, visible, near-IR, mid-IR and far-IR bands [33].One of the most commonly used techniques for the composition of multispectral images is the co-registration of the bands of interest.The images captured by the multispectral cameras show significant band misregistration effects due to lens distor-tion and the varying viewing angles of each lens or sensor [41].To obtain accurate spectral and geometrical information, a precise geometric distortion correction and band-toband co-registration method is necessary [42].Multispectral imaging, with the advantage of light hardware and faster calculation speed, is emerging as the successor to hyperspectral technology.
Thermal infrared sensors help to capture the temperature of the objects, generate images and display the same based on the information collected.Infrared sensors and optical lenses are used in thermal cameras to capture thermal energy [43].The development of higher-resolution thermal imaging systems compatible with unmanned aerial vehicles (UAVs) has facilitated the practical application of thermal imaging in agriculture.The use of thermal measurement, in conjunction with other sensor measurements, such as hyperspectral, visible and optical distance, has proven to be more effective in field-scale crop phenotyping [44].When combined with deep learning, remote heat sensing technology is able to recognize crops and weeds and crop stress assessment [45].
LiDAR, which stands for Light Detection and Ranging, is a highly advanced and dependable sensor that has been widely used in the fields of crop row detection and robotic navigation.This sensor is famous for its high precision, wide range and strong immunity to interference [46].LIDAR works on the principle that the transmitting system emits visible or near-infrared light waves.These light waves are then reflected off the target and detected by the receiving system.The data obtained are subsequently processed to generate parametric information, including distance.LiDAR sensors have been utilized in crop row detection to provide highly accurate and detailed 3D maps of crop canopies [47].Additionally, LiDAR sensors have the capability to penetrate vegetation and capture ground surface data, facilitating the detection of crop rows, even in densely vegetated fields [22].LiDAR can be used in intensive agricultural scenarios.RGB-red, green, blue; NIR-near infrared.

Preprocessing
After acquiring data from various sources, it is essential to prepare the data for the training, testing and validation of models.Raw data may not always be suitable for deep learning (DL) models.Approaches for dataset preparation include the application of various image processing techniques, data labeling, utilization of image enhancement methods to augment the input data and introduce variations, as well as the generation of synthetic data for training.The commonly used image processing techniques are removal of background, resizing of captured images, green component segmentation, removal of motion blur, denoising, image enhancement, extraction of color vegetation indices and alteration in color models [58].Table 2 demonstrates the effect of different image enhancement techniques on segmentation.

Image Resizing
Achieving good accuracy with lower patch sizes proved to require less training time for the model.To expedite processing and reduce computational complexity, many studies performed image resizing operations on the dataset before feeding it into the deep learning (DL) model.Following the collection of field images, their resolution was adjusted to meet the DL network's requirements [58].Julien Champ et al. [56] resized the image so that their shorter edge was 1200 pixels and the longest one 2048 pixels.This allowed the model to be run in a reasonable time on a standard graphics processing unit.Reenul Reedha et al. [28] extracted the crop and weed image patches from the bounding boxes.Then, the image patches were resized to 64 × 64 pixels.This choice of image size aligned with the dimensions of the bounding boxes, possibly corresponding to the altitude at which the UAV was flown and the size of the crops in the study field.Images with high resolution are sometimes split into a number of patches to reduce the computational complexity.Ramirez et al. [60] captured only five images at high resolution using a drone, which were then segmented into non-overlapping chunks and chunks with overlap.These adjustments to the image pixel size can reduce the computational com-plexity and decrease the computational duration of the DL model to achieve optimal results.

Image Enhancement and Denoising
Image enhancement and denoising such strategies can effectively enhance the accuracy of algorithm recognition.Reenul Reedha et al. [28] utilized data augmentation strategies to enrich datasets, including random resized crop, color dithering and rand augments.This technology is achieved using Keras ImageDataGenerator, which instantly generates enhanced images.As a result, the basic ViT B-16 model reached a recognized accuracy of 99.4%.The use of data augmentations aimed to enhance the model's robustness and generalization capabilities.Aichen Wang et al. [59] assessed the performance of the DL model based on the input representation of images.They applied many image preprocessing operations, such as histogram equalization, automatic adjustment of the contrast of images and deep photo enhancement.Babu et al. [27] performed image enhancement through CLAHE, which allowed for better visual interpretation of images.CLAHE has superior contrast limiting compared to ordinary adaptive histogram equalization.In conventional adaptive histogram equalization, the noise in the near-constant regions in images is magnified.The CLAHE algorithm improves the image contrast and limits the amplification, improving the quality of the image.CLAHE is widely used for enhancing medical imagery, satellite images, etc. Dmitrii Vypirailenko et al. [61] utilized two methods for data enhancement.The first was to resize the image to 128 × 128 and then enhance the data by horizontal and vertical flipping, panning and rotating.They also used random contrast correction as an enhancement method to ensure the effectiveness of the enhanced image.Another approach was to use random affine transformation.Each enhancement method was applied to an image when it passed to the model.They also used weights in cross-entropy loss to overcome the imbalance in the dataset.The result of the enhancement should be similar to the image taken at the real site.In conclusion, effective enhancement and denoising of images have a significant impact on the recognition of the algorithm.

Background Removal
Background removal has an important role in weed identification.The aim of segmentation is to extract plant ROI by segregating the background (i.e., soil, stones, etc.) from the vegetation (i.e., leaves of different weeds).Zhaoxia Lou et al. [37] extracted the vegetation canopy spectra for the acquired images.The contrast between the vegetation canopy and the soil background was improved using OSAVI.Mask images of soil background and vegetation canopy were generated through a threshold segmentation method, effectively eliminating the soil portion from the digital surface model (DOM) image, retaining only the vegetation canopy area.Li et al. [34] developed a threshold segmentation algorithm involving spectral data extraction with a threshold of 0.19 at a 950 nm wavelength.The mask generated in this way was multiplied by the original HSL image, and the resulting image contained only plants.At the same time, they used a simple linear iterative clustering algorithm to segment the plant images into hyperpixels.This was accomplished by taking the similarity in spectral and spatial domains into account when grouping pixels into clusters.The results show that the separation of a crop from the background can be achieved by spectral characterization and threshold adjustment.MLP developed using Sp data is a more robust and reliable method compared to traditional classification methods.Similarly, a threshold adjustment in the color is utilized to achieve separation of the crop from the background.Borja Espejo-Garcia et al. [54] first normalized the R, G and B channels in the image for the green channel and then used the ExG (excess green) values to be indexed by the initial vegetation segmentation, followed by OTSU thresholding of the grayscale image to obtain a binary mask.Based on this method in threshold segmentation, many algorithms also obtained more than 95% accuracy in weed identification.Gee, C. et al. [62] proposed a new vegetation index called MetaIndex, which combined the advantages of six vegetation indices.The method refined the results by geodesic segmentation and obtained a black-and-white vegetation image, also known as a black-and-white vegetation mask.

Feature Extraction of Weeds
In agriculture, there are four groups of descriptive features: visual textures, spatial contexts, spectral features and biological morphological features [63].

Visual Texture Feature
For textural features, humans can judge them through their senses, such as identifying whether soft or hard, rough or fine, horizontally or vertically corrugated, etc. [64].Research on texture-aware properties has its origins in computer vision as well as cognitive science.In computer vision-based approaches, visual textures have played a key role in image understanding.And because the texture of the local image descriptors is pooled in an unordered manner, the texture of the image is represented by computing the intensity of the clustered pixels in the space, and six common variability directions are identified [65].Figure 2 is a sample image of texture-based segmentation using a Gabor filter.The Gabor filter, which is a group of Gabor wavelets, automatically determines the boundaries between tobacco and non-tobacco objects (weeds) based on their texture characteristics.The extracted Gabor texture features are input to a k-means clustering algorithm.This classifies textured regions of tobacco from other texture classes (weeds), as shown in Figure 2. It is evident from (b) that the tobacco plant has prominent texture features, as compared to the surrounding objects.Table 3 shows a table of deep learning recognition based on texture feature weed identification.GLCM is a way to define the texture of images using the information of intensity values that co-occur spatially.The technique used texture features derived from a graylevel co-occurrence matrix (GLCM).The next step was the extraction of four texture features from GLCM.These features include contrast, correlation, energy and homogeneity, with 73% accuracy using the Radial Basis Function (RBF) kernel in the support vector machine (SVM) [66].The Gabor wavelet transform enables the analysis of image scenes both in spatial and frequency domains.It is important to note that the wavelet transform of an image is a well-established multi-resolution filtering technique for extracting texture features.Each derived (preprocessed) image was filtered with a bank of Gabor wavelet filters computed with designated lower (Ul) and higher (Uh) frequencies selected to be 0.1 and 0.5, respectively.Four levels of orientation and ten levels of scale were chosen [67].
Yajun Chen et al. [3] identified six texture features, comprising a histogram of oriented gradient features, rotation-invariant local binary pattern (LBP) feature, Hu invariant moment feature, Gabor feature, gray-level co-occurrence matrix, and gray-level-gradient co-occurrence matrix.These six feature descriptors were combined to create a set of 18 feature combinations.For the problem of image size normalization, they proposed a strategy that kept the shape of the leaves unchanged and supplemented 0 pixels in the blank area of the normalized size.Lei Zhang et al. [68] proposed a weed recognition method for support vector machines using any combination of three sets of texture features, including oriented gradient histogram features, rotation-invariant local binary pattern (LBP) features, and grayscale co-occurrence matrix (GLCM).The application of six different texture features for weed identification is enumerated in Table 2.For hybrid feature extraction, the accuracy obtained using machine learning is greater than single feature extraction, and the accuracy of deep learning is greater than machine learning.For the study of deep learning in crop and weed recognition, hybrid texture features can be utilized.

Spatial Context Feature
Plant discrimination based on morphological and spectral properties is prone to variations in plant appearance, exhibiting significant differences in the field, across fields and during the growing season.This variability makes the detection method less stable.In contrast, the sowing pattern of crops is relatively stable, as most crops are sown or planted in rows following a predetermined pattern.Leveraging spatial contexts or posi-tion information can enhance discrimination accuracy [73].In a crop field, crops are often planted regularly in the field, and spatial coordinates can be used to discriminate between crops and weeds.Weeds and crops can also be identified by spatial features [64].Considering that most crops are sown or planted in rows with a predetermined pattern, spatial contexts or position information can contribute to improving discrimination accuracy.For cereals, the detection of inter-row weeds can be effectively achieved by identifying the centerline and edge of crop rows between adjacent crop plants.Figure 3 shows a sample image of bean and spinach based on spatial feature recognition.The Hough transform is a widely employed method for identifying linear features in an image.It works by representing a straight line as a spike in parameter space, where the parameters correspond to the characteristics of the line.In addition, the linear Hough transform can be utilized for detecting or analyzing arbitrary (non-parametric) curves by examining the shape of peaks or their locations in the parameter space [74].Teplyakov et al. [75] proposed a lightweight Artificial Neural Network for line detection with several convolutional layers and a fast Hough transform layer that can be trained in an end-to-end manner.They proposed to use fast Hough transform (FHT) with O(N2logN) complexity.The FHT approximated the lines with dyadic patterns and utilized an efficient solution for summation.In complex backgrounds, the model of YOLOv5s was more accurate than the detection of Hoff variations and was faster.In order to solve the problems of large memory overhead, long time consumption and low recognition accuracy of offset Hough transform, slam, N. et al. [76] proposed an efficient circle localization algorithm based on multi-resolution segmentation (two-step optimized Hough transform).First, the target circle was obtained by adaptive image preprocessing to determine the location of the effective search area.Then, high-quality images were separated by shape quality inspection to be used as accurate data sources.Finally, the location accuracy is improved to the sub-pixel level using least squares circle fitting.The effects of burrs, misalignments, defects and contamination are also reduced.The extraction of spatial features can also be used as an auxiliary recognition criterion when the UAV is flying overhead.

Spectral Feature
Spectroscopy is used to acquire spectral information over a wide spectral range, in which specific frequencies of vibrations can be perceived that match the jump energy of a key or group.Spectroscopy is also categorized in many ways; the common ones are Point Spectroscopy, RGB and hyperspectral imaging, fluorescence spectroscopy and multispectral Imaging.The theoretical basis for using spectral detection is that weed rivalry leads to changes in plant physiology that alter light-absorbing and crown reflectance properties [37].Figure 4 shows the regions of interest for corn seedlings and weeds on hyperspectral images and the corresponding average spectral curves.Table 4 shows shows a sample image of hawkweed flowers based on spatial feature recognition.Islam et al. [77] employed RGB images captured by RGB cameras mounted on a drone.They extracted the reflectance of red, green and blue bands and subsequently calculated vegetation indices, including normalized red band, normalized green band and normalized blue band.The purpose of this normalization was to reduce the effects of different lighting conditions on the color channels.Moreover, in addition to RGB data, Fawakherji et al. [78] took into account near-infrared (NIR) information, generating four channel multispectral synthetic images.They extracted the plant cover from the entire image cover.The plant cover was a binary image where the plant pixels to be learned were set to 1, and the other pixels were set to 0. The plant cover was then mapped to a realistic multispectral image, and the resulting image was used for data enhancement.The use of an NIR channel helps to enhance the accuracy of the activity for which vegetation inspection is required.Photosynthesis in healthy green plants leads to the absorption of more solar energy in the visible spectrum, resulting in a low reflectance level in the RGB channels.Similarly, the reflectance of the NIR spectrum is affected by the same phenomena with opposite results, with a high reflectance level in the NIR channel, where generally 10% or less of radiation is absorbed [78,79].Jinya Su et al. [38] studied that the triangular greenness index (TGI) consisting of green-NIR was the most discriminative SI.Its recognition accuracy was 93.0%.Utilizing thermal measurements in conjunction with other sensor data, such as hyperspectral, visible and optical distance, has demonstrated increased effectiveness in field-scale crop phenotyping [80][81][82].

Biological Morphological Features
Biological morphological features are five characteristics represented by the shape, structure, size, pattern and color of an organism.In agriculture, biomorphic traits can identify the biomorphic characteristics of weeds and crops, although they are more susceptible to leaf-folding or shading problems.They also have a high accuracy rate after training.This current deep learning algorithm approach based on biomorphic feature recognition is innovative [64].Figure 5 illustrates a schematic of biometric extraction through leaves.Color features are extracted from the pixels of images, with advantages of stable features after rotation, scale and translation changes [86].Weeds and crop seedlings are the same green color.It is difficult to distinguish them by color alone [30].The extraction of color features requires the use of color moments, which provide unique features for distinguishing objects based on their color.Color moments are founded on the probability distribution of image intensities, characterized by statistical moments, like mean, variance and skewness.These three are the central moments of intensity distribution and can be easily found for all color spaces, such as RGB, HSV and L*a*b [6].Apart from these color features, there are other shape descriptors/features proposed by researchers.Tannouche et al. [87] used a region-based adjacencies descriptor to discriminate between Dicot and Monocot weeds.The proposed descriptor calculated two numbers of adjacencies between a given original pixel and their adjacent pixels.The first was the number of horizontal and vertical adjacencies, and the second one was the number of diagonal adjacencies.Shape factors that were generated by transformations typically required the use of information about the boundaries or contours of the segmented region and required complex calculations, compared with region-based shape measurements and indices.Therefore, they are often referred to as region-based shape descriptors.Hu's moment invariants (MIs) are popular shape descriptors, which are normalized functions created based on the information of both shape boundary and interior region [88].Weed detection using machine vision relies on features, like plant color, leaf texture, shape and patterns.Drought stress can impact leaf color and morphological features in plants, potentially affecting the reliability of machine vision-based weed detection [89].But they still lack universal segmentation capabilities for different crop varieties with varying leaf shapes and canopy structures.Designing a universal 3D segmentation method for different varieties at multiple growth stages is the current research frontier of plant phenotyping [90].Biomorphic feature extraction has the advantages of strong interpretability, high stability and wide versatility in weed recognition, and it is especially suitable for scenarios that require the identification of different types of plants.
In deep learning, hybrid feature extraction refers to the simultaneous use of multiple levels, sources or types of features for model training and recognition.Different levels and types of features contain different levels of abstraction and semantic information.Hybrid feature extraction captures this diverse information and enables the model to represent the input data more richly.Single feature extraction may ignore or lose some critical information.Using features from multiple sources can make the model more robust and better adaptable to variations and noise in the input data.

Applications for Weed/Crop Discrimination
Deep learning algorithms have developed rapidly over the past few years, leading the possibility of smart farms.Many scientists have studied the problem of applying deep learning algorithms to smart agricultural equipment to recognize weeds and crops.

Learning Algorithm
Deep Neural Networks (DNNs) aim to replicate the communication between biological neurons through layers of nodes, comprising input, hidden and output layers [91].Deep Neural Networks (DNNs) extend the complexity, number of connections and hidden layers of Artificial Neural Networks (ANNs).A convolutional neural network (CNN), a type of DNN, assigns learnable weights and biases to different aspects and objects within input images to distinguish and classify objects, such as weeds [1].Unlike traditional machine learning algorithms that require manual feature selection and classifier choice, deep learning algorithms automatically extract features through self-learning from errors.This automatic feature extraction sets deep learning apart from the broader field of machine learning [1,92,93].To train and evaluate a deep CNN model, each input image undergoes a sequence of convolution layers with filters, followed by flattening, pooling layers and fully connected layers.CNNs autonomously capture the spatial and temporal dependencies within the input image using relevant filters, resulting in enhanced and more efficient image processing.This is achieved with a significantly reduced number of estimable parameters and processing time.Due to potential slight jittering in the graphical information formed by adjacent positions, the pooling operation extracts essential information from the upper feature map.Common pooling operations include maximum pooling and average pooling.The model maintains translation and rotation invariance while preserving crucial features [94].
The attention mechanism is becoming a key concept in the deep learning field.The inspiration for attention comes from the human perception process, where individuals naturally concentrate on specific information, simultaneously neglecting other perceptible details.This attention mechanism has significantly influenced the realm of natural language processing, particularly in prioritizing a subset of crucial words.The selfattention paradigm has evolved from the attention concepts, demonstrating enhancements in the performance of deep networks [95].The utilization of the self-attention mechanism enables the establishment of global references during both model training and prediction.This significantly reduces the training time required to attain high accuracy [96,97].The self-attention mechanism is a crucial element in transformers, explicitly modeling interactions among all entities in a sequence for structured prediction tasks.Essentially, a self-attention layer updates each element of a sequence by consolidating global information from the entire input sequence.In contrast to the fixed K × K neighborhood grid of convolution layers, the self-attention's receptive field encompasses the entire image.This expanded receptive field of self-attention enhances its capability compared to CNN, all without introducing the computational costs associated with excessively large kernel sizes.Moreover, self-attention remains invariant to permutations and variations in the number of input points.Consequently, it can seamlessly operate on irregular inputs, in contrast to standard convolution that necessitates grid structures [98].
Overall, the attention mechanism has some advantages in improving model performance, processing sequence data and improving interpretability.However, for specific tasks, the attention mechanism is not always necessarily superior to the traditional deep neural network structure but, rather, the appropriate model structure should be selected according to the specific application scenario and task requirements.

Recognition Applications
The above describes the part of deep learning collecting images and pre-processing.The following is a review of the latest applications of these techniques to recognize weeds and weed control in smart agricultural equipment.Table 5 demonstrates the accuracy of different deep learning algorithm models for crop/weed recognition.

Spot Photographic Image Recognition
Spotting refers to taking images with a cell phone or camera at a fixed location.This method of acquiring images is relatively simple but requires a great deal of labor to take them.Fixed-point photography usually occurs in relatively fixed environments, which means that the images are relatively consistent in terms of background, lighting and camera angle.This consistency helps train deep learning models to better adapt to specific environments and conditions.It also facilitates the labeling of the images, making it easy to improve the training efficiency of deep learning.
Taskeen Ashraf et al. [66] sought to classify images based on grass density into three classes.The first approach utilized texture features extracted from the gray-level cooccurrence matrix (GLCM) with a Radial Basis Function (RBF) kernel in a support vector machine (SVM), achieving an accuracy of 73%.Another technique employed scale and rotation-invariant moments to classify grass density.The second technique outperformed the first, achieving an accuracy of 86% with a Random Forest classifier.This kind of quantitative agricultural spraying for different densities of weeds can effectively reduce the use of pesticides.To improve weed recognition, some scientists have combined machine learning with deep learning.Tao T.et al. [99] proposed a deep convolutional neural network with a support vector machine classifier aimed at improving the classification accuracy of winter oilseed rape seeding and field weeds.They used a VGG network model with true-color images (224 × 224 pixels) of oilseed rape/weeds as input.The proposed VGG-SVM model obtained a higher classification accuracy, greater robustness and real time.Borja Espejo-Garcia et al. [54] proposed a novel crop/weed identification system.The method involved fine-tuning pre-trained convolutional networks, such as Xception, Inception-Resnet, VGNets, Mobilenet and Densenet.These networks were combined with "traditional" machine learning classifiers, like support vector machines, XGBoost, and Logistic Regression.These classifiers were trained with features extracted from deep learning models.The aim of this approach was to prevent overfitting and achieve a robust and consistent performance.Attention mechanisms have become increasingly popular in recent years and can greatly increase the rate of recognition.Helong Yu et al. [29] introduced a soybean field weed recognition model named Swin-DeepLab.This model was built upon an enhanced DeepLabv3+ model, incorporating a Swin transformer as the feature extraction backbone.Furthermore, a convolution block attention module (CBAM) was integrated after each feature fusion to improve the model's utilization of focused information within the feature maps.The proposed network can further address the problem of weed recognition in intensive agricultural scenarios.

Satellite Photo Image Recognition
In recent decades, substantial progress has been achieved in sensing technologies, wireless communication, autonomous systems and artificial intelligence through collaborative research efforts worldwide [110].Agricultural satellites use remote sensing techniques, including visible, infrared and microwave radiation, to capture information about the Earth's surface.These satellites can provide high-resolution images that can be used to monitor different aspects of agricultural land.Some civil satellites in agriculture, combined with high-performance sensors, have produced a large number of images of farmland with various temporal, spatial and spectral resolutions.Among other things, these images are of great significance to farmers for seeding scheduling, pest and disease tracking and weed control [111].Satellite remote sensing image acquisition, although providing large spatial coverage, has limited its development in the field of smart agriculture due to fixed and long revisit intervals and problems such as cloud cover [112].
Anita Sabat-Tomala et al. [83] conducted a comparison between two machine learning algorithms, support vector machine (SVM) and Random Forest (RF), for the identification of Solidago spp., Calamagrostis epigejos and Rubus spp. on HySpex hyperspectral aerial images.The classifications were performed on 430 spectral bands and on the most informative 30 bands extracted using the Minimum Noise Fraction (MNF) transformation.While satellite images are less suitable for weed recognition, semantic segmentation of remotely sensed images proves to be more effective.In the realm of digital agricultural services, there is a growing need for farmers or their advisors to provide digital records of field boundaries.Automatic extraction of field boundaries from satellite imagery would reduce the reliance on manual input of these records, which is time consuming and would underpin the provision of remote products and services [113,114].

Application of Drone Weed Identification
An unmanned aerial vehicle (UAV) is a powered flying vehicle that operates without a human operator.It can fly autonomously or be controlled remotely, equipped with various payloads.UAVs are rapidly advancing due to their benefits in flexible data acquisition and high spatial resolution.They offer a potent technical solution for numerous applications in precision agriculture (PA) [115,116].For better acquisition of image data, the flight altitude of the UAV is an important parameter as it has a great impact on the resolution of the image, the flight time and the computational cost of image processing [117].Moreover, UAVs have the flexibility to carry various payloads tailored to specific purposes.In precision agriculture (PA), UAVs are commonly equipped with remote sensors, such as RGB imaging, multispectral and hyperspectral imaging sensors, thermal infrared sensors, Light Detection and Ranging (LiDAR) and Synthetic Aperture Radar (SAR) to capture agricultural information [112,118].UAVs are currently used for surveillance [119], disease detection [120] and weed management [20,24].The use of these data allows for the identification of specific spatial features and time-varying in-formation on crop characteristics as well as the targeted spraying of pesticides and fertilizers, resulting in a reduction in pests and diseases and an increase in crop yields and quality [115,121].Figure 6 shows some of the uses of drones in smart agriculture.
Hile Narmilan Amarasingam et al. [40] studied the potential of machine learning (ML) algorithms for the detection of mouse-ear grass leaves and flowers from multispectral (MS) images acquired by unmanned aerial vehicles (UAVs) at different spatial resolutions and compared different machine learning.The highest machine learning recognition was achieved with 100% accuracy.Jinya Su et al. [38] analyzed and mapped blackgrass in wheat fields by incorporating unmanned aerial vehicles (UAVs), multispectral imagery and machine learning techniques.Eighteen widely used techniques were produced from five raw spectral bands.Various feature selection algorithms were then used to refine the simplicity and experience interpretation of the model.The selection of these raw spectral segments and the selection of vegetation indices (VIs) were important for weed identification in multispectral images.Mohd Anul Haq et al. [103] proposed a novel CNNLVQ model to detect weeds in soybean crop images and distinguish between grassy weeds and broadleaf weeds.The uniqueness of their study lies in the development of this innovative CNNLVQ model, meticulous hyperparameter optimization and the utilization of authentic datasets.Faster R-CNN stands out as a deep learning approach incorporating a region proposal network (RPN).This network, formed by merging convolutional features with a classification network, facilitates training and testing through a seamless process.It results in a fast detection rate and outperforms other conventional object detection methods.Shahbaz Khan et al. [24] optimized the architecture of the traditional Faster-R-CNN.Residual Network 101 (ResNet-101) was deployed as a convolutional neural network instead of the normally used Visual Geometry Group 16 (VGG16).Anchors are classified using a traditional SoftMax classifier.In addition, Saad Abouzahir et al. [102] used HOG blocks as key points to generate visual words based on the Bag of Visual Words (BOVW) method and feature vectors as histograms of these visual words.And a backpropagation neural network was used to detect weeds and classify plants from three different crop fields (sugar beet, carrot, soybean).The algorithm had 97.7%, 93% and 96.6% accuracy in weed and crop differentiation.
Drones have an important role in identifying weeds in fields and spraying pesticides in real time.Shahbaz KhanI et al. [116] developed a deep learning-based real-time recognition system for drones.The capability of the system is achieved through a twostep process where the target recognizer part is based on a CNN model.The developed deep learning system achieved an average F1 score of 0.955, while the classifier recognition average computation time was 3.68 ms.This deep learning model can effectively solve the problem of real-time pesticide spraying by UAVs to recognize weeds.Meanwhile, Gunasekaran Raja et al. [71] proposed a UAV-assisted weed detection method using a modified multi-channel gray-scale covariance matrix (GLCM-M) and normalized difference index with red threshold (NDIRT) index (DA-WDGN) to assist the weed detection process.In DA-WDGN, the UAV incorporates information and communication techniques to capture far-field data and accurately detect weeds.The accurate detection of weeds limits the need for pesticides and helps to protect the environment.Reenul Reedha et al. [28] investigated a Visual Transformer (ViT) and applied it to plant classification in unmanned aerial vehicle (UAV) images.They utilized the strategy of migration algorithms to increase the effectiveness of the test set while reducing the training set.The ViT algorithm is able to efficiently process large-scale image data, thus better adapting to the large number of images produced by UAVs in aerial photography.This efficient image processing capability helps to improve the speed and accuracy of weed identification.Moreover, the ViT algorithm is based on the self-attention mechanism, which is able to capture global information in the image and not only limited to local features.This feature gives ViT and UAVs a huge advantage in the future development of weed recognition.

Application of Agricultural Robotics for Weed Recognition
Agricultural robots represent an important trend in modern agricultural automation.By combining machines, sensors and autonomous navigation technologies, they are revolutionizing agricultural production.Agricultural robots can include modified tractors, small ground robots and aerial robots [13].Modern agricultural equipment integrates advanced technologies, such as artificial intelligence, navigation, sensing systems and communication, to increase agricultural productivity and promote smart agriculture [22,122,123].Among the information, navigation data, image recognition data, etc., require the work of sensors, including monocular cameras, binocular cameras, RGB cameras, panorama cameras and spectral imaging systems [22].In the early days of precision agriculture, most image data from fields were collected using ground cameras either mounted on unmanned ground vehicles (UGVs) or fixed next to vegetation patches [21].Through image recognition, agricultural robots can perform laser weeding [124][125][126], spraying pesticides for weed control [13], spot picking [127,128], fertilizer application and other tasks.Figure 7 shows some of the agricultural robots used in smart agriculture.
Yajun Chen et al. [4] trained the classifier based on SVM comparing single features of six features with different fusion strategies.The highest classification accuracy was obtained by fusion feature combining rotationally invariant LBP features with a gray gradient co-occurrence matrix based on an SVM classifier and accurately detected various weeds and maize seedlings.Tufail M et al. [5] presented a machine learning-based crop/weed detection system for tractor boom sprayers to spot spray tobacco crops in the field and proposed an SVM classifier with carefully selected combinations of tobacco plant features (texture, shape and color) with a classification accuracy of 96%.Julien Champ et al. [56] trained and evaluated an instance segmentation convolutional neural network designed to segment and identify each plant specimen visible in an agricultural robot image.And they adjusted the hyperparameters of a mask region-based convolutional neural network (R-CNN) to this specific task and evaluated the resulting training model.Data augmentation via Generative Adversarial Networks (GANs) can add entire synthetic scenes to the training data, thus expanding and enriching their information content [78].
There has been a lot of progress in image recognition using smart agricultural robots, and a number of scientists are working on smart weeding by agricultural robots to reduce the burden on farmers.Yayun Du et al. [5] provided a complete process, from model training at maximum efficiency to deploying TensorRT-optimized models to single-board computers.And the performance of five different CNN models was tested.They deployed MobileNetV2 on a small autonomous robot, SAMBot, for real-time weed detection.In a previously unseen scenario in a flax field (row spacing of 0.2-0.3m), with crops and weeds, distortions, blurring and shadows, 90% accuracy was achieved.Paolo Rommel Sanchez et al. [107] developed a modular precision sprayer that distributes the high computational load of CNNs to parallel low-cost, low-power vision computing devices.The sprayer employed a customized precision spray algorithm based on SSD-MobileNetV1 running on a Jetson Nano 4 GB.The model achieved 76% mAP0.5 at 19 fps in detecting weeds and soybeans in a widely planted field.Muhammad Shahab Alam et al. [55] developed and deployed a vision-based robotic spraying system.By using the vision system in combination with speed sensors, flow sensors and pressure sensors, the technology detected and categorized tobacco plants and weeds in real time.The use of targeted pesticide spraying technology has reduced the use of pesticides, and environmental pollution has been effectively controlled, but laser-targeted weed control is underway in terms of more environmentally friendly future development.Huibin Zhu et al. [16] designed a weeding robot based on the YOLOX convolutional neural network for removing weeds from corn seedling fields.They verified the feasibility of a blue laser as a non-contact weeding tool.Similarly Azmat Hussain et al. [126] designed a laser weeding robot based on the YOLOV5 convolutional neural network.The field trials demonstrated that the robot took approximately 23.7 h at a linear velocity of 0.07 m/s for the weeding of one acre plot.It included 5 s of laser to kill one weed plant.They proposed an innovative weeding operation method, applying herbicides after causing mechanical damage to weeds, and designed a composite intelligent in-row weeding robot based on this method.Based on the YOLOv5 algorithmic model, the detection accuracy reached 93.33% under real operating conditions.The machine was more efficient at weeding compared to simple machines and reduced the amount of pesticides used compared to chemical pesticide spraying robots [129].

Discussions
In the context of the development of artificial intelligence, smart agriculture is the development direction of a large agricultural country.The development of intelligent agriculture is inseparable from the development of intelligent agricultural equipment.In recent years, agricultural robots, agricultural drones, satellites and other booming developments for the development of intelligent agriculture have provided a new program.The development of all three types of smart agricultural equipment is the mainstream of the future, and all have great potential for application in the development of smart farms.Satellites, as part of smart farming equipment, play an important role in delineating farm boundaries for effective farm management.However, it is slightly lacking in weed and crop identification.Waldner, F. et al. proposed a method to facilitate the extraction of site boundaries from satellite images [113].The use of satellite technology to segment and monitor sites in agriculture has a number of benefits that can help farmers to plan land use more accurately.This includes identifying the most suitable locations for specific crops, avoiding overuse of land and increasing the sustainable utilization of agricultural land.And it allows for better allocation of resources, such as water, fertilizers and pesticides, reducing waste of resources and environmental pollution.As an overhead drone, it has an integral role in smart agriculture.For example, high-resolution image acquisition provides a dataset for the training and learning of deep learning algorithms; the detection and identification of crops using sensors allow for the precise application of medicine and irrigation [24,38].Combining deep learning with drones allows for weed crop identification and targeted pesticide spraying.Based on RGB camera sensing, CNN has more than 92% accuracy in weed recognition.It has a higher accuracy rate compared to machine learning.And the results shown by Vit provide the possibility of real-time recognition of pesticide spraying by drones in the future.Weeds can be dealt with more efficiently and with less wastage of resources.Deep learning-based agricultural robots are essentially 95% accurate in weed recognition.The use of agricultural robots in agriculture is not only in data collection and weed identification and processing, as it also allows for precise picking and harvesting of crops.Overall, the combination of deep learning and smart agricultural equipment has been widely used in weed/crop identification research.In smart agriculture scenarios, deep learning has been used to solve the problem of crop and weed identification.Deep learning has four steps in weed/crop detection: data collection, dataset preparation, weed detection and weed/crop localization and classification.First of all, for the collection of datasets, with the help of intelligent agricultural equipment, the collection of images is no longer a problem.Moreover, a variety of sensors have improved the quality of image acquisition.Multispectral cameras have some advantages over RGB cameras and hyperspectral cameras in that they can improve more spectral bands than RGB cameras and are cheaper than hyperspectral cameras, which can be utilized in smart agriculture to reduce the cost and improve the quality of collected images [38,39].Thermal measurements from thermal infrared sensors can complement measurements from other sensors, such as hyperspectral, visible and optical distance, and have also been shown to be more effective in field crop phenotyping [44].For training datasets, manual labeling by researchers is still required, which is a very labor-intensive task.However, semi-supervised learning algorithms and unsupervised learning algorithms are a worthwhile solution for the future, as they can perform labeling during iterations, greatly reducing the human workload.Feature extraction of weeds and crops is an important part of the recognition process, and the main features are texture features, spectral features, spatial features and biomorphic features.All four features have a great role in weed recognition by deep learning, but the current trend in recognition is hybrid feature extraction of spectral features, texture features and biomorphic features.The similarity between weeds and crops makes using a single image feature to detect weeds and crops almost impossible.The commonly used image features can achieve the purpose of weed detection, but the experimental accuracy is low, and the stability is poor in a nonideal environment due to the complex interference factors in the actual field.Acquired images need to be preprocessed for better recognition and classification.The scientists segmented the crop and background by threshold segmentation and color segmentation and performed noise reduction on the images [34,131].
The performance of different deep learning algorithm models in weed/crop identification is influenced by a variety of factors.The main factor is the network structure.In general, lightweight CNN models are less accurate in weed recognition compared to CNN models.However, lightweight CNN models are usually designed to be more concise, using fewer parameters and computational resources, and they require relatively less memory space [104,109].Some of the lightweighting techniques include network pruning, quantization and depth-separable convolution, which aim to minimize the size of the model while maximizing the retention of its representational power.Due to the performance improvement in the Faster R-CNN architecture, it is possible to perform target detection, image classification and instance segmentation simultaneously in a single neural network.The researcher improved the Mask R-CNN by adding an attention mechanism and deep separable convolution.This approach improves the model's ability to represent weed-related features and reduces the number of model parameters, increasing computational speed [132].In addition to this, the performance of deep learning algorithms is greatly influenced by the training strategy used.The training strategy involves the training process of the model, selection of hyperparameters, data augmentation, etc.For example, batch normalization of deep learning models by some researchers accelerates training and improves the generalization performance of the model [54].In addition, the input dataset is key to training deep learning models as it is the basic source of information.The accuracy of deep learning is improved by data augmentation of sample images, as stated in Section 2 of this article.Algorithmic models such as Swin transformer and DeepLabv3+ also excel in weed identification.

Challenges for Weed Recognition in Smart Farming Equipment and Future Trends
In terms of future development, the combination of sensor and drone technology can effectively increase the efficiency of identification.Among the recent innovations, unmanned aerial vehicles (UAVs) or drones have demonstrated their suitability for the timely tracking and assessment of vegetation status due to several advantages, as follows: (1) They can operate at low altitudes to provide aerial imagery with ultra-high spatial resolution, allowing for the detection of fine details of vegetation.(2) The flights can be scheduled with great flexibility according to critical moments imposed by vegetation progress over time.(3) They can use diverse sensors and perception systems, acquiring different ranges of the vegetation spectrum (visible, infrared, thermal).( 4) This technology can also generate digital surface models (DSMs) with three-dimensional (3D) measurements of vegetation by using highly overlapping images and applying photoreconstruction procedures with the structure-from-motion (SfM) technique [23,35,44].
The future of agricultural robotics promises more developments in weed removal: (1) Increased intelligence and autonomy: Future agricultural robots will be more intelligent, with highly autonomous decision-making capabilities.Combined with artificial intelligence and deep learning technology, the robot can analyze farmland images and data in real time, make intelligent weed identification and weeding decisions, without human intervention, and improve operational efficiency.(2) The integration of multimodal sensing technology: Agricultural robots will integrate a variety of sensors, including vision, infrared, ultrasonic and other multimodal sensors, to obtain richer and more accurate information about the farmland.This will help identify weeds more accurately and adapt to different farmland environments.(3) Efficient and precise weeding technology: Future agricultural robots will use more precise and efficient weeding technology.This will require more advanced weeding systems and automated control technologies.Although laser mowing is currently very advantageous, there are still issues to consider, such as whether mowing is safe and whether it can cause fires [124,125].
Deep learning also faces several challenges in weed and crop recognition.First, due to the small visual differences between weeds and crops, there are large similarities between categories, which leads to models that are prone to confusion.In addition, there are variations in weeds and crops such as growth stages and environmental differences, and the models need to have good generalization capabilities to accommodate these variations [73].In addition, datasets are costly to annotate, especially when collected and labeled in a large-scale farmland environment.This poses certain difficulties in model training.To overcome these challenges, future research can be expanded in the following aspects: First, further improve the robustness and generalization ability of deep learning models, and design more effective feature extraction methods and classification algorithms for the similarities between weeds and crops.Second, develop larger-scale datasets containing samples from different times, locations and farming conditions to enhance the generalization ability of the model.At the same time, techniques such as augmented learning and transfer learning are reasonably utilized to achieve better results with fewer data.In addition, combining sensors and smart agricultural equipment tech-nologies for the real-time identification of weeds and crops contributes to intelligent and precise decision making in agricultural production.Proper dosage of plant protection products is one of the key issues in agricultural production.Using advanced sensor technology, crop growth can be monitored more accurately.This technology allows for the timely dosing of weeds or diseases.Spraying the right amount of insecticide will neither cause contamination by using too much nor reduce crop yields by using too little.

Conclusions
This review concentrates on the forefront applications of intelligent agricultural equipment, specifically emphasizing crop and weed identification, pivotal components in the trajectory of smart agriculture.The integration of sensors into smart agricultural equipment assumes a critical role in data acquisition, capturing extensive sets of highdimensional images that serve as foundational training data for deep learning algorithms.Various preprocessing techniques are employed to refine the algorithmic processes, encompassing noise reduction, background effect elimination and image resizing.Deep learning algorithms emerge as powerful tools capable of analyzing complex, high-dimensional data with distinct characteristics compared to the training set, facilitating accurate crop identification.The adoption of hybrid feature extraction techniques underscores the inherent advantages of leveraging multiple features in tandem, contributing significantly to the efficacy of weed and crop identification processes.In the realm of machine learning and deep learning, the attention mechanism stands out as a particularly valuable and promising learning algorithm.Renowned for its high accuracy and expedited processing time, the attention mechanism proves advantageous in the context of crop and weed identification.These attributes position it as a formidable asset for smart agricultural equipment engaged in real-time weeding operations within agricultural fields.The emphasis on attention mechanisms reflects a forward-looking perspective, acknowledging their potential to augment the efficiency and accuracy of smart agricultural practices, particularly in the domain of weed management.

Figure 1 .
Figure 1.General workflow of image processing-based weed detection.

Figure 3 .
Figure 3. From left to right: line detection in bean (a) and spinach (b) fields.Detected lines are in blue.In the spinach field, inter-row distance and the crop row orientation are not regular.The detected lines are mainly located in the center of the crop rows [17].

Table 1 .
Examples of public datasets.

Table 2 .
Effect of different image enhancements on image segmentation.

Table 3 .
Deep learning recognition based on texture feature weed identification.

Table 4 .
Deep learning recognition based on spectral feature weed identification.

Table 5 .
Classification of weeds and crops with regard to algorithms.