Next Article in Journal
A Bibliometric Analysis of Organic Farming and Voluntary Certifications
Previous Article in Journal
Bibliometric Network Analysis of Crop Yield Gap Research over the Past Three Decades
Previous Article in Special Issue
The Path to Smart Farming: Innovations and Opportunities in Precision Agriculture
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Editorial

Application of Vision Technology and Artificial Intelligence in Smart Farming

1
College of Artificial Intelligence, Nanjing Agricultural University, Nanjing 210031, China
2
Faculty of Applied Science, University of British Columbia, Kelowna, BC V1V 1V7, Canada
3
School of Applied Meteorology, Nanjing University of Information Science and Technology, Nanjing 210044, China
4
Faculty of Engineering and Information Technology, University of Technology Sydney, Sydney, NSW 2007, Australia
*
Author to whom correspondence should be addressed.
Agriculture 2023, 13(11), 2106; https://doi.org/10.3390/agriculture13112106
Submission received: 10 October 2023 / Revised: 18 October 2023 / Accepted: 31 October 2023 / Published: 6 November 2023
With the rapid advancement of technology, traditional farming is gradually transitioning into smart farming. Smart farming is an agricultural production system that harnesses modern technologies such as artificial intelligence (AI), big data, and automation technology, which assists farmers in effectively managing and optimizing the production process through data and image analysis. The ultimate goal of smart farming is to achieve precision agriculture and enhance efficiency and quality in agricultural production.
This Special Issue focuses on the application of visual technology and artificial intelligence in smart farming. Researchers from Asia and Europe have contributed a total of fourteen papers, including thirteen articles and one review, to this issue. The key information for each paper is shown in Table 1. These papers encompass a broad spectrum of technologies, such as image recognition algorithms, machine learning techniques, remote sensing technology, and 3D point cloud technology. The objective is to employ these technologies for monitoring the phenotypes of plants and animals, as well as their growth environments. By offering theoretical and technical support, personalized agricultural management solutions are made accessible to farmers.
Machine learning is a widely used modeling technique that leverages large quantities of data to acquire knowledge and make predictions. It can be applied to forecast trends in the growth environment of animals and plants. For example, Ren et al. [1] developed a prediction model for relative soil moisture (RHs 10 cm) in the 0–10 cm soil layer using the extreme gradient boosting (XGBoost) algorithm based on atmospheric and soil factors, which is capable of reasonably predicting the development process of drought events. Analyzing and predicting the growth environment of crops using data analysis methods can provide preventive measures for potential hazards. This method is equally effective for poultry farming.
Deep learning is a specialized machine learning field based on artificial neural networks. It involves learning and training through multi-layered neural networks to extract complex features and patterns from data, enabling more precise predictions. It is particularly suitable for handling complex, large-scale data and high-dimensional features and has wide applications in the field of computer vision. Applying deep learning models to analyze RGB images, remote sensing images, and 3D point cloud data in agriculture can enable more accurate monitoring of phenotypes of animals and plants, facilitating precision farming management. Jiao et al. [2] developed predictive models for the initial flowering period of Platycladus orientalis (Chinese thuja) using recurrent neural networks (RNN), long short-term memory networks (LSTM), and gated recurrent units (GRU). Shapely Additive Explanation (SHAP) was used to analyze the contribution rates of meteorological factors. The accuracy of all three models was significantly higher than that of a regression model based on the accumulated temperature. Among them, the GRU model performed the best, with an average accuracy exceeding 98%. Guo et al. [3] proposed an improved YOLOv5 object detection model, integrating the coordinate attention module and the deformable convolution module for accurately detecting mature Zanthoxylum on a mobile picking platform, addressing the issues of irregular shape and occlusion caused by branches and leaves. Li et al. [4] proposed a lightweight wheat growth stage detection model and a dynamic migration algorithm, which utilizes edge computing to migrate the detection model to the wireless network edge server for processing, improving efficiency significantly compared to the local implementation. By accurately monitoring the growth trends of animals and plants through deep learning and computer vision technologies, they can effectively improve production efficiency. In addition, this technology can also be applied to classification tasks to achieve personalized management of the same type of subjects. Zhang et al. [5] proposed a method for identifying individual dairy cattle in large-scale dairy farms. They used the DeepOtsu model to binarize the body pattern image for primary classification and the EfficientNet-B1 model for secondary classification, and the overall identification accuracy reached 98.5%. Cui et al. [6] proposed an improved CNN-LSTM model for classifying high-yield and low-yield cow udders that have undergone fine-grained segmentation using the SOLOv2 method, which could allocate them to different production groups.
Furthermore, visual technology can detect diseases based on the phenotypic information of animals and plants, providing early warnings for farmers to reduce economic losses. Therefore, the accuracy of model recognition is particularly important for the practical application. Several authors in this issue have researched model improvement. Ding et al. [7] proposed a novel model called RFCA ResNet, which incorporates multi-scale feature extraction and a dual attention mechanism for classifying and recognizing apple leaf diseases. Additionally, the adverse effects of imbalanced datasets on classification accuracy were effectively minimized using the class balance technique in conjunction with focal loss. Hao et al. [8] proposed an improved Faster R-CNN network characterized by the fusion of a 101 layers-deep residual network (ResNet101) and Path Aggregation Network (PAN) for monitoring the feeding behavior of hens, and the ability of the model to extract features is greatly enhanced according to the visualization results of the feature map output by the convolutional layer at each stage of the network. Lee et al. [9] proposed an image-processing method based on a keypoint detection algorithm and image-matching algorithms for detecting small-sized honey mites attached to bees, which can result in economic losses. Additionally, they employed Contrast Limited Adaptive Histogram Equalization (CLAHE) based on the RGB color model to enhance image quality. Their method demonstrated effective performance when applied to the measured 300 mm data. Yu et al. [10] proposed a stored grain pest identification method based on a triple-attention module (FCS), namely, frequency domain attention (FAM), channel attention (CAM), and spatial attention (SAM) to solve pest-detection and segmentation tasks.
However, the improvement in the recognition accuracy of models relying solely on single-image information is limited, making it difficult to capture image features and abstract concepts in complex tasks, which leads to less accurate or complete processing results. Some models require a large amount of annotated data for training, and the lack of sufficient data can affect the effectiveness and generalization ability. When image processing techniques are combined with other technologies, including multidimensional images, data fusion, etc., it enables more precise monitoring of subjects in farming. He et al. [11] proposed a novel decision-making method based on a multimodal fusion detection model, and multiple models were used to predict the rice seed varieties according to 2D images and 3D point cloud datasets to calculate a comprehensive score vector. Finally, the predicted probabilities from 2D and 3D were jointly weighted to obtain the final predicted probability, which could combine the advantages of different data modalities and significantly improve the final prediction results. Sun et al. [12] used multi-view stereoscopic technology (MVS) to reconstruct the entire growth period (13 stages) of five different soybean varieties in three dimensions, constructed a 3D dataset named Soybean-MVS with the labels of the entire soybean growth period, and used RandLA-Net and BAAF-Net two point cloud semantic segmentation models to verify its usability, which can provide usable basic data support for the 3D crop model segmentation models. Xu et al. [13] proposed a reconstruction algorithm based on 3D information for the detection of maize phenotypic traits, utilizing a multi-view registration algorithm and iterative closest point (ICP) algorithm for the global 3D reconstruction of maize seedling populations, which contributes to precise and intelligent early management of maize.
The works received for this Special Issue demonstrate the feasibility of applying artificial intelligence and visual technologies in smart farming. Karunathilake et al. [14] provide a comprehensive overview of the recent innovations in smart farming technology, such as drones, sensors, and automation. The agricultural environment is practically diverse and challenging, with variations in soil conditions, climate patterns, and so on. To develop effective models for smart farming, it is essential to create algorithms that can generalize well across different environments. Moreover, the health and growth of plants and animals are influenced by a range of factors, so the integration of multimodal data and the fusion of multiple features hold great potential for improving the accuracy of predictions and identifications in smart farming applications. Despite the potential benefits of smart farming technology, there are challenges to promoting and popularizing its application. The technology must not only meet practical application requirements, but also address issues such as cost and farmer acceptance. On the one hand, hardware, software, and maintenance costs should be affordable for small and medium-sized farmers. On the other hand, the widespread implementation of smart farming technology requires a strong digital infrastructure and connectivity. In many rural areas, the lack of reliable internet and mobile networks can hinder the deployment of smart farming solutions, and this issue may be addressed well by cloud-edge coordinated computing, achieving more efficient computing and data processing. This Special Issue contains papers related to the above innovative research and will hopefully stimulate further research in these areas.

Acknowledgments

We are thankful to Sunyuan Wang, Qiuhong Zhang, Xinyu Zhang and Tao Liu, who have contributed to helping with paper information sorting.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ren, Y.; Ling, F.; Wang, Y. Research on provincial-level soil moisture prediction based on extreme gradient boosting model. Agriculture 2023, 13, 927. [Google Scholar] [CrossRef]
  2. Jiao, G.; Shentu, X.; Zhu, X.; Song, W.; Song, Y.; Yang, K. Utility of deep learning algorithms in initial flowering period prediction models. Agriculture 2022, 12, 2161. [Google Scholar] [CrossRef]
  3. Guo, J.; Xiao, X.; Miao, J.; Tian, B.; Zhao, J.; Lan, Y. Design and experiment of a visual detection system for zanthoxylum-harvesting robot based on improved YOLOv5 model. Agriculture 2023, 13, 821. [Google Scholar] [CrossRef]
  4. Li, Y.; Liu, H.; Wei, J.; Ma, X.; Zheng, G.; Xi, L. Research on winter wheat growth stages recognition based on mobile edge computing. Agriculture 2023, 13, 534. [Google Scholar] [CrossRef]
  5. Zhang, R.; Ji, J.; Zhao, K.; Wang, J.; Zhang, M.; Wang, M. A cascaded individual cow identification method based on DeepOtsu and EfficientNet. Agriculture 2023, 13, 279. [Google Scholar] [CrossRef]
  6. Cui, G.; Qiao, L.; Li, Y.; Chen, Z.; Liang, Z.; Xin, C.; Xiao, M.; Zou, X. Division of cow production groups based on SOLOv2 and Improved CNN-LSTM. Agriculture 2023, 13, 1562. [Google Scholar] [CrossRef]
  7. Ding, J.; Zhang, C.; Cheng, X.; Yue, Y.; Fan, G.; Wu, Y.; Zhang, Y. Method for classifying apple leaf diseases based on dual attention and multi-scale feature extraction. Agriculture 2023, 13, 940. [Google Scholar] [CrossRef]
  8. Hao, H.; Fang, P.; Jiang, W.; Sun, X.; Wang, L.; Wang, H. Research on laying hens feeding behavior detection and model visualization based on convolutional neural network. Agriculture 2022, 12, 2141. [Google Scholar] [CrossRef]
  9. Lee, H.; Kim, M.; Kim, S.; Lee, S.; Lee, H.; Sin, J.; Mo, C. Identifying an image-processing method for detection of bee mite in honey bee based on keypoint analysis. Agriculture 2023, 13, 1511. [Google Scholar] [CrossRef]
  10. Yu, J.; Shen, Y.; Liu, N.; Pan, Q. Frequency-enhanced channel-spatial attention module for grain pests classification. Agriculture 2022, 12, 2046. [Google Scholar] [CrossRef]
  11. He, X.; Cai, Q.; Zou, X.; Feng, X.; Yin, W.; Qian, Y. Multi-modal late fusion rice seed variety classification based on an improved voting method. Agriculture 2023, 13, 597. [Google Scholar] [CrossRef]
  12. Sun, Y.; Zhang, Z.; Sun, K.; Li, S.; Yu, J.; Miao, L.; Zhang, Z.; Li, Y.; Zhao, H.; Hu, Z.; et al. Soybean-MVS: Annotated three-dimensional model dataset of whole growth period soybeans for 3D plant organ segmentation. Agriculture 2023, 13, 1321. [Google Scholar] [CrossRef]
  13. Xu, N.; Sun, G.; Bai, Y.; Zhou, X.; Cai, J.; Huang, Y. Global reconstruction method of maize population at seedling stage based on Kinect sensor. Agriculture 2023, 13, 348. [Google Scholar] [CrossRef]
  14. Karunathilake, E.; Le, A.; Heo, S.; Chung, Y.; Mansoor, S. The path to smart farming: Innovations and opportunities in precision agriculture. Agriculture 2023, 13, 1593. [Google Scholar] [CrossRef]
Table 1. The key information for each paper.
Table 1. The key information for each paper.
AuthorsObjectsModelsContributions
Ren et al. [1]Soil MoistureExtreme Gradient BoostingEstablish prediction models of soil relative humidity
Jiao et al. [2]Platycladus OrientalisGated Recurrent UnitPredict the initial flowering period
Guo et al. [3]ZanthoxylumYOLOv5Identify mature Zanthoxylum fruits
Li et al. [4]WheatA Lightweight CNNIdentify wheat growth stages
Zhang et al. [5]CowDeepOtsu
EfficientNet
Identify individual cows
Cui et al. [6]CowSOLOv2
CNN-LSTM
Divide cow production groups
Ding et al. [7]AppleRFCA ResNetIdentify apple leaf diseases
Hao et al. [8]HensFaster R-CNNMonitor the feeding behavior of hens
Lee et al. [9]Honey BeeBFMatcherMonitor bee mites and diseases
Yu et al. [10]GrainFcsNetRecognize grain pest species
He et al. [11]Rice SeedMultimodal FusionClassify rice varieties
Sun et al. [12]SoybeanRandLA-Net
BAAF-Net
Construct an annotated three-dimensional model dataset
Xu et al. [13]MaizeMulti-view Registration Algorithm
Iterative Nearest Point Algorithm
Realize early variety selection at the seedling stage
Karunathilake et al. [14]Multi ObjectsMulti ModelsA review of the latest advances in precision agriculture
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zou, X.; Liu, Z.; Zhu, X.; Zhang, W.; Qian, Y.; Li, Y. Application of Vision Technology and Artificial Intelligence in Smart Farming. Agriculture 2023, 13, 2106. https://doi.org/10.3390/agriculture13112106

AMA Style

Zou X, Liu Z, Zhu X, Zhang W, Qian Y, Li Y. Application of Vision Technology and Artificial Intelligence in Smart Farming. Agriculture. 2023; 13(11):2106. https://doi.org/10.3390/agriculture13112106

Chicago/Turabian Style

Zou, Xiuguo, Zheng Liu, Xiaochen Zhu, Wentian Zhang, Yan Qian, and Yuhua Li. 2023. "Application of Vision Technology and Artificial Intelligence in Smart Farming" Agriculture 13, no. 11: 2106. https://doi.org/10.3390/agriculture13112106

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop