Preprint
Review

A Review of CNN Applications in Smart Agriculture Using Multimodal Data

Altmetrics

Downloads

196

Views

101

Comments

0

Submitted:

12 November 2024

Posted:

13 November 2024

You are already at the latest version

Alerts
Abstract
This review explores the applications of Convolutional Neural Networks (CNNs) in smart agriculture, highlighting recent advancements across various applications including weed detection, disease detection, crop classification, water management, and yield prediction. Based on a comprehensive analysis of more than 115 recent studies, this paper contextualizes the use of CNNs within Agriculture 5.0, where technological integration optimizes agricultural efficiency. Key approaches analyzed involve image classification, image segmentation, regression, and object detection methods that use diverse data types ranging from RGB and multispectral images to radar and thermal data. By processing UAV and satellite data with CNNs, real-time and large-scale crop monitoring can be achieved, supporting advanced farm management. A comparative analysis shows how CNNs perform with respect to other techniques that involve traditional machine learning and recent deep learning models in image processing, particularly when applied to high-dimensional or temporal data. Future directions point toward integrating IoT and cloud platforms for real-time data processing and leveraging large language models for regulatory insights. Potential research advancements emphasize improving increased data accessibility and hybrid modeling to meet the agricultural demands of climate variability and food security, positioning CNNs as pivotal tools in sustainable agricultural practices. A related repository that contains the reviewed articles along their publication links is made available (https://github.com/MohammadElSakka/CNN_in_AGRI).
Keywords: 
Subject: Computer Science and Mathematics  -   Artificial Intelligence and Machine Learning

1. Introduction

Smart agriculture involves the adoption of modern technologies and data-driven solutions to optimize resource usage and to enable real-time monitoring, leading to more sustainable and efficient agricultural practices. Referred to as Agriculture 5.0 (AG5.0), this approach leverages Artificial Intelligence (AI), the Internet of Things (IoT), and renewable energy sources to drive innovation in farming practices [1,2,3].
AG5.0 includes a wide range of applications that transform farming practices. One significant application is crop monitoring where AI-driven sensors enable continuous observation of crop health and growth. Crop identification also helps in optimizing field management by distinguishing between crop types and assessing the composition of fields. Additionally, anomaly detection such as plant diseases and weeds becomes more efficient and precise than traditional solutions [1].
Traditionally, detecting anomalies in agriculture relied heavily on manual and visual inspection, where farmers assessed plant phenotypes—observable characteristics such as leaf color or blight spots—for signs of disease. This method, however, is time-consuming and labor-intensive, particularly as the global demand for agricultural products continues to rise with population growth. According to the Food and Agriculture Organization of the United Nations (FAO), "the global production of primary crop commodities reached 9.5 billion tonnes in 2021, increasing by 54 percent since 2000 and 2 percent since 2020" [4]. As agricultural demands grow, more efficient and scalable methods are needed to ensure the timely identification and management of crop anomalies. AI has become a critical tool in addressing these challenges, providing data-driven solutions that enable real-time decision-making and predictive capabilities based on plant phenotypes [5]. One of the most significant advances in this regard has been in the field of computer vision [6].
Computer vision enables computers to analyze and interpret visual data, transforming the way plant phenotypes are monitored and evaluated. Convolutional Neural Networks (CNNs) have become integral to computer vision, designed to detect and learn patterns from images [7]. Since the introduction of the groundbreaking CNN model AlexNet [8], which achieved an impressive 99% accuracy in the ImageNet classification challenge, CNNs have become essential tools for image analysis in many domains, including agriculture [9]. Recent developments have seen CNNs used as backbones for many generative AI models based on transformers [10]. In some cases, CNNs have even outperformed well-known transformers [11], which demonstrates that CNNs continue to evolve with new improvements and techniques, making them one of the most reliable computer vision algorithms for tasks such as image classification.
Similarly to other supervised machine learning models, CNNs undergo a training and testing process to learn and evaluate their performance. In the training phase, a CNN model processes a large amount of data through its layers using forward propagation, where it makes predictions [12]. These predictions are compared to true values and the model uses back-propagation to adjust its parameters in order to improve its accuracy. After training, the model is evaluated on a separate testing dataset that it has not seen before. During this phase, the model’s ability to generalize to new data is tested, as an estimation to real-world performance [13]. Transfer learning can also significantly enhance this process, by allowing CNNs that have been pretrained on large and diverse datasets to leverage this knowledge to improve their performance on datasets specific to agriculture [14,15].
The input data play a crucial role in both training and evaluating models. The quality and diversity of data influences the model’s ability to learn and generalize in real-world scenarios with high accuracy [16]. Common types of image data include color images taken by cameras which consist of three channels (red, green, and blue, i.e. RGB) that show scenes the way the human eye would see it. However, images are not limited to the visible light spectrum. They can be captured in various other wavelengths, each providing unique information. For instance, infrared images are captured in the infrared part of the spectrum, radar images use radio waves, multispectral and hyperspectral images capture a wide spectrum of wavelengths across visible and invisible light, and other images can be heatmaps that visually represent data such as temperature or moisture levels. Images can be captured by different types of instruments, such as ground-based sensors embedded in soil or attached to plants to collect in-situ data, handheld devices involving digital or multispectral cameras, aerial devices such as Unmanned Aerial Vehicles (UAVs) and satellites, or Unmanned Ground Vehicles (UGVs) [17,18,19,20]. A wide range of data types that are acquired differently exist, each offering a unique insight into various agricultural aspects. The data acquisition instruments also vary significantly, which could enhance the possibilities in agricultural monitoring and management. These advancements present notable research opportunities and have gained significant interest for the last few years. Although numerous reviews have addressed the use of AI in agriculture [21,22,23,24,25,26,27,28,29,30], there is still a lack of comprehensive discussion on data-related aspects and methodologies. This gap in the literature is significant, as effective AI-driven solutions are highly dependent on both data and methodology. To address this gap, we conduct a detailed analysis of recent articles published between 2018 and 2024.
More precisely, this review explores how CNNs are used to process various types of data in smart agriculture, while focusing on the specificity of each data type and their acquisition methods. Section 2 reviews relevant literature, including surveys and reviews on machine learning in agriculture. In Section 3 the methodology for gathering papers for this survey is described. In Section4, a brief background on state-of-the-art models and evaluation metrics is provided. In Section 5 the papers are analyzed and categorized into five fields: weed detection, plant disease detection, crop classification, water management, and yield prediction. Section 6 offers a cross-application analysis and future perspective. Finally, a conclusion is made in Section 7 and an abbreviations table is provided.

2. Related work

AI in smart agriculture is a multidisciplinary topic that is attracting a growing interest from both researchers and engineers. Several recent reviews highlight the advancements and achievements where rapid progress in machine learning made it particularly influential in smart agriculture. For example, Liakos et al. [21] categorized machine learning methods by classifying them based on the issues they addressed. Their review analyzed over 40 studies that they categorized into four areas: crop management, livestock management, soil management, and water management. They concluded that methods such as clustering, decision trees (DTs), regression, neural networks, support vector machines (SVMs), and Bayesian models, are efficient in crop monitoring tasks such as yield prediction, disease detection, weed detection, crop quality, species recognition, tasks related to water, soil, and livestock management.
Kok et al. [22] reviewed the use of SVM in agriculture across the literature. They gathered 60 research articles that used SVM in addition to other machine learning and deep learning models, and then identified which model achieved the best performance. The studies they reviewed covered six key areas of agriculture: nutrient estimation, disease detection, crop classification, yield estimation, quality classification, and weed detection. Their findings indicated that SVM generally performed worse than Random Forest (RF) in certain areas and fell short compared to deep learning methods across all fields.
Kamilaris and Prenafeta-Boldú [23] analyzed 23 studies on how deep learning is used in farming and how methods are evaluated as well as how they compared to other approaches. They also conducted an experiment which goal is to detect missing vegetables in a sugar cane field. By using CNNs, they achieved 79.2% accuracy, which the authors considered low accuracy because of mislabeled images in their dataset.
Kamilaris and Prenafeta-Boldú [24] published a review on deep learning in agriculture, including insights on image preprocessing, image augmentation, and testing. A list of 14 image datasets that are publicly available was also provided.
While some reviews cover a global overview of AI uses in smart agriculture, others focused on more specific use cases. For instance, Liu and Wang [25] presented deep learning methods to detect diseases and pests in plants. Classification methods were detailed along with their advantages and drawbacks. Their findings indicate that some methods such as deep learning, especially CNNs, performed better than others, e.g., K-means, DT, SVM, K-Nearest Neighbors (KNN). Another list of 14 image datasets of plant diseases and pests was also provided.
Saleem et al. [26] provided a list of studies that use CNNs in smart agriculture, in addition to visualization methods, e.g. segmentation maps, heatmaps, saliency maps, that are useful for analyzing decisions made by machine learning algorithms.
Kamarudin et al. [27] conducted a review on the use of deep learning in topics related to water stress such as evapotranspiration, water stress identification, soil moisture estimation, and soil water modeling. The review showed that deep learning models outperform traditional machine learning approaches in these applications. However, they also highlight that the application of deep learning in plant water stress assessment is still relatively new, and further research is needed to improve models.
Several other surveys focused on weed detection in crops using deep learning [28,29,30]. These surveys explored CNN model architectures and provided a list of publicly available datasets that are used in weed management.

3. Materials and Methods

To conduct the literature review, keywords such as "Convolutional Neural Networks", "Deep Learning," and "Agriculture" were used to identify relevant papers published in conferences and journals. The primary databases searched included Google Scholar, Web of Science, and IEEE Xplore. Articles were selected based on their relevance to the topic and the diversity of approaches they presented, ensuring a broad perspective on the application of CNNs in agriculture. The reviewed articles focused on one of five main applications in smart agriculture:
  • Weed detection: Identifying and detecting weeds to enhance removal practices.
  • Disease detection: Early identification of crop diseases through image analysis to minimize damage.
  • Crop classification: Classifying different crop types for better field management.
  • Water management: Monitoring water and moisture levels and optimizing irrigation practices.
  • Yield prediction: Using visual and environmental data to predict crop yield more accurately.
Table 1 summarizes the number of papers that we reviewed for each field. The number of collected papers for each category varies depending on the variety of methods and techniques employed across the studies.

4. Background

CNNs have emerged as a powerful tool in image processing, achieving high performance across various fields from visual data. Since that progress, CNNs have transformed the way computers interpret visual information [12]. Originally developed for feature extraction from images for tasks such as image classification or segmentation, CNNs became widely used across domains requiring high accuracy in computer vision [31,32]. Their success can be attributed to their multilayer structure, where convolutional layers capture spatial features from images [12]. CNNs are typically evaluated using standard performance metrics, such as accuracy and precision, to quantify their effectiveness in tasks such as image classification [33]. Each of these metrics provides a different perspective on model performance. For instance, precision indicates how well the model correctly predicts positive predictions. Other metrics, such as Intersection over Union (IoU) are also often applied in image segmentation and object detection to evaluate the overlap between predicted and ground truth areas [34]. Key metrics that are often used in the studied literature for evaluating CNN models are presented in Table 2.
Table 3 presents an overview of widely adopted CNN architectures such as AlexNet[8] and ResNet [35]. Various state-of-the-art architectures have different designs suited to different tasks. For example, AlexNet was one of the earliest models to achieve high accuracy on image classification, while ResNet introduced a solution to the vanishing gradient problem in deep models. CNN architectures continue to evolve with enhancements, allowing them to have better feature extraction capabilities in order to become more efficient.

5. Convolutional Neural Network Applications in Smart Agriculture

CNNs are designed for processing and analyzing visual data, especially images. Due to their efficiency in feature extraction from images, CNNs have been widely applied across various fields, consistently demonstrating high performances [31,87]. In the context of smart agriculture, CNNs have proven to be particularly valuable, aiding in decision-making processes by analyzing complex agricultural data. By being exposed to large datasets, CNNs are able to learn intricate patterns within images and generalize to new, unseen data, making them well-suited for a range of agricultural applications. Figure 1 illustrates the general pipeline used in smart agriculture, highlighting the various stages of the process. Depending on the specific agricultural needs, data is first acquired, preprocessed, and labeled. Then, an AI model is trained to address the agricultural problem. The ability of CNNs to efficiently analyze visual data has made them widely used in a range of applications in agriculture (Section 3). This section provides an analysis of these key applications.

5.1. Weed Detection

Weeds are undesirable plants in agriculture that compete with crops of interest for essential resources such as water, sunlight, or nutrients, and some types of weeds are toxic. They can also include remnants of previous year’s crops that re-emerge alongside current desired crops. If left unmanaged, they can significantly reduce crop yields and growth. Weed invasions are usually treated with herbicides, and very recently advanced techniques such as laser treatment [88]. Achieving precise herbicide control or laser treatment can involve computer vision techniques and more specifically CNNs for accurate weed detection. Weed detection is the identification and localization of weeds in agricultural fields, which can be achieved with CNNs in different approaches (Table 4).
Image segmentation and object detection as main approaches. A major challenge in weed detection is distinguishing weeds from crops, especially in complex environments where they may overlap or have similar visual features (e.g. color or spectral signatures). Two common machine learning approaches for addressing this challenge are image segmentation and object detection. In image segmentation, each pixel in the image is classified as weed or crop [47,48,78]. For example, Kamath et al. [47] showed that CNNs are efficient for weed segmentation in paddy crops, achieving a 90% weighted mean IoU with the use of the PSPNet architecture. Espejo-Garcia et al. [78] showed that using transfer learning to segment weeds from crops accelerated the training process and improved performance. Specifically, the authors trained a SegNet model on segmenting weeds from carrots and then applied the acquired knowledge to segment onions, and vice versa. Asad and Bais [48] achieved a frequency-weighted mean IoU of 98% in weed segmentation in canola fields using SegNet but noted difficulties in detecting weeds where they overlapped with crop leaves and confusions with plant stems because of high similarity between them.
On the other hand, object detection approaches aim to localize individual weeds or plants within a field using bounding boxes [36,59,76,89,90,91]. Zhang et al. [36] proposed a weed localization and identification method based on object detection, using Faster-RCNN. The authors integrated a Convolutional Block Attention Module that improves the efficiency of CNNs, achieving 99% accuracy at detecting and localizing several types of weeds and soybean seedlings. Chen et al. [76] used a local attention mechanism to effectively detect weeds in a sesame field outperforming other models, such as, Fast-RCNN, SSD, YOLOv4, with a mean average precision of 96% and real-time detection speed. Jabir et al. [59] compared the performance of YOLO, Faster-RCNN, Detectron2, and EfficientDet at detecting weed, and they found that YOLOv5 is a fast and accurate model that could be integrated to embedded systems for weed detection. To address the challenges of complex environments such as overlapping and small weeds, Wu et al. [89] improved the YOLOv4 model by modifying its backbone to include a hierarchical residual model, which improves small object detection. YOLO models are widely used in weed detection tasks due to their speed and efficiency in real-time detection, and the latest version, YOLOv11, further improves accuracy and processing speed[86,92,93,94]. Gao et al. [90] used both synthetic images and real images of weeds in sugarbeet fields to train a CNN model that is capable of weed detection under complex situations, e.g. variation in plant appear, illumination changes, foliage occlusions, and different growth stages. Generating synthetic images by cropping, zooming, flipping, and adjusting the brightness of real images, allowed the model to better generalize on real situations.
Multispectral data and vegetation indexes with RGB for enhanced weed detection. While many studies used RGB images for weed detection or segmentation, others have explored a variety of data types, such as multispectral images or vegetation indexes [40,95,96,97,98,99]. Sahin et al. [96] compared different combinations of RGB, near infrared (NIR), and the Normalized Difference Vegetation Index (NDVI)[82] channels, as inputs to a UNet for weed segmentation. They concluded that Green, NDVI, and NIR filtered with an edge-preserving Gaussian bilateral filter were the best input to their model. Moazzam et al. [97] showed that combining NIR to RGB images performs better than taking each one separately for weed segmentation.
Data acquisition methods using UAVs, UGVs, and handheld cameras. To acquire different types of data, sophisticated methods are employed, including UAVs [41,98,99,100,101,102], UGVs [39,50,103,104], and handheld cameras [105,106,107]. Ong et al. [100] compared the performance of a CNN with a classifier based on RF in weed detection using UAVs. To acquire data, the UAV was equipped with a camera of 20 megapixels resolution and a video resolution of 4K with 60 frames per second. The images were captured at two meters above ground level in JPEG format (RGB). The results showed that the CNN model outperformed the RF classifier by achieving an accuracy of 92% while being less sensitive to class imbalance in the dataset. Gallo et al. [101] demonstrated that it is possible to achieve an acceptable and realistic accuracy using high resolution RGB images captured from UAVs at 65 m above ground level. The employed YOLOv7 model outperformed other models, while also achieving real-time detection speed. Haq [102] compared several machine learning and statistical methods in weed detection using RGB UAV images taken at 4 m above ground level. They found that CNNs outperformed traditional methods like SVM, RF, DT, and AdaBoost, with an accuracy of 99%. Similarly, Osorio et al. [41] used a UAV equipped with a multispectral camera at 2 m above ground level. The images were captured in 4 spectral bands: green (500 nm), red (660 nm), red edge (745 nm), and NIR (790 nm). Osorio et al. [41] also added NDVI as a background estimator to help isolating vegetation from the background. Among the compared models, YOLOv3 and Mask-RCNN outperformed SVM, with F1-scores reaching 94%.
UGVs, or agricultural robots as referred to by some authors, were also used in weed detection [39,40,50,104]. Quan et al. [39] used a field robot to detect maize seedlings and weeds in a maize field under different weather conditions, e.g. sunny, rainy, cloudy. A camera mounted on the field robot captured RGB images of the field with different shooting angles, i.e. 0°, 30°, 75°. By using a Faster-RCNN model, the system achieved 97% precision in the detection of maize seedlings among weeds. Rasti et al. [104] captured field RGB images using a camera mounted on a UGV at 1 m above ground and found that integrating a scatter transform to a CNN could enhance model’s performances. Suh et al. [50] captured field images from a camera mounted at 1 m altitude on UGV with 0° for the angle of shooting. The study compared several CNNs with transfer learning on ImageNet and obtained the highest accuracy of 98% with an AlexNet based model. Lottes et al. [40] used a field robot that captures images in RGB and NIR. The authors achieved more than 93% in F1-scores on weed segmentation using a CNN autoencoder with a spatio-temporal fusion.
Cameras or sensors are not exclusively mounted on unmanned vehicles, but they can also be used as handheld devices. For example, many authors [91,105,106] used digital cameras to capture images in fields during different times of the day and under different weather conditions. Chen et al. [107] used mobile phones to build their image datasets for weed detection. Farooq et al. [95] used two multispectral cameras to detect several types of weeds. The first camera captures 16 bands between 460 and 630 nm, and the second captures 4 bands (green, red, red edge, NIR).

5.2. Disease Detection

Plant disease detection is the process of detecting or identifying diseases in plants. Diseases impact crop health, leading to reduced yield and low quality crops. Left untreated, crop diseases may spread quickly in the fields, which directly affects food safety and agricultural products. By analyzing images of plants, CNNs can identify symptoms of diseases, which enables targeted and fast treatments as a part of crop management (Table 5).
Image classification for effective plant disease detection. In the literature, plant disease detection is mainly solved using image classification, where models are trained to classify RGB images, into mainly two categories, e.g. healthy or unhealthy [51,56,85,108,109,110,111,112,113,114]. For instance, Thakur et al. [85] proposed a lightweight CNN architecture of 6 million parameters based on VGG and Inception that classifies plant diseases. Specifically, they employed their model on five datasets separately, with more than 100 crop diseases. The datasets consist of RGB images of several crop leaves that are either healthy or present multiple types of diseases. The model performed consistently well, reaching an accuracy of 99%. Kalbande and Patil [108] proposed a novel CNN model that uses several pooling techniques that include average pooling, max pooling, and global max pooling. Mixing pooling techniques aims to achieve a "smoothing to sharpening" approach in which average pooling and max pooling are applied to smoothen features extracted by convolutional layers, then global max pooling is applied to sharpen them. The method was applied to perform disease classification on images of diseased and healthy tomato leaves, reaching an accuracy of 95%. Panshul et al. [109] compared a CNN model to other machine learning and statistical algorithms, i.e. RF, SVM, Naive Bayes, Gradient Boosting, DT, KNN, and Multilayer Perceptron (MLP), in disease classification in potato plants. The models were trained and tested on potato leaf images. The CNN outperformed other methods by achieving 98% of accuracy. Zhong et al. [110] proposed a light CNN model suitable for embedded systems. The CNN model uses Phish modules and light residual modules which improves feature extraction while reducing the size of the model. The proposed model outperformed other CNN models such as ResNet and VGG at tomato disease classification, achieving 99% of accuracy while being lighter. Kaya and Gürsoy [111] proposed a novel deep learning method to identify plant diseases. Their approach involves applying image fusion between RGB images of plant leaves and versions of the same images with the background removed. After evaluating the method on 54000 leaf images involving 38 classes, the model achieved 98% of accuracy, outperforming state-of-the-art techniques. Furthermore, Ahad et al. [56] compared six state-of-the-art CNNs - DenseNet121, InceptionV3, MobileNetV2, ResNeXt101, ResNet152, and SEResNeXt101- in rice disease classification. Transfer learning from ImageNet could significantly improve classification accuracy, which rose up to 98% by SEResNeXt101. Similarly, Pajjuri et al. [51] compared AlexNet, GoogLeNet, VGG16, and ResNet50V2 in plant disease classification. VGG16 had the best performance, reaching an accuracy of 98%.
As image segmentation or object detection. While most studies approach plant disease detection as a classification problem, others consider it as an image segmentation problem [68,115,116,117]. For instance, Sharmila et al. [115] and Prashanth et al. [68] used a Mask-RCNN model to segment leaf images. The model successfully separated pixels that show symptoms of diseases from healthy leaves, reaching high performances. Shoaib et al. [83] used image segmentation as a preprocessing step before performing plant disease identification. The authors trained a UNet model to create a segmentation mask which effectively isolates leaves from the background. The isolated leaf images are then passed to an InceptionV1 CNN, which achieves 99% accuracy in determining whether the leaf is healthy or diseased. Kaur et al. [116] proposed a CNN model for tomato leaf disease segmentation. Their model successfully classified pixels showing plant diseases, e.g. early or late blight, achieving an accuracy of 98%. Similarly, Sharma and Sethi [117] used a CNN-based segmentation on wheat leaves to classify potential diseases.
Some authors also suggested object detection methods based on UAV imagery [42,49,53]. Liang et al. [53] proposed a CNN model based on the CenterNet architecture that detects diseases and insect pests in a forest. The method was applied on aerial images taken above a forest and it showed high accuracy and real-time speed, outperforming state-of-the-art methods. Wu et al. [49] used the YOLOv3 model with the complete IoU loss function which is optimized for object detection tasks. The authors used drone images of healthy and sick pine trees, and trained the model on detecting sick trees. The results showed an accuracy of 95%, with an average processing time of less than 0.5s. Sangaiah et al. [42] employed a YOLO implementation in rice leaf disease detection from UAV images. The proposed model is lightweight and capable of being deployed on UAVs, while also having a high performance, reaching 86% of mean average precision (mAP).
Beyond RGB images. RGB data is the most frequently used data type in plant disease detection across the literature due to the visual nature of symptoms. In addition to the visual aspect of plant disease symptoms, acquiring RGB images does not require a lot of material and is cheap. A significant portion of research thus focused on analyzing RGB leaf images. Nevertheless, recent studies are using non-visible imagery in plant disease detection, especially at an early stage, when symptoms are not visible yet [118,119,120]. Duan et al. [118] proposed a CNN model for early detection of blight using multispectral imaging. Blight in pepper leaves could be detected 36 hours before visible symptoms start to appear. The CNN model achieved an accuracy of 91%, which demonstrates the feasibility of using multispectral imaging in early disease detection. De Silva and Brown [119] compared several deep learning techniques at detecting tomato diseases using multispectral images. Combining visible and NIR wavelengths achieved the highest accuracy, reaching 93%. It also showed that Vision Transformers (ViTs) outperformed CNNs, Hybrid ViTs, and Swin Transformers at this task. Reyes-Hung et al. [120] discussed the use of object detection methods based on YOLO to classify crop stress in multispectral images of potato crops. This study highlights the importance of using non-visible light, especially NIR and red edge, to detect plant diseases.

5.3. Crop Classification

Crop classification is the identification and categorization of different crop types. Classifying crops is useful for enhancing agricultural management, e.g. crop distribution, crop rotations, land use, or policymaking. Crop classification problems are often solved using CNNs across the literature, using different approaches (Table 6).
Classifying crop images. Image classification approaches enable the classification of an entire image based on which crop it contains [123,124,125]. Gill et al. [123] proposed a CNN-RNN-LSTM model to classify field images of wheat. The proposed approach classified each image into a wheat variety class with an accuracy reaching 95%. Kaya et al. [124] compared the performances of CNNs using transfer learning and fine-tuning to classify leaf images of different crops. The results showed that transfer learning provides the best outcomes, reaching 99% classification accuracy. Lu et al. [125] proposed a six-layer CNN architecture to classify fruit images into 9 classes. Their CNN model outperformed SVM, genetic algorithms, and feedforward neural networks, reaching 91% accuracy.
Satellite-based remote sensing approaches. Crop classification is not limited to classifying leaves or fruits but it can be employed at larger scales such as satellite images [45,126,127,128,129,130]. Yao et al. [45] used Sentinel-2 time series satellite images obtained on 5 dates to detect tea plantations. The authors proposed a combined model architecture between CNNs and Recurrent Neural Networks (RNNs) and compared its performance to methods such as SVM, Random Forest, CNN, and RNN. The RCNN model was trained on image segmentation by detecting pixels that show tea plantations. Their method outperformed various methods, by achieving an IoU of 79%. Rasheed and Mahmood [126] used Sentinel-2 time-series to identify rice crops among other classes (builtup, crops, rangeland, trees, and water), without the need of in-situ data surveys. The NDVI was also adopted in addition to the multispectral inputs of Sentinel-2. A CNN approach was proposed and compared with various classical methods such as RF, SVM, classification and regression trees, Swin Transformer, HRNet, 2D-CNN, and Long Short-Term Memory (LSTM). Deep learning approaches outperformed traditional machine learning ones. Overall, the suggested approach achieved the highest performance, reaching 93% accuracy. Kou et al. [127] used Sentinel-2 multi-temporal satellite images as input features and ground labels from a survey as output features for crop classification. Their proposed CNN outperformed RF, achieving high capabilities of generalizing over temporal data. Farmonov et al. [128] used hyperspectral images acquired using a spectrometer (DESIS) mounted on the International Space Station. While DESIS images had 235 spectral bands, ranging between 400 and 1000 nm, only 29 bands were selected based on their importance. The authors proposed a method based on wavelet transforms, spectral attention and CNNs to correctly classify pixels into several agricultural crops. Zhao et al. [130] used Sentinel-1 synthetic-aperture radar (SAR) time series for early crop classification. They used time series of VH+VV-polarized backscatter data as inputs to different models that they compared. The results showed that 1D-CNN outperformed RF, LSTM, and GRU-RNN, making it effective at classifying crops at an early stage using SAR satellite imagery.
Fusion of Satellite and UAV Data. Satellite images are often combined with UAV imagery for crop classification [131,132,133]. Yin et al. [131] proposed a ViT model based on 3D convolutional attention modules for crop classification tasks using multi-temporal SAR data from UAV and satellites (RADARSAT2). The attention modules consisted of a polarization module and a temporal-spatial module to effectively learn features from temporal data and polarized data (SAR). The proposed model outperformed CNNs such as ResNet or 3DResNet, reaching 98% and 91% accuracy on UAVSAR and RADARSAT2 respectively. Li et al. [132] proposed a CNN model to segment crop parcels or objects in remote sensing time-series. The time-series consists of a combination of SAR data acquired by a UAVSAR and multi-spectral images by RapideEye. Labels were acquired by the United States Department of Agriculture that contains a wide range of data for agriculture. The authors showed that their method effectively outperformed other techniques.
Applications of UAVs in crop classification. UAV imagery is also used independently in many studies in crop classification [52,134,135]. Pandey and Jain [52] proposed an intelligent system based on CNNs and UAV imagery for crop identification and classification. An RGB camera was mounted on a UAV that flew at 100 meters above ground to capture images. The authors compared the performance accuracy of the proposed CNN with other machine learning methods such as RF, SVM, and CNNs like AlexNet, VGG, and ResNet. Galodha et al. [134] used a UAV equipped with a terrestrial hyperspectral spectroradiometer to capture high resolution images of different crops. The authors compared several CNNs with a different number of layers and kernel sizes. CNNs with 3 or 5 layers achieved almost identical accuracy (87%) when 7x7 kernels were used. Kwak et al. [135] proposed a hybrid CNN-RF model for early crop mapping using limited input data. A CNN-RF model outperform CNNs and RFs because of its ability to leverage the advantages and strengths of both architectures for feature extraction and classification.
Crop classification has significantly benefited from remote sensing data, especially multispectral, hyperspectral and radar. The integration of advanced technology, such as sensors on UAVs and satellites, provides temporal data that is analyzed using CNNs in order to enhance crop monitoring and accurately identify crops.

5.4. Water Management

Water management focuses on optimizing water usage by monitoring irrigation, moisture levels, and potential droughts. By assessing the state of water resources, effective strategies can be developed in order to use water efficiently. Due to watering circumstances, general characteristics of plants vary significantly, e.g., the color, the shape, or the curvature. By exploiting these observable changes, several CNN approaches are used to address this topic, which support sustainable agricultural practices in the long-term (Table 7).
Color and spectral data for plant water stress detection. Many studies use image classification techniques to differentiate between color images showing whether the plants suffer from water stress or not [37,72,136,137,138]. Kamarudin et al. [136] proposed a lightweight CNN based on an attention module for water stress detection. They considered images of plants that were subject to different water treatments, ranging from full irrigation to water deprivation. Their method outperformed state-of-the-art models, reaching 87% of classification accuracy. Gupta et al. [72] and Azimi et al. [137] used images of chickpeas that went through three different watering treatments. Gupta et al. [72] found that ResNet-18 achieved 86% accuracy while Azimi et al. [137] achieved 98% accuracy using a hybrid CNN-LSTM model. Hendrawan et al. [37] compared four CNN models (SqueezeNet, GoogLeNet, ResNet50, AlexNet) at identifying water stress in moss cultures. They used RGB images of moss that received different water treatments, i.e. dry, semi-dry, wet, and soak. The authors found that the ResNet50 model achieved the best accuracy, reaching 87%. Zhuang et al. [138] used images of maize plants that were watered differently. The authors used a CNN feature extractor followed by an SVM classifier to classify images into categories ranging from drought stressed to well watered. Their method achieved a balanced performance between classification time and accuracy.
Color images are advantageous at detecting water stress in plants showing observable symptoms such as change in color or curvature. However, other authors used multispectral and hyperspectral images with the goal of identifying water stress earlier and with less observable symptoms [61,139,140,141]. Kuo et al. [139] used a hyperspectral spectrometer on tomato seedlings in order to detect early drought stress with the absence of visible changes. The authors proposed a 1D-CNN based on ResNet’s residual block and Grad-CAM that achieved 96% of accuracy, outperforming other methods, while also minimizing computation and data collection costs. Spišić et al. [140] analyzed multispectral reads to detect water stress in maize canopies. SVM, 1D-CNN, and MLP achieved comparable performance, with trade-offs between performance and detection speed. Kamarudin and Ismail [61] compared several lightweight CNN models at drought stress identification in RGB and NIR plant images that went through different water treatments. Among MobileNet, MobileNetV2, NasNet mobile, and EfficientNet, EfficientNet achieved the best performance, reaching an accuracy of 88%. Zhang et al. [141] also compared different algorithms at detecting water stress in tomato plants using the visible and NIR spectrum and cloud computing. The MLP and one-vs-rest classifier outperformed 1D-CNNs at processing 1D spectral data.
Complementary data sources. Other studies used thermal and weather data for water stress detection due to its correlation with factors like temperature, humidity, soil moisture, and evapotranspiration rates [142,143,144,145]. Li et al. [142] used thermal imagery acquired with a thermal camera along with RGB images of rice leaves that had different levels of water stress. The proposed method demonstrated that using CNNs to extract features from background temperature, along with plant thermal images, improved classification accuracy. According to the authors, this is due to the importance of air temperature, which directly relates to plant temperatures. Sobayo et al. [143] used thermal imagery to estimate soil moisture. They proposed a CNN based regression model that generalized well over three farm areas, while outperforming traditional neural networks. Nagappan et al. [144] used weather data in order to estimate evapotranspiration in the aim of irrigation scheduling. The authors used data that includes wind speed, and max/min temperature as inputs, and evapotranspiration as labels. The authors demonstrated the effectiveness of a 1D-CNN to analyze 1D time series data. Afzaal et al. [145] compares several techniques for groundwater estimation using stream level, stream flow, precipitation, relative humidity, mean temperature, evapotranspiration, heat degree days, and dew point temperature. The study suggested that Artificial Neural Networks, MLP, LSTM, and CNNs were efficient at groundwater estimation, with MLP and CNN slightly outperforming other algorithms. Vegetation indexes were also used to assist in water stress identification [140,146,147,148]. For instance, Chaudhari et al. [148] showed that it is more common to have good results when using NDVI for drought prediction. Similarly, Spišić et al. [140] and Ge et al. [147] used NDVI in water deficit detection and soil moisture estimation respectively.
Satellite imagery for water stress and soil moisture monitoring. While fixed optical cameras and weather variables are widely used in the literature, satellites showed to also be beneficial [146,147,149,150,151,152]. Liu et al. [146] used a combination of Sentinel-1 radar and Sentinel-2 optical satellite data in order to retrieve soil moisture in farmland areas. The authors used dual polarization radar (VH, VV), elevation and local incidence angle, polarization decomposition features (H,A, α ) from Sentinel-1, and several vegetation indexes (NDVI, Modified Soil Adjusted Vegetation Index (MSAVI), Difference Vegetation Index (DVI) [82]) computed from Sentinel-2’s red and NIR bands, as inputs to different algorithms that were compared. The study showed that a regression CNN outperformed support vector regression (SVR) and generalized regression neural networks, and that MSAVI had the strongest correlation with soil moisture content, followed by NDVI, then DVI, due to the influence of MSAVI by both vegetation and soil. Bazzi et al. [149] used both Sentinel-1 and Sentinel-2 time series for tasks related to water management in agriculture. By using Sentinel-1 VV and VH polarization and red and NIR bands from Sentinel-2 and their derived NDVI, the authors compared different algorithms for mapping irrigated areas. They found that the CNN approach achieved 94% of accuracy, outperforming RF. Ge et al. [147] compared several algorithms at estimating soil moisture from satellite observations. Their data included radar data from SMOS and ASCAT satellites, as well as NDVI retrieved from MODIS NDVI product MYD13C1. Here, CNNs can perform better than tradition neural networks in soil moisture retrieval from temporal satellite observations. Hu et al. [150] used microwave data from the Aqua satellite for soil moisture retrieval using regression. Again, a regression CNN performed better than SVR, while also being significantly faster.
UAV approaches for soil moisture and water stress detection. UAVs were used as well in different water management applications [66,153]. For instance, Wu et al. [153] proposed a method based on UAV remote sensing and deep learning for soil moisture estimation in drip-irrigated fields. Kumar et al. [66] employed a UAV based technique using RGB images for water stress identification. The authors proposed a framework to identify different levels of water stress in a maize crop field. RGB images captured by a camera on a UAV were used to train different models. The proposed CNN outperformed models such as ResNet50, VGG19, and InceptionV3, achieving 93% of accuracy. The authors used multimodal and multitemporal UAV imaging that captures RGB, multispectral, and thermal infrared wavelengths and showed that CNN-LSTM has a higher accuracy than CNN and LSTM models.
In conclusion, water management in agriculture encompasses a wide range of applications, especially, drought and water stress detection, irrigation mapping, and soil moisture estimation. Studies consistently showed that it is possible to efficiently accomplish effective water management using CNN methods with different data types acquired from a variety of sources at different temporal and spatial resolutions.

5.5. Yield Prediction

Yield prediction refers to the process of forecasting the quantity of crops that will be harvested from a field at the end of its growing season. It is an important aspect in agriculture that helps farmers in planning and making decisions regarding resources, supplies, and market strategies. With the emergence of AI and advanced technologies in computer vision, research studies focused on finding methods to correlate past yield values with image data (Table 8).
Advances in crop yield prediction through multimodal data through regression. Predicting or estimating crop yield is often a regression problem because of the scalar nature of the predicted values [154,155,156,157,158,159]. Mia et al. [154] compared different setups for yield prediction. The authors used CNN based methods with UAV multispectral imagery and a combination of monthly, weekly, or no weather data. The best results were obtained with weekly weather data that included precipitation, global solar radiation, temperature, average relative humidity, average wind speed, and vapor pressure data. Tanabe et al. [157] used UAV multispectral imagery for winter wheat yield prediction and showed that CNN models outperform conventional regression algorithms such as linear regression. They also demonstrated that using multitemporal data of different growth stages may not improve the prediction accuracy if the CNN is effectively implemented. The heading stage of growth was sufficient for accurate predictions. Morales et al. [155] also used regression CNNs for winter wheat yield prediction. The authors used remote sensing data that included nitrogen rate, precipitation, slope, elevation, topographic position index, terrain aspect, and Sentinel-1 backscatter coefficients and showed that the proposed CNN method outperformed other techniques such as Bayesian multiple linear regression, standard multiple linear regression, RF, feedforward networks with AdaBoost, and a stacked autoencoder. Terliksiz and Altilar [156] used deep learning to extract features from MODIS multispectral data and land surface temperature time series. The proposed model concatenated features extracted by a CNN branch for multispectral and temperature data, and an LSTM branch for past yield data. This method is simple and efficient at crop yield prediction using multimodal and multitemporal data. Zhou et al. [158] used a CNN-LSTM model for rice yield prediction using time series that include MODIS remote sensing that and several vegetation indexes (Enhanced Vegetation Index (EVI), Soil Adjusted Vegetation Index (SAVI) [82,160]), Gross Primary Productivity, temperature data, spatial heterogeneity, and historical yield data. The proposed model consisted of a CNN block for spatial feature extraction followed by an LSTM block for temporal feature extraction. The proposed method outperformed other CNN and LSTM methods. In another hybrid approach, Saini et al. [159] proposed a CNN-LSTM method for yield prediction. The proposed approach includes a CNN block that extracts relevant spatial features, followed by a Bidirectional LSTM for phonological information. The proposed method outperformed other similar research studies.
Object detection and image segmentation for image-based yield prediction. While regression problems were suitable for crop yield prediction, some studies had different approaches. For instance, object detection methods were used to detect individual crops or fruits in RGB images. These approaches allow to estimate yields by counting fruits and identifying their maturity level [161,162]. CNN models were also trained on detecting crop heads and fruits, i.e. wheat head, cotton bolls, apples, and therefore estimating crop yield [163,164,165]. Maji et al. [43] proposed a combined approach of object detection and image segmentation in yield prediction. The proposed method predicts wheat yield by detecting wheat spikes using bounding boxes in the first place, then classifying wheat pixels in the second step. They reported a mAP of 97%, overcoming difficult conditions such as overlapping and background interference.
Using semantic segmentation, Ilyas and Kim [46] proposed a CNN architecture for strawberry yield prediction. Their method classified strawberry pixels into 3 maturity classes, enabling the yield prediction of each class. Their approach outperformed different DeepLab architectures, scoring a mean IoU of 80%. Yang et al. [38] showed that it is possible to estimate the yield of corn plants based on their growth stage. The authors suggested a CNN based method that classifies hyperspectral images of corn plants in the field into five growth stage. Classification accuracy reached 75% when combining color and spectral information.
Aerial crop yield prediction using UAVs. To efficiently predict yield in crop fields, having a complete view of the field is beneficial in many studies and using UAVs is very common in this domain, as they provide an aerial view of plants [38,57,79,154,157,162,163,166]. Bhadra et al. [57] proposed an end-to-end 3D-CNN that uses multi-temporal UAV based color images for soybean yield prediction. They demonstrated that 3D DenseNet outperformed 3D VGG and 3D ResNet. Increasing the spatio-temporal resolution did not necessarily improve model performance; instead, it added more model complexity. Yu et al. [166] compared CNNs with several machine learning algorithms in maize biomass estimation using drone images. A UAV equipped with both a digital camera and a multispectral camera was used to acquire data from the field. CNNs outperformed traditional models and combining multispectral with RGB data gave the best results. Li et al. [79] used UAV imaging and CNNs to estimate cotton yield from a low altitude (5m). The authors used image segmentation using a SegNet model to classify cotton boll pixels, then they applied linear regression on the segmentation result in order to obtain the yield. The proposed model outperformed SVM and RF.
Large-scale crop yield prediction with satellite data. Satellite imagery has also been used in crop yield prediction studies, offering a broad perspective of the fields over long periods [155,156,158,167,168,169,170,171,172,173,174,175,176,177]. For example, Fernandez-Beltran et al. [170] used monthly Sentinel-2 multispectral images with climate and soil data to estimate rice yields. The authors proposed a 3D-CNN that can extract temporal, spatial, and multispectral features from images. The proposed method proved to be effective at yield estimation, while outperforming state-of-the-art 2D and 3D-CNNs. Qiao et al. [172] also proposed a 3D-CNN for crop yield prediction using satellite time series. They used a multispectral dataset (MOD09A1) and a thermal dataset (MYD11A2) from MODIS satellite; the proposed method outperformed competitive methods such as LSTM, SVM, RF, DT, or 2D-CNNs. 3D-CNNs are often used to process temporal images, especially in satellite remote sensing, where data is acquired periodically and historical records are available [169,175,178]. In different approaches, hybrid CNN models were used for yield prediction from satellite images. Hybrid models include a combination of CNN and LSTM modules, making them efficient at processing spatial and temporal data [171,173,176]. Other studies compared different algorithms and models at yield prediction using satellite data. For instance, Huber et al. [167] compared XGBoost, CNN, and CNN-LSTM in yield prediction. The study was conducted using time series from MODIS multispectral (MOD09A1) and thermal (MOD11A2) data, along with meteorogical variables (precipitation and vapor pressure). The results showed that XGBoost can be efficient at yield prediction, while outperforming state-of-the-art deep learning methods. Kang et al. [174] compared Lasso, SVR, RF, XGBoost, LSTM, and CNN at yield at maize yield prediction using remote sensing time series. XGBoost outperforms other algorithms, especially LSTM and CNN when datasets involve small feature space.

6. Cross-Application Discussion

6.1. Data Acquisition

Data acquisition plays a critical role in the successful implementation of AI solutions across various agricultural applications. Different methods are employed to capture data depending on the specific task, whether it involves weed detection, disease detection, crop classification, water management, or yield prediction. Each method meets the needs of the specific application in which it was used. For instance, digital cameras are suitable for tasks requiring high resolution imagery at a close range for visual monitoring. They are mainly used in disease detection and weed management, where capturing clear high-resolution images allows the identification of specific symptoms of plant diseases or the presence of invasive weeds. Also, digital cameras are often the low-cost and easy to use and deploy, making them accessible for all farmers.
In contrast, UGVs are an innovative but underused tool that can navigate fields with minimal human intervention. They can be equipped with different types of advanced sensors and cameras that allow them to monitor fields. They have been particularly useful at detecting weeds because of their ability to capture close-up images of crops and their surroundings, allowing for very precise crop monitoring. However, their use in the literature has been relatively limited compared to UAVs, likely due to their higher operational and maintenance costs, and the potential for causing damage to plants or soil.
UAVs have become highly effective for agricultural monitoring by offering a flexible way to collect aerial data over large areas. They can capture high resolution aerial images, while giving farmers access to real-time data. Like UGVs, UAVs can be equipped with various sensors and cameras (e.g. thermal, multispectral) to gather detailed insights particularly useful in many fields of agriculture, especially disease detection, weed management, crop classification, and yield prediction.
Acquiring weather variables and in-situ measurements is essential for agricultural monitoring in farms and fields. By collecting data from weather stations, remote sensing and IoT devices, users can obtain real-time information about environmental conditions. This type of data can be used to assist decision-making in smart agriculture and enhance productivity.
Satellites provide a broad view of the landscape, making them essential for large-scale monitoring. Satellite data is mainly used for yield prediction, crop classification, and water management due to its wide coverage and ability to capture changes over time. Moreover, satellites provide historical and periodic data, making them valuable for long-term analysis. With advancements in space technology, satellite imagery is improving in resolution and benefiting from shorter revisit times, allowing more accurate and timely insights. As a result, the integration of satellite data with modern agriculture technology enhances decision-making and optimizes practices by leveraging past data to predict future outcomes and assess current conditions across large areas.
Figure 2 illustrates how each data acquisition method is applied based on the specific requirements of different agricultural fields. For instance, some fields like water management and yield prediction require complementary data sources (e.g. weather data and satellites), while others, such as disease detection, can typically be addressed using a single method of data acquisition (e.g. digital cameras).

6.2. Data Types

Different data acquisition technologies capture different data types that are suited to specific applications in agriculture.
RGB images capture standard color images, which are widely used for disease detection and weed management. Clear visual features provided in color images allow the detection of visible symptoms on crops, such as discoloration or unusual growth patterns, making them suitable for plant disease detection. Additionally, different shapes can be identified in color images, enabling to differentiate between plants and weeds, as well as classifying various types of crops. Moreover, the ability to detect objects in images helps in counting plants or fruits, thereby predicting crop yield.
In contrast, multispectral and hyperspectral images extend beyond the visible light spectrum. Their ability to capture multiple wavelengths in multispectral images or a continuous spectrum of hundreds of bands in hyperspectral images makes them valuable for applications where no visible features are present. This is particularly efficient in early plant disease detection or water stress. Healthier plants also lead to higher yields, so evaluating plant health also helps in predicting yield. Since different crops have distinct spectral signatures, multispectral and hyperspectral images are also effective for crop classification and weed detection. Additionally, they are used to compute vegetation indexes which provide more insights on crops.
Vegetation indexes are important information in agriculture that are calculated from multispectral imagery, mainly to assess plant health, biomass, chlorophyll content, or nutrient deficiencies. Indexes like NDVI and SAVI are essential for estimating the amount of vegetation in a given area, making them valuable for distinguishing vegetation from other objects. This ability provides additional knowledge that help in weed detection, disease detection, yield prediction, and crop classification. Moreover, indexes related to water, such as the Normalized Difference Water Index [179], are useful for managing water content and soil moisture in agriculture.
SAR images offer unique advantages for water management and yield prediction. SAR wavelengths can penetrate clouds and provide consistent data under any weather conditions, making them ideal for long-term and periodic monitoring, which is beneficial for predicting yields. Additionally, SAR is effective in measuring surface roughness and capturing information about soil moisture and plant structure, which also supports water related tasks and crop classification. This ability makes SAR a valuable data type, especially when combined with optical data.
Weather and in-situ data is features used in agriculture that provide information on environmental conditions that directly influence crops. This data is particularly important for assessing water stress and predicting crop yield, as they are both heavily affected by factors like temperature, precipitation, and humidity. Other important in-situ variables that are measured in the field include pH, soil salinity, and soil moisture, all of which are significant in agriculture. By leveraging weather and in-situ data, farmers can optimize their production by anticipating climate variability and natural disasters.
Thermal data also plays a crucial role in agriculture. In particular, surface temperature levels and variations over time provide valuable information to predict extreme conditions, which directly affect agricultural yields. Additionally, heatwaves can lead to droughts, therefore, monitoring thermal data can help in planning irrigation strategies.
Figure 3 illustrates how each data type is used in different fields of agriculture. Some fields, such as disease detection and weed management, mainly do not require high-dimensional data and can be effectively addressed using simple RGB images. In contrast, other fields, such as water management, yield prediction, and crop classification, rely more on multimodal approaches to find solutions (e.g. hyperspectral and in-situ or weather data).

6.3. CNN Relevance in Smart Agriculture

CNNs play an important role in smart agriculture by analyzing visual data that are found in various agricultural applications. These applications range from weed detection and disease detection to crop classification, water management, and yield prediction. CNNs solve agricultural challenges in different ways due to their capability of processing diverse data types collected from various instruments. The choice of CNN architecture and algorithm, however, is important to achieve high accuracy and efficiency.
For tasks using RGB images, simpler CNN architectures are commonly used due to their ability to extract color based visual features. More specifically, architectures such as AlexNet, as well as deeper models like VGG, or ResNet are efficient at feature extraction. This is especially true since versions of these models with pretrained weights on large datasets are available, specifically designed for 3-channel inputs (RGB). For real-time object detection, e.g. on UAV systems, YOLO architectures are designed to be fast and accurate models, while Mask-RCNN and UNet are effective for pixel-level segmentation. Together, these approaches allow accurate agricultural monitoring using RGB images.
In contrast, for more complex data such as multispectral and hyperspectral images, which consist of multiple wavelength bands, advanced architectures are more efficient to process both the spatial and spectral information simultaneously. In cases where time is an additional dimension, e.g. when collecting satellite data over growth season, spatio-temporal data processing becomes essential. While spatio-temporal data mainly comes from satellites, it can also come from UAVs or other sensors which offer temporal observations. This category of data also includes any data type, for instance, SAR or thermal. 3D CNN models are effective for handling both the spatial and spectral dimensions because they extend the traditional 2D CNN structure to process volumetric data. Moreover, hybrid models such as CNN-LSTM or CNN-RNN are also well suited to extract temporal features from time series. These models combine CNNs for spatial feature extraction with LSTM or RNN layers to capture temporal dependencies, which improves predictions when the data involves changes over time. ViTs are particularly useful in tasks where long range dependencies in both space and time need to be captured. Unlike traditional CNNs, which rely on convolutions to extract local features, ViTs use a self-attention mechanism that allows them to determine the importance of each data feature relative to other features in that data [180]. When multimodal data combines images with tabular and scalar data, e.g. weather or in-situ, additional approaches such as feature fusion are required.
Finally, comparisons made across the literature between CNNs, traditional machine learning models, and hybrid models have shown significant differences in performances across various applications in agriculture. In fact, CNNs consistently outperform traditional machine learning techniques such as SVM, DT, and RF, particularly in tasks involving image classification and object detection. However, in more complex tasks that include several data modalities, hybrid models such as CNN-RF and CNN-LSTM, or ViTs have outperformed simple CNN models. Similarly, traditional and statistical methods tend to be a more suitable choice when only small datasets are available, which is a common limitation of CNNs in general, especially that in agriculture gathering sufficient data can be an intensive process.

6.4. Potential and Future Directions

Extensive research has been made on smart agriculture in the recent years and many studies have achieved high performances in various tasks. One of the future directions is to complete the gap between theoretical research and real-world applications, ensuring that smart solutions become available for everyday use in different agricultural settings.
Real-time performance in agriculture has been achieved in the literature. The future potential lies in enhancing the scalability of intelligent IoT systems in agriculture. In particular, further expanding the integration of IoT with cloud platforms for real-time data processing. Such a development is made possible by the improvements in sensor technology and GPU hardware.
Advancements in satellite technology offer shorter revisit periods and a higher spatial and spectral resolution. Improved spatio-temporal-spectral analysis opens up new opportunities for further advancement in remote sensing and smart agriculture. Additionally, the increasing accessibility of satellite data to a wider audience provides users (i.e. researchers, farmers, organizations) with valuable data for making informed decisions. As more users use this data, the potential for the application of satellite technology in smart agriculture continues to grow.
Large Language Models (LLMs) have also gained significant popularity due to their widespread use for text generation and their ability to process big amounts of data efficiently. Future research could focus on integrating text modalities to smart agriculture, enabling AI systems to deliver more comprehensive insight that would be context aware. For example, this integration would help users make informed decisions while taking into account different regulations and laws, especially with the recent advancements of LLMs in the legal domain [181] and the fusion of text with images[182].
Furthermore, AI can assist in decision-making during the farming process, even at its earliest stages. Potential directions in smart agriculture could also focus on determining the probability and potentiality of agricultural lands before planting begins, leading to more strategic farming decisions.
Finally, Augmented Reality and Virtual Reality solutions could become critical tools in the future, especially with the ongoing progress in these fields. By providing visually enhanced data on farmlands, Augmented Reality can offer users a better understanding and management of their crops. Additionally, simulated environments, created through Virtual Reality can be crucial for experimenting various farming practices without consequences on real crops.

7. Conclusions

This paper has provided a comprehensive review of the various applications of CNNs in smart agriculture. The literature consistently proves the effectiveness of these models across a wide range of agricultural tasks, including weed management, crop disease detection, crop classification, water management, and yield prediction. CNNs can efficiently process and extract meaningful features from diverse data sources, such as RGB or multispectral images, as well as SAR data, weather variables, and in-situ measurements. CNN-based approaches demonstrated to outperform other machine learning techniques such as SVM, DT, and RF, especially in tasks involving image classification and object detection. However, the success of simple CNN models may become limited when considering specific requirements. For instance, hybrid models showed promising results, especially CNNs combined with LSTMs and RNNs for sequential data such as time series, or 3D CNNs for volumetric data that includes spectral or temporal information.

Funding

The work was funded by the European Union. The AI4AGRI project entitled “Romanian Excellence Center on Artificial Intelligence on Earth Observation Data for Agriculture” received funding from the European Union’s Horizon Europe research and innovation programme under the grant agreement no. 101079136. The Défi Région Occitanie “Observation de la Terre et Territoire en Transition” also supported this work.Preprints 139326 i001

Abbreviations

The following abbreviations are used in this manuscript:
AG5.0 Agriculture 5.0
AI Artificial Intelligence
CNN Convolutional Neural Network
DT Decision Tree
DVI Difference Vegetation Index
EVI Enhanced Vegetation Index
IoT Internet of Things
IoU Intersection over Union
KNN K-Nearest Neighbors
LLM Large Language Models
LSTM Long Short-Term Memory
MLP Multilayer Perceptron
MSAVI Modified Soil Adjusted Vegetation Index
NDVI Normalized Difference Vegetation Index
NIR Near Infrared
RF Random Forest
RGB Red Green Blue
RNN Recurrent Neural Networks
SAR Synthetic-aperture radar
SVM Support Vector Machine
SVR Support Vector Regression
UAV Unmanned aerial vehicle
UGV Unmanned ground vehicle
ViT Vision Transformers

Conflicts of Interest

The authors declare that they have no financial or non-financial competing interest. Funded by the European Union. Views and opinions expressed are however those of the author(s) only and do not necessarily reflect those of the European Union. Neither the European Union nor the granting authority can be held responsible for them.

References

  1. Ivanovici, M.; Olteanu, G.; Florea, C.; Coliban, R.M.; Ștefan, M.; Marandskiy, K. Digital Transformation in Agriculture. In Digital Transformation: Exploring the Impact of Digital Transformation on Organizational Processes; Springer, 2024; pp. 157–191.
  2. Ragazou, K.; Garefalakis, A.; Zafeiriou, E.; Passas, I. Agriculture 5.0: A New Strategic Management Mode for a Cut Cost and an Energy Efficient Agriculture Sector. Energies 2022, 15. [Google Scholar] [CrossRef]
  3. Latief Ahmad, F.N. Agriculture 5.0: Artificial Intelligence, IoT and Machine Learning; CRC Press, 2021.
  4. FAO. Agricultural production statistics 2000–2020, 2022.
  5. Lee, U.; Chang, S.; Putra, G.A.; Kim, H.; Kim, D.H. An automated, high-throughput plant phenotyping system using machine learning-based plant segmentation and image analysis. PloS one 2018, 13, e0196615. [Google Scholar] [CrossRef] [PubMed]
  6. Chai, J.; Zeng, H.; Li, A.; Ngai, E.W. Deep learning in computer vision: A critical review of emerging techniques and application scenarios. Machine Learning with Applications 2021, 6, 100134. [Google Scholar] [CrossRef]
  7. Khan, S.; Rahmani, H.; Shah, S.A.A.; Bennamoun, M.; Medioni, G.; Dickinson, S. A guide to convolutional neural networks for computer vision; Springer, 2018.
  8. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Advances in neural information processing systems 2012, 25. [Google Scholar] [CrossRef]
  9. El Sakka, M.; Mothe, J.; Ivanovici, M. Images and CNN applications in smart agriculture. European Journal of Remote Sensing 2024, 57, 2352386. [Google Scholar] [CrossRef]
  10. Sun, G.; Yang, W.; Ma, L. BCAV: a generative ai author verification model based on the integration of Bert and CNN. Working Notes of CLEF 2024. [Google Scholar]
  11. Liu, Z.; Mao, H.; Wu, C.Y.; Feichtenhofer, C.; Darrell, T.; Xie, S. A convnet for the 2020s. Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2022, pp. 11976–11986.
  12. Gu, J.; Wang, Z.; Kuen, J.; Ma, L.; Shahroudy, A.; Shuai, B.; Liu, T.; Wang, X.; Wang, G.; Cai, J.; others. Recent advances in convolutional neural networks. Pattern recognition 2018, 77, 354–377. [Google Scholar] [CrossRef]
  13. Hossain, M.A.; Sajib, M.S.A. Classification of image using convolutional neural network (CNN). Global Journal of Computer Science and Technology 2019, 19, 13–14. [Google Scholar] [CrossRef]
  14. Niu, S.; Liu, Y.; Wang, J.; Song, H. A decade survey of transfer learning (2010–2020). IEEE Transactions on Artificial Intelligence 2020, 1, 151–166. [Google Scholar] [CrossRef]
  15. Ma, Y.; Chen, S.; Ermon, S.; Lobell, D.B. Transfer learning in environmental remote sensing. Remote Sensing of Environment 2024, 301, 113924. [Google Scholar] [CrossRef]
  16. Wujek, B.; Hall, P.; Günes, F. Best practices for machine learning applications. SAS Institute Inc 2016, p. 3.
  17. D’Aniello, M.; Zampella, M.; Dosi, A.; Rownok, A.; Delli Veneri, M.; Ettari, A.; Cavuoti, S.; Sannino, L.; Brescia, M.; Donadio, C.; Longo, G. RiverZoo: A Machine Learning Framework for Terrestrial and Extraterrestrial Drainage Networks Classification Using Clustering Techniques and Fuzzy Reasoning. 2024. [CrossRef]
  18. Adams, S.; Friedland, C.; Levitan, M. Unmanned aerial vehicle data acquisition for damage assessment in hurricane events. Proceedings of the 8th international workshop on remote sensing for disaster management, Tokyo, Japan, 2010, Vol. 30.
  19. Ouchra, H.; Belangour, A. Satellite image classification methods and techniques: A survey. 2021 IEEE International Conference on Imaging Systems and Techniques (IST). IEEE, 2021, pp. 1–6.
  20. Awad, M.M. A New Winter Wheat Crop Segmentation Method Based on a New Fast-UNet Model and Multi-Temporal Sentinel-2 Images. Agronomy 2024, 14, 2337. [Google Scholar] [CrossRef]
  21. Liakos, K.G.; Busato, P.; Moshou, D.; Pearson, S.; Bochtis, D. Machine learning in agriculture: A review. Sensors 2018, 18, 2674. [Google Scholar] [CrossRef]
  22. Kok, Z.H.; Shariff, A.R.M.; Alfatni, M.S.M.; Khairunniza-Bejo, S. Support vector machine in precision agriculture: a review. Computers and Electronics in Agriculture 2021, 191, 106546. [Google Scholar] [CrossRef]
  23. Kamilaris, A.; Prenafeta-Boldú, F.X. A review of the use of convolutional neural networks in agriculture. The Journal of Agricultural Science 2018, 156, 312–322. [Google Scholar] [CrossRef]
  24. Kamilaris, A.; Prenafeta-Boldú, F.X. Deep learning in agriculture: A survey. Computers and electronics in agriculture 2018, 147, 70–90. [Google Scholar] [CrossRef]
  25. Liu, J.; Wang, X. Plant diseases and pests detection based on deep learning: a review. Plant Methods 2021, 17, 1–18. [Google Scholar] [CrossRef]
  26. Saleem, M.H.; Potgieter, J.; Arif, K.M. Plant disease detection and classification by deep learning. Plants 2019, 8, 468. [Google Scholar] [CrossRef] [PubMed]
  27. Kamarudin, M.H.; Ismail, Z.H.; Saidi, N.B. Deep learning sensor fusion in plant water stress assessment: A comprehensive review. Applied Sciences 2021, 11, 1403. [Google Scholar] [CrossRef]
  28. Hasan, A.M.; Sohel, F.; Diepeveen, D.; Laga, H.; Jones, M.G. A survey of deep learning techniques for weed detection from images. Computers and electronics in agriculture 2021, 184, 106067. [Google Scholar] [CrossRef]
  29. Wu, Z.; Chen, Y.; Zhao, B.; Kang, X.; Ding, Y. Review of weed detection methods based on computer vision. Sensors 2021, 21, 3647. [Google Scholar] [CrossRef]
  30. Hu, K.; Wang, Z.; Coleman, G.; Bender, A.; Yao, T.; Zeng, S.; Song, D.; Schumann, A.; Walsh, M. Deep Learning Techniques for In-Crop Weed Identification: A Review. arXiv 2021. arXiv preprint arXiv:2103.14872. [CrossRef]
  31. Zhao, X.; Wang, L.; Zhang, Y.; Han, X.; Deveci, M.; Parmar, M. A review of convolutional neural networks in computer vision. Artificial Intelligence Review 2024, 57, 99. [Google Scholar] [CrossRef]
  32. Krichen, M. Convolutional neural networks: A survey. Computers 2023, 12, 151. [Google Scholar] [CrossRef]
  33. Naidu, G.; Zuva, T.; Sibanda, E.M. A review of evaluation metrics in machine learning algorithms. Computer Science On-line Conference. Springer, 2023, pp. 15–25.
  34. Rainio, O.; Teuho, J.; Klén, R. Evaluation metrics and statistical tests for machine learning. Scientific Reports 2024, 14, 6086. [Google Scholar] [CrossRef] [PubMed]
  35. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 770–778.
  36. Zhang, X.; Cui, J.; Liu, H.; Han, Y.; Ai, H.; Dong, C.; Zhang, J.; Chu, Y. Weed identification in soybean seedling stage based on optimized Faster R-CNN algorithm. Agriculture 2023, 13, 175. [Google Scholar] [CrossRef]
  37. Hendrawan, Y.; Damayanti, R.; Al Riza, D.F.; Hermanto, M.B. Classification of water stress in cultured Sunagoke moss using deep learning. Telkomnika (Telecommunication Computing Electronics and Control) 2021, 19, 1594–1604. [Google Scholar] [CrossRef]
  38. Yang, W.; Nigon, T.; Hao, Z.; Paiao, G.D.; Fernández, F.G.; Mulla, D.; Yang, C. Estimation of corn yield based on hyperspectral imagery and convolutional neural network. Computers and Electronics in Agriculture 2021, 184, 106092. [Google Scholar] [CrossRef]
  39. Quan, L.; Feng, H.; Lv, Y.; Wang, Q.; Zhang, C.; Liu, J.; Yuan, Z. Maize seedling detection under different growth stages and complex field environments based on an improved Faster R–CNN. Biosystems Engineering 2019, 184, 1–23. [Google Scholar] [CrossRef]
  40. Lottes, P.; Behley, J.; Milioto, A.; Stachniss, C. Fully convolutional networks with sequential information for robust crop and weed detection in precision farming. IEEE Robotics and Automation Letters 2018, 3, 2870–2877. [Google Scholar] [CrossRef]
  41. Osorio, K.; Puerto, A.; Pedraza, C.; Jamaica, D.; Rodríguez, L. A deep learning approach for weed detection in lettuce crops using multispectral images. AgriEngineering 2020, 2, 471–488. [Google Scholar] [CrossRef]
  42. Sangaiah, A.K.; Yu, F.N.; Lin, Y.B.; Shen, W.C.; Sharma, A. UAV T-YOLO-Rice: An Enhanced Tiny Yolo Networks for Rice Leaves Diseases Detection in Paddy Agronomy. IEEE Transactions on Network Science and Engineering 2024. [Google Scholar] [CrossRef]
  43. Maji, A.K.; Marwaha, S.; Kumar, S.; Arora, A.; Chinnusamy, V.; Islam, S. SlypNet: Spikelet-based yield prediction of wheat using advanced plant phenotyping and computer vision techniques. Frontiers in plant science 2022, 13, 889853. [Google Scholar] [CrossRef] [PubMed]
  44. Costa, L.d.F. Further generalizations of the Jaccard index. arXiv preprint arXiv:2110.09619 2021. [CrossRef]
  45. Yao, Z.; Zhu, X.; Zeng, Y.; Qiu, X. Extracting Tea Plantations from Multitemporal Sentinel-2 Images Based on Deep Learning Networks. Agriculture 2022, 13, 10. [Google Scholar] [CrossRef]
  46. Ilyas, T.; Kim, H. A deep learning based approach for strawberry yield prediction via semantic graphics. 2021 21st International Conference on Control, Automation and Systems (ICCAS). IEEE, 2021, pp. 1835–1841.
  47. Kamath, R.; Balachandra, M.; Vardhan, A.; Maheshwari, U. Classification of paddy crop and weeds using semantic segmentation. Cogent engineering 2022, 9, 2018791. [Google Scholar] [CrossRef]
  48. Asad, M.H.; Bais, A. Weed detection in canola fields using maximum likelihood classification and deep convolutional neural network. Information Processing in Agriculture 2020, 7, 535–545. [Google Scholar] [CrossRef]
  49. Wu, Y.; Yang, H.; Mao, Y. Detection of the Pine Wilt Disease Using a Joint Deep Object Detection Model Based on Drone Remote Sensing Data. Forests 2024, 15, 869. [Google Scholar] [CrossRef]
  50. Suh, H.K.; Ijsselmuiden, J.; Hofstee, J.W.; van Henten, E.J. Transfer learning for the classification of sugar beet and volunteer potato under field conditions. Biosystems engineering 2018, 174, 50–65. [Google Scholar] [CrossRef]
  51. Pajjuri, N.; Kumar, U.; Thottolil, R. Comparative evaluation of the convolutional neural network based transfer learning models for classification of plant disease. 2022 IEEE International Conference on Electronics, Computing and Communication Technologies (CONECCT). IEEE, 2022, pp. 1–6.
  52. Pandey, A.; Jain, K. An intelligent system for crop identification and classification from UAV images using conjugated dense convolutional neural network. Computers and Electronics in Agriculture 2022, 192, 106543. [Google Scholar] [CrossRef]
  53. Liang, D.; Liu, W.; Zhao, L.; Zong, S.; Luo, Y. An improved convolutional neural network for plant disease detection using unmanned aerial vehicle images. Nature Environment and Pollution Technology 2022, 21, 899–908. [Google Scholar] [CrossRef]
  54. Duan, K.; Bai, S.; Xie, L.; Qi, H.; Huang, Q.; Tian, Q. Centernet: Keypoint triplets for object detection. Proceedings of the IEEE/CVF international conference on computer vision, 2019, pp. 6569–6578.
  55. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. Proceedings of the IEEE conference on computer vision and pattern recognition, 2017, pp. 4700–4708.
  56. Ahad, M.T.; Li, Y.; Song, B.; Bhuiyan, T. Comparison of CNN-based deep learning architectures for rice diseases classification. Artificial Intelligence in Agriculture 2023, 9, 22–35. [Google Scholar] [CrossRef]
  57. Bhadra, S.; Sagan, V.; Skobalski, J.; Grignola, F.; Sarkar, S.; Vilbig, J. End-to-end 3D CNN for plot-scale soybean yield prediction using multitemporal UAV-based RGB images. Precision Agriculture 2024, 25, 834–864. [Google Scholar] [CrossRef]
  58. Wu, Y.; Kirillov, A.; Massa, F.; Lo, W.Y.; Girshick, R. Detectron2. https://github.com/facebookresearch/detectron2, 2019.
  59. Jabir, B.; Falih, N.; Rahmani, K. Accuracy and efficiency comparison of object detection open-source models. International Journal of Online & Biomedical Engineering 2021, 17. [Google Scholar]
  60. Tan, M.; Pang, R.; Le, Q.V. Efficientdet: Scalable and efficient object detection. Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2020, pp. 10781–10790.
  61. Kamarudin, M.; Ismail, Z.H. Lightweight deep CNN models for identifying drought stressed plant. IOP Conference Series: Earth and Environmental Science. IOP Publishing, 2022, Vol. 1091, p. 012043. [CrossRef]
  62. Tan, M.; Le, Q. Efficientnet: Rethinking model scaling for convolutional neural networks. International conference on machine learning. PMLR, 2019, pp. 6105–6114.
  63. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. IEEE transactions on pattern analysis and machine intelligence 2016, 39, 1137–1149. [Google Scholar] [CrossRef]
  64. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. Proceedings of the IEEE conference on computer vision and pattern recognition, 2015, pp. 1–9.
  65. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the inception architecture for computer vision. Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 2818–2826.
  66. Kumar, A.; Shreeshan, S.; Tejasri, N.; Rajalakshmi, P.; Guo, W.; Naik, B.; Marathi, B.; Desai, U. Identification of water-stressed area in maize crop using uav based remote sensing. 2020 IEEE India geoscience and remote sensing symposium (InGARSS). IEEE, 2020, pp. 146–149.
  67. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask r-cnn. Proceedings of the IEEE international conference on computer vision, 2017, pp. 2961–2969.
  68. Prashanth, K.; Harsha, J.S.; Kumar, S.A.; Srilekha, J. Towards Accurate Disease Segmentation in Plant Images: A Comprehensive Dataset Creation and Network Evaluation. Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, 2024, pp. 7086–7094.
  69. Howard, A.G. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv preprint arXiv:1704.04861 2017. [CrossRef]
  70. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.C. Mobilenetv2: Inverted residuals and linear bottlenecks. Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 4510–4520.
  71. Zhao, H.; Shi, J.; Qi, X.; Wang, X.; Jia, J. Pyramid scene parsing network. Proceedings of the IEEE conference on computer vision and pattern recognition, 2017, pp. 2881–2890.
  72. Gupta, E.; Azimi, S.; Gandhi, T.K. Characterizing Water Deficiency induced stress in Plants using Gabor filter based CNN. 2022 IEEE IAS Global Conference on Emerging Technologies (GlobConET). IEEE, 2022, pp. 91–95.
  73. Xie, S.; Girshick, R.; Dollár, P.; Tu, Z.; He, K. Aggregated residual transformations for deep neural networks. Proceedings of the IEEE conference on computer vision and pattern recognition, 2017, pp. 1492–1500.
  74. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 7132–7141.
  75. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, October 11–14, 2016, Proceedings, Part I 14. Springer, 2016, pp. 21–37.
  76. Chen, J.; Wang, H.; Zhang, H.; Luo, T.; Wei, D.; Long, T.; Wang, Z. Weed detection in sesame fields using a YOLO model with an enhanced attention mechanism and feature fusion. Computers and Electronics in Agriculture 2022, 202, 107412. [Google Scholar] [CrossRef]
  77. Badrinarayanan, V.; Kendall, A.; Cipolla, R. Segnet: A deep convolutional encoder-decoder architecture for image segmentation. IEEE transactions on pattern analysis and machine intelligence 2017, 39, 2481–2495. [Google Scholar] [CrossRef] [PubMed]
  78. Espejo-Garcia, B.; Mylonas, N.; Athanasakos, L.; Fountas, S.; Vasilakoglou, I. Towards weeds identification assistance through transfer learning. Computers and Electronics in Agriculture 2020, 171, 105306. [Google Scholar] [CrossRef]
  79. Li, F.; Bai, J.; Zhang, M.; Zhang, R. Yield estimation of high-density cotton fields using low-altitude UAV imaging and deep learning. Plant Methods 2022, 18, 55. [Google Scholar] [CrossRef]
  80. Iandola, F.N. SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and<0.5 MB model size. arXiv preprint arXiv:1602.07360 2016. [CrossRef]
  81. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. Medical image computing and computer-assisted intervention–MICCAI 2015: 18th international conference, Munich, Germany, October 5-9, 2015, proceedings, part III 18. Springer, 2015, pp. 234–241.
  82. Bannari, A.; Morin, D.; Bonn, F.; Huete, A. A review of vegetation indices. Remote sensing reviews 1995, 13, 95–120. [Google Scholar] [CrossRef]
  83. Shoaib, M.; Hussain, T.; Shah, B.; Ullah, I.; Shah, S.M.; Ali, F.; Park, S.H. Deep learning-based segmentation and classification of leaf images for detection of tomato plant disease. Frontiers in plant science 2022, 13, 1031748. [Google Scholar] [CrossRef]
  84. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 2014. [CrossRef]
  85. Thakur, P.S.; Sheorey, T.; Ojha, A. VGG-ICNN: A Lightweight CNN model for crop disease identification. Multimedia Tools and Applications 2023, 82, 497–520. [Google Scholar] [CrossRef]
  86. Vijayakumar, A.; Vairavasundaram, S. Yolo-based object detection models: A review and its applications. Multimedia Tools and Applications 2024, 1–40. [Google Scholar] [CrossRef]
  87. Bhatt, D.; Patel, C.; Talsania, H.; Patel, J.; Vaghela, R.; Pandya, S.; Modi, K.; Ghayvat, H. CNN variants for computer vision: History, architecture, application, challenges and future scope. Electronics 2021, 10, 2470. [Google Scholar] [CrossRef]
  88. Andreasen, C.; Scholle, K.; Saberi, M. Laser weeding with small autonomous vehicles: Friends or foes? Frontiers in Agronomy 2022, 4, 841086. [Google Scholar] [CrossRef]
  89. Wu, H.; Wang, Y.; Zhao, P.; Qian, M. Small-target weed-detection model based on YOLO-V4 with improved backbone and neck structures. Precision Agriculture 2023, 24, 2149–2170. [Google Scholar] [CrossRef]
  90. Gao, J.; French, A.P.; Pound, M.P.; He, Y.; Pridmore, T.P.; Pieters, J.G. Deep convolutional neural networks for image-based Convolvulus sepium detection in sugar beet fields. Plant methods 2020, 16, 1–12. [Google Scholar] [CrossRef] [PubMed]
  91. Yang, J.; Wang, Y.; Chen, Y.; Yu, J. Detection of weeds growing in Alfalfa using convolutional neural networks. Agronomy 2022, 12, 1459. [Google Scholar] [CrossRef]
  92. Khanam, R.; Hussain, M. YOLOv11: An Overview of the Key Architectural Enhancements. arXiv preprint arXiv:2410.17725 2024. [CrossRef]
  93. Redmon, J. You only look once: Unified, real-time object detection. Proceedings of the IEEE conference on computer vision and pattern recognition, 2016.
  94. Hussain, M. Yolov1 to v8: Unveiling each variant–a comprehensive review of yolo. IEEE Access 2024, 12, 42816–42833. [Google Scholar] [CrossRef]
  95. Farooq, A.; Jia, X.; Hu, J.; Zhou, J. Transferable convolutional neural network for weed mapping with multisensor imagery. IEEE Transactions on Geoscience and Remote Sensing 2021, 60, 1–16. [Google Scholar] [CrossRef]
  96. Sahin, H.M.; Miftahushudur, T.; Grieve, B.; Yin, H. Segmentation of weeds and crops using multispectral imaging and CRF-enhanced U-Net. Computers and Electronics in Agriculture 2023, 211, 107956. [Google Scholar] [CrossRef]
  97. Moazzam, S.I.; Khan, U.S.; Qureshi, W.S.; Tiwana, M.I.; Rashid, N.; Alasmary, W.S.; Iqbal, J.; Hamza, A. A patch-image based classification approach for detection of weeds in sugar beet crop. IEEE Access 2021, 9, 121698–121715. [Google Scholar] [CrossRef]
  98. Ramirez, W.; Achanccaray, P.; Mendoza, L.; Pacheco, M. Deep convolutional neural networks for weed detection in agricultural crops using optical aerial images. 2020 IEEE Latin American GRSS & ISPRS Remote Sensing Conference (LAGIRS). IEEE, 2020, pp. 133–137.
  99. Xu, B.; Fan, J.; Chao, J.; Arsenijevic, N.; Werle, R.; Zhang, Z. Instance segmentation method for weed detection using UAV imagery in soybean fields. Computers and Electronics in Agriculture 2023, 211, 107994. [Google Scholar] [CrossRef]
  100. Ong, P.; Teo, K.S.; Sia, C.K. UAV-based weed detection in Chinese cabbage using deep learning. Smart Agricultural Technology 2023, 4, 100181. [Google Scholar] [CrossRef]
  101. Gallo, I.; Rehman, A.U.; Dehkordi, R.H.; Landro, N.; La Grassa, R.; Boschetti, M. Deep object detection of crop weeds: Performance of YOLOv7 on a real case dataset from UAV images. Remote Sensing 2023, 15, 539. [Google Scholar] [CrossRef]
  102. Haq, M.A. CNN based automated weed detection system using UAV imagery. Computer Systems Science & Engineering 2022, 42. [Google Scholar]
  103. Smith, L.N.; Byrne, A.; Hansen, M.F.; Zhang, W.; Smith, M.L. Weed classification in grasslands using convolutional neural networks. Applications of Machine Learning. SPIE, 2019, Vol. 11139, pp. 334–344. [CrossRef]
  104. Rasti, P.; Ahmad, A.; Samiei, S.; Belin, E.; Rousseau, D. Supervised image classification by scattering transform with application to weed detection in culture crops of high density. Remote Sensing 2019, 11, 249. [Google Scholar] [CrossRef]
  105. Rahman, A.; Lu, Y.; Wang, H. Performance evaluation of deep learning object detectors for weed detection for cotton. Smart Agricultural Technology 2023, 3, 100126. [Google Scholar] [CrossRef]
  106. Jin, X.; Liu, T.; McCullough, P.E.; Chen, Y.; Yu, J. Evaluation of convolutional neural networks for herbicide susceptibility-based weed detection in turf. Frontiers in Plant Science 2023, 14, 1096802. [Google Scholar] [CrossRef]
  107. Chen, D.; Lu, Y.; Li, Z.; Young, S. Performance evaluation of deep transfer learning on multi-class identification of common weed species in cotton production systems. Computers and Electronics in Agriculture 2022, 198, 107091. [Google Scholar] [CrossRef]
  108. Kalbande, K.; Patil, W.V. The convolutional neural network for plant disease detection using hierarchical mixed pooling technique with smoothing to sharpening approach. International Journal of Computing and Digital Systems 2023, 14, 1–1. [Google Scholar] [CrossRef]
  109. Panshul, G.S.; Pushadapu, D.; Reddy, G.E.K.K.; Abhishek, S.; Anjali, T. Deeptuber: Sequential cnn-based disease detection in potato plants for enhanced crop management. 2023 5th International Conference on Inventive Research in Computing Applications (ICIRCA). IEEE, 2023, pp. 380–386.
  110. Zhong, Y.; Teng, Z.; Tong, M. The Convolutional Neural Network for Plant Disease Detection Using Hierarchical Mixed Pooling Technique with Smoothing to Sharpening Approach. Frontiers in Plant Science 2023, 14, 1166296. [Google Scholar] [CrossRef]
  111. Kaya, Y.; Gürsoy, E. A novel multi-head CNN design to identify plant diseases using the fusion of RGB images. Ecological Informatics 2023, 75, 101998. [Google Scholar] [CrossRef]
  112. Sunitha, G.; Madhavi, K.R.; Avanija, J.; Reddy, S.T.K.; Vittal, R.H.S. Modeling convolutional neural network for detection of plant leaf spot diseases. 2022 3rd International Conference on Electronics and Sustainable Communication Systems (ICESC). IEEE, 2022, pp. 1187–1192.
  113. Pandian, J.A.; Kanchanadevi, K.; Kumar, V.D.; Jasińska, E.; Goňo, R.; Leonowicz, Z.; Jasiński, M. A five convolutional layer deep convolutional neural network for plant leaf disease detection. Electronics 2022, 11, 1266. [Google Scholar] [CrossRef]
  114. Narayanan, K.L.; Krishnan, R.S.; Robinson, Y.H.; Julie, E.G.; Vimal, S.; Saravanan, V.; Kaliappan, M. Banana plant disease classification using hybrid convolutional neural network. Computational Intelligence and Neuroscience 2022, 2022, 9153699. [Google Scholar] [CrossRef] [PubMed]
  115. Sharmila, R.; Kamalitta, R.; Singh, D.P.; Chauhan, A.; Acharjee, P.B. ; others. Weighted Mask Recurrent-Convolutional Neural Network based Plant Disease Detection using Leaf Images. 2023 7th International Conference on Intelligent Computing and Control Systems (ICICCS). IEEE, 2023, pp. 681–687.
  116. Kaur, P.; Harnal, S.; Gautam, V.; Singh, M.P.; Singh, S.P. Performance analysis of segmentation models to detect leaf diseases in tomato plant. Multimedia Tools and Applications 2024, 83, 16019–16043. [Google Scholar] [CrossRef]
  117. Sharma, T.; Sethi, G.K. Improving Wheat Leaf Disease Image Classification with Point Rend Segmentation Technique. SN Computer Science 2024, 5, 244. [Google Scholar] [CrossRef]
  118. Duan, Z.; Li, H.; Li, C.; Zhang, J.; Zhang, D.; Fan, X.; Chen, X. A CNN model for early detection of pepper Phytophthora blight using multispectral imaging, integrating spectral and textural information. Plant Methods 2024, 20, 115. [Google Scholar] [CrossRef] [PubMed]
  119. De Silva, M.; Brown, D. Tomato Disease Detection Using Multispectral Imaging with Deep Learning Models. 2024 International Conference on Artificial Intelligence, Big Data, Computing and Data Communication Systems (icABCD). IEEE, 2024, pp. 1–9.
  120. Reyes-Hung, L.; Soto, I.; Majumdar, A.K. Neural Network-Based Stress Detection in Crop Multispectral Imagery for Precision Agriculture. 2024 14th International Symposium on Communication Systems, Networks and Digital Signal Processing (CSNDSP). IEEE, 2024, pp. 551–556.
  121. Bansal, P.; Kumar, R.; Kumar, S. Disease detection in apple leaves using deep convolutional neural network. Agriculture 2021, 11, 617. [Google Scholar] [CrossRef]
  122. Guan, X. A novel method of plant leaf disease detection based on deep learning and convolutional neural network. 2021 6th International conference on intelligent computing and signal processing (ICSP). IEEE, 2021, pp. 816–819.
  123. Gill, H.S.; Bath, B.S.; Singh, R.; Riar, A.S. Wheat crop classification using deep learning. Multimedia Tools and Applications 2024, 1–17. [Google Scholar] [CrossRef]
  124. Kaya, A.; Keceli, A.S.; Catal, C.; Yalic, H.Y.; Temucin, H.; Tekinerdogan, B. Analysis of transfer learning for deep neural network based plant classification models. Computers and electronics in agriculture 2019, 158, 20–29. [Google Scholar] [CrossRef]
  125. Lu, S.; Lu, Z.; Aok, S.; Graham, L. Fruit classification based on six layer convolutional neural network. 2018 IEEE 23rd International conference on digital signal processing (DSP). IEEE, 2018, pp. 1–5.
  126. Rasheed, M.U.; Mahmood, S.A. A framework base on deep neural network (DNN) for land use land cover (LULC) and rice crop classification without using survey data. Climate Dynamics 2023, 61, 5629–5652. [Google Scholar] [CrossRef]
  127. Kou, W.; Shen, Z.; Liu, D.; Liu, Z.; Li, J.; Chang, W.; Wang, H.; Huang, L.; Jiao, S.; Lei, Y.; others. Crop classification methods and influencing factors of reusing historical samples based on 2D-CNN. International Journal of Remote Sensing 2023, 44, 3278–3305. [Google Scholar] [CrossRef]
  128. Farmonov, N.; Amankulova, K.; Szatmári, J.; Sharifi, A.; Abbasi-Moghadam, D.; Nejad, S.M.M.; Mucsi, L. Crop type classification by DESIS hyperspectral imagery and machine learning algorithms. IEEE Journal of selected topics in applied earth observations and remote sensing 2023, 16, 1576–1588. [Google Scholar] [CrossRef]
  129. Seydi, S.T.; Arefi, H.; Hasanlou, M. Crop-Net: A Novel Deep Learning Framework for Crop Classification using Time-series Sentinel-1 Imagery by Google Earth Engine 2023.
  130. Zhao, H.; Chen, Z.; Jiang, H.; Jing, W.; Sun, L.; Feng, M. Evaluation of three deep learning models for early crop classification using sentinel-1A imagery time series—A case study in Zhanjiang, China. Remote Sensing 2019, 11, 2673. [Google Scholar] [CrossRef]
  131. Yin, Q.; Lin, Z.; Hu, W.; López-Martínez, C.; Ni, J.; Zhang, F. Crop classification of multitemporal PolSAR based on 3-D attention module with ViT. IEEE Geoscience and Remote Sensing Letters 2023, 20, 1–5. [Google Scholar] [CrossRef]
  132. Li, H.; Tian, Y.; Zhang, C.; Zhang, S.; Atkinson, P.M. Temporal Sequence Object-based CNN (TS-OCNN) for crop classification from fine resolution remote sensing image time-series. The Crop Journal 2022, 10, 1507–1516. [Google Scholar] [CrossRef]
  133. Chamundeeswari, G.; Srinivasan, S.; Bharathi, S.P.; Priya, P.; Kannammal, G.R.; Rajendran, S. Optimal deep convolutional neural network based crop classification model on multispectral remote sensing images. Microprocessors and Microsystems 2022, 94, 104626. [Google Scholar] [CrossRef]
  134. Galodha, A.; Vashisht, R.; Nidamanuri, R.R.; Ramiya, A.M. Convolutional Neural Network (CNN) for Crop-Classification of Drone Acquired Hyperspectral Imagery. IGARSS 2022-2022 IEEE International Geoscience and Remote Sensing Symposium. IEEE, 2022, pp. 7741–7744.
  135. Kwak, G.H.; Park, C.w.; Lee, K.d.; Na, S.i.; Ahn, H.y.; Park, N.W. Potential of hybrid CNN-RF model for early crop mapping with limited input data. Remote Sensing 2021, 13, 1629. [Google Scholar] [CrossRef]
  136. Kamarudin, M.H.; Ismail, Z.H.; Saidi, N.B.; Hanada, K. An augmented attention-based lightweight CNN model for plant water stress detection. Applied Intelligence 2023, 53, 20828–20843. [Google Scholar] [CrossRef]
  137. Azimi, S.; Wadhawan, R.; Gandhi, T.K. Intelligent monitoring of stress induced by water deficiency in plants using deep learning. IEEE Transactions on Instrumentation and Measurement 2021, 70, 1–13. [Google Scholar] [CrossRef]
  138. Zhuang, S.; Wang, P.; Jiang, B.; Li, M. Learned features of leaf phenotype to monitor maize water status in the fields. Computers and electronics in agriculture 2020, 172, 105347. [Google Scholar] [CrossRef]
  139. Kuo, C.E.; Tu, Y.K.; Fang, S.L.; Huang, Y.R.; Chen, H.W.; Yao, M.H.; Kuo, B.J. Early detection of drought stress in tomato from spectroscopic data: A novel convolutional neural network with feature selection. Chemometrics and Intelligent Laboratory Systems 2023, 239, 104869. [Google Scholar] [CrossRef]
  140. Spišić, J.; Šimić, D.; Balen, J.; Jambrović, A.; Galić, V. Machine learning in the analysis of multispectral reads in maize canopies responding to increased temperatures and water deficit. Remote Sensing 2022, 14, 2596. [Google Scholar] [CrossRef]
  141. Zhang, W.; Zhang, W.; Yang, Y.; Hu, G.; Ge, D.; Liu, H.; Cao, H.; others. A cloud computing-based approach using the visible near-infrared spectrum to classify greenhouse tomato plants under water stress. Computers and Electronics in Agriculture 2021, 181, 105966. [Google Scholar]
  142. Li, M.W.; Chan, Y.K.; Yu, S.S. Use of CNN for Water Stress Identification in Rice Fields Using Thermal Imagery. Applied Sciences 2023, 13, 5423. [Google Scholar] [CrossRef]
  143. Sobayo, R.; Wu, H.H.; Ray, R.; Qian, L. Integration of convolutional neural network and thermal images into soil moisture estimation. 2018 1st International Conference on Data Intelligence and Security (ICDIS). IEEE, 2018, pp. 207–210.
  144. Nagappan, M.; Gopalakrishnan, V.; Alagappan, M. Prediction of reference evapotranspiration for irrigation scheduling using machine learning. Hydrological Sciences Journal 2020, 65, 2669–2677. [Google Scholar] [CrossRef]
  145. Afzaal, H.; Farooque, A.A.; Abbas, F.; Acharya, B.; Esau, T. Groundwater estimation from major physical hydrology components using artificial neural networks and deep learning. Water 2019, 12, 5. [Google Scholar] [CrossRef]
  146. Liu, J.; Xu, Y.; Li, H.; Guo, J. Soil moisture retrieval in farmland areas with sentinel multi-source data based on regression convolutional neural networks. Sensors 2021, 21, 877. [Google Scholar] [CrossRef] [PubMed]
  147. Ge, L.; Hang, R.; Liu, Y.; Liu, Q. Comparing the performance of neural network and deep convolutional neural network in estimating soil moisture from satellite observations. Remote Sensing 2018, 10, 1327. [Google Scholar] [CrossRef]
  148. Chaudhari, S.; Sardar, V.; Rahul, D.; Chandan, M.; Shivakale, M.S.; Harini, K. Performance analysis of CNN, Alexnet and vggnet models for drought prediction using satellite images. 2021 Asian Conference on Innovation in Technology (ASIANCON). IEEE, 2021, pp. 1–6.
  149. Bazzi, H.; Baghdadi, N.; Ienco, D.; El Hajj, M.; Zribi, M.; Belhouchette, H.; Escorihuela, M.J.; Demarez, V. Mapping irrigated areas using Sentinel-1 time series in Catalonia, Spain. Remote Sensing 2019, 11, 1836. [Google Scholar] [CrossRef]
  150. Hu, Z.; Xu, L.; Yu, B. Soil moisture retrieval using convolutional neural networks: Application to passive microwave remote sensing. The International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences 2018, 42, 583–586. [Google Scholar] [CrossRef]
  151. Xue, M.; Hang, R.; Liu, Q.; Yuan, X.T.; Lu, X. CNN-based near-real-time precipitation estimation from Fengyun-2 satellite over Xinjiang, China. Atmospheric Research 2021, 250, 105337. [Google Scholar] [CrossRef]
  152. Sankararao, A.U.; Priyanka, G.; Rajalakshmi, P.; Choudhary, S. Cnn based water stress detection in chickpea using uav based hyperspectral imaging. 2021 IEEE International India Geoscience and Remote Sensing Symposium (InGARSS). IEEE, 2021, pp. 145–148.
  153. Wu, Z.; Cui, N.; Zhang, W.; Yang, Y.; Gong, D.; Liu, Q.; Zhao, L.; Xing, L.; He, Q.; Zhu, S.; others. Estimation of soil moisture in drip-irrigated citrus orchards using multi-modal UAV remote sensing. Agricultural Water Management 2024, 302, 108972. [Google Scholar] [CrossRef]
  154. Mia, M.S.; Tanabe, R.; Habibi, L.N.; Hashimoto, N.; Homma, K.; Maki, M.; Matsui, T.; Tanaka, T.S. Multimodal deep learning for rice yield prediction using UAV-based multispectral imagery and weather data. Remote Sensing 2023, 15, 2511. [Google Scholar] [CrossRef]
  155. Morales, G.; Sheppard, J.W.; Hegedus, P.B.; Maxwell, B.D. Improved yield prediction of winter wheat using a novel two-dimensional deep regression neural network trained via remote sensing. Sensors 2023, 23, 489. [Google Scholar] [CrossRef] [PubMed]
  156. Terliksiz, A.S.; Altilar, D.T. A Simple and Efficient Deep Learning Architecture for Corn Yield Prediction. 2023 11th International Conference on Agro-Geoinformatics (Agro-Geoinformatics). IEEE, 2023, pp. 1–6.
  157. Tanabe, R.; Matsui, T.; Tanaka, T.S. Winter wheat yield prediction using convolutional neural networks and UAV-based multispectral imagery. Field Crops Research 2023, 291, 108786. [Google Scholar] [CrossRef]
  158. Zhou, S.; Xu, L.; Chen, N. Rice yield prediction in hubei province based on deep learning and the effect of spatial heterogeneity. Remote Sensing 2023, 15, 1361. [Google Scholar] [CrossRef]
  159. Saini, P.; Nagpal, B.; Garg, P.; Kumar, S. CNN-BI-LSTM-CYP: A deep learning approach for sugarcane yield prediction. Sustainable Energy Technologies and Assessments 2023, 57, 103263. [Google Scholar] [CrossRef]
  160. Jiang, Z.; Huete, A.R.; Didan, K.; Miura, T. Development of a two-band enhanced vegetation index without a blue band. Remote sensing of Environment 2008, 112, 3833–3845. [Google Scholar] [CrossRef]
  161. MacEachern, C.B.; Esau, T.J.; Schumann, A.W.; Hennessy, P.J.; Zaman, Q.U. Detection of fruit maturity stage and yield estimation in wild blueberry using deep learning convolutional neural networks. Smart Agricultural Technology 2023, 3, 100099. [Google Scholar] [CrossRef]
  162. Chen, Y.; Lee, W.S.; Gan, H.; Peres, N.; Fraisse, C.; Zhang, Y.; He, Y. Strawberry yield prediction based on a deep neural network using high-resolution aerial orthoimages. Remote Sensing 2019, 11, 1584. [Google Scholar] [CrossRef]
  163. Sun, J.; Yang, K.; Chen, C.; Shen, J.; Yang, Y.; Wu, X.; Norton, T. Wheat head counting in the wild by an augmented feature pyramid networks-based convolutional neural network. Computers and Electronics in Agriculture 2022, 193, 106705. [Google Scholar] [CrossRef]
  164. Tedesco-Oliveira, D.; da Silva, R.P.; Maldonado Jr, W.; Zerbato, C. Convolutional neural networks in predicting cotton yield from images of commercial fields. Computers and Electronics in Agriculture 2020, 171, 105307. [Google Scholar] [CrossRef]
  165. Häni, N.; Roy, P.; Isler, V. Apple counting using convolutional neural networks. 2018 IEEE/RSJ international conference on intelligent robots and systems (IROS). IEEE, 2018, pp. 2559–2565.
  166. Yu, D.; Zha, Y.; Sun, Z.; Li, J.; Jin, X.; Zhu, W.; Bian, J.; Ma, L.; Zeng, Y.; Su, Z. Deep convolutional neural networks for estimating maize above-ground biomass using multi-source UAV images: A comparison with traditional machine learning algorithms. Precision Agriculture 2023, 24, 92–113. [Google Scholar] [CrossRef]
  167. Huber, F.; Yushchenko, A.; Stratmann, B.; Steinhage, V. Extreme Gradient Boosting for yield estimation compared with Deep Learning approaches. Computers and Electronics in Agriculture 2022, 202, 107346. [Google Scholar] [CrossRef]
  168. Sagan, V.; Maimaitijiang, M.; Bhadra, S.; Maimaitiyiming, M.; Brown, D.R.; Sidike, P.; Fritschi, F.B. Field-scale crop yield prediction using multi-temporal WorldView-3 and PlanetScope satellite data and deep learning. ISPRS journal of photogrammetry and remote sensing 2021, 174, 265–281. [Google Scholar] [CrossRef]
  169. Khaki, S.; Pham, H.; Wang, L. Simultaneous corn and soybean yield prediction from remote sensing data using deep transfer learning. Scientific Reports 2021, 11, 11132. [Google Scholar] [CrossRef] [PubMed]
  170. Fernandez-Beltran, R.; Baidar, T.; Kang, J.; Pla, F. Rice-yield prediction with multi-temporal sentinel-2 data and 3D CNN: A case study in Nepal. Remote Sensing 2021, 13, 1391. [Google Scholar] [CrossRef]
  171. Gastli, M.S.; Nassar, L.; Karray, F. Satellite images and deep learning tools for crop yield prediction and price forecasting. 2021 International Joint Conference on Neural Networks (IJCNN). IEEE, 2021, pp. 1–8.
  172. Qiao, M.; He, X.; Cheng, X.; Li, P.; Luo, H.; Tian, Z.; Guo, H. Exploiting hierarchical features for crop yield prediction based on 3-d convolutional neural networks and multikernel gaussian process. IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing 2021, 14, 4476–4489. [Google Scholar] [CrossRef]
  173. Qiao, M.; He, X.; Cheng, X.; Li, P.; Luo, H.; Zhang, L.; Tian, Z. Crop yield prediction from multi-spectral, multi-temporal remotely sensed imagery using recurrent 3D convolutional neural networks. International Journal of Applied Earth Observation and Geoinformation 2021, 102, 102436. [Google Scholar] [CrossRef]
  174. Kang, Y.; Ozdogan, M.; Zhu, X.; Ye, Z.; Hain, C.; Anderson, M. Comparative assessment of environmental variables and machine learning algorithms for maize yield prediction in the US Midwest. Environmental Research Letters 2020, 15, 064005. [Google Scholar] [CrossRef]
  175. Terliksiz, A.S.; Altỳlar, D.T. Use of deep neural networks for crop yield prediction: A case study of soybean yield in lauderdale county, alabama, usa. 2019 8th international conference on Agro-Geoinformatics (Agro-Geoinformatics). IEEE, 2019, pp. 1–4.
  176. Sun, J.; Di, L.; Sun, Z.; Shen, Y.; Lai, Z. County-level soybean yield prediction using deep CNN-LSTM model. Sensors 2019, 19, 4363. [Google Scholar] [CrossRef]
  177. Tiwari, P.; Shukla, P. Crop yield prediction by modified convolutional neural network and geographical indexes. International Journal of Computer Sciences and Engineering 2018, 6, 503–513. [Google Scholar] [CrossRef]
  178. Nevavuori, P.; Narra, N.; Linna, P.; Lipping, T. Crop yield prediction using multitemporal UAV data and spatio-temporal deep learning models. Remote Sensing 2020, 12, 4000. [Google Scholar] [CrossRef]
  179. Xu, H. Modification of normalised difference water index (NDWI) to enhance open water features in remotely sensed imagery. International journal of remote sensing 2006, 27, 3025–3033. [Google Scholar] [CrossRef]
  180. Vaswani, A. Attention is all you need. Advances in Neural Information Processing Systems 2017. [Google Scholar]
  181. Hammami, E.; Boughanem, M.; Faiz, R.; Dkaki, T. Intermediate Hidden Layers for Legal Case Retrieval Representation. International Conference on Database and Expert Systems Applications. Springer, 2024, pp. 306–319.
  182. Neptune, N.; Mothe, J. Automatic annotation of change detection images. Sensors 2021, 21, 1110. [Google Scholar] [CrossRef]
Figure 1. General workflow of smart agriculture. From identifying agricultural needs to deploying solutions, both data and models play a crucial role in developing effective solutions.
Figure 1. General workflow of smart agriculture. From identifying agricultural needs to deploying solutions, both data and models play a crucial role in developing effective solutions.
Preprints 139326 g001
Figure 2. Data sources used in various fields in smart agriculture. This Sankey diagram illustrates the flow of data from various sources to different fields in agriculture.
Figure 2. Data sources used in various fields in smart agriculture. This Sankey diagram illustrates the flow of data from various sources to different fields in agriculture.
Preprints 139326 g002
Figure 3. Data types used in various fields in smart agriculture. This Sankey diagram illustrates the various data types used in smart agriculture and their distribution across different agricultural fields.
Figure 3. Data types used in various fields in smart agriculture. This Sankey diagram illustrates the various data types used in smart agriculture and their distribution across different agricultural fields.
Preprints 139326 g003
Table 1. The number of selected papers for each field in smart agriculture. This number varies depending on the variety and diversity of methods employed across the studies.
Table 1. The number of selected papers for each field in smart agriculture. This number varies depending on the variety and diversity of methods employed across the studies.
Application About Number of papers
Weed detection Identifying unwanted plants within target crops 26
Disease detection Diagnosing and assessing plant diseases to prevent their spread 23
Crop classification Categorizing crop and plantation varieties 15
Water management Managing water resources or detecting water scarcity 22
Yield prediction Estimating future crop production levels 29
Table 2. Common performance metrics for machine learning models. This table provides an overview of standard metrics frequently used to evaluate model performances. Each metric includes its formula and a brief description.
Table 2. Common performance metrics for machine learning models. This table provides an overview of standard metrics frequently used to evaluate model performances. Each metric includes its formula and a brief description.
Metric Formula Description
Accuracy [36,37,38]
Accuracy = # C o r r e c t p r e d i c t i o n s # A l l p r e d i c t i o n s
Ratio of correct predictions to total predictions.
Precision [39]
Precision = T P T P + F P
True positives over predicted positives; accuracy of positive predictions.
Recall[40]
Recall = T P T P + F N
True positives over actual positives; ability to identify relevant instances.
F1-Score [40,41]
F 1 = 2 · Precision · Recall Precision + Recall
Harmonic mean of precision and recall; balances both metrics.
Mean Average Precision [42,43]
mAP = 1 N i = 1 N AP i
Average of AP scores across classes.
Intersection over Union (Jaccard index)[44,45]
IoU = A B A B
Ratio of overlap area to union area; used in image segmentation and object detection.
Mean Intersection over Union[46]
mIoU = 1 N i = 1 N IoU i
Average IoU across classes for multi-class evaluation.
Weighted Mean Intersection over Union [47,48]
wIoU = i = 1 N w i · IoU i i = 1 N w i
mIoU with class weights w i to emphasize importance.
Processing Time[49]
T = time taken for inference
Total time for model to process data and produce predictions.
Table 3. State-of-the-art CNN models. This table lists popular CNN models frequently used in the literature, along with links to their open-source implementations.
Table 3. State-of-the-art CNN models. This table lists popular CNN models frequently used in the literature, along with links to their open-source implementations.
Model Name Implementation URL
AlexNet [8,50,51,52], https://github.com/amir-saniyan/AlexNet
CenterNet [53,54] https://github.com/xingyizhou/CenterNet
DenseNet [55,56,57] https://github.com/titu1994/DenseNet
Detectron2 [58,59] https://github.com/facebookresearch/detectron2
EfficientDet [59,60,61] https://github.com/rwightman/efficientdet-pytorch
EfficientNet [61,62] https://github.com/lukemelas/EfficientNet-PyTorch
Faster-RCNN [36,39,63] https://github.com/trzy/FasterRCNN
GoogLeNet [37,51,64] https://github.com/conan7882/GoogLeNet-Inception
InceptionV3 [56,65,66] https://www.kaggle.com/code/yasserh/inception-v3-implementation
Mask-RCNN [41,67,68] https://github.com/matterport/Mask_RCNN
MobileNet [61,69] https://github.com/cyrilminaeff/MobileNet
MobileNetV2 [56,61,70] https://github.com/ShowLo/MobileNetV2
PSPNet [47,71] https://github.com/Lextal/pspnet-pytorch
ResNet [35,56,72] https://github.com/JayPatwardhan/ResNet-PyTorch
ResNeXt [56,73] https://github.com/titu1994/Keras-ResNeXt
SEResNeXt [56,74] huggingface.co/docs/timm/en/models/seresnext
SSD [75,76] https://github.com/amdegroot/ssd.pytorch
SegNet [77,78,79] https://github.com/vinceecws/SegNet_PyTorch
SqueezeNet [37,80] https://github.com/cmasch/squeezenet
UNet [81,82,83] https://github.com/zhixuhao/unet
VGG [51,84,85] https://github.com/Lornatang/VGG-PyTorch
YOLO [86] https://github.com/srebroa/awesome-yolo
Table 4. Summary of weed detection approaches in agriculture.
Table 4. Summary of weed detection approaches in agriculture.
Category Approaches Purpose
Detection Techniques Image segmentation (e.g. PSPNet, SegNet, UNet) Classifies pixels as weeds or not (e.g. crops, background) [40,41,47,48,50,96,97,98,99]
Object detection (e.g. YOLO, Faster-RCNN, Mask-RCNN) Efficient at detecting and locating zones that contain weeds in an image [36,39,41,76,91,101,105]
Input Data Type RGB Most common data type in weed detection due to the different shapes of weeds that makes possible their identification [36,39,47,48,50,76,78,89,91,99,100,101,103,104,105,106,107]
Multispectral Improves weed detection performances [40,41,95,96,97,98]
Vegetation indexes (e.g. NIR, NDVI) Assists in distinguishing vegetation from non-vegetation [41,78,96,97,98,99]
Data Acquisition UAV Useful for aerial weed detection at different altitudes (1 - 65 m) [41,89,98,99,100,101,102]
UGV Efficient for close-range weed detection [40,50,96,103,104,107]
Handheld devices (e.g. cameras, mobile phones) Suitable for small scale weed detection [36,39,47,48,76,78,91,105,106]
Table 5. Summary of disease detection approaches in agriculture.
Table 5. Summary of disease detection approaches in agriculture.
Category Approaches Purpose
Detection Techniques Image classification (e.g. VGG, Inception, DenseNet, ResNet) Efficient at identifying disease types in leaf images [51,56,85,108,109,110,111,112,113,114]
Object detection (e.g. YOLO, CenterNet) Used to detect areas in plants or fields that show disease symptoms [42,49,53]
Image segmentation (e.g. Mask R-CNN) Helpful at classifying diseased pixels in crops and leaves [68,115,116,117]
Input Data Type RGB Most commonly used for detecting visible symptoms [51,53,56,85,108,109,110,111,112,113,114,121,122]
Multispectral images Provides a high potential for early detection of diseases, without apparent symptoms [118,119,120]
Data Acquisition UAV Efficient for large-scale monitoring and real-time disease detection [42,49,53]
Handheld devices Used for close-range and on-ground images. Useful for quick data collection on fields and in controlled laboratory environment [51,56,85,108,109,110,111,112,113,114,121,122]
Table 6. Summary of crop classification approaches in agriculture.
Table 6. Summary of crop classification approaches in agriculture.
Category Approaches Purpose
Detection Techniques Image classification (e.g. CNN-RNN-LSTM, CNN,) Efficient at classifying images of leaves, plants, or fruits [123,124,125]
Image segmentation (e.g. 1D-CNN, 3D-CNN, ViT, Recurrent CNN, HRNet) High performances on classifying each pixel into the corresponding crop type [45,128,129,132]
Input Data Type RGB images Mostly used for leaf and plant classifications [52,124,125,135]
Multispectral and hyperspectral images Captures unique spectral signatures that are crop specific, assisting land cover crop classification [45,127,128,132,133]
SAR data Acquires detailed surface information. Useful to capture crop structure while being unaffected by weather conditions (e.g. clouds) [129,130,132]
Data Acquisition Satellites (e.g. Sentinel-1, Sentinel-2, RADARSAT2) Provides historical and periodical data for large-scale crop classification (e.g. land cover) [45,126,127,128,129,130,131,132,133]
UAV Captures high resolution aerial images that can be combined with satellite images to improve crop classification [52,131,132,133,134,135]
Handheld devices Close-range imaging for small scale classification [124,125]
Table 7. Summary of water management approaches in agriculture.
Table 7. Summary of water management approaches in agriculture.
Category Approaches Purpose
Detection Techniques Image classification High accuracy in detecting water stress, predicting droughts, and classifying different irrigation treatments [37,61,72,136,137,138,139,140,141]
Regression Accurately estimate soil moisture content, evapotranspiration, and groundwater content [143,144,146,150,151]
Input Data Type RGB images Effective for detecting visible changes (e.g. color, curvature) in plants under water stress [37,72,136,137,138]
Multispectral and hyperspectral Useful for early detection of water stress, even before visible symptoms [61,139,140,141]
Vegetation indexes (e.g. NDVI, MSAVI) Commonly used to assist in drought prediction and soil moisture estimation [140,146,147,148]
Thermal Helps in detecting water stress and estimating soil moisture [142,143]
SAR Beneficial in soil moisture estimation [146,147,149]
Weather and in-situ data Effective in estimating evapotranspiration, groundwater, and soil moisture content [144,145,147,150]
Data Acquisition Satellites (e.g. Sentinel-1, Sentinel-2, RADARSAT2) Provides high spatio-temporal data, useful for soil moisture and irrigation mapping[146,147,149,150,151,152]
UAV Captures high resolution imagery, mostly used in water stress detection and soil moisture estimation [66,153]
Handheld devices Allows ground-level data acquisition for close-range water stress detection [37,137,138]
Other sensors (e.g. tensiometers, thermometer) Gathers data for better water management and ground truth labels [61,139,140,141,142,143,146,147,152]
Table 8. Summary of yield prediction approaches in agriculture.
Table 8. Summary of yield prediction approaches in agriculture.
Category Approaches Purpose
Detection Techniques Image classification Used in identifying crop growth stages, which correlates with yield [38,154,155,159,167,168,169,172,173,174,176,177]
Image segmentation (instance and semantic segmentation) Used for crop segmentation and maturity classification, which helps in crop counting and yield estimation [43,165]
Regression Most commonly used technique for yield prediction [154,155,157,158,159,167,168,169,170,172,173,174,176,177]
Object detection Applied for detecting individual crop heads, fruits, or plants [161,162,163,164,165]
Input Data Type RGB images Effective for identifying crop growth stages based on different visible traits [43,161,162,163,164,165,168]
Multispectral and hyperspectral images Useful for detecting crop health and predicting yield [38,154,157,167,168,169,170,172,173,174,176]
Vegetation indexes (e.b. NDVI, SAVI, EVI) Helps in biomass estimation [157,158,168,170,172,174,177]
Thermal Improves yield prediction performances when combined with spectral data [167,172,174,176]
Weather and in-situ data Provides more features to help in yield prediction [154,170,174]
Data Acquisition Satellites (e.g. MODIS, Sentinel-1, Sentinel-2) Used for large-scale yield prediction based on multitemporal and historical data [155,156,158,167,168,169,170,171,172,173,174,175,176,177]
UAV Provides high resolution imagery for yield estimation [38,57,79,154,157,162,163,166]
Handheld devices Offers localized data for specific crops with limited coverage but effective for small farms [43,161,163,164,165]
surveys and land cover Mostly used as labels for ground truth [38,154,155,157,158,159,167,168,169,170,172,173,176]
Other sensors (e.g. in-situ sensors) Provides weather and soil related measurements, assisting yield prediction [38,155,158,159,170,174]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

© 2024 MDPI (Basel, Switzerland) unless otherwise stated