Preprint
Article

Groundnut (ARACHIS HYPOGAEA L.) Seed Defect Classification Using Ensemble Deep Learning Techniques

Altmetrics

Downloads

87

Views

23

Comments

0

A peer-reviewed article of this preprint also exists.

This version is not peer-reviewed

Submitted:

13 August 2024

Posted:

13 August 2024

You are already at the latest version

Alerts
Abstract
Groundnut is an oil seed cash crop, that can be consumed directly by humans and animals. It is also used as an intelligence in the industrial production of butter and other products. However, Groundnut seeds can become infected with fungi, viruses, pests, physical damage, and high heat, which can negatively impact crop yield, quality, and economic value, and pose health hazards. To address these challenges, computer vision technology detects and classifies defects. This study introduced an ensemble deep learning defected classification model that combines VGG16 and InceptionV3 using seed images. To enhance model performance, collected images are preprocessed using novel techniques tailored to different image types. Preprocessing techniques are chosen based on image-quality evaluation metrics. Watershed segmentation is applied followed by detecting the region of interest in the image using YOLOv3. The image dataset is augmented and balanced using a Generative Adversarial Network (GAN). The model development involves a combination of classical and deep-based features, comparing features extracted with (HOG and GLCM) to those extracted with InceptionV3 and VGG16. The ensemble model achieves an accuracy of 96.25% with a split ratio of 10% for testing, 10% for validation, and 80% for training set. This research benefits farmers looking to improve yield, consumers of groundnut products, and researchers studying seed defects. The work provides valid insights for future researchers regarding dataset creation, augmentation, methods, and feature selection for modeling. However, the model experienced misclassifications due to the similar appearance of sample images in different classes. Future researchers should address these challenges and consider factors such as oil content and defect grading.
Keywords: 
Subject: Computer Science and Mathematics  -   Artificial Intelligence and Machine Learning

Introduction

Agriculture is essential in crop production for food security and economic development, serving domestic and foreign trade exchanges. According to a study by (Jenber et al., 2022) more than 85% of Ethiopia's population engaged in the agricultural sector to produce various agricultural products including crops. This evidence proved that agriculture including crop production such as maize, wheat, barley, sesame, bean, groundnut, cotton, etc., is the backbone of the Ethiopian economy.
Groundnut is one of the nutrition-rich agricultural oil seed cash crops included (H. Yang et al., 2021). Several food preparations incorporate groundnuts to improve the protein level, taste, and flavors and byproduct of its industry used for animal foods (Asibuo et al., 2008). However, groundnut seeds are vulnerable to defects of pests, and fungi, damaged by mechanical force and high heat before and after harvesting leading to loss of expected nutrients, agricultural yield, and health risks (Qi et al., 2021).
The groundnut market is currently facing challenges due to aflatoxin contamination leading to a halt in exports and a decrease in demand (Guchi, 2015). Insect pests are a major factor in reducing groundnut yields and quality through direct damage or as vectors of virus diseases. This ultimately leads to a loss of oil content and quality. Detecting defects in groundnut seeds is a crucial and challenging issue to prevent disease caused by infected seeds and improve production yield for farmers (Bajia et al., 2017).
Classification and packaging of seeds based on their quality is a challenging task for farmers and agro-industries due to this segregation of seeds at the harvesting site is important (Kundu & Rani., 2021). One of the fascinating research areas in agriculture is the identification of defects in healthy seeds, with technologies like computer vision and machine learning by analyzing digital image features of crops that may indicate defects such as cracks, spots, and deformities (Gómez-Camperos et al, 2021).
However, the selected feature extraction and classification techniques have a major role in the performance of the identification model. This study aims to classify defects in groundnut seeds using a deep ensemble learning approach, a subclass of machine learning. The ensemble-teaching model combines the advantages of various deep learning models as base learners, resulting in a final model with proven generalization performance (Renda et al., 2019).
The rest of the paper is organized as follows: Section 2 describes related works, which discuss groundnut and related seed defect detection state of the art. The methodology and the proposed model are presented in Section 3. In section four, we discussed the experiment results and discussion. Conclusion and future work in Section 5 following the acknowledgments and references.

2. Related Works

Scholars (Huang et al., 2019; Szczypinski et al., 2017) run their study on crop defect recognition and identification during storage using computer technology. This section briefly discusses existing research approaches and methodologies in the context of machine learning-based defect inspection and classifications of groundnut seeds. So many researchers have done on groundnut seeds using images from different perspectives such as defect detection(Guchi, 2015), variety classifications (H. Yang et al., 2021) as well as genetic characters inheritance (Aworinde et al., 2007; District, 1935), and oil content quality (Sarvamangala et al., 2011; Shasidhar et al., 2017), etc. before and after harvesting. This study is limited to groundnut seed defect classifications after harvesting.
The study (Ziyaee et al., 2021) investigated effective image processing along with computational techniques for detecting a defect with an aflatoxin-producing fungus as well as the growth of molds using three machine learning algorithms which are support vector machine(SVM), Artificial neural network(ANN), and the adaptive neuron-fuzzy inference system (ANFIS) using 1500 images. The study was conducted on a small dataset and labeling of defects is subjective. Similarly (Manhando et al., 2021) employed an optical coherence tomography (OCT) imaging technique for the detection of mold-contaminated peanuts and the growth of molds during the incubation period with SVM trained on features extracted using a pre-trained Deep Convolutional Neural Network (DCNN). Conversely, hyperspectral imaging technology (Manhando et al., 2021; Zou et al.,2022) was used for non-destructive variety classification and mold detection of peanut seeds by stack ensemble learning mechanism. They achieved high detection performance on five varieties of peanut seed, more over the study (Liu et al., 2022) verified that augmentation can improve the detection accuracy on K-nearest neighbors (KNN), support vector machines (SVM), and MobileViT-xs models.
However, none of them addresses other peanut defect types such as defects with a virus, physical damage, infection with pests, and seed burning with high heat were not considered. The researcher(Y. Wang et al., 2023) developed a classical Faster Recurrent Convolutional Neural Network(R-CNN) model to detect defects on 1300 groundnut seed images based on textural features. The finding was achieved using texture enhancement and texture-based attention modules on the neural network of the backbone network VGG16. However, Faster R-CNN with a VGG16 backbone predominantly focuses on learning and extracting deep convolutional features rather than explicitly focusing on texture features. Nevertheless, considering a combination of features, including texture, shape, context, and semantic information, to achieve robust and accurate classifications in Faster R-CNN is important. The work by (Y. Wang et al., 2023) used Faster R-CNN for the detection of groundnut defects including moldy, mechanically damaged, and germinated using texture features sensitive to change in scale. As a scientific contribution, the findings provide valuable insights, novel approaches, and techniques for seed defect classifications as listed below.
  • In the preprocessing phase, techniques are chosen based on image quality evaluation metrics value to assess the impact of preprocessing techniques.
  • The GAN-based Augmentation technique was developed to increase dataset size and diversity for model performance and generalization.
  • The deep ensemble classification model was developed utilizing VGG16 and InceptionV3, incorporating classical and deep-based features through concatenation.
  • We collected 6050 images that served as defect grading for future research.

3. Methodology

The performance of the proposed model is based on the adjustment of the independent variables. This experimental research design involves the detail of dataset preparation, preprocessing, implementation, and performance evaluation as the working architecture of Figure 1 stated below.
The proposed architecture in Figure 1:1 outlines a classification model design consisting of four stages. The stages are preprocessing, feature extraction, model development, and model evaluation. Each stage involves specific techniques to ensure the successful completion of the task.

3.1. Preprocessing

The collected groundnut seed image suffered from burling low insight intensity and had varying sizes with a limited imbalanced collection. These challenges required noise filtering, resizing, enhancement, and other preprocessing techniques. To choose the most effective preprocessing technique image quality evaluation metrics, such as, Mean Square Error (MSE) Structural Similarity Index (SSI), and Peak Signal Noise Ratio (PSNR) were utilized.
Image quality evaluation: In Preprocessing, the result of the image may lead to a loss of important information. Image quality evaluation (Padilla et al., 2020) can be subjective to relying on human judgment, or objective utilizing image evaluation techniques. The study (Hu et al., 2020) suggested that the subjective method is efficient in cases like multimedia images, still for other image-based applications, it may be time-consuming and result in a loss of important information (Z. Wang et al., 2004). Instead, objective methods are best suited to compare and evaluate processed images with original images or image processing algorithms. Some of these evaluation metrics include Mean Squared Error (MSE), Signal to Noise Ratio (SNR), Peak Signal-to-Noise Ratio (PSNR), Structural Content (SC), etc.
Mean squared Error (MSE): The MSE is a method used to assess the dissimilarity between two images by calculating the average of the squared differences between corresponding pixel values in a reference (original) image and a distorted (processed) image. A lower MSE value indicates a higher similarity between the distorted and the original image. MSE is a mathematical measure that does not always correlate well with human perception, especially for view and small differences between images (Hu et al., 2020). It can be expressed as follows.
M S E = ( 1 / N ) Σ [ I o r g x , y I d i s t x , y ] 2
where N is the total number of pixels in the image, I_org(x, y) is the pixel value at coordinates (x, y) in the original image, and I_pro(x, y) is the pixel value at coordinates (x, y) in the processed image.
PSNR: It assesses and calculates the ratio between the maximum signal power and the MSE between the original and distorted images. It is typically expressed in decibels (dB) and serves as an indicator of the amount of noise or distortion present in the image. A higher PSNR value suggests that the image has a reduced amount of noise or distortion. PSNR can be calculated as
P S N R = 10 l o g 10 ( ( m a x _ p i x e l _ v a l u e ^ 2 ) / M S E )
where l o g 10 Function converts the ratio of the peak signal power to MSE into a logarithmic scale and multiplying by 10 scales it to decibels (dB).
Structural Similarity Index: It assesses how similar the structures in two images are unlike conventional metrics that, only consider pixel differences on factors like luminance, contrast, and the overall structure of the image (Horé & Ziou, 2010). It captures important visual characteristics instead of just individual pixel values. When the SSIM value is higher, it signifies that the original and distorted images have a stronger resemblance in terms of their structure, implying a higher level of perceptual quality. Index similarity can be calculated as.
S S I M x , y = L x , y C x , y S x , y
where: x and y represent the corresponding pixel values in the original and processed images. L(x, y) is the luminance similarity, which measures the similarity in mean pixel intensities between x and y. C(x, y) is the contrast similarity, which evaluates the similarity in variances of pixel intensities between x and y. S(x, y) is the structure similarity, which calculates the similarity in covariance of pixel intensities between x and y. The values of L(x, y), C(x, y), and S(x, y) are typically calculated using the mean, variance, and covariance of pixel intensities in the corresponding patches of the images. The final SSIM value is obtained by taking the product of the three similarity components.

3.1.1. Histogram Equalization

It is an image processing technique that aims to enhance contrast and improve the visual quality of an image. Rajendran et al. (2015) illustrated that a histogram is used to segment an image into several regions by thresholding and redistributing the intensity values across the entire range. We applied Adaptive Histogram Equalization to address brightness issues in the collected image.

3.1.2. Resizing

Image resizing is necessary to increase or decrease the total number of pixels. Making all experimental images uniform in dimensions helps to further preprocessing speeds in image retrieval (Gordo et al., 2017). Images are resized to 224 × 224 pixels with an aspect ratio of 2:1 by comparing resizing techniques, which include Nearest Neighbor, Bilinear Interpolation, Bicubic Interpolation, and Lanczos based on image evaluation metrics.
However, selecting the optimum resizing technique using experimental compression of objective evaluation metrics is the best strategy rather than relying on subjective evaluation in a large dataset. Table 1 shows the values of the four resizing techniques in a single image comparison. Based on the results, the bicubic interpolation technique has been selected.
Figure 2. Resized image quality evaluation metrics value.
Figure 2. Resized image quality evaluation metrics value.
Preprints 115091 g002

3.1.3. Noise Reduction

Noise removal is a technique used to eliminate unwanted detail (Motwani et al., 2004) that depreciates and destroys the original image quality. Similar to resizing techniques we have selected noise reduction techniques based on the experimental result of evaluation matrices. Based on noise reduction experimental information and evaluation results. Gaussian Adaptive Bilateral Filter (GABF) is selected from the comparison result of mean, bilateral, Gaussian, and Gaussian adaptive bilateral filtering techniques. The evaluation metrics result of PSNR and MSE in Gaussian filtering is larger than others and bilateral filtering has smaller values of MSE and PSNR from this the combination of the two has an optimum value than other techniques.

3.1.4. Segmentation

Image segmentation includes the process of digital images partitioning into meaningful multiple regions (Kunaver & Tasič, 2005) offering easy dimensional reduction. It represents some kind of information to the user in the form of color, intensity, or texture offering easier ways (Yogamangalam & Karthikeyan, 2013)for further analysis and comparison of images. Three segmentation techniques are compared through experiments on three segmentation techniques, which are watershed, threshold, and edge-based segmentation. Watershed segmentation is selected based on the characteristics of the images, which are irregular shapes and clear intensity, images with minimal overlap, and the noise is filtered by good noise reduction techniques in the previous preprocessing.

Detection

The regions of interest in the captured images are identified using YOLOv3. The detection process involves taking an input image learning the class probabilities and bounding box coordinates, and dividing the image into a grid of S x S, which predicts N bounding, boxes. The confidence level indicates the accuracy of the bounding box determining whether it contains an object (regardless of its class). Additionally, YOLOv3 predicts the classification score for each box in every image during training. The total predicted probability for all boxes is calculated as S×S×N.

GAN-Based Augmentation

Generative Adversarial Networks (GAN)--based augmentation was applied after preprocessing the image to increase and balance the number of instances in each class. GAN is an artificial neural network, that generates a new image based on discriminator and generator networks (Frid-Adar et al., 2018) as shown below.
Figure 1. Work of GAN-based Augmentations.
Figure 1. Work of GAN-based Augmentations.
Preprints 115091 g003
As shown in the Figure above, once the generator network is trained, it takes random noise as input and generates new synthetic images as training datasets. These images are fed into the discriminator network with real images from the training dataset. The discriminator then attempts to distinguish between real and synthetic images and provides feedback on the realism of the generated images. The generator network uses this feedback to enhance the quality of the synthetic images. Compared to traditional image augmentation methods such as rotation and flipping, GAN-based image augmentation can generate highly diverse and realistic images. These images can enhance the performance of deep ensemble learning models trained on limited datasets.
Table 1. summaries of augmented images.
Table 1. summaries of augmented images.
Class Name Images in each class after augmentation Number of Augmented Images File type
Healthy 1560 94 Jpg,
Virus defect 1560 541 Jpg
Fungal defect 1560 3 Jpg
physical damage 1560 558 Jpg
Pest defect 1560 554 Jpg
Total image 7800

Feature Extraction

The groundnut seed image features are extracted using classical feature extraction techniques with HOG and deep learning CNN(VGG16 &Inception V3). Classical feature extraction with HOG involves the local shape and edge image characteristics to capture and divide it into cells and calculate the magnitude and orientation of gradients at each pixel within these cells. It is a preprocess quantizing the gradient orientations and accumulating them into histograms. HOG represents the image's structural information as a feature vector a one-dimensional array that evolves the image's characteristics and captures the distribution of gradient orientations in different local regions. The size of the feature vector is determined by the HOG parameters such as the number of cells, orientation bins, and block size. This process allows HOG to capture important features related to edges and shapes from the groundnut images.
The Gray Level Co-Occurrence Matrix (GLCM) can extract second-order features from an image (Mohanaiah et al., 2013). In two pixels, where the gray levels are i and j, the probability that the positional direction is θ, and the distance d is denoted as
d i s t a n c e d = p ( i , j , d , Θ )
where Θ has one of four values 0°, 45°, 90°, and 135°. The entire image computing without reducing the scale is computationally intensive to gain the ideal gray-level image with efficient computing time and capability.
In this experiment, the Gray Level Co-Occurrence Matrix (GLCM) extracts five features: contrast, dissimilarity, homogeneity, energy, and correlation from the image of a groundnut seed. The pairs of pixels considered in the GLCM matrix are separated by a distance of 1, 2, or 3 pixels that is indicated by the distance parameter set to [1, 2, 3]. The angles parameter specifies the direction or orientation in which the pairs of pixels are considered: horizontal, diagonal, vertical, and diagonal directions, represented by radians. These angles are annotated in radians np.pi is the constant representing the value of pi (3.14159...), and np.pi / 4, np.pi / 2, and 3*np.pi/4 corresponds to angles of 45 degrees, 90 degrees, and 135 degrees, respectively.
The Visual Geometry Group (VGG) is a CNN architecture created by Oxford University's Visual Geometry Group (Veni & Manjula, 2022). VGG has either 16 or 19 convolutional layers followed by fully connected layers. One of the significant developments in VGG is its use of small receptive fields (3x3) in convolutional layers, which has contributed to a high degree of precision in image classification. The VGG16 architecture is a 16-layer deep neural network that includes (Qassim et al., 2016) convolutional neural network features with small convolution filters. It consists of three fully connected layers, 13 convolutional layers, and 5 max pooling layers.
Figure 1. Network architecture of VGG16.
Figure 1. Network architecture of VGG16.
Preprints 115091 g004
In the network architecture shown in Figure 1:2, the input shape and size for the VGG16 network may vary depending on the implementation, as it can be trained on different input sizes. The VGG16 processes preprocessed images through a sequence of convolutional and pooling layers. Each convolutional layer applies learned filters to detect visual patterns at various scales and orientations. Activation functions introduce non-linearity, and max pooling reduces spatial dimensions while retaining important features. Fully connected layers are typically excluded for feature extraction. The output of convolutional layers represents the presence and characteristics of visual features, with depth increasing in deeper layers. Global average pooling converts feature maps into a fixed-size feature vector. By passing images through VGG16, the model learns to extract complex and informative features, capturing different levels of visual patterns.
Deep learning based (InceptionV3): The input image is resized to a fixed scale (299, 299) pixels and passed to the stem layer which consists of a combination of convolutional and pooling layers that perform basic feature extraction, such as detecting edges and textures. The convolutional layers generate feature maps that represent visual features. Global average pooling then reduces these feature maps to a fixed-size feature vector, summarizing the image's content.
To determine, the best feature extraction techniques based on the evaluation results of the model performance, three experiments were conducted on training and testing splits of 20% and 80%. A total of 6000 images were used for this experiment with a learning rate of 0.001 using the Adam Optimizer, 16 samples per gradient, and 10 iterations of the entire dataset—the first experiment conducted a classical feature concatenated with VGG16 & InceptionV3 features. In the second experiment, GLCM features were concatenated with deep-based features. In the last experiment, only deep-based features with InceptionV3 and VGG16 were used to train the model. The comparison results of the three experiments are shown in the table below.
Table 1. The result of model performance on feature concatenation.
Table 1. The result of model performance on feature concatenation.
Classical feature Deep-based Feature used Base models from concatenation feature Accuracy of a base model Accuracy of the ensemble model
HOG feature VGG16 feature VGG16 94% 95.44%
InceptionV3 feature Inception V3 93.2%
GLCM feature VGG16 feature VGG16 model 94.51% 95.2%
InceptionV3 feature Inception V3 model 91.31%
Not used VGG16-feature VGG16 model 93.56% 93.86%
Inception V3 feature Inception V3 model 90.32
The table above indicates that the model outperforms the combination of classical features with HOG and deep-based features.

Ensemble Strategy

Due to the availability of vast amounts of data deep learning (Y. Yang et al., 2022) has a dominant position in various artificial intelligence applications (Livieris et al., 2021; Wyatt et al., 2022). Currently, ensemble learning techniques based on deep learning have shown significant performance improvements in learning systems. Ensemble learning (Dong & Yu, 2020; Schwenker, 2013) involves using multiple learning algorithms on a set of extracted features to gain knowledge and achieve better performance than individual weak predictive models through voting schemes. The robustness of the learning system in DEL comes from having multiple learners (neural network) (Rokibul & Nazmul, 2020; Ganaie et al., 2022) that aim to provide accurate results by combining the output of individual train neural networks instead of selecting the best one. This offers solutions to machine learning challenges by entertaining various ensemble models (Al-sahaf et al., 2019) such as averaging, stacking, boosting, etc.
Ensemble Strategies are a machine learning method that enhances a model's predictive performance by addressing the weaknesses or biases of individual models known as base learners, through the aggregation of multiple models. Effective Ensemble Strategies result in decreased overfitting, increased accuracy, and improved generalizability to unseen data. Some popular ensemble strategies include bagging, boosting, majority voting, weighted voting, averaging, and stacking. Bagging reduces variance by averaging the output of multiple models while boosting reduces bias by combining various models to enhance accuracy. Stacking involves training numerous models and using their outputs to train a final model that provides a more accurate prediction.
Majority voting is the method used to combine predictions from many individual models into a final prediction. In this method, the class labels that are most frequently predicted among the models are selected as the final prediction. It is a straightforward and powerful technique for integrating predictions from many models. It is effective in enhancing prediction accuracy across various applications.
Weighted voting: - this strategy combines predictions of several models by assigning different weights to each model according to their purposes. In this technique, each model is assigned a weighted and predicted probabilities are weighted accordingly and the final prediction is based on the weighted sum of the individual models’ predictions.
Averaging also known as voting aggregation techniques of ensemble learning is a linear combination of multiple models. The pertained model is classified, and later an average of the results is used to make the final prediction. In the proposed, the mean approach is calculated as the following aggregated function.
y = 1 n i = 1 n P i x
where: y is the result of the given list of models w is the prediction of the ith model on the x dataset and n is the number of ensemble models involved in the given problem. We use a deep ensemble learning strategy to develop averaging methods on two pre-trained models: convolutional neural network (CNN) which are VGG16, and InceptionV3 for classifications with specific ensemble strategy.

4. Experimental Result

After collecting and preparing the images; each model is individually trained with end-to-end feature extraction. After training and evaluating each model; the results show varying classification abilities. The base model uses voting aggregation to develop an ensemble model with the same data. The model is trained and evaluated using three-split ratios of 70%-15%-15%, and 80%-10%-10% for training; testing; and validation respectively with hyperparameter setup using grid search optimization. The model is trained with both GAN-based and classical augmentation techniques. However; we achieved the best result using 80%-10%-10% split ratio using GAN-based augmentation. The total number of instances used for model development is 7800 images. The validation set was calculated to be 780; while the testing set also consists of 780 image instances. The remaining 6240 image instances are allocated to the training set. Each class in the validation and testing set contained 156 images; while the training set had 1248 images for each class. The accuracy of the base and ensemble model using both augmentations is summarized in the following table
Table 1. The result of augmented data.
Table 1. The result of augmented data.
Augmentation Base models Base model accuracy Ensemble model accuracy
GAN based VGG16 model 94.51% 96.25%
Inception V3 model 91.31%
Classical VGG16 model 93.56% 94.2%
Inception V3 model 90.32
No augmentation VGG16 model 89% 87%
Inception V3 model 83%
The result of the augmented data has significant differences for both the base model and ensemble model as shown in the above table. The accuracy and loss graphs of the experimental results with augmented data are represented in Figure 1.4 (a) and (b).
Figure 3. Accuracy and loss graph of ensemble model with GAN-based and classical augmentation.
Figure 3. Accuracy and loss graph of ensemble model with GAN-based and classical augmentation.
Preprints 115091 g005
The model achieved 97.22%, 92.05%, and 96.25% train, validation, and test accuracy respectively. The model with GAN-based augmentation has better generalization than the classical one when we have seen the accuracy and loss graph of the model.

5. Error Analysis

The ensemble learning classification model achieved high accuracy by correctly classifying 6067 instances of training images, 718 validation images, and 751 testing images. However, misclassifications amounted to 173 images in the training dataset, 62 validation images dataset, and 29 images in the testing dataset. Most errors and misclassifications in the training, validation, and test data were due to cracked classes. The model often mistook cracked images as pest defects, classifying them as pest defective instead of pest defect images. Another error occurred with two classes, where pest-defected images were learned as cracked, and virus-defected images were considered fungal defective instead of their actual classification. To improve the model's accuracy and robustness, we suggested removing images that resemble other classes and finding additional feature extraction techniques to differentiate them. Additionally, the VGG16 pre-trained model exhibited weaker classification ability compared to InceptionV3.

6. Conclusions

In this study, Groundnut seed image defects are classified into five categories virus, fungus, pest damage, physical damage, and health using a deep ensemble model. The collected images are enhanced with Global Histogram Equalization (GHE), and noise filtered using Gaussian adaptive Bilateral Filtering (GABF) techniques. The images are resized to 224 × 224 using bicubic interpolation with an aspect ratio of 2:1. Preprocessing techniques are chosen based on image evaluation metrics. Regions of interest are detected using YOLOv3. The initial dataset of 6050 images is augmented using traditional and GAN-based methods, resulting in a total of 7800 images. The classification model is trained using features extracted from a comparison of classical techniques (GLCM & HOG) with deep-based techniques (VGG16 & InceptionV3). The experiment showed that different base learners performed differently depending on the feature concatenation with InceptionV3 performing well on deep-based and HOG features. The model evaluation without augmentation, GAN-based augmentation, and traditional augmentation achieved test accuracies of 87%, 96.25%, and 94.2%, respectively. The model’s performance is limited by the similar appearance of different defects and the weaknesses of the classification base learners. The challenges make it difficult for researchers to claim performance improvements and the effectiveness of the model.

Funding

The authors did not receive any financial support for the research, and publication of this article.

Data Availability Statement

Code Availability: The code used in this study is available from the corresponding author upon reasonable request. The data that support the findings of this study are available from the corresponding author upon reasonable request.

Acknowledgments

We would like to express our gratitude to the anonymous reviewers for their valuable comments and constructive suggestions. This research did not receive specific grants from funding agencies in the public, commercial, or not-for-profit sectors.

Conflict of Interest

The author declares no conflicts of interest for the manuscript and its forthcoming publication process.

References

  1. Al-sahaf, H., Bi, Y., Chen, Q., Lensen, A., Mei, Y., Sun, Y., Xue, B., & Zhang, M. (2019). A survey on evolutionary machine learning. Journal of the Royal Society of New Zealand, 49(2), 205–228. [CrossRef]
  2. Asibuo, J., Akromah, R., Adu-Dapaah, H., & Safo-Kantanka, O. (2008). Evaluation of nutritional quality of groundnut (Arachis Hypogaea L.) from Ghana. African Journal of Food, Agriculture, Nutrition and Development, 8(2), 133–150. [CrossRef]
  3. Bajia, R., Singh, S. K., Bairwa, B., & Padwal, K. G. (2017). MAJOR INSECT PESTS OF GROUNDNUT ( Arachis hypogaea L.),. August. [CrossRef]
  4. District, S. A. (1935). THE INHERITANCE OF CHARACTERS IN THE. 1(8).
  5. Dong, X., & Yu, Z. (2020). A survey on ensemble learning. 14(2), 241–258.
  6. Frid-Adar, M., Diamant, I., Klang, E., Amitai, M., Goldberger, J., & Greenspan, H. (2018). GAN-based synthetic medical image augmentation for increased CNN performance in liver lesion classification. Neurocomputing, 321, 321–331. [CrossRef]
  7. Ganaie, M. A., Hu, M., Malik, A. K., Tanveer, M., & Suganthan, P. N. (2022). Ensemble deep learning: A review. Engineering Applications of Artificial Intelligence, 115. [CrossRef]
  8. Gordo, A., Almazán, J., Revaud, J., & Larlus, D. (2017). End-to-End Learning of Deep Visual Representations for Image Retrieval. International Journal of Computer Vision, 124(2), 237–254. [CrossRef]
  9. Guchi, E. (2015). Aflatoxin Contamination in Groundnut (Arachis hypogaea L.) Caused by Aspergillus Species in Ethiopia. Journal of Applied & Environmental Microbiology, 3(1), 11–19.
  10. Horé, A., & Ziou, D. (2010). Image quality metrics: PSNR vs. SSIM. Proceedings - International Conference on Pattern Recognition, 2366–2369. [CrossRef]
  11. Hu, B., Li, L., Wu, J., & Qian, J. (2020). Subjective and objective quality assessment for image restoration: A critical survey. Signal Processing: Image Communication, 85(June), 1–50. [CrossRef]
  12. Huang, S., Fan, X., Sun, L., Shen, Y., & Suo, X. (2019). Research on Classification Method of Maize Seed Defect Based on Machine Vision. 2019(1).
  13. enber, A., Olalekan, A., Ashagrie, M., & Bitew, M. (2022). Informatics in Medicine Unlocked Development of a chickpea disease detection and classification model using deep learning. Informatics in Medicine Unlocked, 31(May), 100970. [CrossRef]
  14. Kunaver, M., & Tasič, J. F. (2005). Image feature extraction - An overview. EUROCON 2005 - The International Conference on Computer as a Tool, I, 183–186. [CrossRef]
  15. Kundu, N., & Rani, G. (n.d.). Seeds Classification and Quality Testing Using Deep Learning and YOLO v5.
  16. Li, Z., Gong, B., & Yang, T. (2016). Improved dropout for shallow and deep learning. Advances in Neural Information Processing Systems, Nips, 2531–2539.
  17. Liu, Z., Jiang, J., Li, M., Yuan, D., Nie, C., Sun, Y., & Zheng, P. (2022). Identification of Moldy Peanuts under Different Varieties and Moisture Content Using Hyperspectral Imaging and Data Augmentation Technologies. Foods, 11(8). [CrossRef]
  18. Livieris, I. E., Iliadis, L., & Pintelas, P. (2021). On ensemble techniques of weight-constrained neural networks. Evolving Systems, 12(1), 155–167. [CrossRef]
  19. Manhando, E., Zhou, Y., & Wang, F. (2021). Early Detection of Mold-Contaminated Peanuts Using Machine Learning and Deep Features Based on Optical Coherence Tomography. AgriEngineering, 3(3), 703–715. [CrossRef]
  20. Mohanaiah, P., Sathyanarayana, P., & Gurukumar, L. (2013). Image Texture Feature Extraction Using GLCM Approach. International Journal of Scientific & Research Publication, 3(5), 1–5.
  21. Motwani, M., Motwani, R., & Harris, F. (2004). Survey of image denoising techniques Survey of Image Denoising Techniques. January.
  22. Padilla, R., Netto, S. L., & Silva, E. A. B. (2020). A Survey on Performance Metrics for Object-Detection Algorithms. July. [CrossRef]
  23. Qi, H., Liang, Y., Ding, Q., & Zou, J. (2021). Automatic Identification of Peanut-Leaf Diseases Based on S tack Ensemble.
  24. Rajendran, S., Dorothy, R., Joany, R. M., Rathish, R. J., Santhana Prabha, S., & Rajendran, S. (2015). Image enhancement by Histogram equalization Image enhancement by Histogram equalization Image enhancement by Histogram equalization. Int. J. Nano. Corr. Sci. Engg, 2(4), 21–30.
  25. Rokibul, K., & Nazmul, A. (2020). A dynamic ensemble learning algorithm for neural networks. Neural Computing and Applications, 32(12), 8675–8690. [CrossRef]
  26. Sarvamangala, C., Gowda, M. V. C., And, & Varshney, R. K. (2011). This is the author's version of the postprint archived in the official repository of ICRISAT Identification of quantitative trait loci for protein content, oil content, and oil quality for groundnut ( Arachis hypogaea L.). 122(1), 49–59.
  27. Schwenker, F. (2013). Ensemble Methods: Foundations and Algorithms [Book Review]. IEEE Computational Intelligence Magazine, 8(February), 77–79. [CrossRef]
  28. Shasidhar, Y., Vishwakarma, M. K., Pandey, M. K., Janila, P., Variath, M. T., Manohar, S. S., Nigam, S. N., Guo, B., & Varshney, R. K. (2017). Molecular mapping of oil content and fatty acids using dense genetic maps in groundnut (Arachis hypogaea L.). Frontiers in Plant Science, 8(May), 1–14. [CrossRef]
  29. State, O., & State, O. (2007). Fruit Morphological Characterization among Some Varieties of. 1(2), 155–160.
  30. Szczypinski, P. M., Klepaczko, A., & Kociolek, M. (2017). Barley defects identification. International Symposium on Image and Signal Processing and Analysis, ISPA, September, 216–219. [CrossRef]
  31. Veni, N., & Manjula, J. (2022). High-performance visual geometric group deep learning architectures for MRI brain tumor classification. Journal of Supercomputing, 78(10), 12753–12764. [CrossRef]
  32. Wang, Y., Ding, Z., Song, J., Ge, Z., Deng, Z., Liu, Z., Wang, J., Bian, L., & Yang, C. (2023). Peanut Defect Identification Based on Multispectral Image and Deep Learning. Agronomy, 13(4). [CrossRef]
  33. Wang, Z., Bovik, A. C., Sheikh, H. R., & Simoncelli, E. P. (2004). Image quality assessment: From error visibility to structural similarity. IEEE Transactions on Image Processing, 13(4), 600–612. [CrossRef]
  34. Wyatt, M., Radford, B., Callow, N., Bennamoun, M., & Hickey, S. (2022). Using ensemble methods to improve the robustness of deep learning for image classification in marine environments. 2022(February), 1317–1328. [CrossRef]
  35. Yang, H., Ni, J., Gao, J., Han, Z., & Luan, T. (2021). A novel method for peanut variety identification and classification by Improved VGG16. Scientific Reports, 11(1), 1–17. [CrossRef]
  36. Yang, Y., Lv, H., & Chen, N. (2022). A Survey on ensemble learning under the era of deep learning.
  37. Yogamangalam, R., & Karthikeyan, B. (2013). Segmentation techniques comparison in image processing. International Journal of Engineering and Technology, 5(1), 307–313.
  38. Ziyaee, P., Ahmadi, V. F., Bazyar, P., & Cavallo, E. (2021). Comparison of different image processing methods for segregation of peanut (Arachis hypogaea L.) seeds infected by aflatoxin-producing fungi. Agronomy, 11(5). [CrossRef]
  39. Zou, Z., Chen, J., Wang, L., Wu, W., Yu, T., Wang, Y., Zhao, Y., Huang, P., Liu, B., Zhou, M., Lin, P., & Xu, L. (2022). Nondestructive detection of peanuts mildew based on hyperspectral image technology and machine learning algorithm. Food Science and Technology (Brazil), 42, 1–11. [CrossRef]
Figure 1. proposed model architecture.
Figure 1. proposed model architecture.
Preprints 115091 g001
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

© 2024 MDPI (Basel, Switzerland) unless otherwise stated