Next Article in Journal
Tracing the Evolution of the SEPALLATA Subfamily across Angiosperms Associated with Neo- and Sub-Functionalization for Reproductive and Agronomically Relevant Traits
Previous Article in Journal
Responses of Forest Net Primary Productivity to Climatic Factors in China during 1982–2015
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Tomato Leaf Disease Recognition on Leaf Images Based on Fine-Tuned Residual Neural Networks

by
Paul Shekonya Kanda
,
Kewen Xia
*,
Anastasiia Kyslytysna
and
Eunice Oluwabunmi Owoola
School of Electronics and Information Engineering, Hebei University of Technology, Tianjin 300401, China
*
Author to whom correspondence should be addressed.
Plants 2022, 11(21), 2935; https://doi.org/10.3390/plants11212935
Submission received: 15 August 2022 / Revised: 19 September 2022 / Accepted: 28 September 2022 / Published: 31 October 2022
(This article belongs to the Section Plant Modeling)

Abstract

:
Humans depend heavily on agriculture, which is the main source of prosperity. The various plant diseases that farmers must contend with have constituted a lot of challenges in crop production. The main issues that should be taken into account for maximizing productivity are the recognition and prevention of plant diseases. Early diagnosis of plant disease is essential for maximizing the level of agricultural yield as well as saving costs and reducing crop loss. In addition, the computerization of the whole process makes it simple for implementation. In this paper, an intelligent method based on deep learning is presented to recognize nine common tomato diseases. To this end, a residual neural network algorithm is presented to recognize tomato diseases. This research is carried out on four levels of diversity including depth size, discriminative learning rates, training and validation data split ratios, and batch sizes. For the experimental analysis, five network depths are used to measure the accuracy of the network. Based on the experimental results, the proposed method achieved the highest F1 score of 99.5%, which outperformed most previous competing methods in tomato leaf disease recognition. Further testing of our method on the Flavia leaf image dataset resulted in a 99.23% F1 score. However, the method had a drawback that some of the false predictions were of tomato early light and tomato late blight, which are two classes of fine-grained distinction.

1. Introduction

We rely on edible plants in the same way that we rely on oxygen. There is no food without crops, and there is no life without food. It is no coincidence that the invention of agriculture coincided with the rise of human civilization [1]. The tomato (Solanum lycopersicum L.), family Solanaceae, originated in the Andean region of South America has, according to the Food and Agriculture Organization Statistics (FAOSTAT), in the past fifty years become one of the most important and extensively grown horticultural crops in the Mediterranean region and throughout the world. Currently, it is the world’s second most cultivated vegetable crop after the potato, with approximately 181 million tonnes from 5 Mha [2]. With 0.2 Mha, it is the highest-yielding vegetable in Southern Europe, and the major producers in the Mediterranean basin are Turkey, Egypt, Italy, Spain, and Morocco [2,3].
The tomato is susceptible to a variety of plant diseases caused by pathogens such as fungal, bacterial, phytoplasma, virus, and viroid pathogens due to its genetic properties as shown in Table 1. Not only is its genetic inheritance critical to managing the numerous tomato pathogens, but so are current climate changes, recently revised phytopathological control measures, and seed industry globalization [4]. One of the common diseases affecting tomato yield, the Septoria leaf spot, is caused by a fungal pathogen. Septoria has emerged as a major emerging pathogen as a result of climatic change and widespread variability. The pathogen’s disease severity ranges between 35 and 65 percent in both cultivated and non-cultivated crops, posing a serious threat in the future [5]. The pathogen’s complex adaptability combined with its cosmopolitan nature makes it more vigorous by targeting new susceptible hosts and coupling this with increased viability within its infection cycle [5]. Table 1 lists some tomato plant pathogens present in the Mediterranean basin [4].
The cultivated tomato has a low genetic diversity due to its intensive selection and severe genetic bottlenecks that arose during evolution and domestication [10,11,12]. For these reasons, the tomato is more prone to a high disease incidence, and during the cultivation and post-harvest period, it can be affected by more than 200 diseases caused by different pathogens throughout the world [13,14]. In this paper, we propose a method for effective recognition of diseases affecting the tomato, that are mainly reflected on the leaves. We propose a residual neural network algorithm for this, which is the state-of-the-art, most recent deep learning image recognition algorithm. The efficient recognition of such can inform the farmers of the presence of such diseases on their crops and allow them to carry out control measures currently authorized in the EU that allow growers to achieve a successful and eco-sustainable disease management of this vegetable crop, fundamental for the Mediterranean diet.
One of the most important research areas in precision agriculture is disease identification using images of plant leaves [15]. Artificial intelligence, image processing, and graphical processing unit advancements have the potential to broaden and improve the practice of precise plant protection and growth. Most plant diseases produce a variety of visible symptoms; thus, learning models should be able to adequately observe and identify the distinctive symptoms of any disease [16].
Tomato leaf disease identification falls under the purview of computational agriculture [17,18]. Traditional methods for identifying tomato leaf disease at an early stage frequently use global features (such as color, texture, and shape) to describe the characteristics of disease spots in crop leaf disease images. The methods used served to separate the diseased and normal parts of the tomato leaf, and the area ratio of the two parts was used as the criterion for identification [19].
Recent advancements in deep learning provide obvious benefits in feature extraction and recognition, such as the convolutional neural network (CNN), which automatically trains the network to extract data features by introducing local connections and weight sharing in the training process. Additionally, significant progress has been made in the identification of diseases in plants such as apples, bananas, cucumbers, and tomatoes [20,21].
However, the recognition effect varies due to the structural differences of the various recognition models that make up the state-of-the-art model. Additionally, there exists the challenge of choosing suitable hyperparameters for training networks on these leaf datasets while leveraging the richly learned weights that some state-of-the-art models have gotten while learning from very large datasets. Additionally, different layers need to learn at different learning rates, for an optimal learning experience.
As a result, the primary aim of this research is to achieve high recognition accuracy for tomato leaf diseases; to achieve this, a residual neural network will be utilized, and the following will be studied and modified for the proposed network:
  • A double form of data augmentation, using image transformations and the implementation of CutMix as a secondary form of data augmentation for model generalization.
  • The effect of the train/test data split size ratio of our dataset on the network model for disease recognition on tomato leaf images. Train/test data split ratios of sizes 40/60, 50/50, 60/40, 70/30, and 80/20 were adopted and studied.
  • The effect of varying batch sizes in training our network to correctly recognize tomato leaf diseases. According to the capacity of the GPU available, batch sizes of 40, 50, 60, 70, 80, 90, and 100 were adopted.
  • The role of network depth in the effective recognition of tomato leaf disease. Residual networks with varying depths of 18, 34, 50, 101, and 152 layers were studied.
  • The effect of tuning the learning rate while training the network and identification of a threshold to obtain suitable learning rates to effectively train the network to recognize tomato leaf diseases. The implementation of a discriminative learning rate for efficient training of residual models.

2. Related Work

Durmus et al. [22] used AlexNet [23] and SqueezeNet [24] models to classify and recognize 10 different types of tomato diseases in the PlantVillage dataset. The experiment discovered that while AlexNet’s classification accuracy is slightly higher than SqueezeNet’s, the size of the model and the time required are doubled.
Aravind et al. [25] used AlexNet and VGG16 [26] in conjunction with transfer learning to identify seven types of tomato diseases; the experiment revealed that the accuracies were 97.29 and 97.49 percent, respectively. Although transfer learning can accelerate model convergence and improve recognition performance, it is constrained by the original network structure.
Karthik et al. [27] proposed an attention-based deep residual network for detecting the type of tomato leaf infection. The PlantVillage dataset was used in the experiment, with 95,999 images used as training models and 24,001 images used for validation. The diseases included in the dataset were the tomato early blight, late blight, and leaf mold. The experimental results demonstrated that the proposed attention-based residual network can use CNN learning features at different processing levels and achieve 98 percent overall accuracy on the validation set in five-fold cross-validation.
Anand et al. [28] proposed an image processing and machine learning-based technique for diagnosing brinjal leaf disease. They used a K-means clustering technique to segment brinjal leaf diseases with some remarkable performance. In 2018, Zhang et al. [29] proposed a K-means clustering and PHOG algorithms-based fusion of super-pixel clustering-based leaf segmentation. Their technique performed admirably in the segmentation and recognition of plant leaf diseases. By extracting features based on color and texture and feeding them to a multiclass SVM classifier, Rani et al. [30] also proposed a K-means clustering-based leaf disease and classification technique. On average, they recorded a classification accuracy of 95%. In the same year, Kumari et al. [31] proposed an image processing-based leaf spot recognition system, with the four stages of image acquisition, image segmentation, feature extraction, and classification. To compute the disease features, they also utilized the K-means algorithm. They achieved an accuracy of 90% and 80% for bacterial leaf spot and cotton leaf disease target spot, respectively. Liu et al. [32] proposed a leaf disease identification model based on generative adversarial networks. This model employed DenseNet and instance normalization to recognize actual and false disease images, as well as the feature extraction capability on grape leaf lesions. Finally, the approach applied a deep regret gradient penalty to stabilize the training process. The findings revealed that the GAN-based data augmentation strategy may effectively overcome the overfitting problem in disease identification while simultaneously improving accuracy. A leaf disease detection approach based on the AlexNet architecture was proposed by Lv et al. [33] in 2020. First, they created a maize leaf feature enhancement framework, which improved the capability of feature extraction combined with dilated convolution and multiscale convolution in a complex environment. After that, a DMS-Robust AlexNet architecture network was created, which improved the capability of feature extraction combined with dilated convolution and multiscale convolution in a complex environment. The disease features on tomato leaves, such as spot blight, late blight, and yellow leaf curl disease, were extracted using a deep learning method by Jiang et al. [34]. After continuous iterative learning, the proposed technique correctly predicted the disease category for each disease, with accuracy increases of 0.6 percent and 2.3 percent in the training and test sets, respectively. Waheed et al. [35] proposed an optimized DenseNet-based maize leaf recognition model with few parameters to boost job efficiency. The results of the experiments demonstrated that this technology is effective at detecting corn leaf disease. Huang et al. [36] proposed an end-to-end plant disease diagnostic model-based deep neural network, which can reliably classify plant types and plant diseases. Their model consists of two components: the leaf segmentation part that separates the leaves in the original image from the background; and the plant disease classifier, which is based on a two-headed network that classifies plant diseases using features extracted by multiple common pre-trained models. Experimental results show that this method can achieve a plant classification accuracy of 0.9807 and a disease recognition accuracy of 0.8745. In 2020, [37] proposed a combination of ABCK-BWTR and B-ARNet models for the identification of tomato leaf disease, consisting of a channel attention module in a ResNet50 [38], using the dual channel filter to extract the primary leaf features.
Sethy et al. [39] used different deep learning models for extracting rich features and applied an SVM classifier to classify them. They achieved their highest performance accuracy with a combination of a ResNet50 model with an SVM classifier. Oyewola et al. [40] in their work proposed using plain CNNs(PCNN) and deep residual network (DRNN) in identifying five different cassava plant diseases; their results showed that PCNN was outperformed by DRNN by a margin of +9.25%. Zeng et al. [41], on the other hand, proposed a self-attention convolutional neural network (SACNN) to identify several crop diseases. To examine the robustness of their model, the authors introduced noise at different levels in the test images. Diseases prone to affect rice and maize leaves were identified by Chen et al. [42] using an INC-VGGN method. They replaced the last convolutional layer of a VGG19 model with two inception layers and a global average pooling layer. Maize, apple, and grape crop diseases were identified by Yang et al. [43] using a shallow CNN (SCNN) embedded with SVM and RF classifiers. A transfer-learning approach was adopted by Ramacharan et al. [44] to identify three diseases and two pest-damage types that plague cassava plants. The authors further extended their work by implementing a smartphone-based CNN model for the identification of cassava plant diseases and recorded an 80.6% accuracy [45]. Adedoja et al. [46] proposed a deep CNN architecture based on NASNet to identify diseases on some plant leaves with an accuracy of 93.82%. However, there is still a need for improvement in the accuracy of plant disease recognition.
In 2022, an attention-based method was proposed by Devi et al. [47] where they used the Salp Swarm algorithm in the classification of tomato leaf diseases. Their method achieved an accuracy of 97.56% in predicting five types of tomato leaf diseases from leaf images taken from the plant village dataset. Apart from the computational complexity of the method, it is also limited in performance score. A method that utilized a lightweight attention-based CNN [48] to classify tomato leaf diseases achieved a model accuracy of 99.34% but with a slightly higher time complexity than conventional methods. Also in 2022, Zhao et al. [49] developed a method that utilized a spatial attention mechanism with CNN for real-time leaf disease detection. However, this method achieved a 95.20% accuracy and did not generalize well. With the aim of improving performance and generalizability, our method was researched. We also tested the proposed method on another plant leaf benchmark dataset that is different from our target dataset.

3. Evaluation Metrics, Results, and Discussion

This section presents the metrics used in evaluating the results of this research, the detailed results, and relevant discussion.

3.1. Evaluation Metrics

The accuracy, precision, recall, and f1-score of the proposed method were all evaluated. The proposed plant recognition system’s accuracy has been calculated using the following expression, which incorporates numerical details such as true positive (TP) (the number of correctly identified leaf images), false positive (FP) (the number of incorrectly detected leaves), true negative (TN) (the number of correctly detected leaf images), and false negative (FN) (it is a parameter for representation of the number of leaf images that are correctly recognized).
  • Accuracy: Accuracy is the number of right predictions that are made by the model with respect to the total number of predictions that were made. It is mathematically represented by Equation (1).
Accuracy = TP + TN TP + TN + FP + FN  
  • Precision: Precision is defined as the number of true positive results (TP) divided by the number of positive results (TP + FP) that are predicted by the model. The range of the precision is between 0 and 1 and is calculated using Equation (2). It is used to find the proportion of positive identifications that is true.
  • Recall: The recall is the number of true positives (TP) divided by the number of all relevant sample data (TP + FN). Equation (3) represents the mode of calculation of the recall. It is used to determine the proportion of actual positives that were correctly identified. These concepts are represented mathematically by Equations (2) and (3), respectively:
Precision   = TP TP + FP  
Recall   = TP TP + FN  
  • F1 Score: Being one of the widely used metrics for the performance evaluation of machine learning algorithms, the F1 score is the harmonic mean of precision and recall. The range of the F1 score is between 0 and 1, and it is calculated as shown by Equation (4). It reflects the number of instances that are correctly classified by the learning model.
F 1 = 2 × P × R P + R  

3.2. Results and Discussion

3.2.1. Results on Varied Network Depth

Given five different network depths adopted in this research, 18, 34, 50, 101, and 152 layers, respectively, this section reports the results and discusses the findings on the relationship between the network depth and the performance of the proposed network. The results of the F1 score based on the different depths of the proposed residual neural network are displayed in Figure 1, while Figure 2, Figure 3, Figure 4 and Figure 5 show the performance via confusion matrices of the various network depths.
Figure 1 displays the plot of epoch over F1 score for a network of the five varying depths (18, 34, 50, 101, and 152) on the maximum batch size used in this research, having a value of 100. The results show that a train-validation split ratio of 80/20 recorded the highest performance in the F1 score. After 29 epochs, the result shows how the performance score reached a peak of 98% and a minimum of 94% based on different train-validation data split ratios as indicated on the plot image.
It was observed that the network depth affected the network performance, though not at a very high value. However, the network depth of 152, being the highest depth used in this research, had the highest performance score of 99.51% as shown in Figure 1e above.

3.2.2. Results on Varied Train-Validation Data Split Ratios

Different train-validation data split ratios were tested on this proposed network, ranging from 40/60, 50/50, 60/40/ 70/30, and 80/10 for training and validation data, respectively. This section displays and discusses the results obtained.
Table 2 shows the relationship between batch size and performance on the network on different train-validation data split ratios.
The results of the F1 score for different train-validation data split ratios on the proposed residual neural network are displayed in Figure 6 and Figure 7 below.
The results in Figure 6 and Figure 7 show the relationship between the train-validation data split ratio and the F1 score of the networks. These results suggest that the split ratio had a great impact on the performance of the network. Figure 7 shows a clear distinction in the performance value as the train samples are increased.
For a batch size of 100 images, being the highest batch size value adopted for this research, train set values of 40, 50, 60, 70, and 80% of the entire dataset resulted in a performance of 0.9775, 0.9848, 0.9884, 0.9921, and 0.9937, respectively, out of a total value of 1, for a network depth of 152. The results on other network depths also show a similar pattern in performance with such data split ratios. This suggests that a split ratio of 80/20 is a good choice for plant leaf image recognition.

3.2.3. Results on Different Batch Sizes

The effect of different batch sizes on the various models was studied and the results are described here. The results of the F1 score for different train-validation data split ratios on the proposed residual neural network are displayed in Figure 8 below. It displays the plot of the epoch over the F1 score for a batch size of 100, 90, 80, 70, 60, 60, and 50 images, respectively.
The results of the different batch sizes above do not show much difference in the overall performance of the F1 score at the end of the number of training epochs; however, the time taken was greatly influenced as is displayed in the next section.

3.2.4. Results on Computing Time

The results of time on the proposed residual neural network for plant leaf recognition are displayed in Figure 9, and Table 3 shows the time taken on various train-validation split ratios and batch sizes.
From Figure 9 above, the fastest of the networks was that of a depth of 18 layers, with a train-validation data split ratio of 40/60. Having more data for validation than for training made the network training and testing procedure faster, albeit our goal was not just for speed but also improved performance. Hence, the result of performance with the fastest time was not the optimized result we have. The split ratio of 80/20 was rather that with the highest performance in training and testing, as recorded in Table 2.
Figure 10 shows the plot of the test split ratio and layer depth over time. This displays a more elaborate view of the relationship between time and other parameters such as split ratio and network depth.
The F1 score, being a more robust metric for recognition, combined both the recall and precision levels of the network. Here, we show how the network tried to maintain a considerably consistent F1 score for most of the training process with a +0.15 and −0.25 interval. The learning rate on this dataset had been seen to do well from 10−5 to 10−3. The best position to train a model has been estimated to fall along that axis. Afterward, as can be seen in the plot, as the learning rate increased, the model loss increased. Table 4 compares the performance of the proposed network on the PlantVillage dataset on various training and validation data split ratios. Figure 11 shows the areas of mistake recorded by the model in this research: (a) the plot loss for a network depth of 152; (b) the images the model predicted wrongly for a network depth of 152, detailing the predicted, ground truth, loss, and probability values.
The network, though having an outstanding performance, was not 100% perfect. From Figure 11 above, classes that were wrongly recognized are displayed. The first image shows how the model predicted a late blight leaf as an early blight class. Table 4 shows the result of the network on various train-validation split ratios. The early blight and late blight diseased leaves had a very striking resemblance and as such, most of the wrong predictions recorded by our method happened to fall in between the two leaf classes. Additionally, some of the spots that could be found on leaves were so close that they could constitute a challenge in perfect distinction, giving rise to an imperfect recognition, even to the human eye. However, more research on fine-grained distinct images is needed to get to that point.
Figure 12 below shows the plot of loss against epoch for the network of depth 152 layers on a train-validation split ratio of 40/60.
Table 5, Table 6, Table 7, Table 8 and Table 9 detail the performance of the proposed network on the PlantVillage dataset with various model depths of the residual neural network architecture on six different training and validation data split ratios. Table 5 compares the performance of our model on various model depths of the same architecture on a validation data split ratio of 60.
Table 6 compares the performance of our model on various model depths of the same architecture on a validation data split ratio of 50.
Table 7 compares the performance of our model on various model depths of the same architecture on a validation data split ratio of 40.
Table 8 compares the performance of our model on various model depths of the same architecture on a validation data split ratio of 30.
Table 9 compares the performance of our model on various model depths of the same architecture on a validation data split ratio of 20.

3.2.5. Benchmark against Other Models

A summary of the related work carried out on the Flavia plant leaf image datasets and our result comparison is shown in Table 10.
A summary of the related work on plant disease identification based on leaf images and the result comparison of some of them is shown in Table 11. As can be seen from both Table 11 and Figure 13, our model outperformed the previous models, surpassing that of Li et al. [58], which was the closest in performance, with a +0.75% performance gain. Whereby some authors used the accuracy metric to measure their performance, we recorded a much higher accuracy but chose to benchmark against our F1 score value, which is regarded as a much better form of performance measure for classification problems, since it combines both the precision and recall of the model in question.
Figure 13 below shows the chart of a comparison of our proposed technique and other techniques. A detailed summary of the technique, dataset and year of publication is just as displayed in Table 11 above. More discussion on the image is given in the Discussion section of this article.

4. Materials and Methods

4.1. Data Acquisition and Pre-Processing

4.1.1. Datasets

  • The Flavia leaf dataset
The Flavia leaf dataset (download link: http://flavia.sourceforge.net/ (accessed on 13 November 2021)), introduced by Wu et al. [75], contains 1907 leaf images of size (1600 × 1200 pixels) obtained from 32 plant species on a white background, containing about 50–77 images per class. Figure 14 shows random sample images from the Flavia dataset used for this research work.
2.
The tomato leaf dataset
The tomato leaf dataset that was used in this research consists of images of diseased and healthy tomato plant leaves that were obtained from the publicly available PlantVillage [1] dataset, which is an open-access public resource for agriculture-related content. The entire dataset contains 54,306 images of plant leaves, which have a spread of 38 class labels assigned to them. However, the experiments in this research were narrowed down to only images of tomato plant leaves, which include nine types of tomato leaf diseases, and some healthy tomato leaves, making a total of 10 different categories for our research out of the entire 38 which are obtainable in the entire PlantVillage database. Our crop-specific data contained 18,160 images of tomato plant leaves, and each class is defined either with the corresponding name of the disease affecting the leaves in it or categorized as part of the healthy class. Figure 15 shows sample images from the PlantVillage dataset used for this research work.
From Figure 15, the various diseases can be categorized either as fungi, bacteria, mold, viruses, or mites. Four of the diseases, namely, early blight, late blight, leaf spot, and target spots are caused by fungi; the bacterial spot is caused by bacteria, the leaf mold is the cause of a mold disease, while both the tomato yellow leaf curl and the tomato mosaic are viral infections, and the spider mite is a mite disease. A brief description of each of these diseases is given below:
  • Early blight is a fungal infection, and symptoms start as oval-shaped lesions with a yellow chlorotic region across the lesion; concentric leaf lesions may be seen on infected leaves.
  • Late blight, being another fungal infection, affects all aerial parts of the tomato plant; initial symptoms of the disease appear as water-soaked green to black areas on leaves which rapidly change to brown lesions; fluffy white fungal growth may appear on infected areas and leaf undersides during wet weather.
  • The leaf spot is another fungal infection. Infected plants exhibit bronzing or purpling of the upper sides of young leaves and develop necrotic spots; leaf spots may resemble those caused by bacterial spots, but a bacterial ooze test will be negative; leaves may cup downwards, and shoot tips may begin to die back.
  • Septoria leaf spot is yet another fungal disease. Symptoms may occur at any stage of tomato development and begin as small, water-soaked spots or circular grayish-white spots on the underside of older leaves; spots have a grayish center and a dark margin, and they may coalesce.
  • Leaf mold is still another fungal infection. The older leaves exhibit pale greenish to yellow spots (without distinguishable margins) on the upper surface, whereas, the lower portion of these spots exhibits green to brown velvety fungal growth. As the disease progresses, the spots may coalesce and appear brown. The infected leaves wither and die but stay attached to the plant.
  • Bacterial spots are bacterial diseases, and lesions start as small water-soaked spots; lesions become more numerous and coalesce to form necrotic areas on the leaves giving them a blighted appearance; leaves drop from the plant, and severe defoliation can occur leaving the fruit susceptible to sunscald; mature spots have a greasy appearance and may appear transparent when held up to a light source; centers of lesions dry up and fall out of the leaf; blighted leaves often remain attached to the plant and give it a blighted appearance.
  • Spider mites (two-spotted spider mites). Leaves stippled with yellow; leaves may appear bronzed; webbing covering leaves; mites may be visible as tiny moving dots on the webs or underside of leaves, best viewed using a hand lens; usually not spotted until there are visible symptoms on the plant; leaves turn yellow and may drop from the plant.
  • Target spot is also caused by a fungus. The fungus infects all parts of the plant. Infected leaves show small, pinpoint, water-soaked spots initially. As the disease progresses, the spots enlarge to become necrotic lesions with conspicuous concentric circles, dark margins, and light brown centers. Whereas the fruits exhibit brown, slightly sunken flecks in the beginning, later the lesions develop a large, pitted appearance.
  • Tomato mosaic virus is a viral infection. Symptoms can occur at any growth stage and any part of the plant can be affected; infected leaves generally exhibit a dark green mottling or mosaic; some strains of the virus can cause yellow mottling on the leaves; young leaves may be stunted or distorted; severely infected leaves may have raised green areas; dark necrotic streaks may appear on the petioles’ leaves.
  • Tomato yellow leaf curl disease is another viral infection. The infected leaves become reduced in size, curl upward, appear crumpled, and show yellowing of veins and leaf margins.

4.1.2. Data Pre-Processing

1.
Data Augmentation 1—Image Transformations
Pre-processing steps are applied to cleanse and organize data before being fed into the model. By introducing a few distorted images into the training dataset, image transformations are used to increase the number of images in the dataset and reduce the chances of the model overfitting. The augmented images for the training data are created using standard image augmentation techniques such as flipping, Gamma correction, noise injection, PCA color augmentation, rotation, and scaling transformations. The images are each further resized to 128 before being fed to the training model.
2.
Data Augmentation 2—CutMix
Let x     W × H × C and y denote both the training image and the corresponding label of our tomato leaf image, respectively. The goal of the CutMix augmentator is to generate a new training sample ( x ˜ , y ˜ ) through the combination of two training samples, ( x A , y A ) and ( x B , y B ) . The newly generated sample ( x ˜ , y ˜ ) is then used in training the model with its original loss function.
We define the combining operation as
x ˜ = M   ʘ   x A + ( 1 M )   ʘ   x B y ˜ = λ y A + ( 1 λ ) y B ,  
where M ∈ {0, 1}W×H denotes a binary mask indicating where to drop out and fill in from two images, 1 is a binary mask filled with ones and is an element-wise multiplication, and the combination ratio λ between two data points is sampled from the beta distribution Beta(α, α).
CutMix replaces an image region with a patch from another training image and generates locally natural images. CutMix is simple and incurs a negligible computational overhead as with existing data augmentation techniques; we can efficiently utilize it to train any network architecture. To sample the binary mask M, the bounding box coordinates B = ( r x , r y , r w , r h ) are first sampled indicating the cropping regions on x A and x B . The region B in x A is removed and filled in with the patch cropped from B of x B . The box coordinates are uniformly sampled according to:
r x   Unif     ( 0 ,   W ) ,       r w = W 1 λ , r y   Unif     ( 0 ,   H ) ,       r h = H 1 λ ,  
making the cropped area ratio:
r w   r h W   H = 1 λ  
With the cropping region, the binary mask M ∈ {0, 1}W×H is decided by filling with 0 within the bounding box B, otherwise 1. In each training iteration, a CutMix-ed sample ( x ˜ , y ˜ ) is generated by combining two randomly selected training samples in a mini-batch according to Equation (5). Figure 16 shows a visualization of the CutMix operation on the Flavia dataset. The mixture of patch images on different class images can be seen.

4.2. Our Proposed Method

4.2.1. Convolutional Neural Networks

A typical convolutional neural network architecture will contain some basic building blocks, some of which are referred to as layers. These make up the building block for our proposed network approach and are described in the following subsection.
1.
Convolutional Layer
In this layer, convolutional operations are performed on the input to learn useful features. To this effect, a convolutional kernel slides along the input image with a certain stride and outputs convolution plus a bias generally known as a feature map. The input to this layer could be either an RGB image or the output feature of a preceding layer for a multilayer network. A convolutional kernel means that given an input image when it is processed, the weighted average of pixels in a small area of the input image becomes each corresponding pixel in the output image, and the weight is defined by a function, and hence, they share weights to reduce parameters in the network. This process can be expressed mathematically as:
X j l = f ( i M j Y i l 1 K i j l + b j l )
where Y i l 1 is the output of the i-th feature map in the l 1 layer, and X j l is the input of the j t h feature map in the l layer. K i j and b j l are the convolutional kernel and bias in the l layer. f ( . ) is the activation function. Higher-level unique features can be identified through a series of increased convolution layers; hence, the need to go deeper.
2.
Pooling Layer
After the convolution operation, there is a need to reduce the dimensions of the image for further processing. This process is known as down sampling or simply a pooling operation. This process can be expressed mathematically as:
X j l = f ( β j l down s ( X j l 1 ) + d j l )
where X j l represents the j t h feature map in the l layer. β j l and d j l are the multiplicative factor and bias, respectively. down s represents an under-sampling function. Under-sampling can be done in many forms, some of which are average pooling, maximal pooling (max pool), minimal pooling operation, and so on. In our work, we employed max pooling.
3.
Fully Connected Layer
Each neuron in the fully connected layer is connected to all neurons in the feature map of the previous layer, and the output can be expressed as:
h W , b ( x ) = f ( W T + b )
where h W , b ( x ) is the output, and W represents the corresponding weights of the network. The inputs to the fully connected layer are mainly features extracted from the preceding layer. Each feature in the former layer represents different semantic information that is unique and important for the next layer.

4.2.2. Transfer Learning Approach

The goal of transfer learning is to improve the target learners’ performance on target domains by transferring useful knowledge [76,77] from disparate but related source domains to the target at a lower computational cost. The reliance on a large number of target domain data for constructing target learners can thus be reduced. It has emerged as a popular and promising area of machine learning due to its wide range of application possibilities, especially in solving real-world problems [78,79,80,81] in a cheaper and more reliable method. The sphere of use of transfer learning is not few, coupled with its record of high results [82,83].
An approach in transfer learning is that the last few layers of the pre-trained network are replaced with new layers, such as a fully connected layer and a softmax classification layer, with the number of classes set to be equivalent to that of the new target dataset, which in our research is 10 for the number of tomato leaf classes. In our research, all the networks used were pre-trained on the imagenet dataset before they were re-trained to learn the features of the tomato leaf image dataset in order to correctly recognize them based on their different classes.
However, the challenge of maximizing its use with regard to modifying the learning rate to suit the target data still exists. It is discovered that one learning rate being used across all the layers of a model in transfer learning is not a good practice for transfer learning [84]. In each model, we unfreeze the layer and add a stack of one activation layer, one batch-normalization layer, and one dropout layer. All models are tested with the same dropout values, different learning rates, and varying batch sizes.

4.2.3. Overall Architecture of the Proposed Method

The proposed research network architecture consists of five networks designed similarly but with a different number of layers. Each layer consists of both an identity and a convolutional block. The identity block is the standard block used in ResNets and corresponds to the case where the input activation and output activation have the same dimensions. Figure 17 shows the entire workflow for the tomato leaf disease recognition system.
The identity block is as shown in Figure 18, in which two hidden layers are skipped to avoid the vanishing gradient problem. The purpose of this block is to match input and output dimensions. The purpose of the identity block is to resize the input to a different dimension. Figure 18 shows the residual identity mapping for the residual neural network which is the basic building block for the network used in this research, while Figure 19 shows a simplified block diagram for the arrangement of the various blocks for the residual neural network [60] network used in this research.
Training deep learning architectures, whether it is a new model from scratch or via transfer learning is not without its difficulties. Learning algorithms based on artificial neural networks, and particularly deep learning for computer vision, may appear to include a plethora of bells and whistles, popularly known as hyperparameters [85]. To successfully train and debug them, more hyperparameters should be adjustable, and this has made it possible for one to obtain more interesting results. Making sure we have the right learning rate is one of the most important things we can do when training a model. If our learning rate is too low, training our model may take many, many epochs. For this research, all our five networks were pre-trained with the ImageNet dataset and then fine-tuned in the manner described below for re-training on our tomato leaf images.
We implement discriminative learning rates [86] in the process of training our model on plant leaf images. Discriminative fine-tuning is a fine-tuning strategy introduced by universal language model fine-tuning (ULMFiT) [86] for the implementation of natural language processing (NLP). Instead of the same learning rate being used across all the layers of our model, this allows us to tune each layer with a different learning rate that is efficient and suits it better than any other learning rate. Since the deepest layers of pre-trained models may not require so high a learning rate as the final ones will require, we will use different learning rates for them as described by [84,87].
To preserve the quality of those pre-trained weights even after unfreezing them, some of the added parameters are tuned for a few epochs, and naturally, we would not expect the best learning rate for those pre-trained parameters to be as high as the best learning rate for the randomly added parameters. Keep in mind that the pre-trained weights were trained on millions of images over hundreds of epochs, so they have learned rich features from being previously trained on large datasets. Our experiment was carried out with a few specific detailed goals in mind, they are:
  • To research the role of depth on the performance of such models, various residual convolutional neural networks, having different layers, such as ResNet-18, ResNet-34, ResNet-50, ResNet-102, and ResNet-152 were tested.
  • Various train/test data split ratios were experimented on to determine the optimum value of the train/test split ratio for such a research area.
  • Different batch sizes were selected based on the capacity of the GPU system obtainable in the laboratory to test for the influence of batch size on the training and if so on the result of the training and test processes of the model.
  • The discriminative learning process was studied to determine the best learning rates to select in re-training the various models to achieve an optimal training process from one domain of data to a new domain of datasets: specific to this research is the tomato plant leaf dataset for disease recognition.

4.3. Training Procedure

4.3.1. Tuning the Learning Rate Schedule

We began with an extremely low learning rate, and used it for one mini-batch, then determined the losses and increased the learning rate by a certain percentage, i.e., doubling it each time. Then, we ran another mini-batch and repeated the procedure above until the loss worsened rather than improved. This is the point where we realized we had gone too far. Then, we chose a learning rate that was slightly lower than this point. We used discriminative learning rates and a gradual unfreezing of the network to re-train the network using the tomato leaf images. The learning rate tool in fast.ai (fastai is a deep learning library which provides practitioners with high-level components [88]) was employed to return a plot of learning rate versus loss (cross-entropy) and to identify the optimal learning rates to use with the Adam optimizer. Figure 20 shows the plot of the learning rate curve as we trained the network. We used smaller learning rates in the early layers to permit the weights in those layers to change in a slower pattern than those from the later layers: 1 × 10−4, 1 × 10−3, and 1 × 10−2 for the first, middle, and last layers, respectively. While training, for the first three epochs, only the final layers were trained, while all prior layers were frozen. Subsequently, all layers were unfrozen and trained for an additional 27 epochs.
We can see from plot Figure 20a that in the range of 1 × 10−6 to 1 × 10−4, nothing happened and the model did not train. Then, the loss started to decrease until it reached a minimum at 1 × 10−1, and then it increased again. This showed that a learning rate greater than 1 × 10−1 was high and will make training diverge, but a learning rate of 1 × 10−1 was already too high for training; hence, we needed to select a better rate. Figure 20 shows us four different points from our learning rate finder: valley, slide, steep, and minimum points, respectively. With such, we can select an appropriate value for the learning rate. However, that is not the end of the training procedure. Our goal is to train the whole network without breaking the pre-trained weights.
Figure 21 shows activations of a convolutional neural network by layers, visually demonstrating what is learned by the different layers of a model [89].
Because different layers in a neural network capture different types of information, there is a need for them to be fine-tuned to different extents. Instead of using the same learning rate for all layers of the model, the practice of discriminative fine-tuning allows for the tuning of each layer in a neural network with different learning rates. For a proper context, the regular SGD update of a model’s parameters θ at time step t looks can be represented by the following:
θ t = θ t 1 η   ·   θ J ( θ )  
where η is the learning rate and ∇θJ(θ) represents the gradient with regard to the objective function of the model in question.
For discriminative fine-tuning, we split the parameters θ into {θ1,...,θL}, where θl contains the parameters of the model at the l-th layer and L is the number of layers in the entire model.
Similarly, we obtain {η1,...,ηL}, where ηl signifies the learning rate of the l-th layer.
The SGD update with discriminative fine-tuning can then be represented by the following:
θ t l = θ t l 1 η l   ·   θ l J ( θ )  
Empirically, it has been found to be the best practice to first of all choose the learning rate ηL of the last layer by fine-tuning only the last layer and using ηl−1 = ηl/2.6 as the learning rate for lower layers.

4.3.2. Unfreezing and Re-Tuning the Learning Rate Schedule

As we went about fine-tuning the model, the deepest layers of our pre-trained ResNet model did not need as high a learning rate as the last ones, so we should probably use different learning rates for those—this is known as using discriminative learning rates. It is based on the idea that we use a lower learning rate for the early layers of the neural network, and a higher learning rate for the later layers, especially the randomly added layers that suit the problem to be solved. It derives its foundation from the insights drawn from the proposal of Jason Yosinski et al. [87] as summarized in Figure 22 below, which shows the impact different layers make and training methods on the transfer learning approach in applications of deep learning models.
The learning rate finder selects a minibatch of images, determines the gradient for the minibatch, and then steps the weights based on the learning rate and the gradient. Leslie Smith initially set a very low learning rate for the first minibatch before gradually raising it. We chose the steep or minimal point divided by a value of 10 because the minimum was where learning stops, but the steepest point was where learning was most effective. The early layers did not need to be modified at all; however, the later layers did need to be adjusted to fit the tomato leaf images. Therefore, we used discriminative learning rates and set a smaller learning rate for the early layers and a greater learning rate for the latter layers.

4.4. Experimental Setup

The training and testing processes in this experiment were implemented on Pytorch using fastai library [88]; the scikit-learn, pillow, and OpenCV libraries were used, all of which are written in Python. The model was trained and tested on an NVIDIA DGX-1 V100 with 8X Tesla V100 GPUs with a performance of one petaFLOP.
To carry out this research, a computer system was used to carry out the analysis processes. The computer system has the following specifications: Windows 10, 64-bit, Intel Core i7-4720 CPU @ 2.60 GHz, RAM 32 GB and GPU Nvidia GeForce GTX 1050 4 GB dedicated memory, and Python 3.7 on Anaconda. The PC was used on a Linux-based DELL PowerEdge T640 Tower Server with CUDA-based video cards GTX 1080TI, each GPU Video memory was 11Gb, with a storage memory of 10TB Hard Drive and 3320 GB SSD. Table 12 and Table 13 detail the setup of the system used and the parameters used in running the model.

5. Discussion

Using data from ten different classes, the effect of different depths, and batch sizes (depth = 18, 34, 50, 101, and 152 and BS = 40, 50, 60, 70, 80, 90, and 100) and parameter values on their performance in adequately recognizing healthy tomatoes and other diseased classes were investigated in this study. The research was carried out using transfer learning and a pre-trained residual neural network. They performed well on the training and test data, according to the results. However, the steady state in the test data was observed to be delayed as the batch size value increased. The highest recognition accuracy value was 99.51% for depth = 152, with a train/validation data split size of 80/20 and a batch size of 40. According to the findings, the batch size value has no significant effect on overall performance but increasing the batch size value delays obtaining stable results.
Table 11 shows a detailed comparison of our proposed technique with other state-of-the-art models. As can be seen from Table 11 and Figure 13, our model outperformed the previous models, surpassing that of Li et al. [58], which was the closest in performance, with a +0.75% performance gain, and achieved a favorable performance with a +3.79% improvement from that of Paymode et al. [69] and other previous models. However, most recent models, such as Islam et al. [84] and Tarek et al. [71] outperformed our method, recording 100% and 99.81% accuracy. Whereby some authors used the accuracy metric to measure their performance, we recorded a much higher accuracy but chose to benchmark against our F1 score value, which is regarded as a much better form of performance measure for classification and recognition problems, since it combines both the precision and recall of the model in question.

6. Conclusions

Discriminative learning was implemented in training our proposed networks, having diverse layers, such as 18, 34, 50, 102, and 152 layers, respectively. The results show that the network with the highest depth produced the best performance, suggesting that deeper networks could still be better in deep learning. Five train/test data split ratios of 40/60, 50/50, 60/40, 70/30, and 80/20 were experimented on to determine the optimum value of the train/test split ratio for such a research area. The split ratio of 80/20 resulted in the optimal solution, with 70/30 coming after it. Different batch sizes were selected based on the capacity of the GPU system obtainable in the laboratory to test for the influence of batch size on the training, and if so on the result of the training and test processes of the model. The charts show a slight increase in performance for lesser batch sizes, while the most notable effect of the batch sizes was in speeding up the training process. The discriminative learning process was studied to determine the best learning rates to select in re-training the network during transfer learning, and the results show how effective it is to pre-train, determine a range of suitable learning rates, and re-train the entire network based on the discriminative learning rates to achieve a faster and more efficient performance. The data augmentation also proved to be effective, as the results were a little bit improved with the augmentation process.

Author Contributions

Conceptualization, P.S.K.; methodology, P.S.K.; software, P.S.K., A.K.; validation, K.X.; formal analysis, P.S.K., A.K.; resources, K.X.; data curation, P.S.K.; writing—original draft preparation, P.S.K.; writing—review and editing, E.O.O., A.K.; visualization, E.O.O., A.K.; supervision, K.X.; funding acquisition, K.X. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China No. U1813222 and No. 42075129, the Hebei Province Natural Science Foundation No. E2021202179, Key Research and Development Project from Hebei Province No. 19210404D, No. 20351802D, and No. 21351803D.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The PlantVillage dataset is publicly available online at https://github.com/spMohanty/PlantVillage-Dataset (accessed on 13 November 2021).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Hughes, D.P.; Salathe, M. An Open Access Repository of Images on Plant Health to Enable the Development of Mobile Disease Diagnostics. arXiv 2015, arXiv:1511.08060. [Google Scholar]
  2. Food and Agriculture Organization of the United Nations. Plant Pests and Diseases. Available online: https://www.fao.org/emergencies/emergency-types/plant-pests-and-diseases/en/ (accessed on 13 November 2021).
  3. FAOSTAT. 2019. Available online: http://www.fao.org/faostat/en/#data/QC (accessed on 13 November 2021).
  4. Panno, S.; Davino, S.; Caruso, A.G.; Bertacca, S.; Crnogorac, A.; Mandić, A.; Noris, E.; Matić, S. A Review of the Most Common and Economically Important Diseases That Undermine the Cultivation of Tomato Crop in the Mediterranean Basin. Agronomy 2021, 11, 2188. [Google Scholar] [CrossRef]
  5. Das, S. Over View of Septoria Diseases on Different Crops and Its Management. Int. J. Agric. Environ. Biotechnol. 2020, 13, 361–370. [Google Scholar] [CrossRef]
  6. Gilardi, G.; Matic, S.; Guarnaccia, V.; Garibaldi, A.; Gullino, M.L. First Report of Fusarium Clavum Causing Leaf Spot and Fruit Rot on Tomato in Italy. Plant Dis. 2021, 105, 2250. [Google Scholar] [CrossRef]
  7. Infantino, A.; Loreti, S. Malattie. In Il Pomodoro; Angelini, R., Ed.; Collana Coltura & Cultura, Bayer CropScience: Bologna, Italy, 2010; pp. 194–219. [Google Scholar]
  8. Barba, M.; Martelli, G.; Tomassoli, L.; Galllitelli, D.; Di Serio, F.; Pasquini, G. Virosi e Fitoplasmosi. In Proceedings of the AA.VV. Il Pomodoro; Bayer CropScience: Bologna, Italy, 2010; pp. 220–235. [Google Scholar]
  9. Zaagueri, T.; Mnari-Hattab, M.; Moussaoui, N.; Accotto, G.P.; Noris, E.; Marian, D.; Vaira, A.M. Chickpea Chlorotic Dwarf Virus Infecting Tomato Crop in Tunisia. Eur. J. Plant Pathol. 2019, 154, 1159–1164. [Google Scholar] [CrossRef]
  10. Williams’, C.E.; Clair, D.A.S.; Williams, C.E.; St, N.D.; Clair, D.A. Phenetic Relationships and Levels of Variability Detected by Restriction Fragment Length Polymorphism and Random Amplified Polymorphic DNA Analysis of Cultivated and Wild Accessions of Lycopersicon Esculentum. Genome 2011, 36, 619–630. [Google Scholar] [CrossRef]
  11. Bai, Y.; Im, P.; In, L.; Ut, D.O. Domestication and Breeding of Tomatoes: What Have We Gained and What Can We Gain in the Future? Ann. Bot. 2007, 100, 1085–1094. [Google Scholar] [CrossRef]
  12. Blanca, J.; Montero-Pau, J.; Sauvage, C.; Bauchet, G.; Illa, E.; Díez, M.J.; Francis, D.; Causse, M.; van der Knaap, E.; Cañizares, J. Genomic Variation in Tomato, from Wild Ancestors to Contemporary Breeding Accessions. BMC Genom. 2015, 16, 257. [Google Scholar] [CrossRef] [Green Version]
  13. King, K.C.; Lively, C.M. Does Genetic Diversity Limit Disease Spread in Natural Host Populations? Heredity 2012, 109, 199–203. [Google Scholar] [CrossRef] [Green Version]
  14. Singh, V.K.; Singh, A.K.; Kumar, A. Disease Management of Tomato through PGPB: Current Trends and Future Perspective. 3 Biotech 2017, 7, 255. [Google Scholar] [CrossRef]
  15. Khirade, S.D.; Patil, A.B. Plant Disease Detection Using Image Processing. In Proceedings of the 2015 International Conference on Computing Communication Control and Automation, Pune, India, 26–27 February 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 768–771. [Google Scholar]
  16. Bharate, A.A.; Shirdhonkar, M.S. A Review on Plant Disease Detection Using Image Processing. In Proceedings of the Proceedings of the International Conference on Intelligent Sustainable Systems, ICISS 2017, Palladam, India, 7–8 December 2017; Institute of Electrical and Electronics Engineers Inc.: Piscataway, NJ, USA, 2018; pp. 103–109. [Google Scholar] [CrossRef]
  17. Song, Z.; Cai, C.; Yan, Y. Automatic Identification of Grape Varieties Based on Leaf Morphological Characteristics. Comput. Simul. 2012, 29, 307–310. [Google Scholar]
  18. Wang, L.I.; Shao-Xian, T. Review of Crop Diseases Recognition Based on Image Processing. Hunan Agric. Mach. 2012, 1, 176–178. [Google Scholar]
  19. Tao, H.; Zhao, L.; Xi, J.; Yu, L.; Wang, T. Fruits and Vegetables Recognition Based on Color and Texture Features. Trans. Chin. Soc. Agric. Eng. 2014, 30, 305–311. [Google Scholar]
  20. Sladojevic, S.; Arsenovic, M.; Anderla, A.; Culibrk, D.; Stefanovic, D. Deep Neural Networks Based Recognition of Plant Diseases by Leaf Image Classification. Comput. Intell. Neurosci. 2016, 2016, 1–11. [Google Scholar] [CrossRef] [Green Version]
  21. Zhou, G.; Zhang, W.; Chen, A.; He, M.; Ma, X. Rapid Detection of Rice Disease Based on FCM-KM and Faster R-CNN Fusion. IEEE Access 2019, 7, 143190–143206. [Google Scholar] [CrossRef]
  22. Durmus, H.; Gunes, E.O.; Kirci, M. Disease Detection on the Leaves of the Tomato Plants by Using Deep Learning. In Proceedings of the 2017 6th International Conference on Agro-Geoinformatics, Fairfax, VA, USA, 7–10 August 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 1–5. [Google Scholar] [CrossRef]
  23. Antonellis, G.; Gavras, A.G.; Panagiotou, M.; Kutter, B.L.; Guerrini, G.; Sander, A.C.; Fox, P.J. Shake Table Test of Large-Scale Bridge Columns Supported on Rocking Shallow Foundations. J. Geotech. Geoenviron. Eng. 2015, 141, 04015009. [Google Scholar] [CrossRef]
  24. Hu, J.; Shen, L.; Sun, G. Squeeze-and-Excitation Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar] [CrossRef] [Green Version]
  25. Rangarajan, A.K.; Purushothaman, R.; Ramesh, A. Tomato Crop Disease Classification Using Pre-Trained Deep Learning Algorithm. Procedia Comput. Sci. 2018, 133, 1040–1047. [Google Scholar] [CrossRef]
  26. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. In Proceedings of the 3rd International Conference on Learning Representations, ICLR 2015—Conference Track Proceedings, San Diego, CA, USA, 7–9 May 2015; pp. 1–14. [Google Scholar]
  27. Karthik, R.; Hariharan, M.; Anand, S.; Mathikshara, P.; Johnson, A.; Menaka, R. Attention Embedded Residual CNN for Disease Detection in Tomato Leaves. Appl. Soft Comput. 2020, 86, 105933. [Google Scholar] [CrossRef]
  28. Anand, R.; Veni, S.; Aravinth, J. An Application of Image Processing Techniques for Detection of Diseases on Brinjal Leaves Using K-Means Clustering Method. In Proceedings of the 2016 International Conference on Recent Trends in Information Technology (ICRTIT), Chennai, India, 8–9 April 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 1–6. [Google Scholar] [CrossRef]
  29. Zhang, S.; Wang, H.; Huang, W.; You, Z. Plant Diseased Leaf Segmentation and Recognition by Fusion of Superpixel, K-Means and PHOG. Optik 2018, 157, 866–872. [Google Scholar] [CrossRef]
  30. Rani, F.A.P.; Kumar, S.; Fred, A.L.; Dyson, C.; Suresh, V.; Jeba, P. K-Means Clustering and SVM for Plant Leaf Disease Detection and Classification. In Proceedings of the 2019 International Conference on Recent Advances in Energy-Efficient Computing and Communication (ICRAECC), Nagercoil, India, 7–8 March 2019; Fred, A.L., Dyson, C., Suresh, V., Eds.; IEEE: Piscataway, NJ, USA, 2019; pp. 1–4. [Google Scholar] [CrossRef]
  31. Kumari, C.U.; Jeevan Prasad, S.; Mounika, G. Leaf Disease Detection: Feature Extraction with K-Means Clustering and Classification with ANN. In Proceedings of the 2019 3rd International Conference on Computing Methodologies and Communication (ICCMC), Erode, India, 27–29 March 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 1095–1098. [Google Scholar]
  32. Liu, B.; Tan, C.; Li, S.; He, J.; Wang, H. A Data Augmentation Method Based on Generative Adversarial Networks for Grape Leaf Disease Identification. IEEE Access 2020, 8, 102188–102198. [Google Scholar] [CrossRef]
  33. Lv, M.; Zhou, G.; He, M.; Chen, A.; Zhang, W.; Hu, Y. Maize Leaf Disease Identification Based on Feature Enhancement and DMS-Robust Alexnet. IEEE Access 2020, 8, 57952–57966. [Google Scholar] [CrossRef]
  34. Jiang, D.; Li, F.; Yang, Y.; Yu, S. A Tomato Leaf Diseases Classification Method Based on Deep Learning. In Proceedings of the Proceedings of the 32nd Chinese Control and Decision Conference, CCDC 2020, Hefei, China, 22–24 August 2020; Institute of Electrical and Electronics Engineers Inc.: Piscataway, NJ, USA, 2020; pp. 1446–1450. [Google Scholar] [CrossRef]
  35. Waheed, A.; Goyal, M.; Gupta, D.; Khanna, A.; Hassanien, A.E.; Pandey, H.M. An Optimized Dense Convolutional Neural Network Model for Disease Recognition and Classification in Corn Leaf. Comput. Electron. Agric. 2020, 175, 105456. [Google Scholar] [CrossRef]
  36. Huang, S.; Liu, W.; Qi, F.; Yang, K. Development and Validation of a Deep Learning Algorithm for the Recognition of Plant Disease. In Proceedings of the 2019 IEEE 21st International Conference on High Performance Computing and Communications; IEEE 17th International Conference on Smart City; IEEE 5th International Conference on Data Science and Systems (HPCC/SmartCity/DSS), Zhangjiajie, China, 10–12 August 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 1951–1957. [Google Scholar] [CrossRef]
  37. Chen, X.; Zhou, G.; Chen, A.; Yi, J.; Zhang, W.; Hu, Y. Identification of Tomato Leaf Diseases Based on Combination of ABCK-BWTR and B-ARNet. Comput. Electron. Agric. 2020, 178, 105730. [Google Scholar] [CrossRef]
  38. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 770–778. [Google Scholar] [CrossRef] [Green Version]
  39. Sethy, P.K.; Barpanda, N.K.; Rath, A.K.; Behera, S.K. Deep Feature Based Rice Leaf Disease Identification Using Support Vector Machine. Comput. Electron. Agric. 2020, 175, 105527. [Google Scholar] [CrossRef]
  40. Oyewola, D.O.; Dada, E.G.; Misra, S.; Damaševičius, R. Detecting Cassava Mosaic Disease Using a Deep Residual Convolutional Neural Network with Distinct Block Processing. PeerJ Comput. Sci. 2021, 7, e352. [Google Scholar] [CrossRef]
  41. Zeng, W.; Li, M. Crop Leaf Disease Recognition Based on Self-Attention Convolutional Neural Network. Comput. Electron. Agric. 2020, 172, 105341. [Google Scholar] [CrossRef]
  42. Chen, J.J.; Chen, J.J.; Zhang, D.; Sun, Y.; Nanehkaran, Y.A. Using Deep Transfer Learning for Image-Based Plant Disease Identification. Comput. Electron. Agric. 2020, 173, 105393. [Google Scholar] [CrossRef]
  43. Li, Y.; Nie, J.; Chao, X. Do We Really Need Deep CNN for Plant Diseases Identification? Comput. Electron. Agric. 2020, 178, 105803. [Google Scholar] [CrossRef]
  44. Ramcharan, A.; Baranowski, K.; McCloskey, P.; Ahmed, B.; Legg, J.; Hughes, D.P. Deep Learning for Image-Based Cassava Disease Detection. Front. Plant Sci. 2017, 8, 1852. [Google Scholar] [CrossRef] [Green Version]
  45. Ramcharan, A.; McCloskey, P.; Baranowski, K.; Mbilinyi, N.; Mrisho, L.; Ndalahwa, M.; Legg, J.; Hughes, D.P. A Mobile-Based Deep Learning Model for Cassava Disease Diagnosis. Front. Plant Sci. 2019, 10, 272. [Google Scholar] [CrossRef] [Green Version]
  46. Adedoja, A.; Owolawi, P.A.; Mapayi, T. Deep Learning Based on NASNet for Plant Disease Recognition Using Leave Images. In Proceedings of the 2019 International Conference on Advances in Big Data, Computing and Data Communication Systems (icABCD), Winterton, South Africa, 5–6 August 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 1–5. [Google Scholar] [CrossRef]
  47. Devi, S.N.; Muthukumaravel, A. A Novel Salp Swarm Algorithm With Attention-Densenet Enabled Plant Leaf Disease Detection And Classification. In Precision Agriculture, Proceedings of the 2022 International Conference on Advanced Computing Technologies and Applications (ICACTA), Coimbatore, India, 4–5 March 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 1–7. [Google Scholar] [CrossRef]
  48. Bhujel, A.; Kim, N.-E.; Arulmozhi, E.; Basak, J.K.; Kim, H.-T. A Lightweight Attention-Based Convolutional Neural Networks for Tomato Leaf Disease Classification. Agriculture 2022, 12, 228. [Google Scholar] [CrossRef]
  49. Zhao, Y.; Sun, C.; Xu, X.; Chen, J. RIC-Net: A Plant Disease Classification Model Based on the Fusion of Inception and Residual Structure and Embedded Attention Mechanism. Comput. Electron. Agric. 2022, 193, 106644. [Google Scholar] [CrossRef]
  50. Keivani, M.; Mazloum, J.; Sedaghatfar, E.; Tavakoli, M.B. Automated Analysis of Leaf Shape, Texture, and Color Features for Plant Classification. Traitement du Signal 2020, 37, 17–28. [Google Scholar] [CrossRef]
  51. Li, J.; Yang, J. Supervised Classification of Plant Image Based on Attention Mechanism. In Proceedings of the 2021 7th International Conference on Systems and Informatics (ICSAI), Chongqing, China, 13–15 November 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 1–6. [Google Scholar] [CrossRef]
  52. Kanda, P.S.; Xia, K.; Sanusi, O.H. A Deep Learning-Based Recognition Technique for Plant Leaf Classification. IEEE Access 2021, 9, 162590–162613. [Google Scholar] [CrossRef]
  53. Thanikkal, J.G.; Dubey, A.K.; Thomas, M. Importance of Image Morphological Features in Continues Learning. In Proceedings of the 2022 International Mobile and Embedded Technology Conference (MECON), Noida, India, 10–11 March 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 372–376. [Google Scholar] [CrossRef]
  54. Twum, F.; Missah, Y.M.; Oppong, S.O.; Ussiph, N. Textural Analysis for Medicinal Plants Identification Using Log Gabor Filters. IEEE Access 2022, 10, 83204–83220. [Google Scholar] [CrossRef]
  55. Gajjar, V.K.; Nambisan, A.K.; Kosbar, K.L. Plant Identification in a Combined-Imbalanced Leaf Dataset. IEEE Access 2022, 10, 37882–37891. [Google Scholar] [CrossRef]
  56. Goyal, N.; Gupta, K.; Kumar, N. Clustering-Based Hierarchical Framework for Multiclass Classification of Leaf Images. IEEE Trans. Ind. Appl. 2022, 58, 4076–4085. [Google Scholar] [CrossRef]
  57. Ganguly, S.; Bhowal, P.; Oliva, D.; Sarkar, R. BLeafNet: A Bonferroni Mean Operator Based Fusion of CNN Models for Plant Identification Using Leaf Image Classification. Ecol. Inform. 2022, 69, 101585. [Google Scholar] [CrossRef]
  58. Li, M.; Zhou, G.; Chen, A.; Yi, J.; Lu, C.; He, M.; Hu, Y. FWDGAN-Based Data Augmentation for Tomato Leaf Disease Identification. Comput. Electron. Agric. 2022, 194, 106779. [Google Scholar] [CrossRef]
  59. Too, E.C.; Yujian, L.; Njuki, S.; Yingchun, L. A Comparative Study of Fine-Tuning Deep Learning Models for Plant Disease Identification. Comput. Electron. Agric. 2019, 161, 272–279. [Google Scholar] [CrossRef]
  60. Hu, G.; Wu, H.; Zhang, Y.; Wan, M. A Low Shot Learning Method for Tea Leaf’s Disease Identification. Comput. Electron. Agric. 2019, 163, 104852. [Google Scholar] [CrossRef]
  61. Wang, G.; Sun, Y.; Wang, J. Automatic Image-Based Plant Disease Severity Estimation Using Deep Learning. Comput. Intell. Neurosci. 2017, 2017, 1–8. [Google Scholar] [CrossRef] [PubMed]
  62. Agarwal, M.; Gupta, S.K.; Biswas, K.K. Development of Efficient CNN Model for Tomato Crop Disease Identification. Sustain. Comput. Inform. Syst. 2020, 28, 100407. [Google Scholar] [CrossRef]
  63. Gandhi, R.; Nimbalkar, S.; Yelamanchili, N.; Ponkshe, S. Plant Disease Detection Using CNNs and GANs as an Augmentative Approach. In Proceedings of the 2018 IEEE International Conference on Innovative Research and Development (ICIRD), Bangkok, Thailand, 11–12 May 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 1–5. [Google Scholar] [CrossRef]
  64. Elhassouny, A.; Smarandache, F. Smart Mobile Application to Recognize Tomato Leaf Diseases Using Convolutional Neural Networks. In Proceedings of the 2019 International Conference of Computer Science and Renewable Energies (ICCSRE), Agadir, Morocco, 22–24 July 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 1–4. [Google Scholar] [CrossRef]
  65. Darwish, A.; Ezzat, D.; Hassanien, A.E. An Optimized Model Based on Convolutional Neural Networks and Orthogonal Learning Particle Swarm Optimization Algorithm for Plant Diseases Diagnosis. Swarm Evol. Comput. 2020, 52, 100616. [Google Scholar] [CrossRef]
  66. Mishra, S.; Sachan, R.; Rajpal, D. Deep Convolutional Neural Network Based Detection System for Real-Time Corn Plant Disease Recognition. Procedia Comput. Sci. 2020, 167, 2003–2010. [Google Scholar] [CrossRef]
  67. Lamba, M.; Gigras, Y.; Dhull, A. Classification of Plant Diseases Using Machine and Deep Learning. Open Comput. Sci. 2021, 11, 491–508. [Google Scholar] [CrossRef]
  68. Zhao, S.; Peng, Y.; Liu, J.; Wu, S. Tomato Leaf Disease Diagnosis Based on Improved Convolution Neural Network by Attention Module. Agriculture 2021, 11, 651. [Google Scholar] [CrossRef]
  69. Paymode, A.S.; Malode, V.B. Transfer Learning for Multi-Crop Leaf Disease Image Classification Using Convolutional Neural Network VGG. Artif. Intell. Agric. 2022, 6, 23–33. [Google Scholar] [CrossRef]
  70. Islam, M.S.; Sultana, S.; Al Farid, F.; Islam, M.N.; Rashid, M.; Bari, B.S.; Hashim, N.; Husen, M.N. Multimodal Hybrid Deep Learning Approach to Detect Tomato Leaf Disease Using Attention Based Dilated Convolution Feature Extractor with Logistic Regression Classification. Sensors 2022, 22, 6079. [Google Scholar] [CrossRef]
  71. Tarek, H.; Aly, H.; Eisa, S.; Abul-Soud, M. Optimized Deep Learning Algorithms for Tomato Leaf Disease Detection with Hardware Deployment. Electronics 2022, 11, 140. [Google Scholar] [CrossRef]
  72. Tej, B.; Nasri, F.; Mtibaa, A. Detection of Pepper and Tomato Leaf Diseases Using Deep Learning Techniques. In Proceedings of the 2022 5th International Conference on Advanced Systems and Emergent Technologies (IC_ASET), Hammamet, Tunisia, 22–25 March 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 149–154. [Google Scholar] [CrossRef]
  73. Ozbilge, E.; Ulukok, M.K.; Toygar, O.; Ozbilge, E. Tomato Disease Recognition Using a Compact Convolutional Neural Network. IEEE Access 2022, 10, 77213–77224. [Google Scholar] [CrossRef]
  74. Mukherjee, G.; Chatterjee, A.; Tudu, B. Identification of the Types of Disease for Tomato Plants Using a Modified Gray Wolf Optimization Optimized MobileNetV2 Convolutional Neural Network Architecture Driven Computer Vision Framework. Concurr. Comput. Pract. Exp. 2022, 34, e7161. [Google Scholar] [CrossRef]
  75. Wu, S.G.; Bao, F.S.; Xu, E.Y.; Wang, Y.X.; Chang, Y.F.; Xiang, Q.L. A Leaf Recognition Algorithm for Plant Classification Using Probabilistic Neural Network. In Proceedings of the ISSPIT 2007—2007 IEEE International Symposium on Signal Processing and Information Technology, Giza, Egypt, 15–18 December 2007; IEEE: Piscataway, NJ, USA, 2007; pp. 11–16. [Google Scholar] [CrossRef] [Green Version]
  76. Zhuang, F.; Qi, Z.; Duan, K.; Xi, D.; Zhu, Y.; Zhu, H.; Xiong, H.; He, Q. A Comprehensive Survey on Transfer Learning. Proc. IEEE 2021, 109, 43–76. [Google Scholar] [CrossRef]
  77. Neyshabur, B.; Sedghi, H.; Zhang, C. What Is Being Transferred in Transfer Learning? Adv. Neural Inf. Process. Syst. 2020, 33, 512–523. [Google Scholar]
  78. Saini, D.; Chand, T.; Chouhan, D.K.; Prakash, M. A Comparative Analysis of Automatic Classification and Grading Methods for Knee Osteoarthritis Focussing on X-ray Images. Biocybern. Biomed. Eng. 2021, 41, 419–444. [Google Scholar] [CrossRef]
  79. Cai, C.; Wang, S.; Xu, Y.; Zhang, W.; Tang, K.; Ouyang, Q.; Lai, L.; Pei, J. Transfer Learning for Drug Discovery. J. Med. Chem. 2020, 63, 8683–8694. [Google Scholar] [CrossRef]
  80. Achicanoy, H.; Chaves, D.; Trujillo, M. StyleGANs and Transfer Learning for Generating Synthetic Images in Industrial Applications. Symmetry 2021, 13, 1497. [Google Scholar] [CrossRef]
  81. Zhu, C. Pretrained language models. In Machine Reading Comprehension; Elsevier: Amsterdam, The Netherlands, 2021; pp. 113–133. ISBN 978-0-323-90118-5. [Google Scholar]
  82. Bashath, S.; Perera, N.; Tripathi, S.; Manjang, K.; Dehmer, M.; Streib, F.E. A Data-Centric Review of Deep Transfer Learning with Applications to Text Data. Inf. Sci. 2022, 585, 498–528. [Google Scholar] [CrossRef]
  83. Hussain, M.; Bird, J.J.; Faria, D.R. A Study on CNN Transfer Learning for Image Classification. In Advances in Computational Intelligence Systems, Proceedings of the Contributions Presented at the 18th UK Workshop on Computational Intelligence, Nottingham, UK, 5–7 September 2018; Springer: Cham, Switzerland, 2019; Volume 840, pp. 191–202. [Google Scholar] [CrossRef]
  84. Howard, J.; Gugger, S. Deep Learning for Coders with Fastai and PyTorch; O’Reilly Media, Inc.: Sebastopol, CA, USA, 2020; Volume 66, ISBN 978-1-492-04552-6. [Google Scholar]
  85. Bengio, Y. Practical Recommendations for Gradient-Based Training of Deep Architectures. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Berlin/Heidelberg, Germany, 2012; Volume 7700, pp. 437–478. ISBN 9783642352881. [Google Scholar]
  86. Howard, J.; Ruder, S. Universal Language Model Fine-Tuning for Text Classification. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics, Melbourne, Australia, 15–20 July 2018; Association for Computational Linguistics: Stroudsburg, PA, USA, 2018; Volume 1, pp. 328–339. [Google Scholar] [CrossRef] [Green Version]
  87. Yosinski, J.; Clune, J.; Bengio, Y.; Lipson, H. How Transferable Are Features in Deep Neural Networks? In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, USA, 8–13 December 2014; Neural Information Processing Systems Foundation: La Jolla, CA, USA, 2014; Volume 4, pp. 3320–3328. [Google Scholar] [CrossRef]
  88. Howard, J.; Gugger, S. Fastai: A Layered API for Deep Learning. Information 2020, 11, 108. [Google Scholar] [CrossRef] [Green Version]
  89. Zeiler, M.D.; Fergus, R. Visualizing and Understanding Convolutional Networks. In Computer Vision—ECCV 2014, Proceedings of the3th European Conference, Zurich, Switzerland, 6–12 September 2014; Springer: Cham, Switzerland, 2014; Volume 8689, pp. 818–833. [Google Scholar] [CrossRef]
Figure 1. Showing the results on different train/test split ratios on various models used in our research: (a) showing epoch over F1 score for ResNet-18; (b) showing epoch over F1 score for ResNet-34; (c) showing epoch over F1 score for ResNet-50; (d) showing epoch over F1 score for ResNet-101; (e) showing epoch over F1 score for ResNet-152; (f) showing epoch over F1 score for ResNet-152 on a batch size of 80, this is where the best F1 score of our research is located.
Figure 1. Showing the results on different train/test split ratios on various models used in our research: (a) showing epoch over F1 score for ResNet-18; (b) showing epoch over F1 score for ResNet-34; (c) showing epoch over F1 score for ResNet-50; (d) showing epoch over F1 score for ResNet-101; (e) showing epoch over F1 score for ResNet-152; (f) showing epoch over F1 score for ResNet-152 on a batch size of 80, this is where the best F1 score of our research is located.
Plants 11 02935 g001aPlants 11 02935 g001b
Figure 2. Confusion matrix for network depth = 34.
Figure 2. Confusion matrix for network depth = 34.
Plants 11 02935 g002
Figure 3. Confusion matrix for network depth = 50.
Figure 3. Confusion matrix for network depth = 50.
Plants 11 02935 g003
Figure 4. Confusion matrix for network depth = 101.
Figure 4. Confusion matrix for network depth = 101.
Plants 11 02935 g004
Figure 5. Confusion matrix for network depth = 152.
Figure 5. Confusion matrix for network depth = 152.
Plants 11 02935 g005
Figure 6. Split ratio against F1 score.
Figure 6. Split ratio against F1 score.
Plants 11 02935 g006
Figure 7. Batch size against F1 score for various validation split ratios.
Figure 7. Batch size against F1 score for various validation split ratios.
Plants 11 02935 g007
Figure 8. Showing the results on different train/test split ratios on various models used in our research: (a) showing epoch over F1 score for ResNet-18; (b) showing epoch over F1 score for ResNet-34; (c) showing epoch over F1 score for ResNet-50; (d) showing epoch over F1 score for ResNet-101; (e) showing epoch over F1 score for ResNet-152; (f) showing epoch over F1 score for ResNet-152 on a batch size of 80, this is where the best F1score of our research is located.
Figure 8. Showing the results on different train/test split ratios on various models used in our research: (a) showing epoch over F1 score for ResNet-18; (b) showing epoch over F1 score for ResNet-34; (c) showing epoch over F1 score for ResNet-50; (d) showing epoch over F1 score for ResNet-101; (e) showing epoch over F1 score for ResNet-152; (f) showing epoch over F1 score for ResNet-152 on a batch size of 80, this is where the best F1score of our research is located.
Plants 11 02935 g008
Figure 9. Epoch over time for depth of (a) 18, and (b) 152.
Figure 9. Epoch over time for depth of (a) 18, and (b) 152.
Plants 11 02935 g009
Figure 10. A plot of validation split ratio and layer depth over time.
Figure 10. A plot of validation split ratio and layer depth over time.
Plants 11 02935 g010
Figure 11. Showing the mistakes recorded by the model in our research: (a) showing the plot loss for ResNet-152; (b) showing the images the model predicted wrongly for ResNet-152, showing the predicted, ground truth, loss, and probability values.
Figure 11. Showing the mistakes recorded by the model in our research: (a) showing the plot loss for ResNet-152; (b) showing the images the model predicted wrongly for ResNet-152, showing the predicted, ground truth, loss, and probability values.
Plants 11 02935 g011
Figure 12. Showing the model losses and error rates for: (a) 152, (b) 101, (c) 50, (d) 34, and (e) 18 layers respectively.
Figure 12. Showing the model losses and error rates for: (a) 152, (b) 101, (c) 50, (d) 34, and (e) 18 layers respectively.
Plants 11 02935 g012aPlants 11 02935 g012b
Figure 13. Benchmark against other models [27,47,48,49,58,59,60,61,62,63,64,65,66,67,68,69,70,71,72,73,74].
Figure 13. Benchmark against other models [27,47,48,49,58,59,60,61,62,63,64,65,66,67,68,69,70,71,72,73,74].
Plants 11 02935 g013
Figure 14. Samples from the Flavia leaf dataset.
Figure 14. Samples from the Flavia leaf dataset.
Plants 11 02935 g014
Figure 15. Samples from the tomato leaf dataset.
Figure 15. Samples from the tomato leaf dataset.
Plants 11 02935 g015
Figure 16. CutMix results from the Flavia dataset.
Figure 16. CutMix results from the Flavia dataset.
Plants 11 02935 g016
Figure 17. Workflow for the tomato leaf disease recognition system.
Figure 17. Workflow for the tomato leaf disease recognition system.
Plants 11 02935 g017
Figure 18. Showing the residual identity mapping for a residual neural network.
Figure 18. Showing the residual identity mapping for a residual neural network.
Plants 11 02935 g018
Figure 19. Simplified block diagram for the residual neural network.
Figure 19. Simplified block diagram for the residual neural network.
Plants 11 02935 g019
Figure 20. Showing results of the learning rate finder on the network: (a) shows the learning rate find result before unfreezing for the network of depth 152; (b) shows the result of the second learning rate find after unfreezing the model and training briefly on the first selected learning rate.
Figure 20. Showing results of the learning rate finder on the network: (a) shows the learning rate find result before unfreezing for the network of depth 152; (b) shows the result of the second learning rate find after unfreezing the model and training briefly on the first selected learning rate.
Plants 11 02935 g020
Figure 21. Activations of a convolutional neural network by layers.
Figure 21. Activations of a convolutional neural network by layers.
Plants 11 02935 g021aPlants 11 02935 g021b
Figure 22. Layer depth against performance in training methods on transfer learning.
Figure 22. Layer depth against performance in training methods on transfer learning.
Plants 11 02935 g022
Table 1. List of tomato plant pathogens present in the Mediterranean basin.
Table 1. List of tomato plant pathogens present in the Mediterranean basin.
Pathogen GroupPathogen NameReference
FungiAlternaria solani, Botrytis cinerea, Cladosporium fulvum, Colletotrichum coccodes, Fusarium oxysporum, Fusarium clavum, Leveillula taurica, Oidium lycopersici, Pseudoidium neolycopersici, Pyrenochaeta lycopersici, Rhizoctonia solani, Septoria lycopersici, Sclerotinia sclerotiorum, Sclerotium rolfsii, Stemphylium spp., Verticillium dahliae[4,6]
OomycetesPhytophthora infestans, Phytophthora nicotianae, Phytophtora cryptogea, Pythium debaryanum, Pythium sylvaticum[7]
BacteriaClavibacter michiganensis subsp. michiganensis, Erwinia carotovora subsp. carotovora, Pseudomonas corrugata, Pseudomonas mediterranea, Pseudomonas syringae pv. tomato, Ralstonia solanacearum, Xanthomonas axonopodis pv. vesicatoria[7]
PhytoplasmaCandidatus Phytoplasma solani[8]
VirusesAlfalfa mosaic virus, Chickpea chlorotic dwarf virus, Cucumber mosaic virus, Eggplant mottled dwarf virus, Parietaria mottle virus, Pelargonium zonate spot virus, Pepino mosaic virus, Potato virus Y, Southern tomato virus, Tobacco mosaic virus, Tomato brown rugose fruit virus, Tomato chlorosis virus, Tomato infectious chlorosis virus, Tomato leaf curl New Delhi virus, Tomato mosaic virus, Tomato spotted wilt virus, Tomato torrado virus, Tomato yellow leaf curl virus, Tomato yellow leaf curl Sardinia virus[4,9]
ViroidsPotato spindle tuber viroid, Tomato apical stunt viroid[8]
Table 2. Performance on various split ratios.
Table 2. Performance on various split ratios.
Batch SizePerformance (%)
40/6050/5060/4070/3080/20
1000.9774470.9848160.9883780.9921390.993655
900.979570.9879270.9930780.992410.99521
800.9813490.9878480.9901030.9926410.995611
700.9847340.9869580.9887160.9944960.994071
600.9843380.9875950.9943160.9934440.995579
500.9871150.98930.9935660.9947690.994317
400.9852950.9900850.9930450.9940830.995048
Table 3. Time taken.
Table 3. Time taken.
Batch SizeTime (s)
40/6050/5060/4070/3080/20
100162196221234249
90179188218237253
80169193222235247
70176188211232240
60172193217236259
50188208221242275
40200204226248281
Table 4. Results for various train-validation split ratios.
Table 4. Results for various train-validation split ratios.
Train
Split (%)
Validation Split (%)Train
Loss
Valid
Loss
AccuracyRecallPrecisionF1 Score
90100.0522910.079080.9760460.9728970.9723380.97166
80200.0495480.0715330.9765970.9739580.9738450.97321
70300.0453660.0812450.971090.9674640.9669350.966463
60400.0422450.0702980.9757710.9723070.972890.971938
50500.0336660.0496220.9848570.9811650.9822640.981518
40600.0023240.0143660.9964210.9957810.9954510.995611
Table 5. Network depth on validation split = 60.
Table 5. Network depth on validation split = 60.
Network DepthError RateF1 ScoreTime (s)
180.0378120.95362835
340.0280840.96455253
500.0258810.9691683
1010.018080.977994126
1520.0189060.977447160
Table 6. Network depth on validation split = 50.
Table 6. Network depth on validation split = 50.
Network DepthError RateF1 ScoreTime (s)
180.0274230.96730640
340.0245590.96895959
500.0186120.97801791
1010.0154190.981207139
1520.0132160.984816187
Table 7. Network depth on validation split = 40.
Table 7. Network depth on validation split = 40.
Network DepthError RateF1 ScoreTime (s)
180.0213380.97490341
340.0156940.97990463
500.0155560.9817297
1010.0101870.988444152
1520.0096370.988378217
Table 8. Network depth on validation split = 30.
Table 8. Network depth on validation split = 30.
Network DepthError RateF1 ScoreTime (s)
180.0161530.97828343
340.0110130.98572869
500.0111970.986037105
1010.0097280.987897160
1520.0066080.992139228
Table 9. Network depth on validation split = 20.
Table 9. Network depth on validation split = 20.
Network DepthError RateF1 ScoreTime (s)
180.0154190.98140747
340.0115640.98621273
500.0090860.988667112
1010.0068830.991295179
1520.0055070.993655243
Table 10. Comparison of the proposed technique with the existing prevalent approaches on the Flavia dataset.
Table 10. Comparison of the proposed technique with the existing prevalent approaches on the Flavia dataset.
TechniqueYearObjective# ImagesMethodsAccuracy (%)
Keivani et al. [50]2020Flavia dataset1907Decision Tree98.58
Li et al. [51]2021Flavia dataset1907DenseNet20198.69
Kanda et al. [52]2021Flavia dataset1907DL + Logistic Regression99.0
Thanikkal et al. [53]2022Flavia dataset1907DL99.0
Twum et al. [54]2022Flavia dataset1907Log Gabor Filters97.0
Gajjar et al. [55]2022Flavia dataset1907Extreme learning machines99.10
Goyal et al. [56]2022Flavia dataset1907Hierarchical cluster96.24
Ganguly et al. [57]2022Flavia dataset1907ResNet + Bonferroni mean operator98.7
Proposed Network2022Flavia dataset1907ResNet + Discriminative Learning99.23
Table 11. Comparison of the proposed technique with the existing prevalent approaches on the tomato leaf dataset.
Table 11. Comparison of the proposed technique with the existing prevalent approaches on the tomato leaf dataset.
TechniqueYearObjective# ImagesMethodsAccuracy (%)
Too et al. [59]2019Plant leaf disease54,306VGG1681.83
Gensheng et al. [60]2019Tea leaf disease4980VGG1690
Wang et al. [61]2017Plant leaf disease54,306VGG1690.4
Agarwal et al. [62]2020Tomato leaf disease18,160VGG1693.5
Wang et al. [61]2017Plant leaf disease54,306Inception-V380
Gandhi et al. [63]2018Plant leaf disease56,000Inception-V388.6
Agarwal et al. [62]2020Tomato leaf disease18,160Inception-V377.5
Elhassouny & Smarandache [64]2019Tomato leaf disease7176MobileNet88.4
Gandhi et al. [63]2018Plant leaf disease56,000Mobilenet92
Agarwal et al. [62]2020Tomato leaf disease18,160Mobilenet82.6
Darwish et al. [65]2019Maize leaf disease15,408VGG1998.2
Karthik et al. [27]2020Tomato leaf disease5452
(4 classes)
ResNet + DenseNet98
Mishra et al. [66]2020Corn leaf disease3703CNN98.4
Lamba et al. [67]2021Tomato leaf disease16,012CNN98.2
Agarwal et al. [62]2020Tomato leaf disease18,160CNN98.4
Zhao et al. [68]2021Tomato leaf disease18,160
(10 classes)
ResNet50 + SeNet96.81
Li et al. [58]2022Tomato leaf disease4240FWDGAN + B-ARNet98.75
Paymode et al. [69]2022Tomato leaf disease VGG1695.71
Devi et al. [47]2022Tomato leaf disease9281DensNet + Attention mechanism97.56
Bhujel et al. [48]2022Tomato leaf disease19,510Lightweight Attention-Based CNN99.34
Zaho et al. [49]2022Tomato leaf disease18,160Spatial attention with CNN95.20
Islam et al. [70]2022Tomato leaf disease15,989cGAN + CNN + Logistic Regression100
Tarek et al. [71]2022Tomato leaf disease16,004MobileNetV399.81
Tej et al. [72]2022Pepper and Tomato leaf diseases488CNN98.85
Özbılge et al. [73]2022Tomato leaf disease18,160Compact CNN98.49
Mukherjee et al. [74]2022Tomato leaf disease10,839
(7 classes)
Gray Wolf + MobileNetV298
Proposed Technique2022Tomato leaf disease18,160ResNet + Discriminative Learning99.51
Table 12. Configuration of the machine used.
Table 12. Configuration of the machine used.
NameParameter
Memory32 GB
ProcessorIntel(R) Xeon(R) Silver 4114 CPU @ 2.20 GHz
Server modelDELL PowerEdge T640 Tower Server
GraphicsCUDA-based video cards 4X 1080TI; GPU Video memory of 11 Gb
OSLinux
LanguagePython 3
FrameworkPytorch
Table 13. Configuration of the machine used.
Table 13. Configuration of the machine used.
NameParameter
Solver typeAdam
Batch sizes20, 30, 40, 50, 60, 70, 80, 90, 100
Image input size256 × 256
Train/Test-split ratio40/60, 50/50, 60/40, 70/30, 80/20
Learning rateDiscriminative ranges
Drop out0.5
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Kanda, P.S.; Xia, K.; Kyslytysna, A.; Owoola, E.O. Tomato Leaf Disease Recognition on Leaf Images Based on Fine-Tuned Residual Neural Networks. Plants 2022, 11, 2935. https://doi.org/10.3390/plants11212935

AMA Style

Kanda PS, Xia K, Kyslytysna A, Owoola EO. Tomato Leaf Disease Recognition on Leaf Images Based on Fine-Tuned Residual Neural Networks. Plants. 2022; 11(21):2935. https://doi.org/10.3390/plants11212935

Chicago/Turabian Style

Kanda, Paul Shekonya, Kewen Xia, Anastasiia Kyslytysna, and Eunice Oluwabunmi Owoola. 2022. "Tomato Leaf Disease Recognition on Leaf Images Based on Fine-Tuned Residual Neural Networks" Plants 11, no. 21: 2935. https://doi.org/10.3390/plants11212935

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop