Research Paper Updated 4.0
Research Paper Updated 4.0
Disease Detection
Manoj Kumar Singh1,a), Ali Sher Khan2, b) , Abbas Akbar3, c) , Ananya Lamba4, d)
,Prakriti Gupta5,e)
1
Associate Professor, CS&E Department, Moradabad Institute of Technology, Moradabad, India
2,3,4,5
BTech 4th Year, CS&E Department, Moradabad Institute of Technology, Moradabad, India
a)
[email protected]
b)
[email protected]
c)
[email protected]
d)
[email protected]
e)
[email protected]
Abstract. Farmers around the world face significant challenges in managing crop diseases, which
can drastically reduce yield and quality, leading to economic losses. Traditional methods of
disease detection, relying on manual inspection by specialists, are time-consuming, prone to
human error, and often unable to detect diseases at early stages. To address these issues, advanced
technologies like deep learning offer promising solutions for automated, accurate, and rapid
disease detection. This study presents the application of deep learning models for the detection of
leaf diseases across pepper bell, tomato, and potato crops. Specifically, we evaluate the
effectiveness of several well-established models, including VGG19, DenseNet201, ResNet50, and
AlexNet, for classifying 15 disease types using the PlantVillage dataset. In addition to these pre-
trained models, we introduce our new convolutional neural network, Plant Scan, designed to
address the unique challenges of plant disease detection. Plant Scan optimizes the balance
between computational efficiency and high classification accuracy, making it a practical tool for
real-world agricultural applications. The models were evaluated on various performance metrics
such as precision, recall, F1 score, and classification accuracy, using Python along with Keras and
TensorFlow frameworks on a GPU-equipped system. The results provide valuable insights into the
strengths and limitations of these models, helping to create reliable tools for early disease
detection, ultimately improving crop management and enhancing agricultural productivity.
Keywords: Leaf disease detection, pre-trained networks, convolution neural networks, deep
learning.
I. INTRODUCTION
In recent years, the agricultural sector has increasingly embraced advanced technologies to address
challenges such as crop diseases, which significantly impact yield and quality. Traditional methods of
disease detection, which rely on manual inspection by experts, are time-consuming and prone to
human error. As deep learning technologies have evolved, they have emerged as a powerful tool for
automating disease detection, enabling more efficient and accurate identification of plant diseases.
This study introduces Plant Scan, a new deep learning model specifically designed to tackle the
challenges of leaf disease detection in crops such as pepper bell, tomato, and potato. Plant Scan
leverages convolutional neural networks (CNNs) [3] to efficiently extract features and classify images
of diseased leaves. By combining the strengths of established models with novel optimizations, Plant
Scan aims to provide a more accurate, computationally efficient solution for real-world agricultural
applications. The model is evaluated on the PlantVillage dataset, which contains images of 20
different plant diseases, to assess its performance in terms of classification accuracy, precision, recall,
and F1 score. Through this study, we aim to demonstrate how Plant Scan can contribute to the
development of reliable, scalable tools for disease management in agriculture, ultimately helping to
improve productivity and reduce losses within the farming sector.
B. METRICS
To evaluate the performance of these algorithms and architectures, several metrics are commonly
used. These include top performance rates [6]-[9], precision and recall [10]-[11], F1 score [12][13],
training and validation accuracy [13], loss metrics [14], classification accuracy [15][16], and mean
average precision. These metrics help in providing a comprehensive understanding of the model's
effectiveness, highlighting its strengths and limitations in various evaluation contexts.
III. METHODOLOGY
The core of our approach lies in the Plant Scan model, a new deep learning architecture designed for
efficient leaf disease detection in agricultural crops such as pepper bell, tomato, and potato. The
model is built upon the principles of Convolutional Neural Networks (CNNs), which have shown
exceptional performance in computer vision tasks. CNNs are particularly effective for image
processing tasks due to their ability to automatically learn spatial hierarchies of features, making them
ideal for detecting complex patterns in images like diseased plant leaves.
A. KEY COMPONENTS
a) CONVOLUTIONAL LAYER
The convolutional layers are responsible for extracting local patterns from the input images. In the
Plant Scan architecture, several convolutional layers are stacked to gradually learn more abstract and
complex features. Each convolutional operation applies a set of filters (or kernels) to the input image,
performing an element-wise multiplication followed by a sum. These learned filters can detect edges,
textures, and other key features crucial for identifying plant diseases. Mathematically, the convolution
operation can be expressed as:
S(i,j)=(I∗K)(i,j)=m∑n∑I(m,n)⋅K(i−m,j−n)
where I represents the input image, K is the kernel, and S(i,j) is the resulting feature map.
b) POOLING LAYER
Following the convolutional layers, the Plant Scan model utilizes MaxPooling layers to downsample
the feature maps. Pooling reduces the spatial dimensions of the feature maps, making the model more
computationally efficient and robust to minor translations and distortions in the input data.
MaxPooling selects the maximum value from a specific region of the feature map, capturing the most
prominent features while reducing computational complexity.
d) ACTIVATION FUNCTIONS
ReLU (Rectified Linear Unit) is used as the activation function in most layers. ReLU introduces
non-linearity to the model, enabling it to learn complex patterns in the data. The activation function
plays a key role in making the network adaptive to a wide range of input data distributions.
e) OUTPUT LAYER
The output layer in Plant Scan consists of as many neurons as there are disease categories. This layer
uses a SoftMax function to produce the probability distribution over the potential classes. The class
with the highest probability is chosen as the final predicted disease type.
B. ARCHITECTURE
Input Layer: The model accepts images of size 256×256256 \times 256256×256 pixels,
which are RGB images of diseased plant leaves.
Convolutional Layers: A series of convolutional layers (using 3×33 \times 33×3 kernels) are
applied to the input to extract hierarchical features. The number of filters increases as we move
deeper into the network to capture more complex patterns.
Max Pooling Layers: These layers reduce the spatial dimensions of the feature maps, aiding
in computational efficiency and preventing overfitting.
Fully Connected Layers: After feature extraction, the network uses fully connected layers to
aggregate features and make the final classification decision.
SoftMax Output Layer: The final layer predicts the probability distribution over 20 disease
categories, using a Softmax activation function to output the class with the highest probability.
Figure 1: Architecture
The Plant Scan model is a new-designed Convolutional Neural Network (CNN) optimized for
detecting plant diseases. Below is a detailed explanation of its components and their roles in achieving
efficient and accurate classification.
1. Input Layer : The model's input layer processes RGB plant leaf images of size 256 × 256 × 3,
ensuring uniformity and compatibility with the neural network. The three colour channels preserve
crucial visual details, like discoloration and texture variations, which help identify disease-related
patterns. This input format enables the model to effectively distinguish between healthy and diseased
leaves.
2. Convolutional Layer : The model's feature extraction relies on convolutional layers that identify
patterns like edges, textures, and disease-related spots. The first layer uses 32 filters (3 × 3) with
ReLU activation to capture basic features like edges. The second layer, with 64 filters, detects more
complex patterns such as leaf venation and disease spots. The third layer, with 128 filters, extracts
abstract features like blotches and fungal patches. These layers progressively capture and refine
disease-relevant visual cues.
3. Pooling Layer: The model uses max pooling layers to reduce the spatial dimensions of feature
maps, enhancing computational efficiency and ensuring invariance to minor distortions. Each max
pooling layer uses a 2 × 2 window to downsample the feature maps, selecting the largest value in each
region. The first two pooling layers reduce the spatial dimensions, retaining essential features, while
the final pooling layer further simplifies the data for the flattening step. This approach effectively
maintains critical features while reducing data complexity for efficient processing.
4. Fully Connected Layer : Following feature extraction, the model transitions to fully connected
layers to perform classification. The process begins with a flattening step, where the 2D feature maps
are converted into a 1D array, enabling their input into dense layers. These dense layers, equipped
with learnable weights, map the extracted features to specific output classes. By combining and
interpreting the learned features, the dense layers generate meaningful predictions, completing the
classification task.
5. Output Layer : The output layer of the model consists of 20 neurons, each representing one of the
20 disease classes in the dataset. A softmax activation function is applied in this layer to convert the
raw scores into probabilities, ensuring that the sum of all outputs equals 1. The class associated with
the highest probability is then selected as the predicted disease, providing the final classification
result.
The model is trained using the PlantVillage dataset, which comprises images of both diseased and
healthy plant leaves. The dataset is divided into two subsets: a training set and a testing set. The
training set is utilized to optimize the model's weights through iterative learning, while the testing set
is employed to assess the model's ability to generalize and perform accurately on unseen data.
Performance Metrics
Accuracy: Measures overall correctness of predictions.
F1 Score: Balances precision and recall, particularly useful for imbalanced datasets.
A. DISEASE DATASETS
For this study, we utilized the PlantVillage Dataset, a comprehensive collection of images featuring
diseased plant leaves. Although the dataset's size varied across different machine learning
experiments, with approximately more than 80,000 images used, it consistently included 38 distinct
disease classes. These images were coloured and had a resolution of 256 × 256 pixels. The dataset
covered a range of plant diseases, including Potato Late Blight, Potato Early Blight, Potato Healthy,
Tomato Bacterial Spot, Tomato Early Blight, Tomato Late Blight, Tomato Leaf Mold, Tomato
Septoria Leaf Spot, Tomato Spider Mites (Two-Spotted Spider Mite), Tomato Target Spot, Tomato
Yellow Leaf Curl Virus, Tomato Mosaic Virus, and Tomato Healthy, Grape Black Spot, Grape Black
Measles, Apple scab, Apple Black rot, Apple cedar, Cherry Powdery mildew etc. Each class
corresponds to a specific plant disease affecting crops, primarily focusing on cherry, apple, grape,
peppers, potatoes, and tomatoes, strawberry , peach. The PlantVillage Dataset served as the
foundation for training and evaluating our deep learning models, helping us improve disease detection
and classification in agricultural settings and contribute to advancements in crop management and
agricultural technology.
Figure 2: Tomato Septoria leaf spot Figure 3: Tomato yellow leaf curl virus
In our study, we evaluated pre-trained deep learning models for identifying leaf diseases in crops like
pepper bell, tomato, and potato. We compared models like VGG19, DenseNet201, ResNet50, and
AlexNet using the Plant Village dataset. Our analysis focused on factors such as computational
efficiency, training time, and classification accuracy. Overall, our results showed that all models
performed well in identifying leaf diseases, with variations in efficiency and accuracy. We found
similarities in performance trends across models, indicating the effectiveness of deep learning for
large-scale dataset analysis. Our study highlights the strengths and weaknesses of each model, aiding
decision-makers in selecting appropriate models for agricultural applications. Ultimately, our research
aims to enhance leaf disease detection in crops and contribute to agricultural resilience and
productivity.
The epoch table illustrates the superior performance of PlantScan over VGG19,
DenseNet201, ResNet50, and AlexNet across varying epochs. At 10 epochs, PlantScan
achieves 97.23%, surpassing DenseNet201 (84.12%) and ResNet50 (70.45%). By 25 epochs,
PlantScan improves to 98.14%, while DenseNet201 and ResNet50 reach 87.56% and 79.34%.
At 50 epochs, PlantScan leads with 98.5%, outperforming DenseNet201 (91.78%) and
ResNet50 (85.45%). Even at 80 epochs, PlantScan excels with 98.9%, maintaining its
superiority over DenseNet201 (93.45%) and AlexNet (79.23%). These results highlight
PlantScan’s efficiency and strong learning capability.
V. CONCLUSION
In recent years, the agricultural sector has increasingly relied on machine learning models for plant
disease detection to ensure early intervention and effective management. While several pre-trained
models have demonstrated effectiveness in image classification tasks, they often struggle with
domain-specific datasets like those in agriculture.The proposed PlantScan model proves to be a
superior solution for plant disease detection compared to pre-trained models like VGG19,
DenseNet201, ResNet50, and AlexNet. While these models are effective for general image
classification, they face limitations such as higher computational demands and suboptimal
performance on agricultural datasets. PlantScan, designed specifically for plant disease detection,
achieves an impressive accuracy of 99.3%, outperforming DenseNet201 (98.3%), ResNet50 (97.5%),
and AlexNet (96.2%). It excels in precision (98.9%), recall (99.1%), and F1 score (99.0%), ensuring
reliable and thorough disease identification. With an optimized architecture, PlantScan maximizes
feature extraction while minimizing computational overhead, making it ideal for resource-constrained
environments. Unlike pre-trained models, which struggle with domain-specific datasets, PlantScan’s
new design ensures accuracy and practicality for agricultural applications. By enabling early and
precise detection, PlantScan addresses critical challenges in farming, reducing crop losses and
improving productivity, solidifying its role as a transformative tool in modern agriculture.
VI. REFERENCES
[1] M. A. Ebrahimi, M. H. Khoshtaghaza, S. Minaei, and B. Jamshidi, ‘‘Vision-based pest detection
based on SVM classification method,’’ Comput. Electron. Agricult., vol. 137, pp. 52–58, May 2017.
[2] M. Türkoglu and D. Hanbay, ‘‘Plant disease and pest detection using deep learning-based
features,’’ TURKISH J. Electr. Eng. Comput. Sci., vol. 27, no. 3, pp. 1636–1651, May 2019.
[3] Simonyan, K., Zisserman, A.: Very Deep Convolutional Networks for Large-Scale Image
Recognition. In: 3rd International Conference on Learning Representations, pp. 1-14. dblp, San
Diego, CA, USA (2015).
[4] W. S. McCulloch and W. Pitts, ‘‘A logical calculus of the ideas immanent in nervous activity,’’
Bull. Math. Biophys., vol. 5, no. 4, pp. 115–133, Dec. 1943.
[5] W. Xinshao and C. Cheng, ‘‘Weed seeds classification based on PCANet deep learning baseline,’’
in Proc. Asia–Pacific Signal Inf. Process. Assoc. Annu. Summit Conf. (APSIPA), Hong Kong, Dec.
2015, pp. 408–415.
[6] K. He, X. Zhang, S. Ren, and J. Sun, ‘‘Deep residual learning for image recognition,’’ in Proc.
IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), Sacramento, CA, USA, Jun. 2016, pp. 770–778.
[7] K. Simonyan and A. Zisserman, ‘‘Very deep convolutional networks for large-scale image
recognition,’’ in Proc. Int. Conf. Learn. Repr., London, U.K., Apr. 2014, pp. 1–14.
[8] M. Dyrmann, H. Karstoft, and H. S. Midtiby, ‘‘Plant species classification using deep
convolutional neural network,’’ Biosystems Eng., vol. 151, pp. 72–80, Nov. 2016.
[9] Z.-X. Guan, J. Tang, B.-J. Yang, Y.-F. Zhou, D.-Y. Fan, and Q. Yao, ‘‘Study on recognition
method of rice disease based on image,’’ Chin. J. Rice Sci., vol. 24, no. 5, pp. 497–502, May 2010.
[10] S. P. Mohanty, D. P. Hughes, and M. Salathé, ‘‘Using deep learning for image-based plant
disease detection,’’ Frontiers Plant Sci., vol. 7, p. 1419, Sep. 2016.
[11] J. Amara, B. Bouaziz, and A. Algergawy, ‘‘A deep learning-based approach for banana leaf
diseases classification,’’ in Proc. Datenbanksys. Für Bus., Technol. Web (BTW), Workshopband, CA,
USA, Jul. 2017, pp. 1–24.
[12] A. Krizhevsky, I. Sutskever, and G. E. Hinton, ‘‘ImageNet classification with deep convolutional
neural networks,’’ in Proc. Adv. Neural Inf. Process. Syst., vol. 25, Sep. 2012, pp. 1097–1105.
[13] F. Fina, P. Birch, R. Young, J. Obu, B. Faithpraise, and C. Chatwin, ‘‘Automatic plant pest
detection and recognition using k-means clustering algorithm and correspondence filters,’’ Int. J. Adv.
Biotechnol. Res., vol. 4, no. 2, pp. 189–199, Jul. 2013