Detection and Prediction of Rice Leaf Disease Using A Hybrid CNN-SVM Model
Detection and Prediction of Rice Leaf Disease Using A Hybrid CNN-SVM Model
Abstract—Agriculture is one of India’s greatest money makers and a measure of financial growth. Rice
is one of India’s most widely grown crops as a staple diet. Rice crops have been shown to be heavily
afflicted by illnesses, resulting in significant losses in agriculture. Rice leaf diseases not only cause a
loss of revenue for farmers, but they also decrease the quality of their final output. External appear-
ances of diseased rice leaves can be subjected to image processing processes. On the other hand, dis-
ease sickness may vary depending on the different leaves. Each disease has its own distinct features,
some of leaves have the same colour but various shapes, while others have different colours but the
same shapes. Farmers are sometimes confused and unable to make an accurate judgement when it
comes to pesticide choosing. To solve this problem, a hybrid CNN (Inception-ResNet)-SVM model
for detecting and treating damaged rice leaves has been developed. In this designed model, the images
are collected and gathered by capturing rice leaf using camera at the agricultural field. These images
are refined to improve image quality and visibility for reliable estimation, and then segregated using
Grab-Cut algorithm to eliminate undesired sections of image. Features of the segmented images are
extracted and classified using hybrid CNN (Inception-Resnet V2)-SVM algorithm. The developed
model’s study results are analysed and discussed to recent techniques. The suggested model achieved
accuracy, precision, recall, and error values of 0.97, 0.93 and 0.03 accordingly. As a conclusion, sug-
gested model outperforms revious methodologies.
Keywords: rice leaf disease, contrast limited adaptive histogram equalization, wiener filter, CNN-
SVM, grab-cut algorithm
DOI: 10.3103/S1060992X2301006X
1. INTRODUCTION
India is a rapidly developing country with agriculture serving as the country’s backbone. However,
there are other difficulties in the agricultural area, including significant productivity losses [1]. Rice dis-
eases have substantial negative effect on quality and amount of rice crops, and are major threat to food
security. In contemporary agriculture, collecting details concerning real-time rice health and growth is
prioritized [2]. Rice illnesses can be predicted before they emerge throughout the rice production process
using data from several sources [3]. Visual findings by skilled farmers or rice specialists on field used to
substantially in conventional approach of detecting rice diseases. This necessitates continual skilled mon-
itoring, that prohibitively expensive in big farms [4]. Farmers in certain emerging regions have to travel
long way to get agronomic specialists, which is moment, costly, and uncertain.
A new technique for detecting and treating plant diseases has developed as result of rapid advancement
of image processing and pattern identification systems [5]. Image recognition utilizing deep learning is an
essential research topic because it may aid monitor massive areas of plants and identify diseases on plant
leaves manually [6]. Several previous studies conducted to evaluate image identification utilizing a specific
classifier that distinguishes between image types [7]. Due to improvements in digital cameras and risen
computation power [8] auto image processing innovation shown strong performance in current decades
and successfully deployed in a number of fields, such as medical image analysis, food interpretation, bio-
metrics, factory output, and expression learning, and other.
Most plant illnesses may be identified automatically using effective image recognition techniques and
the primary source of detection is usually the plant leaves [9]. The rice leaf plant diseases Leaf smut, Rice
burst, Bacterial blight, and brown spot are identified utilising hybrid CNN (Inception-Resnet v2)-SVM
39
40 CHAUDHARI, MALATHI
with actual pictures from farm areas in this suggested model. Every disease has own distinct features; for
example, some diseases have similar colour but various shapes, while others have different colours but
similar shapes [10]. So, to classify the disease accurately from the captured images some of the pre-pro-
cessing techniques are used to enhance the image. The enhanced images consist of unwanted portions of
the leaf that should be segmented to avoid wrong predictions and low computational time. Major contri-
butions of the proposed model are:
• Detection and prediction of rice leaf plant disease using Hybrid CNN-SVM model.
• Raw data captured from rice field are pre-processed using image resizing, Wiener filter and image
quality increase using Contrast Limited Adaptive histogram equalization (CLAHE) approaches.
•The GCA is used to divide pre-processed photos and eliminate undesired regions.
• To obtain features from image for categorization, a hybrid CNN depending on type of Inception-
Resnet V2 technique is utilised.
• Extracted features from the hybrid CNN algorithm is classified using the SVM algorithm.
Section 2 describes particular research reports relevant to current solutions utilised for rice leaf disease
detection. Section 3 provides a brief overview of suggested methodology. The suggested framework’s out-
comes and performance measures are detailed in Section 4. The entire study project is wrapped up in Sec-
tion 5.
2. LITERATURE SURVEY
Many different methods are designed for predicting the leaf disease with various techniques and clas-
sifiers. Some of the existing research papers are reviewed below.
Jiang F. et al. [11] had performed deep learning and support vector machines were used to identify leaf
diseases in pictures. The rice leaf disease images are first extracted using convolution neural networks
(CNNs) in this model. After that, the SVM algorithm is used to classify and predict the disease. The opti-
mal SVM model values are identified utilising 10-fold cross validation method. The test results reveal that
deep learning and SVM-based rice disease recognition model has higher average proper recognition rate.
Shrivastava V. et al. [12] had designed rice plant illness categorization using deep CNN transfer learn-
ing especially for Oryza Sativa disease. Photos of sick symptoms in leaves and stems were acquired on rice
field. In field, 619 images of rice plant injury were collected and grouped into four categories: Healthy
Leaves (HL), Sheat Blight (SB), Rice Blast (RB) and Bacterial Leaf Blight (BLB). A pre-trained deep
CNN model is utilized as feature extractor, while SVM is used for categorization. This technique of
advanced rice disease detection used both prophylactic and an early warning system.
Atole R.R. and Park D. [13] had performed for detection of frequent rice plant irregularities, a multi-
class deep CNN classifier was used. The use of deep CNN in categorization of rice plants based on images
of their leaves for their health status is examined. The training employed 600 photos of rice plants to rep-
resent the different classes. Agricultural technicians confirmed training and validation dataset photos were
acquired from rice fields all over district. A three-class classifier depicting normal, diseased, and snail-
infested plants is created utilizing transfer learning from an AlexNet deep network.
Krishnamoorthy N. et al. [14] had designed deep neural networks with learning algorithms were used
to forecast rice leaf diseases. Rice is a key grain crop that provides energy to half of world’s population as
staple meal. Soil fertility, precipitation, pests, temperature, virus bacteria, and other abiotic and biotic fac-
tors impact rice grain harvest production amount and quality. The advancement of agriculture technical
support has made much easier to automatically detect harmful organisms in rice plant leaves. Inception-
ResNetV2 is sort of CNN model used to recognise illnesses in rice leaf pictures using transfer learning
technique.
Larijani M.R. et al. [15] had developed KNN algorithm enhancement utilizing k-means used to exam-
ine an image analysis strategy for detecting rice blast disease in field situations. In order to define disease
patches on rice leaves, colour images made utilizing image processing method and K-means-improved
KNN algorithm used to divide images in Lab colour space. The Otsu approach used to produce an adap-
tive thresholding histogram of photos based on shape or to reduce grey level in binary images. Squared
classification depends on Euclidean distance. Finally specificity, total accuracy and sensitivity of sug-
gested algorithm are evaluated to determine its efficacy.
Most of existing techniques for designing rice leaf plant disease detection are done by using CNN-SVM,
CNN (AlexNet)-SVM, CNN (Alexnet), CNN (Inception-ResNet) and KNN. From the mentioned
existing techniques, the proposed CNN (Inception-ResNet)-SVM model performs better.
Data
collection
Pre-processing
Segmentation
Classification
Predicted output
Bacterial
Rice Blast Leaf Smut Brown Spot
Blight
3. PROPOSED METHODOLOGY
Rice is an essential grain since it feeds humans with energy. The identification and prediction of Rice
leaf diseases in farming fields is an important process for appropriate management and reduction of Rice
damage. Rice leaf infections can impact leaf at any time during planting and harvesting, resulting in sig-
nificant losses in rice farming and market price. Rice leaf diseases not only waste farmer’s income, but also
they reduce quality of their final product. Hence, Image processing operations can be applied on external
appearances of infected rice leaf to detect and predict disease at early stage.
Figure 1 depicts suggested model’s process flow. In this designed model, the data are collected by cap-
turing images of the rice leaf using camera at the agricultural field. These collected images are pre-pro-
cessed using the image resizing, wiener filter and CLAHE techniques. The pre-processed images are seg-
mented using the Grab-Cut algorithm for segmenting the rice leaf form unwanted portions of the image.
Features of the segmented images are extracted and classified using hybrid CNN (Inception-Resnet V2)-
SVM algorithm.
3.2. Pre-Processing
Data collection consists of raw data that cannot be classified with high accuracy by a classifier. In order
to get highest precision, specific pre-processing procedures are used.
3.2.1. Image resizing. Process of resizing source input image known to as image resizing. The original
input image is scaled into the size of 256 × 256 pixels [16]. This type of picture size makes deep learning
techniques like CNN to employ easily. Bi-cubic interpolation produces a sharper edge result than prior
methods such as bilinear interpolation. The mathematical term used to represent image resizing is given
below (1).
3 3
p ( A, B ) = x A B .
i j
ij (1)
i =0 j =0
From Eq. (1) B signifies the height in pixel, A denotes the width in pixel and xij denotes the squared
area in the image. The noise is removed using a filtering algorithm using the scaled image as an input.
3.2.2. Wiener filter. The Wiener filter is low-pass filter used to recover noise-degraded signals in variety
of situations. The filter defined on statistical approach where signal and noise are considered to stationary
linear random processes with specified spectral properties [17]. This approach outperforms others because
it achieves best balance in bias–variance trade-off. In terms, Wiener filter is an adaptive filter that evalu-
ates neighbourhood’s mean before providing little flattening when variance is large and greater flattening
when variance is low.
The filter reduces difference between evaluated and actual signals much as feasible. The error ratio for
untainted image f and anticipated image f is as regards:
e2 = E {( f − f ) }. 2
(2)
Where E{∙} denotes arguments anticipated value. Obtaining lowest of quadratic error function becomes
difficulty of getting an approximated image. The frequency domain utilised to do this, using following
indicators: noise and image are nonstationary, noise or picture has a mean of 0, and intensity levels in pre-
dicted image reduced by linear function. The error functions minimal is provided while following factors
closely:
H * (u, v ) v S f (u, v )
Fˆ (u, v ) = G (u, v ) . (3)
S f (u, v ) H (u, v ) + S f (u, v )
2
Which Fˆ (u, v ) is predicted picture in spectral domain, H(u, v) is deterioration function’s transform, G(u, v)
is reduced picture’s transform, H * (u, v ) is complex conjugate of H(u, v), and S f (u, v ) = |F(u, v)|2 is non-
degraded picture’s power distribution. The combination of complicated value and its inverse is magnitude
of complicated number doubled, as per general principle of filter. Thus,
H (u, v )
2
Fˆ (u, v ) = 1 G (u, v ) , (4)
H (u, v ) H (u, v ) + Sη (u, v ) S f (u, v )
2
where the noise power spectrum is Sη (u, v ) = N (u, v ) . Since spectrogram of non-degraded image is
2
nential, Uniform, or Rayleigh distributions are used to create contrast transform function. The CLAHE
method comprises several steps:
Its first phase to partition actual image across M × N tiles and compute histogram for every zone based
on image grey levels. Then, for each region, compute contrast-limited histogram:
N avg = ( N x × NY ) N grey . (5)
To compute mean value of pixels, N grey by number of grey levels in region and umber of pixels within
X and Y axes.
N avrgrey = N clip N grey . (6)
NCL represents actual clip limit CL, while N clip is standardized CL in [0, 1] range. If pixel value is more
than NCL , pixels are clipped. The median of remaining pixels and overall number of clipped pixels N clip .
Hregionclip (i ) = Hregion (i ) + NCL. (7)
The clipped histogram is Hregionclip (i ) and original histogram is H region (i ) of every region at ith grey level.
Step= N grey N remain . (8)
Step is positive integer higher than or equal to 1, where N remain is number of clipped pixels left. Using
preceding step, proceed from lowest to highest grey level. Allocate one image to grey level if number of
pixels in grey level is less than N CL . To compute new path and create a new search session when all pixels
will not dispersed at end of search. When all of remaining pixels were distributed, stop. Using transform,
increase intensity values in each zone (e.g., Rayleigh).The Rayleigh transform is provided by:
y (i ) = ymin + 2α ln 1
2
, (9)
1 − pinput (i )
( y (i ) − ymin )
( y (i ) − ymin )2
p ( y (i )) = exp − for y (i ) ≥ ymin . (10)
α2 2α2
Transform clipped histogram into a continuous possibility pinput (i ) to generate a transfer function.
Assuming, α is Rayleigh scaling parameter and ymin is bottom bound of pixel value.
The linear contrast extension used to eliminate suddenly shifting effects.
x (i ) − xmin
y (i ) = . (11)
xmax − xmin
Reduce border distortions by using bilinear interpolation to determine grey level attribution of pixels.
3.3. Segmentation
Segmentation is mainly useful for eliminating the unwanted portions of the image for reducing the
computational time and wrong predictions. The GAC used for classification in this model.
3.3.1. Grab Cut algorithm. Grab Cut is method for image separation that is repetitive and automated.
To create perfect image segmentation, that graph is constructed by employing a small cost reduction func-
tion. The image pixels are used to build graph vertices. In other terms, graph node represents each pixel in
image. Two additional nodes, sink and source, are inserted to graph in contrast to such nodes. Every pixel
in image, or node in graph, is linked to these two nodes. The foreground connection point is defined by
source node, while background connection point is represented by sink node. The graph’s edge weights
are generated using cost function based on region and limit details in image. The graph segmented utiliz-
ing Min-Cut/Max-Flow method. The Grab Cut approach utilizes Gaussian Mixture Models (GMMs) to
determine area information from an image’s colour information. The algorithm’s mathematical statement
is as follows. I, p = (p1, p2, ..., pN) of N pixels where pi = (Ri, Gi, Bi), i ∈ [1, ..., N] in the RGB colour space.
The categorization is represented as an array s = (S1, S2, ..., SN), Si ∈ {0, 1} with label provided to each
pixel of image expressing its connection to background or foreground. The user starts procedure by semi-
automatically creating rectangle (R) contains area to be divided. The image is separated into three sections
termed RB, RF, and RU, which reflect initial values of background, foreground, and unknown pixels,
respectively, once user specifies R. The images on outside of R is designated RB, while images inside are
marked RU. Grab Cut determines RU images are in foreground or background. It uses colour data pre-
sented by GMMs for this goal. For foreground pixels (si = 1), complete covariance GMMs of C parts is
generated, and for background pixels (si = 0), limited as follows:
θ = {π ( s, c ) , ω ( s, c ) , ( s, c ) , s ∈ {0,1} , c = 1,.., C} . (12)
Where ω denotes GMMs’ means, P represents model’s covariance matrices and π denotes weight of
model. The array c = (c1, ..., ci, ... cN), ci ∈ (1, ..., C), i ∈ [1, ..., N] indicates component of foreground or
background GMMs to which pixel pi belongs. For segmentation, the energy function revealed is:
E ( s, c, θ, p) = U ( s, c, θ, p) + V ( s, p) . (13)
Based on probability distributions pr (.) of GMMs, U is likelihood capacity.
U ( s, c, θ, p ) = − log pr ( p
i
i si , ci , θ ) − log π ( si , ci ) . (14)
Where V is regularising prior that predicts segmented regions to scolour consistent, with neighbourhood
C about each pixel.
V ( s, p) = γ [Sn ≠ Sm ] exp ( −β pm − pn
2
). (15)
{ }
m,n ∈C
The picture with specific base rectangle is subjected to energy minimization approach.
()
2
2 l k 2 falling the computation rate with a quarter. However, overall dimensionality of demonstration
2
()
2
reduces to l k minus communicative network, this produces representative bottlenecks. Instead of
2
using this, another option is suggested that decreases the calculation rate even more by reducing the rep-
resentative bottleneck.
3.4.7. Inception-v2. Now, the above-mentioned concepts are integrated to present a new structural
design as shown in figure 4. It has three traditional inception sections 3535 by 288 filters in network’s
Inception section. Using grid reduction approach described in Section 2.3.3, it reduced to 17 × 17 grid
with 768 filters. The grid reduction method can reduce it to an 8 × 8 × 1280 grid. Two Inception compo-
nents are present at coarse 8 × 8 level, due to concatenated result filter bank size of 2048 for each tile. It
is observed that the network quality is relatively stable.
3.4.8. Inception-ResNet v2. In comparison to classical Inception network, residual version of Incep-
tion network uses a simpler Inception block. Each Inception phase is followed by a filter-expansion layer
that is used to enhance depth of filter bank before it used to map depthless of input. The dimensionality
minimization achieved by Inception block must be returned. Figure 5 depicts structure of Inception-Res-
Net v2 model.
Where Y is feature vectors with kernel value X, n denotes number of network layers, bias value a and Mj is
set of each layer feature maps and ∗ denotes convolution operation.
• The input image’s dimensions are (H × W × D).
• ( Fh × Fw × D) is the kernel dimension.
• (( H − Fh + 1) × (W − Fw + 1) × 1) is output dimension of after convolution.
Resize the supplied images to 229 × 229 pixels utilizing three colour channels. The basic goal of using
neural networks is to lower image’s width and height while increasing its depth (store all relevant informa-
tion in depth of image). The depth of image is affected by number of convolution layers applied in each
sector; thus, network contains 825 convolution layers for extracting features.
Filter Concat
3×3
stride 2
3×3 3×3
stride 1 stride 2
Pool stride
1×1 1×1 2
Base
17 × 17 × 640
Concat
17 × 17 × 320 17 × 17 × 320
Conv Pool
35 × 35 × 320
3.5. Classification
Classification is technique for predicting class of supplied data points. Classes described utilizing
phrases such as targets, labels, and categories. Classification predictive modelling (y) is process of estimat-
ing a mapping function (f) from discrete input variables (X) to discrete output variables.
3.5.1. SVM Classifier. SVM is machine learning method for identifying data and recognising patterns
in categorise it [21]. It includes assigning set of input data to one of two classes. A training set required to
learn support vector machines, with each element of set specifying which class it belongs to. SVM model,
data from training set is divided by border with broadest possible margin, i.e. the distance from this hyper-
3 × 3 conv Maxpoll
(192 V) (stride = 2V)
3 × 3 conv
(96 V)
3 × 3 conv
(96 V) 1 × 7 conv
(64)
1 × 1 conv
7 × 1 conv
(64)
(64)
l × l conv
(64)
3 × 3 Maxpool 3 × 3 conv
(stride 2 V) (96 stride 2 V)
3 × 3 conv
(64)
147 × 147 × 64
3 × 3 conv
(32 V)
147 × 147 × 32
3 × 3 conv
(32 stride 2 V)
149 × 149 × 32
Input
(299 × 299 × 3)
299 × 299 × 3
plane. The Support Vector Machine (SVM) is well-known binary classification learning technique. The
main purpose is to find the best hyper plane for splitting data into two categories. Combining multiple
binary SVMs to achieve multiclass classification was recently achieved. The basic architecture of SVM
classifier is given as Fig. 6.
Let x is considered as input, j be the training instances {x j , y j } , j = 1,… m and for each occurrence
consist x j as well as label yi ∈ {−1,1} . A weight vector (w), a bias (b) both are used to parameterized every
hyper plane which is expressed as Eq. (4) as given as below.
w.x + b = 0. (17)
The hyper plane function that divides training and testing data can be described by following Eq. (5),
f ( x ) = sgn ( w.x + b ) . (18)
When interacting with kernel functions, prior function can be expressed as an Eq. (6),
N
f ( x ) = sgn
α y k ( x , x ) + b .
j =1
j j j (19)
Where x j is the input of training instance, yi is its corresponding class label, b is a bias, K ( x j , x ) is
employed kernel function that maps input vectors into an extended features space, and N is number of
training instances. The variables α j are determined by two constraints (7) and (8).
0 ≤α j , j = 1,.., N , (20)
N
α y
j =1
j j = 0. (21)
Thus, following equations are utilised to train and test disease prediction feature retrieved from rice leaf
dataset. The algorithm 1 contains pseudocode for predicting rice leaf disease.
Bias b
L(X, X1)
J1
L(X, X2) J2
Fully connected layer
J3 6 Output (Y)
(X)
L(X, X3) J4
4.1. Dataset
For rice leaf disease prediction, data set used in this model was acquired from [22]. This dataset
includes photos for diseases like blast, leaf smut, brown spot, and blight, and also images for testing and
training.
50 50 50 50
250
100 100 100 100
500
150 150 150 150
750 200 200 200 200
50 50 50 50
50
100 100 100 100
100 150 150 150 150
50 50 50 50
250
100 100 100 100
500
150 150 150 150
750 200 200 200 200
0.8
True positive rate
0.6
0.4
0.2
0 75 5 0 0
1 1 79 0 0
True label
2 0 0 36 4
3 0 0 3 37
0 1 2 3
Predicted label
1.0
0.8
0.6
Aceuracy
0.4
0.2
0
(Inception VGG16 MobileNetv1 ResNet AlexNet
ResNetv2)
+SVM
0.8
0.6
Precision
0.4
0.2
0
(Inception VGG16 MobileNetv1 ResNet AlexNet
ResNetv2)
+SVM
0.8
0.6
Recall
0.4
0.2
0
(Inception VGG16 MobileNetv1 ResNet AlexNet
ResNetv2)
+SVM
has higher accuracy. Figure 11 explains the comparison of precision metrics, in this the precision value for
(Inception-ResNet v2) + SVM is 0.93. Which is greater compared to VGG16, Mobilenet v1, ResNet and
Alexnet whose values are 0.90, 0.89, 0.83 and 0.7.
The recall metrics for suggested and existing approaches are compared in Fig. 12. Recall metrics for
Inception-ResNet v2 + SVM, VGG16, Mobilenet V1, ResNet and Alexnet 0.93, 0.84, 0.70, 0.71 and 0.68.
When compared to existing methodologies, suggested method has higher recall value. Figure 13 explains
the comparison of error metrics, in this the acquired error metrics for Inception-ResNet v2 + SVM is 0.03,
which is smaller compared to VGG16, Mobilenet V1, ResNet and Alexnet whose values are 0.11, 0.13, 0.18
and 0.26.
0.25
0.20
0.15
Error
0.10
0.05
0
(Inception VGG16 MobileNetv1 ResNet AlexNet
ResNetv2)
+SVM
1.0
0.8
Specificity
0.6
0.4
0.2
0
(Inception VGG16 MobileNetv1 ResNet AlexNet
ResNetv2)
+SVM
The comparison parameters for proposed and existing techniques are explained in Fig. 14. Specificity
metrics for Inception-ResNet v2 + SVM, VGG16, Mobilenet V1, ResNet and Alexnet 0.98, 0.93, 0.83,
0.86 and 0.80. In comparison to existing approaches, specificity value of suggested method is higher. Fig-
ure 15 explains the comparison of F1_score metrics, in this the acquired error metrics for Inception-Res-
Net v2 + SVM is 0.93, which is greater compared to VGG16, Mobilenet V1, ResNet and Alexnet whose
values are 0.90, 0.84, 0.77 and 0.60.
The NPV comparison measures for proposed and existing approaches are described in Fig. 16. The
NPV metrics for Inception-ResNet v2 + SVM, VGG16, Mobilenet V1, ResNet and Alexnet is 0.98, 0.93,
0.83, 0.90 and 0.82. When compared to existing methodologies, proposed method has higher NPV. Figure 17
0.8
0.6
F1_Score
0.4
0.2
0
(Inception VGG16 MobileNetv1 ResNet AlexNet
ResNetv2)
+SVM
1.0
0.8
Negative predictive value
0.6
0.4
0.2
0
(Inception VGG16 MobileNetv1 ResNet AlexNet
ResNetv2)
+SVM
explains the comparison of False Negative Rate (FNR) metrics, in this the acquired FNR metrics of
Inception-ResNet v2 + SVM is 0.06, which is smaller compared to VGG16, Mobilenet V1, ResNet and
Alexnet, whose values are 0.10, 0.14, 0.28 and 0.63.
Figure 18 depicts False Positive Rate (FPR) comparative metrics for proposed and existing
approaches. The FPR metrics for Inception-ResNet v2 + SVM, VGG16, Mobilenet V1, ResNet and
Alexnet is 0.019, 0.025, 0.038, 0.130 and 0.192. In comparison to existing methodologies, proposed
method has lower FPR value. The suggested model’s performance measures are examined and compared
to existing methodologies.
0.6
0.5
0.3
0.2
0.1
0
(Inception VGG16 MobileNetv1 ResNet AlexNet
ResNetv2)
+SVM
Fig. 17. Comparison of FNR for proposed and existing method.
0.200
0.175
0.150
Fslse positive rate
0.125
0.100
0.075
0.050
0.025
0
(Inception VGG16 MobileNetv1 ResNet AlexNet
ResNetv2)
+SVM
Fig. 18. Comparison of FPR for proposed and existing method.
5. CONCLUSIONS
Fast and precise disease prediction is crucial in agricultural disease management. In this designed
model, the rice leaf images are captured by camera placed in the agricultural filed, these images are col-
lected and gathered as a dataset. The collected images are pre-processed for enhancing the accuracy of the
proposed model. These pre-processed images are segmented using the grab-cut algorithm for removing
the unwanted portions of the rice leaf. The features of the segmented portions are extracted with the CNN
(Inception-ResNet V2) and SVM is used for classifying the rice leaf disease. The suggested model’s per-
formance measures are examined and compared to existing methodologies. The accuracy, precision,
recall, error, specificity, F1-Score, NPV, FNR, and FPR obtained values of 0.97, 0.93, 0.03, 0.98, 0.93,
0.98, 0.068, and 0.019, respectively. In comparison to existing methodologies, suggested model performs
better. Multiple disease prediction system for rice leaf employing better deep leaning techniques will be
included in future scope of this suggested model.
CONFLICT OF INTEREST
The authors declare that they have no conflicts of interest.
REFERENCES
1. Kitpo, N. and Inoue, M., Early rice disease detection and position mapping system using drone and IoT architecture,
in 2018 12th South East Asian Technical University Consortium (SEATUC), IEEE, 2018, vol. 1, pp. 1–5.
2. Sharma, M., Kumar, C.J., and Deka, A., Early diagnosis of rice plant disease using machine learning tech-
niques, Arch. Phytopathol. Plant Prot., 2022, vol. 55, no. 3, pp. 259–283.
3. Upadhyay, S.K. and Kumar, A., A novel approach for rice plant diseases classification with deep convolutional
neural network, Int. J. Inf. Technol., 2022, vol. 14, no. 1, pp. 185–199.
4. Nettleton, D.F., Katsantonis, D., Kalaitzidis, A., Sarafijanovic-Djukic, N., Puigdollers, P., and Confalonieri, R.,
Predicting rice blast disease: Machine learning versus process-based models, BMC Bioinf., 2019, vol. 20, no. 1,
pp. 1–16.
5. Sethy, P.K., Barpanda, N.K., Rath, A.K., and Behera, S.K., Deep feature based rice leaf disease identification
using support vector machine, Comput. Electron. Agric., 2020, vol. 175, pp. 105527.
6. Chen, W.L., Lin, Y.B., Ng, F.L., Liu, C.Y., and Lin, Y.W., RiceTalk: Rice blast detection using Internet of
Things and artificial intelligence technologies. IEEE Internet Things J., 2019, vol. 7, no. 2, pp. 1001–1010.
7. Agrawal, M.M. and Agrawal, D.S., Rice plant diseases detection and classification using deep learning models:
a systematic review, J. Crit. Rev., 2020, vol. 7, no. 11, pp. 4376–4390.
8. Liang, W.J., Zhang, H., Zhang, G.F., and Cao, H.X., Rice blast disease recognition using a deep convolutional
neural network, Sci. Rep., 2019, vol. 9, no. 1, pp. 1–10.
9. Bari, B.S., Islam, M.N., Rashid, M., Hasan, M.J., Razman, M.A.M., Musa, R.M., and Majeed, A.P.A.,
A real-time approach of diagnosing rice leaf disease using deep learning-based faster R-CNN framework, PeerJ
Comput. Sci., 2021, vol. 7, e432.
10. Rahman, C.R., Arko, P.S., Ali, M.E., Khan, M.A.I., Apon, S.H., Nowrin, F., and Wasif, A., Identification and
recognition of rice diseases and pests using convolutional neural networks, Biosyst. Eng., 2020, vol. 194, pp. 112–
120.
11. Jiang, F., Lu, Y., Chen, Y., Cai, D., and Li, G., Image recognition of four rice leaf diseases based on deep learn-
ing and support vector machine, Comput. Electron. Agricult., 2020, vol. 179, p. 105824.
12. Shrivastava, V.K., Pradhan, M.K., Minz, S., and Thakur, M.P., Rice plant disease classification using transfer
learning of deep convolution neural network, Int. Arch. Photogramm., Remote Sens. Spat. Inf. Sci., 2019, vol. 42,
no. 3, p. W6.
13. Atole, R.R. and Park, D., A multiclass deep convolutional neural network classifier for detection of common
rice plant anomalies, Int. J. Adv. Comput. Sci. Appl., 2018, vol. 9, no. 1, pp. 67–70.
14. Krishnamoorthy, N., Prasad, L.N., Kumar, C.P., Subedi, B., Abraha, H.B., and Sathishkumar, V.E., Rice leaf
diseases prediction using deep neural networks with transfer learning, Environ. Res., 2021, vol. 198, p. 111275.
15. Larijani, M.R., Asli-Ardeh, E.A., Kozegar, E., and Loni, R., Evaluation of image processing technique in iden-
tifying rice blast disease in field conditions based on KNN algorithm improvement by K-means, Food Sci. Nutr.,
2019, vol. 7, no. 12, pp. 3922–3930.
16. Kumar, R., Baloch, G., Pankaj, A.B.B., and Bhatti, J., Fungal blast disease detection in rice seed using machine
learning, Int. J. Adv. Comput. Sci. Appl., 2021, vol. 12, pp. 248–258.
17. Alwazzan, M.J., Ismael, M.A., and Ahmed, A.N., A hybrid algorithm to enhance colour retinal fundus images
using a Wiener filter and CLAHE, J. Digital Imaging, 2021, vol. 34, no. 3, pp. 750–759.
18. Sahu, S., Singh, A.K., Ghrera, S.P., and Elhoseny, M., An approach for de-noising and contrast enhancement
of retinal fundus image using CLAHE, Opt. Laser Technol., 2019, vol. 110, pp. 87–98.
19. Zhang, S., Zhao, Y., and Bai, P., Object localization improved grabcut for lung parenchyma segmentation, Pro-
cedia Comput. Sci., 2018, vol. 131, pp. 1311–1317.
20. Senthilkumar, C. and Kamarasan, M., A novel citrus disease detection and classification using deep learning
based inception resnet v2 model, Adalya J., 2020, vol. 9, no. 2. http://adalyajournal.com/.
21. Dey, A. and Bandyopadhyay, S.K., Automated glaucoma detection using support vector machine classification
method, Br. J. Adv. Med. Med. Res., 2016, vol. 11, no. 12, pp. 1–12.
22. https://www.kaggle.com/datasets/chandrug/riceleafdisease.
23. Malathi, K. and Nedunchelian, R., Efficient Method To Detect And Classify Diabetic Retinopathy Using Ret-
inal Fundus Images, International Journal of Pure and Applied Mathematics, 2017, vol. 116, no. 21, pp. 89–97,
ISSN: 1311–8080.
24. Malathi, K. and Nedunchelian, R., A recursive support vector machine (RSVM) algorithm to detect and classify
diabetic retinopathy in fundus retina images, Biomedical Research, Volume Special Issue, pp. 1–8, ISSN: 0970-
938X.
25. Malathi, K. and Nedunchelian, R., An Automated Detection of Optic Disc From Digital Retinal Fundus Im-
ages Using Region Based Segmentation Technique, International Journal of Applied Engineering Research,
vol. 10, Issue 1, ISSN: 0973-4562.
26. Malathi, K. and Nedunchelian, R., An Efficient Method to Detect Diabetic Retinopathy Using Gaussian-Bi-
lateral and Haar Filters with Threshold Based Image Segmentation, Research Journal of Applied Sciences, Engi-
neering and Technologу, 2014, 8(11), 1389–1395, ISSN: 2040-7459.
27. Malathi, K. and Nedunchelian, R., Comparison of Various Noises and Filters for Fundus Images using Pre—
processing Techniques, International Journal of Pharma and Bio Sciences, vol. 5, Issue 3, ISSN 0975-6299.