Pest Identification and Control Using Deep Learning and Augmented Reality
Pest Identification and Control Using Deep Learning and Augmented Reality
One of the biggest reasons for losses in agriculture is pests. Using various Machine Learning, Deep Learning, and Image
There are several chemical and biological pest management Processing model methods to precisely identify pests is covered in
techniques, but it is advised to take strict monitoring and control this section's literature review. This literature review's main aim is
procedures to ensure they are used to the fullest extent possible. to outline the advantages and potential uses of the various
The younger generations in India rarely receive formal training algorithms and pest identification methods.
in pest detection or pest control techniques. All pests are treated
Thenmozhi Kasinathan et al. [1], Wang and Xie dataset
the same way, with no consideration because some of them may
were experimented for the classification of nine and 24 insect
be beneficial and do not affect the plants, while others may also
classes respectively using machine learning algorithms such
feed on bugs that destroy crops. To solve this issue, we created
Artificial Neural Network (ANN), Support Vector Machines
a smartphone application for identifying pests that also makes
(SVM), K-Nearest Neighbor (KNN), and Naive bayes. The results
suitable control recommendations to satisfy the demands of
shown that the CNN model, using datasets from Wang and Xie,
farmers.
dataset, has the maximum classification accuracy of 91.5 percent
Depending on the pest it has identified, the application will and 90 percent for 9 and 24 classes of insects.
recommend management actions. The outcome will be shown
as augmented reality. Augmented reality is described as the Bo Wang et al. [2] have proposed a deep learning-based model for
simultaneous integration of digital information with the user's identifying crop illnesses and insect pests. Combining image
environment. A fresh layer of information is added to the processing and machine learning technologies improves the
existing world in augmented reality, as opposed to virtual performance of disease and insect pest recognition. The typical
reality, which produces a completely fabricated environment accuracy rate for identification is 96.26 percent.
Authorized licensed use limited to: DRDO-CABS. Downloaded on November 14,2022 at 04:29:41 UTC from IEEE Xplore. Restrictions apply.
Murray B Isman et al. [3], the issues of pest management in the currently included in identification systems, but only
21st century are compiled in this publication, along with the armyworms have real-time forecasting capabilities.
technology that may be employed to address them. The effects
of rising temperatures on the dynamics of pest populations have Cedric et al. [14], a pest detection application based on Cognitive
led to projections of heavy losses of yield of major grain crops computing and machine learning techniques which identifies crop
grown throughout the world, such as maize, wheat, etc. diseases and pests based on the image captured. But there are
certain API usage limits associated with Microsoft API providers.
Jayme Garcia Arnal Barbedo et al. [4] presents an article on Also, the result only displays the pest category and tags associated
Classifying and Detecting Pests Machine Learning and Using with it.
Proximal Images. They came to the conclusion that deep
learning methods particularly has been found to be
III. TRANSFER LEARNING-BASED DEEP
demonstrating a remarkable capacity to accurately detect as
LEARNING MODELS FOR PEST CATEGORIZATION
well as classify pests, whether in traps or natural photos.
Shrikant Salve et al. [5] makes use of augmented reality Pre-trained deep CNN models are used in a transfer learning-based
technology and a mobile camera to examine the crop disease. techniques for identifying pests. These models receive our dataset
The crop leaf image that was captured by the camera matches [15]. For categorization of images in agricultural pest recognition,
the online crop disease database. For crop disease prediction, which needs a sizable dataset coupled with powerful computing,
they suggested a mobile-based augmented reality application. deep learning methods like CNN have proved successful. Large
They have successfully implemented the prototype of a mobile datasets for the agriculture area are typically difficult to get
app that uses augmented reality and can find disease on crop because of regional weather, insect indiscriminacy, and abandoned
leaves. A photo of the crop leaf that was harmed has been used environments for chronic pests. In order to categorize data,
as a marker. researchers employed a transfer learning model that draws on
training data from a huge dataset. The CNNs include a variety of
Mohanty et al. [6], AlexNet [7] and GooLeNet [8] deep learning filters and kernels made of trainable parameters that may spatially
models were used to identify plant diseases using images, convolve on an input picture to recognise characteristics like edges
these two deep learning models have been compared. They and shapes. For first stage we do resizing of 224,224. For second
conducted their experiments using a collection of 54,306 stage, networks like EfficinetNetB7 or Xception is used. For third
images of good as well as damaged leaf images from 14 stage GlobalAveragePooling2D layer is used which is followd by
varieties of crops and 26 types of diseases. The outcome flatten and dense layer with shape 128 and relu activation function.
demonstrates GoogleNet outperforms AlexNet in their research Sixth stage is batch Normalization which is again followed by
job. dense layer with shape 512 amd relu activation. The last layer is
dense layer with shape of 9 and softmax activation function. At
In the study [9], they adjusted a few of the hyperparameters in last, Adam optimizer is used with learning_rate of 1e-5, beta_1 of
the deep learning-based models Cifar10 [11] and GooLeNet in 0.9 and beta_2 of 0.99.
order to increase accuracy. 500 photos of maize leaves are
included in the dataset. When the training-testing ratio is 80-20, Dataset from the kaggle notebook of Simran Volunesia [15] is
for the GoogLeNet model they have reached average accuracy used. According to the model's requirements, we adjusted the pest
levels of 98.9 .For the Cifar10 model, the accuracy reached photos into the desired shape, the pest dataset was then separated
was 98.8 percent. They created CaffeNet, a CNN-based deep into 70% for training and the rest 30% for test. Each model was
learning model that can identify 13 different types of diseases trained over four trials using 10 epochs with a batch size of 16 and
and 2589 pictures, using the Caffe [10] framework. They have a learning rate of 0.01. Python programming language was used
processed the database's images through picture resizing and for all experiments, together with Keras and TensorFlow. Four
enhancement. The 10-fold cross validation approach is used to different types of CNN models were investigated: Xception,
measure accuracy. Using finetuning, it can attain an accuracy MobileNet, MobilenetV2 and EfficientNetB7. The ImageNet
of 96.3 percent on average. Large Scale Visual Recognition Challenge saw these networks
perform quite well (ILSVRC).
Wikilimo et al. [12]—Wikilimo’s is a Lightweight Deep
Learning Model for Automated Pest Detection. It is an app that The feature extraction results show that simple backgrounds have
analyzes pest images to determine its type and the crops it a classification accuracy that is 5.9 percent greater than that
affects. The developers employed deep learning models, which of complicated backgrounds. The accuracy of four distinct transfer
were pre-trained on the IPV102 database and fine-tuned to learning models is contrasted in the accompanying graph. Even
operate in this scenario. The model had low accuracy and could though EfficientNetB7 is significantly more accurate than the
not specify which crops the pest would affect. other three models, all four have comparable accuracy
percentages.
Pest and Crop Management application by UofM Mobile[13] is
a mobile application for identifying pests, insects and weeds.
The app will help users adopt integrated pest control for
Canadian field crops.Insect pests of canola and wheat are
Authorized licensed use limited to: DRDO-CABS. Downloaded on November 14,2022 at 04:29:41 UTC from IEEE Xplore. Restrictions apply.
Table II. DETAILS OF AVAILABLE PEST IDENTIFICATION APPS
Authorized licensed use limited to: DRDO-CABS. Downloaded on November 14,2022 at 04:29:41 UTC from IEEE Xplore. Restrictions apply.
Fig. 3.1 Login Fig. 3.2 English version Fig. 3.3 Hindi version
Fig. 3.4 Kannada Version Fig. 3.5 Expert View Fig. 3.6 Chat Page
Authorized licensed use limited to: DRDO-CABS. Downloaded on November 14,2022 at 04:29:41 UTC from IEEE Xplore. Restrictions apply.
Fig. 3.7 Multi Languagge Weather Forecase Pages
Fig. 3.8. Mites and Beetle Pests Detection through Augmented Reality.
Authorized licensed use limited to: DRDO-CABS. Downloaded on November 14,2022 at 04:29:41 UTC from IEEE Xplore. Restrictions apply.
Fig. 3.9. Stem borer Pest Detection through Augmented Reality.
1211
Authorized licensed use limited to: DRDO-CABS. Downloaded on November 14,2022 at 04:29:41 UTC from IEEE Xplore. Restrictions apply.
IV. CONCLUSION “Identification of plant diseases using convolutional neural
networks”, International Journal of Information Technology. .
The evaluation based on the aforementioned three Int. j. inf. tecnol. 13, 2461–2470, 2021
approaches, namely EfficientNetB7, Xception, Mobilenet [10]. Jia, Yangqing, Evan Shelhamer, Jeff Donahue, Sergey Karayev,
and MobileNetV2, trained using Kaggle dataset [15]. It is Jonathan Long, Ross Girshick, Sergio Guadarrama, and Trevor
found that EfficientNetB7 method has a classification Darrell, "Caffe: Convolutional architecture for fast feature
accuracy of 95.15% for the dataset [15]. When you lack a embedding," In Proceedings of the 22nd ACM international
huge dataset and powerful computing capabilities, you can conference on Multimedia, pp. 675-678. 2014.
employ a transfer-learning-based algorithms. Based on the [11]. Marko Arsenovic, Andras Anderla, Sladojevic, Srdjan, Dubravko
Culibrk, and Darko Stefanovic, "Deep neural networks based
mobile apps survey done for pest identification as depicted in
recognition of plant diseases by leaf image classification,"
table II, currently available apps are inactive, this gives us a
Computational intelligence and neuroscience , Vol. 2016, doi:
need for developing an efficient functional app. 10.1155/2016/3289801, June 2016
The model trained using EfficientNetB7 was then linked to [12]. Wikilimo app, “https://github.com/wikilimo/Wikilimo”, 2019.
the newly created mobile application called Kisaan. It was [13]. UofM Mobiles, “https://appgrooves.com/app/pest-and-crop-
management-by-university-of-manitoba”,
developed using a flutter framework. The images captured
[14]. RajGopal, “https://github.com/rajagopal28/Cedric”,2016.
through the app are passed onto the transfer learning model
[15]. “https://www.kaggle.com/datasets/simranvolunesia/pest-dataset”
to give pest classification output. The outcome of the pest [16]. Jadhav, S.B., Udupi, V.R. & Patil, S.B. Identification of plant
detection is represented by scanning a pest image and the diseases using convolutional neural networks. Int. j. inf. tecnol. 13,
output is then presented in augmented reality form. 2461–2470 (2021). https://doi.org/10.1007/s41870-020-00437-5
REFERENCES
Authorized licensed use limited to: DRDO-CABS. Downloaded on November 14,2022 at 04:29:41 UTC from IEEE Xplore. Restrictions apply.