Plant Disease Recognition A Large-Scale Benchmark Dataset and A Visual Region and Loss Reweighting Approach
Plant Disease Recognition A Large-Scale Benchmark Dataset and A Visual Region and Loss Reweighting Approach
Plant Disease Recognition A Large-Scale Benchmark Dataset and A Visual Region and Loss Reweighting Approach
Abstract— Plant disease diagnosis is very critical for agricul- Index Terms— Plant disease recognition, fine-grained visual
ture due to its importance for increasing crop production. Recent classification, reweighting approach, feature aggregation.
advances in image processing offer us a new way to solve this issue
via visual plant disease analysis. However, there are few works
in this area, not to mention systematic researches. In this paper, I. I NTRODUCTION
we systematically investigate the problem of visual plant disease
recognition for plant disease diagnosis. Compared with other
types of images, plant disease images generally exhibit randomly
distributed lesions, diverse symptoms and complex backgrounds,
P LANT diseases cause severe threats to global food secu-
rity by reducing crop production all over the world.
According to the statistics, about 20%-40% of all crop losses
and thus are hard to capture discriminative information. To facil-
itate the plant disease recognition research, we construct a globally are due to plant diseases [1]. Therefore, plant disease
new large-scale plant disease dataset with 271 plant disease diagnosis is critical to the prevention of spread of plant
categories and 220,592 images. Based on this dataset, we tackle diseases and reduction of economic losses in agriculture.
plant disease recognition via reweighting both visual regions Most of the plant disease diagnosis methods heavily rely on
and loss to emphasize diseased parts. We first compute the either the molecular assay or plant protector’s observation.
weights of all the divided patches from each image based on the
cluster distribution of these patches to indicate the discriminative However, the former is complicated and constrained to cen-
level of each patch. Then we allocate the weight to each loss tralized labs while the latter is time-consuming and prone to
for each patch-label pair during weakly-supervised training to errors. Currently, image-based technologies are being widely
enable discriminative disease part learning. We finally extract applied to various interdisciplinary tasks via deciphering visual
patch features from the network trained with loss reweighting, content, e.g., medical imaging [2], food computing [3] and
and utilize the LSTM network to encode the weighed patch
feature sequence into a comprehensive feature representation. cellular image analysis [4]. Benefitting from recent advances
Extensive evaluations on this dataset and another public dataset in machine learning, especially deep learning [5], we assert
demonstrate the advantage of the proposed method. We expect that plant image analysis and recognition can also provide a
this research will further the agenda of plant disease recognition new way for plant disease diagnosis. Meanwhile, the applica-
in the community of image processing. tions in visual plant disease diagnosis conversely promote the
development of image processing technologies.
Manuscript received March 11, 2020; revised August 28, 2020 and The research and exploration on plant image analysis in this
October 10, 2020; accepted December 28, 2020. Date of publication field have begun to develop, such as aerial phenotyping [6]
January 14, 2021; date of current version January 22, 2021. This work
was supported in part by the National Natural Science Foundation of China and fingerprinting of leaves [1]. However, these methods
under Project 61932003 and Project 61772051; in part by the National Key heavily rely on either expensive devices or complex molecular
Research and Development Plan under Grant 2019YFC1521102; in part technology, and thus are not easily popularized. Recently,
by the Beijing Natural Science Foundation under Grant L182016; in part
by the Beijing Program for International S&T Cooperation Project under some works [7]–[12] adopt deep learning methods for plant
Grant Z191100001619003; and in part by the Shenzhen Research Institute disease recognition. However, most of them directly extract
of Big Data (Shenzhen). The associate editor coordinating the review of deep features from plant disease images without consider-
this manuscript and approving it for publication was Prof. Guo-Jun Qi.
(Corresponding author: Lili Wang.) ing characteristics of the task. In addition, these works are
Xinda Liu and Lili Wang are with the State Key Laboratory of Virtual restricted to small datasets with fewer categories and simple
Reality Technology and Systems, Beijing Advanced Innovation Center for visual backgrounds.
Biomedical Engineering, Beihang University, Beijing 100191, China, and
also with the Peng Cheng Laboratory, Shenzhen 518066, China (e-mail: According to our survey, there are mainly three distinctive
[email protected]; [email protected]). characteristics for plant disease images taken in real-world
Weiqing Min and Shuqiang Jiang are with the Key Laboratory of Intelligent scenarios. (1) Randomly distributed lesions. The foliar
Information Processing, Institute of Computing Technology, Chinese Acad-
emy of Sciences, Beijing 100190, China, and also with the University lesions probably randomly occur in the plant leaves. As shown
of Chinese Academy of Sciences, Beijing 100049, China (e-mail: min- in Fig. 1 (a), the cherry fungal shot hole disease is distributed
[email protected]; [email protected]). in many different parts of the leaf, including the top, left and
Shuhuan Mei is with Beijing Puhui Sannong Technology Company Ltd.,
Beijing 100190, China (e-mail: [email protected]). right positions. Because deep convolutional neural networks
Digital Object Identifier 10.1109/TIP.2021.3049334 trained with image level labels only tend to focus on the
1941-0042 © 2021 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission.
See https://www.ieee.org/publications/rights/index.html for more information.
Authorized licensed use limited to: University of Leeds. Downloaded on August 24,2021 at 22:40:20 UTC from IEEE Xplore. Restrictions apply.
2004 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 30, 2021
Authorized licensed use limited to: University of Leeds. Downloaded on August 24,2021 at 22:40:20 UTC from IEEE Xplore. Restrictions apply.
LIU et al.: PLANT DISEASE RECOGNITION: A LARGE-SCALE BENCHMARK DATASET 2005
TABLE I
S TATISTICS ON E XISTING P LANT D ISEASE D ATASETS
Fig. 2. Disease leaf image samples from various categories of PDD271 (one samples per category). The dataset contains three macro-classes: Fruit Tree,
Vegetable, and Field Crops.
than 75, 000 images belonging to 102 categories for insect recognition methods without high-cost object part locations or
pest recognition. In contrast, PDD271 aims at advancing plant attribute annotations. For example, Yang et al. [32] initialized
disease recognition. We believe that PDD271 and IP102 are many anchors randomly and extracted their features as
very complementary and can jointly promote the development their informativeness using the RPN method, and finally
of intelligent agriculture analysis and understanding in the chose the informative region to improve the classification
image processing and computer vision community. performance. There are also several attention-based methods
proposed for Fine-Grained Visual Classification. For example,
Hu et al. [33] used attention maps to guide the data
B. Fine-Grained Visual Classification augmentation, Peng et al. [34] proposed the object-part
Fine-grained image recognition aims to distinguish attention model to select discriminative regions subjecting to
sub-ordinate categories, such as birds and food. In the early the object-part spatial constraint, and SeNet154 [35] enhance
stage, researchers [25], [26] based on deep learning first the recognition performance with spatial-channel attention.
used strong supervised mechanisms with part bounding box However, attention-based methods probably focus on the
annotations to learn to attend on discriminative parts. Recent most discriminative parts while missing other parts for the
researches [3], [13], [27]–[32] focused on weakly-supervised whole image.
Authorized licensed use limited to: University of Leeds. Downloaded on August 24,2021 at 22:40:20 UTC from IEEE Xplore. Restrictions apply.
2006 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 30, 2021
Authorized licensed use limited to: University of Leeds. Downloaded on August 24,2021 at 22:40:20 UTC from IEEE Xplore. Restrictions apply.
LIU et al.: PLANT DISEASE RECOGNITION: A LARGE-SCALE BENCHMARK DATASET 2007
The whole data construction takes about 2 years. The result- regions and the loss during the weakly-supervised learning to
ing PDD271 contains 220,592 images and 271 categories. emphasize discriminative diseased parts for the purpose of the
As shown in Fig. 3, the minimum number of images per plant disease recognition. As shown in Fig. 6, this framework
category is over 400 and the maximum one is 2000. The mainly consists of three stages, namely Cluster-based Region
balanced distribution ensures the stability of model train- Reweighting (CRR), Training with Loss Reweighting (TLR)
ing. A reliable dataset plays an essential role in developing and Weighted Feature Integration (WFI). CRR takes all the
image processing technologies in a specific area. For example, divided patches from plant disease images as input and sets
HiEve [38] is vital to human-centric analysis, so as ATRW [39] the weight of each patch according to the cluster distribution
to wildlife conservation. Likewise, the proposed dataset of the visual features of these patches. For each patch-label
PDD271 offers a large coverage and diversity of plant dis- pair, TLR allocates the corresponding weight to each loss
eases. It will further the plant disease recognition agenda and during weakly-supervised training in order to enable the
expand the image processing techniques into the agricultural discriminative disease part learning. Based on extracted patch
area. features from TLR and corresponding weights from CRR,
WFI utilizes the LSTM network to encode the weighed patch
IV. F RAMEWORK feature sequence into a comprehensive feature representation.
In this section, we introduce the proposed framework which Section IV-A details CRR, Section IV-B introduces TLR and
explores a multi-scale strategy and reweights both visual Section IV-C presents WFI.
Authorized licensed use limited to: University of Leeds. Downloaded on August 24,2021 at 22:40:20 UTC from IEEE Xplore. Restrictions apply.
2008 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 30, 2021
A. Cluster-Based Region Reweighting Normally, the patches containing similar visual symptoms
are likely to be assigned to the same clusters. In case of small
Many diseases present small and scattered lesions, such
distance among clusters, the visual phenotypes of different
as pumpkin mildew, pear frog-eye leaf spot and actinidia
diseases are similar and hard to distinguish by the deep model.
brown spot. The deep convolutional neural networks trained
Therefore, these clusters are given higher weights to enhance
with image level labels often overlook these lesions while
their influence in follow-up feature learning and integration.
focusing on more salient parts. Considering these situations,
The size of cluster is also an important indicator. There is
we explore a multi-scale strategy by dividing the images into
a highly skewed distribution of different disease patches. For
non-overlapping patches and enlarging every patch to avoid
example, the number of non-diseased patches containing com-
missing diseased patches. However, the disease-independent
plex backgrounds and foliar healthy parts is very large, but the
patches, such as the complex backgrounds and the healthy
number of patches containing cotton eye spot disease is small
parts, are enhanced even more in the above process, which
due to the concentrated symptom of this disease leading to
could lead to severe unbalance between the diseased patches
the poor classification performance. Meanwhile, the distance
and the irrelevant ones. To address this problem, we attempt
between two clusters indicates their visual difference. If one
to use the visual similarity among the same disease to cluster
cluster is far from the other clusters, we can easily obtain
the patches of the same disease. Afterwards, we reweight
discriminative features for this cluster, thus assign a small
the patches based on the clustering result and indicate the
weight to it. Hence we assign these clusters suitable weights
discriminative level of each patch.
to make their influences as balanced as possible.
Formally, all patches from all the original training images
Given all these, we assign the cluster weights according
form a new training set. Let X ∈ Rm×N denotes the visual
to the following rule: the larger size the cluster and the
features of these patches, where m is the dimension of the
farther away from the others, the smaller its weight. We use
visual feature and N is the number of training patches. We then
a monotone decreasing function F = e x/(x−1) to model this
have these patches clustered into k cluster classes c with
change. According to the size of the cluster and the distance
their centroids being {μ1 , μ2 , . . . , μk } ∈ Rm . To compute the
distribution among the cluster centroids, we compute the
weight wx , x ∈ X, the weights of the clusters wc and the
weights of the cluster ci as follows,
probability distribution px of x belonging to over all clusters
are computed. Then, wx is computed as
wci = F(Nci ) × F( d(μi , μ j )), (2)
wx = px · wc , (1) j =i, j ∈1,...,k
where wc = [wc1 , . . . , wci , . . . , wck ] and wci denotes the where Nci is the number of patches in cluster ci and d(μi , μ j )
weight of the cluster ci . is the distance between the centroid μi and μ j .
Authorized licensed use limited to: University of Leeds. Downloaded on August 24,2021 at 22:40:20 UTC from IEEE Xplore. Restrictions apply.
LIU et al.: PLANT DISEASE RECOGNITION: A LARGE-SCALE BENCHMARK DATASET 2009
Authorized licensed use limited to: University of Leeds. Downloaded on August 24,2021 at 22:40:20 UTC from IEEE Xplore. Restrictions apply.
2010 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 30, 2021
TABLE II
Fig. 7. The result of elbow method. The blue line shows that the SSE changes P ERFORMANCE C OMPARISON FOR D IFFERENT T RAINING M ETHODS
with the K , and the orange line is the MA line.
Authorized licensed use limited to: University of Leeds. Downloaded on August 24,2021 at 22:40:20 UTC from IEEE Xplore. Restrictions apply.
LIU et al.: PLANT DISEASE RECOGNITION: A LARGE-SCALE BENCHMARK DATASET 2011
TABLE IV
A BLATION S TUDY ON THE PDD271
Authorized licensed use limited to: University of Leeds. Downloaded on August 24,2021 at 22:40:20 UTC from IEEE Xplore. Restrictions apply.
2012 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 30, 2021
Fig. 10. Qualitative results. From top to bottom, (a) the original image with annotating diseased parts by red boxes, (b) the feature map from the last
convolution layer of VGG16, (c) the feature map from the last convolution layer of ResNet152, (d) the feature map from the last convolution layer of
SeNet154, (e) visualisation of the proposed CRR weights for each patch. The red means high weights and the blue means relatively low weights. For the best
view, we only visualize the weights which are bigger than 0.75. CRR can consider more regions and obtain more characteristics.
TABLE VII Compared with feature maps from typical deep networks,
I MPACT OF O RDERS . T HE ‘ T ’, ‘ B ’, ‘ L’, AND ‘ R ’ D ENOTE THE T OP, we can find that the proposed reweighted maps can cover more
B OTTOM , L EFT AND R IGHT, R ESPECTIVELY. T HE ‘ T 2 B , L 2 R ’ M EANS
THE PATCHES F ROM E ACH I MAGE A RE O RDERED F ROM T OP
discriminative regions. The VGG16 and ResNet152 probably
TO B OTTOM AND L EFT TO R IGHT. T HE ‘ RD ’ D ENOTES THE focus on disease-irrelevant regions, and meanwhile ignores
R ANDOM O RDER . T HE ‘ FIXED ’ D ENOTES T HAT THE O RDER some useful information. Our approach can pay attention
OF THE PATCH L IST FOR E ACH I MAGE I S F IXED , AND
THE ‘ UNFIXED ’ D ENOTES T HAT THE O RDER OF THE
to multiple scattered regions, which is more appropriate for
PATCH L IST FOR E ACH I MAGE I S U NFIXED plant disease recognition. The visualization results of the
PDD271 further demonstrate the effectiveness of the proposed
cluster-based reweighting strategy.
In addition, we further show the confusion matrix of our
method on the PDD271 in Fig. 11, where the vertical axis
shows the ground-truth classes and the horizontal axis shows
the predicted classes. Yellower colors indicate better perfor-
mance. We can see that our method still does not provide per-
fect performance for some plant disease categories. We enlarge
specific regions to highlight the misclassified results and show
some samples from confused categories. We can see that these
plant disease categories are very similar in visual appearance
is diseased no matter where the lesions appear in. Another and texture. Even the humans do not easily distinguish among
possible explanation is that the uncertain order is likely to these disease categories. The probable solution is to design
enhance the power of networks. more fine-grained visual feature learning methods or use
8) Visualization: We visualize different emphasized parts multi-source information from different sensors to classify
in different methods via gradient-weighted class activation these plant disease categories.
heatmap [49]. Fig. 10 shows the visualization results of some
typical deep architectures, such as VGG16 and ResNet152.
The reweighted maps of the proposed cluster-based region C. Experiment on PlantVillage Dataset
reweighting strategy are shown in Fig. 10 (d), where we only Besides the PDD271, we also conduct the evalua-
visualize the weight of the patch x when wx ≥ 0.75. tion on another publicly available benchmark datasets, the
Authorized licensed use limited to: University of Leeds. Downloaded on August 24,2021 at 22:40:20 UTC from IEEE Xplore. Restrictions apply.
LIU et al.: PLANT DISEASE RECOGNITION: A LARGE-SCALE BENCHMARK DATASET 2013
Authorized licensed use limited to: University of Leeds. Downloaded on August 24,2021 at 22:40:20 UTC from IEEE Xplore. Restrictions apply.
2014 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 30, 2021
[15] A. K. Lees, L. Sullivan, J. S. Lynott, and D. W. Cullen, “Development [38] W. Lin et al., “Human in events: A large-scale benchmark for human-
of a quantitative real-time PCR assay for phytophthora infestans and centric video analysis in complex events,” 2020, arXiv:2005.04490.
its applicability to leaf, tuber and soil samples,” Plant Pathol., vol. 61, [Online]. Available: https://arxiv.org/abs/2005.04490
no. 5, pp. 867–876, Oct. 2012. [39] S. Li, J. Li, H. Tang, R. Qian, and W. Lin, “ATRW: A benchmark for
[16] C. H. Bock, G. H. Poole, P. E. Parker, and T. R. Gottwald, “Plant disease Amur tiger re-identification in the wild,” in Proc. 28th ACM Int. Conf.
severity estimated visually, by digital photography and image analysis, Multimedia, Oct. 2020, pp. 2590–2598.
and by hyperspectral imaging,” Crit. Rev. Plant Sci., vol. 29, no. 2, [40] L. van der Maaten and G. Hinton, “Visualizing data using t-SNE,”
pp. 59–107, Mar. 2010. J. Mach. Learn. Res., vol. 9, pp. 2579–2605, Nov. 2008.
[17] F. Ahmad and A. Airuddin, “Leaf lesion detection method using artificial [41] Y. Chen, Y. Bai, W. Zhang, and T. Mei, “Destruction and construction
bee colony algorithm,” in Advances in Computer Science and its learning for fine-grained image recognition,” in Proc. IEEE/CVF Conf.
Applications, vol. 279. Beijing, China: Springer, 2014, pp. 989–995. Comput. Vis. Pattern Recognit. (CVPR), Jun. 2019, pp. 5157–5166.
[18] S. Prasad, P. Kumar, and A. Jain, “Detection of disease using block- [42] K. Simonyan and A. Zisserman, “Very deep convolutional networks for
based unsupervised natural plant leaf color image segmentation,” in large-scale image recognition,” in Proc. Int. Conf. Learn. Represent.,
Swarm, Evolutionary, and Memetic Computing. Beijing, China: Springer, 2015, pp. 1–14.
2011, pp. 399–406. [43] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for
[19] A. Ramcharan, K. Baranowski, P. Mcclowsky, B. Ahmed, and image recognition,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit.
D. P. Hughes, “Using transfer learning for image-based cassava disease (CVPR), Jun. 2016, pp. 770–778.
detection,” Frontiers Plant Sci., vol. 8, p. 1852, Oct. 2017. [44] S. Zagoruyko and N. Komodakis, “Wide residual networks,” in Proc.
Brit. Mach. Vis. Conf., 2016, p. 87.
[20] E. Mwebaze, T. Gebru, A. Frome, S. Nsumba, and J. Tusubira,
[45] G. Huang, Z. Liu, L. Van Der Maaten, and K. Q. Weinberger, “Densely
“iCassava 2019 fine-grained visual categorization challenge,” 2019,
connected convolutional networks,” in Proc. IEEE Conf. Comput. Vis.
arXiv:1908.02900. [Online]. Available: https://arxiv.org/abs/1908.02900
Pattern Recognit. (CVPR), Jul. 2017, pp. 2261–2269.
[21] D. P. Hughes and M. Salathé, “An open access repository of images
[46] C. Szegedy, V. Vanhoucke, S. Ioffe, J. Shlens, and Z. Wojna, “Rethinking
on plant health to enable the development of mobile disease diagnostics
the inception architecture for computer vision,” in Proc. IEEE Conf.
through machine learning and crowdsourcing,” 2015, arXiv:1511.08060.
Comput. Vis. Pattern Recognit. (CVPR), Jun. 2016, pp. 2818–2826.
[Online]. Available: https://arxiv.org/abs/1511.08060
[47] X. Zhang, Z. Li, C. C. Loy, and D. Lin, “PolyNet: A pursuit of structural
[22] H. Yu and C. Son, “Apple leaf disease identification through diversity in very deep networks,” in Proc. IEEE Conf. Comput. Vis.
region-of-interest-aware deep convolutional neural network,” 2019, Pattern Recognit. (CVPR), Jul. 2017, pp. 3900–3908.
arXiv:1903.10356. [Online]. Available: https://arxiv.org/abs/1903.10356 [48] C. Liu et al., “Progressive neural architecture search,” in Proc. Eur. Conf.
[23] C. Xie et al., “Multi-level learning features for automatic classification Comput. Vis., Sep. 2018, pp. 19–35.
of field crop pests,” Comput. Electron. Agricult., vol. 152, pp. 233–241, [49] R. R. Selvaraju, M. Cogswell, A. Das, R. Vedantam, D. Parikh, and
Sep. 2018. D. Batra, “Grad-CAM: Visual explanations from deep networks via
[24] X. Wu, C. Zhan, Y.-K. Lai, M.-M. Cheng, and J. Yang, “IP102: A gradient-based localization,” in Proc. IEEE Int. Conf. Comput. Vis.
large-scale benchmark dataset for insect pest recognition,” in Proc. (ICCV), Oct. 2017, pp. 618–626.
IEEE/CVF Conf. Comput. Vis. Pattern Recognit. (CVPR), Jun. 2019, [50] F. N. Iandola, M. W. Moskewicz, K. Ashraf, S. Han, W. J. Dally,
pp. 8787–8796. and K. Keutzer, “Squeezenet: Alexnet-level accuracy with 50x fewer
[25] J. Donahue et al., “Decaf: A deep convolutional activation feature for parameters and 1mb model size,” 2016, arXiv:1602.07360. [Online].
generic visual recognition,” in Proc. Int. Conf. Mach. Learn., vol. 2014, Available: https://arxiv.org/abs/1602.07360
pp. 647–655. [51] H. Hu and G. Qi, “State-frequency memory recurrent neural networks,”
[26] N. Zhang, J. Donahue, R. Girshick, and T. Darrell, “Part-based R-CNNs in Proc. Int. Conf. Mach. Learn., vol. 70, 2017, pp. 1568–1577.
for fine-grained category detection,” in Proc. Eur. Conf. Comput. Vis., [52] X. Shu, J. Tang, G. Qi, W. Liu, and J. Yang, “Hierarchical long
2014, pp. 834–849. short-term concurrent memory for human interaction recognition,” IEEE
[27] T. Xiao, Y. Xu, K. Yang, J. Zhang, Y. Peng, and Z. Zhang, “The Trans. Pattern Anal. Mach. Intell., early access, Sep. 17, 2018, doi: 10.
application of two-level attention models in deep convolutional neural 1109/TPAMI.2019.2942030.
network for fine-grained image classification,” in Proc. IEEE Conf. [53] K. Chen, Y. Chen, C. Han, N. Sang, and C. Gao, “Hard sample
Comput. Vis. Pattern Recognit. (CVPR), Jun. 2015, pp. 842–850. mining makes person re-identification more efficient and accurate,”
[28] M. Lam, B. Mahasseni, and S. Todorovic, “Fine-grained recognition as Neurocomputing, vol. 382, pp. 259–267, Mar. 2020.
HSnet search for informative image parts,” in Proc. IEEE Conf. Comput. [54] Y. Zhao, Z. Jin, G. Qi, H. Lu, and X. Hua, “An adversarial approach to
Vis. Pattern Recognit. (CVPR), Jul. 2017, pp. 6497–6506. hard triplet generation,” in Proc. Eur. Conf. Comput. Vis., vol. 11213,
[29] T.-Y. Lin, A. RoyChowdhury, and S. Maji, “Bilinear convolutional neural Sep. 2018, pp. 508–524.
networks for fine-grained visual recognition,” IEEE Trans. Pattern Anal.
Mach. Intell., vol. 40, no. 6, pp. 1309–1322, Jun. 2018.
[30] S. Jiang, W. Min, L. Liu, and Z. Luo, “Multi-scale multi-view deep
feature aggregation for food recognition,” IEEE Trans. Image Process.,
vol. 29, pp. 265–276, 2020.
[31] W. Min, L. Liu, Z. Luo, and S. Jiang, “Ingredient-guided cascaded multi-
attention network for food recognition,” in Proc. 27th ACM Int. Conf.
Multimedia, Oct. 2019, pp. 99–107.
[32] Z. Yang, T. Luo, D. Wang, Z. Hu, J. Gao, and L. Wang, “Learning
to navigate for fine-grained classification,” in Proc. Eur. Conf. Comput.
Vis., Sep. 2018, pp. 438–454.
[33] T. Hu and H. Qi, “See better before looking closer: Weakly supervised
data augmentation network for fine-grained visual classification,” CoRR,
vol. abs/1901.09891, 2019.
[34] Y. Peng, X. He, and J. Zhao, “Object-part attention model for fine-
grained image classification,” IEEE Trans. Image Process., vol. 27, no. 3, Xinda Liu received the B.E. degree from the
pp. 1487–1500, Mar. 2018. China University of Mining and Technology, Bei-
[35] J. Hu, L. Shen, and G. Sun, “Squeeze-and-excitation networks,” in jing, China, in 2013, and the M.E. degree from
Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit., Jun. 2018, Ningxia University, Yinchuan, China, in 2016. He is
pp. 7132–7141. currently pursuing the Ph.D. degree with the State
[36] G.-J. Qi, “Hierarchically gated deep networks for semantic segmen- Key Laboratory of Virtual Reality Technology and
tation,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), Systems, School of Computer Science and Engineer-
Jun. 2016, pp. 2267–2275. ing, Beihang University, Beijing. He is also with
[37] G.-J. Qi, X.-S. Hua, Y. Rui, J. Tang, and H.-J. Zhang, “Image classifica- the Peng Cheng Laboratory. His research interests
tion with kernelized spatial-context,” IEEE Trans. Multimedia, vol. 12, include machine learning and image processing.
no. 4, pp. 278–287, Jun. 2010.
Authorized licensed use limited to: University of Leeds. Downloaded on August 24,2021 at 22:40:20 UTC from IEEE Xplore. Restrictions apply.
LIU et al.: PLANT DISEASE RECOGNITION: A LARGE-SCALE BENCHMARK DATASET 2015
Weiqing Min (Member, IEEE) received the B.E. Lili Wang (Member, IEEE) received the Ph.D.
degree from Shandong Normal University, Jinan, degree from Beihang University, Beijing, China. She
China, in 2008, the M.E. degree from Wuhan Uni- is currently a Professor with the School of Computer
versity, Wuhan, China, in 2010, and the Ph.D. Science and Engineering, Beihang University, where
degree from the National Laboratory of Pattern she is also a Researcher with the State Key Labo-
Recognition, Institute of Automation, Chinese Acad- ratory of Virtual Reality Technology and Systems.
emy of Sciences, in 2015. He is currently an She is also with the Beijing Advanced Innovation
Associate Professor with the Key Laboratory of Center for Biomedical Engineering. Her research
Intelligent Information Processing, Institute of Com- interests include virtual reality, augmented reality,
puting Technology, Chinese Academy of Sciences. mixed reality, real-time, and realistic rendering.
His current research interests include multimedia
content analysis, understanding and applications, food computing, and
geo-multimedia computing. He has authored or coauthored more than 40 peer-
referenced papers in relevant journals and conferences, including ACM
Computing Surveys, IEEE T RANSACTIONS ON I MAGE P ROCESSING, IEEE Shuqiang Jiang (Senior Member, IEEE) is cur-
T RANSACTIONS ON M ULTIMEDIA, ACM TOMM, IEEE Multimedia Maga- rently a Professor with the Institute of Computing
zine, ACM Multimedia, AAAI, and IJCAI. He organized several special issues Technology, Chinese Academy of Sciences (CAS),
on international journals, such as IEEE Multimedia Magazine, Multimedia Beijing, and a Professor with University of Chinese
Tools and Applications, as a Guest Editor. He served as a TPC member of Academy of Sciences. He is also with the Key Lab-
many academic conferences, including ACM MM, AAAI, and IJCAI. He was oratory of Intelligent Information Processing, CAS.
a recipient of the 2016 ACM TOMM Nicolas D. Georganas Best Paper Award His research interests include multimedia processing
and the 2017 IEEE Multimedia Magazine Best Paper Award. and semantic understanding, pattern recognition, and
computer vision. He has authored or coauthored
more than 150 articles in the related research topics.
He was supported by the New-Star Program of
Science and Technology of Beijing Metropolis in 2008, the NSFC Excellent
Shuhuan Mei received the M.E. degree from Young Scientists Fund in 2013, and the Young top-notch talent of Ten
the Shandong University of Science and Technol- Thousand Talent Program in 2014. He has also served as a TPC member for
ogy, China. His current research interests include more than 20 well-known conferences, including ACM Multimedia, CVPR,
multimedia content analysis, understanding, and ICCV, IJCAI, AAAI, ICME, ICIP, and PCM. He won the Lu Jiaxi Young
applications. Talent Award from Chinese Academy of Sciences in 2012, and the CCF
Award of Science and Technology in 2012. He is the Senior Member of
CCF, a member of ACM, an Associate Editor of ACM TOMM, IEEE
M ULTI M EDIA, and Multimedia Tools and Applications. He is the Vice Chair
of IEEE CASS Beijing Chapter and ACM SIGMM China chapter. He is the
General Chair of ICIMCS 2015, the Program Chair of ACM Multimedia
Asia2019 and PCM2017.
Authorized licensed use limited to: University of Leeds. Downloaded on August 24,2021 at 22:40:20 UTC from IEEE Xplore. Restrictions apply.