Earth Observation Data Analytics Using Machine and Deep Learning Modern Tools Applications and Challenges
Earth Observation Data Analytics Using Machine and Deep Learning Modern Tools Applications and Challenges
Edited by
Sanjay Garg, Swati Jain, Nitant Dube and
Nebu Varghese
This publication is copyright under the Berne Convention and the Universal Copyright
Convention. All rights reserved. Apart from any fair dealing for the purposes of research
or private study, or criticism or review, as permitted under the Copyright, Designs and
Patents Act 1988, this publication may be reproduced, stored or transmitted, in any
form or by any means, only with the prior permission in writing of the publishers, or in
the case of reprographic reproduction in accordance with the terms of licences issued
by the Copyright Licensing Agency. Enquiries concerning reproduction outside those
terms should be sent to the publisher at the undermentioned address:
While the authors and publisher believe that the information and guidance given in this
work are correct, all parties must rely upon their own skill and judgement when making
use of them. Neither the author nor publisher assumes any liability to anyone for any
loss or damage caused by any error or omission in the work, whether such an error or
omission is the result of negligence or any other cause. Any and all such liability is
disclaimed.
The moral rights of the author to be identified as author of this work have been
asserted by him in accordance with the Copyright, Designs and Patents Act 1988.
1 Introduction 1
Preeti Kathiria, Swati Jain, Kimee Joshi and Nebu Varghese
1.1 Earth observation data 1
1.1.1 Organization 2
1.2 Categories of EO data 2
1.2.1 Passive imaging system 3
1.2.2 Active imaging system 5
1.3 Need of data analytics in EO data 6
1.4 Data analytics methodology 6
1.4.1 Machine learning 7
1.4.2 Deep learning 8
1.5 Data visualization techniques 8
1.5.1 Cartogram map 8
1.5.2 Heat map 9
1.5.3 Choropleth map 9
1.6 Types of inferences from data analytics (application areas) 9
1.6.1 Agriculture 9
1.6.2 Forestry 9
1.6.3 Land cover classification 10
1.6.4 Flooding 12
1.6.5 Maritime 12
1.6.6 Defence and security 13
1.6.7 Wetland 13
1.7 Conclusion 13
References 14
Part III: Tools and technologies for Earth Observation data 175
11 The application of R software in water science 177
Nasrin Fathollahzadeh Attar and Mohammad Taghi Sattari
11.1 Introduction 177
11.1.1 What is hydrology? 178
11.1.2 What is computational hydrology? 178
11.1.3 What is hydroinformatics? 178
11.1.4 Free, open-source software (FOSS) 179
11.1.5 What is GitHub? 179
11.2 Material and methods 179
11.2.1 What is R? What is an integrated development
environment (IDE)? 179
11.2.2 What are R packages? 180
11.2.3 What are cheatsheets? 180
11.2.4 What are R communities? 180
11.2.5 What is RPubs? 181
11.2.6 What are popular conferences in R? 181
11.2.7 What is joss (open source software)? 182
x Earth observation data analytics using machine and deep learning
14 Conclusion 227
Sanjay Garg, Kimee Joshi and Nebu Varghese
14.1 Excerpts from various chapters 227
14.2 Issues and challenges 229
14.2.1 Collecting meaningful and real-time data 229
14.2.2 Data storage 230
14.2.3 Resolution; quality promotion 230
14.2.4 Budget limitations 230
14.2.5 Standardization 231
14.2.6 Lack of ground truth data 231
14.2.7 Processing and analysis 231
References 231
Index 233
This page intentionally left blank
About the editors
Nitant Dube is the group director of MOSDAC Research at the joint Space
Applications Centre (SAC) and ISRO, India. His research fields include satellite
image processing, big data analytics, AI/ML and its applications for Earth obser-
vation data, geo-intelligence and web-based processing. He is involved in the
design and development of software for meteorological and oceanographic appli-
cations. He has been responsible for the design and development of data products
and information processing systems for Indian remote sensing satellites and has
contributed towards the development and operationalization of data processing
systems at Indian and International ground stations. He is an ISRO nominated
member for the CEOS Working Group on Information System and Services
(WGISS). He holds a PhD degree in Computer Science from Nirma University,
Ahmedabad (GJ), India.
Nebu Varghese is an assistant manager (GIS) in the Land and Municipal Service at
Dholera Industrial City Development Limited (DICDL), India. He works in the
areas of GIS systems analysis and prepare design for new GIS methodologies, land
xiv Earth observation data analytics using machine and deep learning
use mapping, land cover mapping, urban land use analysis, spatial data manage-
ment, satellite image processing and analysis, machine learning and deep learning.
Currently, he is working on regional and city-level planning projects, where he
employs the most cutting-edge technologies for building information model (BIM)
to GIS Integration with micro-level asset information of all Infrastructure in city
development. He has been involved in various government DST, ISRO, and IIMA
funded projects and was also involved in the innovation hub Malawi project with
the Deutsche Gesellschaft für Internationale Zusammenarbeit (GIZ), India. He is a
member of ISPRS. He holds a master’s degree in remote sensing & GIS from Sam
Higginbottom University of Agriculture, Technology & Sciences (SHUATS),
Prayagraj (UP), India.
Foreword
Dr M.B. Potdar
Former Scientist (Indian Space Research Organization)
ISRO/Space Applications Centre, Govt. of India
Former Project Director (Bhaskaracharya National Institute for
Space Applications and Geo-informatics) BISAG, Govt. of Gujarat.
Chapter 1
Introduction
Preeti Kathiria1, Swati Jain1, Kimee Joshi1 and
Nebu Varghese2
Sustainable development and climate change are problems that require immediate
solutions, and both are critical to humanity’s present and future well-being [1,2].
Also, human activities are increasing strain on natural resources, which has a global
impact on the environment. Continuous and ongoing monitoring is required to
analyze, comprehend, and minimize these environmental changes [3]. The United
Nations (UN) seeks a sustainable development model for this generation and future
generations, as well as shared prosperity for people and the planet, through the
promotion of its Sustainable Development Agenda and the United Nations
Framework Convention on Climate Change (UNFCCC) [4,5]. The UN has defined
a set of 17 sustainable development goals (SDGs) as a plan of action to reach peace
and prosperity for all people on our planet by 2030. Several benchmarks and
indicators for each of the 17 goals are used to measure, track, and report the
development of every nation. The global framework established by the UN is
designed around 169 targets, and 232 indicators, 71 (42%) of these targets and 30
(13%) of the indicators can be measured directly or indirectly by Earth observation
(EO) [6–8].
EO plays an essential role in advancing many of the SDGs. Addressing sci-
entific issues like global warming and climate change, ecological change, and
reduction effects of habitat and biodiversity deterioration and producing statistics
and indicators that allow the quantification of SD. The UN report has shown the
viability of using EO data to produce official statistics, including SDGs statistics
like agricultural, urban, and land planning, or food security indicators [9]. Data on
the state of the atmosphere [10], oceans [11], crops [12], forests [13], climate [14],
natural disasters [15], natural resources [16], urbanization [17], biodiversity [18],
and human conditions [19] can be provided by EO. The SDGs that benefit from all
EO indicators are zero hunger (SDG 2), clean water and sanitation (SDG 6),
1
Institute of Technology, Nirma University, India
2
Dholera Industrial City Development Limited, India
2 Earth observation data analytics using machine and deep learning
climate action (SDG 13), life below water (SDG 14), and partnership for the goals
(SDG 17). EOs from satellites and airborne and in situ sensors provide accurate and
reliable information on the state of the atmosphere, oceans, coasts, rivers, soils,
crops, forests, ecosystems, natural resources, ice, snow, and building infrastructure
it changes over time. These observations are directly or indirectly required for all
governmental functions, all economic sectors, and nearly all daily societal activities
[7]. EO satellites make real-time observations of the land, ocean, atmosphere,
cryosphere, and carbon cycle from space, which continuously relay this data to the
ground.
1.1.1 Organization
This paper presents EO data along with various applications. The rest of the paper is
organized as follows: Section 1.2 discusses the categories of the EO data, Section 1.3
describes the need of data analytics, Section 1.4 describes the data analytics metho-
dology, Section 1.5 shows a data visualization techniques, Section 1.6 presents the
application areas, and concluding remarks are in Section 1.7.
Earth
observation data
Hyperspectral
Panchromatic
Panchromatic scanners typically capture electromagnetic radiation (EMR) in a
single band that includes all wavelengths from the visible to infrared spectrum. A
grey-scale image, which makes pixels with lower image values appear dark and
those with higher values appear bright, is the most common way to display pan-
chromatic data. Concerning optical imagery, panchromatic channels typically
record low values for water and dense vegetation and high values for urban and
bare areas [21]. Panchromatic sensors produce images with higher spatial resolu-
tion than multispectral scanners [20,21].
Multispectral
The creation of “natural colour” images using measurements from three visible
spectrum bands is a typical example of a multispectral image (narrow bands cen-
tered around the blue, green, and red wavelengths) [20]. As the name suggests,
multispectral scanners (MSS) are a specific form of remote sensing equipment that
detects and digitally records radiation in various, defined wavelength areas of the
visible and infrared parts of the electromagnetic spectrum [21]. Multi-spectral
instruments typically have to collect energy on larger spatial extents to “fill” the
imaging detector, resulting in a lower resolution than for panchromatic images
because the range of wavelengths contributing to the radiation energy detected by
the sensor is reduced.
Hyperspectral
Hyperspectral scanners collect image data for hundreds of spectral channels.
Instead of assigning primary colours (red, green, blue) to each pixel, hyperspectral
imaging (HSI) analyzes a broad spectrum of light. To provide more information on
what is imaged, the light striking each pixel is broken down into many different
spectral bands [22].
In Figure 1.2, the images a and b handled by [23,24] are shown here as an
example of a panchromatic image obtained from SPOT satellite with 10 m resolution
and a multispectral image obtained from plants cope with 3.7 m resolution,
4 Earth observation data analytics using machine and deep learning
Figure 1.2 (a) Panchromatic image. (b) Multispectral image. (c) Hyperspectral image.
respectively. In the same figure, image c shows the picture of a portion of Horseshoe
Bay Village in Xiong’ with 0.5 m spatial resolution reported by the Institute of
Remote Sensing and Digital Earth of the Chinese Academy of Sciences and the
Shanghai Institute of Technical Physics of the Chinese Academy of Sciences [25].
1.2.2.1 Lidar
Light detection and ranging (lidar) is a technique that uses a transmitted laser pulse
to detect the presence of a target and measures the distance to the target based on
the time and intensity of any reflected return pulse. Lidar is an active remote sen-
sing technology that operates in the ultraviolet to near-infrared wavelength range.
Lidar systems are available on a variety of sensing platforms, including satellite,
airborne, and ground-based systems [21].
1.2.2.2 Radar
RADAR is an acronym for RAdio Detection And Ranging, that is, using actively
transmitted radio waves to detect objects and determine their position or ‘range’. For
EO, the X, C, and L bands are the most commonly used [21]. Compared to infrared and
optical sensing devices, the primary goal of radar is to detect distant targets under
adverse weather conditions and determine their distance, range, and precision. The
radar has a transmitter that serves as an illumination source for target placement. It
generally operates in the microwave region of the electromagnetic spectrum, which is
measured in Hertz [30]. The electrical wave vibrations in the transmitted radar signals
can be constrained to a single plane, that is, perpendicular to the wave propagation
direction (rather than vibrating in all directions perpendicular to that of propagation).
The term “polarization” refers to this filtering procedure. In imaging radar, there are
two orthogonal polarization modes known as horizontal (H) and vertical (V), which are
normally transmitted individually [21]. Two types of radar-based systems are com-
monly used for microwave imaging on aircraft and satellite platforms:
1. Side-Looking Airborne Radar (SLAR) or Real Aperture Radar (RAR)
2. Synthetic Aperture Radar (SAR)
SLAR stands for synthetic aperture radar (Side Looking Airborne Radar). The
illumination of both Real Aperture Radar and Synthetic Aperture Radar is typically
perpendicular to the flight path, making them side-looking systems. The along-
track resolution, also known as the azimuth direction, distinguishes one system
from another. Actual Aperture Radars’ Azimuth resolution is based on the antenna
beam width, which is proportional to the distance between the radar and the target
(slant range). Using a series of signals that have been stored in the system memory,
synthetic aperture radar uses signal processing to create an aperture that is hundreds
of times more significant than the actual antenna [31].
SLAR was the first active sensor to produce terrain imagery from back-
scattered microwave radiation. An antenna is mounted beneath the platform in
SLAR to produce a fan beam (wide vertically and narrow horizontally) pointing to
the platform’s side [32].
6 Earth observation data analytics using machine and deep learning
Figure 1.5 (a) Cartogram map [62]. (b) Heat map [63]. (c) Choropleth map [64].
Introduction 9
proportional to the values of the variable being measured. Then, to correspond to its
corresponding value, each segment is assigned a different colour or shade. As a
result, the data is more directly related to the land area to which it refers [60,61].
1.6.1 Agriculture
For planning and decision-making objectives such as distribution and storage of
food grains, governmental policies, pricing, procurement, and food security, among
others, it is necessary to have access to crop statistics. The Ministry of Agriculture
and Farmers’ Welfare successfully employs modern satellite remote sensing tech-
nology in such decision-making [65].
Satellite imagery has the potential to increase revenue generation for agri-
cultural applications by providing information related to crop type, crop insurance
damage assessment, production management techniques, fertilizer application
requirements, yield estimations, re-growth monitoring, illicit crop monitoring, pest
and invasive species monitoring, and irrigation requirements and application,
monitoring agri-environmental measures (such as acreage) to inform subsidy allo-
cations, field boundary management, crop health mapping, field scale mapping, and
storm damage assessment [66] (see Figure 1.7).
1.6.2 Forestry
Over 1.6 billion people rely on forests for their livelihoods and their sources of
food, medicine, and fuel. Forests cover 31% of the total area of the planet. Provide
10 Earth observation data analytics using machine and deep learning
more information for Obtaining information on forest acreage, stand density, spe-
cies composition, age, and condition to be recognized as a For management pur-
poses, a single unit) surveying, assessing, and keeping track of forest health,
Updating forest management plans: tree cutting, delineation, and tracking of par-
cels, estimating biomass, assessing plant health, and plantation surveillance,
Estimating damage from fire, storms, and other extreme weather, Conservation
area planning and protection, “Conducting fuel analysis and locating the locations
where the Fire risk is high, Deforestation mapping, Monitoring of forest con-
servation and regrowth initiatives [66] (see Figure 1.8).
Agriculture
Yield
estimation Crop type
Crop health
monitoring
Precision Regrowth
farming monitoring
Illicit crop
Crop pests monitoring
1.6.4 Flooding
Monitoring soil moisture and water levels, and how they fluctuate over time pro-
vide a solid indicator of how likely it is that flood and drought threats may occur.
Combining high fidelity interferometric synthetic aperture radar (SAR) measure-
ments with ground truth measurements, S-band SAR, medium resolution optical
images, and digital elevation modeling can be used to achieve this. Combining this
with accurate weather forecasts to determine the likelihood and anticipated amount
of rain enables the detection of potential flood and drought concerns [66].
1.6.5 Maritime
Oceans comprise 96.5% of the water on Earth and makeup 70% of its surface. Ship
tracking data from bathymetry is used to create nautical charts and measure beach erosion,
subsidence, and sea levels. Detection of iceberg threats on shipping routes, protecting the
environment in the ocean, oil spill monitoring, detection of an illegal oil discharge,
detection of unlicensed fishing vessels, port surveillance, detection of incoming hostile
objects in maritime piracy, and marine environmental protection [66] (see Figure 1.9).
Ship detection
Detection of illegal ships, vessels
detection, fisheries monitoring, shipping
traffic
Hydrology
Flood extent, inland waterways, water
levels, ice floe, iceberg movements,
glacier and lake monitoring
1.6.5.3 Hydrology
A hydrological survey is essential for understanding the coastal zones and inland
waterways of a territory. Timely and reliable assessments of available water
resources via satellites and models provide critical input for developing strategies
and water management in the country [67]. Coastal and tidal zones that change
frequently due to meteorological influences can be of concern. Using satellite
technology, information such as shallow water depths, mudflat topology, and the
presence or absence of outflow or sediments can be derived. Satellite imagery
provides information on changing bathymetry in coastal zones, which is especially
useful around ports and busy shipping areas. The surveys that can be conducted
from space provide a consistent and accurate overview [66].
1.6.7 Wetland
Wetlands are the most productive ecosystems, with a diverse range of flora and
fauna. Wetland conservation is therefore critical for preserving biological diversity.
Wetlands are currently under stress due to biological diversity loss, deteriorating
water quality, sedimentation and shrinkage in the area, infestation by unwanted
weeds, and other factors. Remote sensing data is the primary source of information
for monitoring and mapping large areas, such as wetland extent, distribution, and
wetland types such as freshwater, peat swamps, and non-forested peatlands, among
others [68].
1.7 Conclusion
In this paper, we have presented detailed information on EO data and various
types of EO data. Then, we have listed various data analytics methods used to
represent the data into meaningful manner. Then, we have listed some data
visualization techniques. Finally, this paper presented types of inferences from
data analytics.
14 Earth observation data analytics using machine and deep learning
References
[1] Steffen W, Richardson K, Rockström J, et al. Planetary boundaries: guiding
human development on a changing planet. Science, 2015;347(6223):1259855;
2015.
[2] Biermann F, Bai X, Bondre N, et al. Down to earth: contextualizing the
Anthropocene. Global Environmental Change. 2016;39:341–350.
[3] Giuliani G, Camara G, Killough B, et al. Earth observation open science:
enhancing reproducible science using data cubes. MDPI. 2019;4:147.
[4] Transforming Our World: The 2030 Agenda for Sustainable Development,
2015. Available from: https://undocs. org/en/A/RES/70/71.
[5] Paris Agreement, 2015. Available from: https://unfccc.int/process-and-
meetings/the-paris-agreement/theparis- agreement.
[6] Persello C, Wegner JD, Hänsch R, et al. Deep learning and earth observation
to support the sustainable development goals: current approaches, open
challenges, and future opportunities. IEEE Geoscience and Remote Sensing
Magazine. 2022;10(2):172–200.
[7] Earth Observations in Support of the 2030 Agenda for Sustainable
Development. Available from: https://www.earthobservations.org/docu-
ments/publications/201703_geo_eo_for_2030_agenda.pdf.
[8] Estoque RC. A review of the sustainability concept and the state of SDG
monitoring using remote sensing. Remote Sensing. 2020;12(11):1770.
[9] Ferreira B, Iten M, and Silva RG. Monitoring sustainable development by
means of earth observation data and machine learning: a review.
Environmental Sciences Europe. 2020;32(1):1–17.
[10] Boyte SP, Wylie BK, Howard DM, et al. Estimating carbon and showing
impacts of drought using satellite data in regression-tree models.
International Journal of Remote Sensing. 2018;39(2):374–398.
[11] Poursanidis D, Topouzelis K, and Chrysoulakis N. Mapping coastal marine
habitats and delineating the deep limits of the Neptune’s seagrass meadows
using very high resolution Earth observation data. International Journal of
Remote Sensing. 2018;39(23):8670–8687.
[12] Reza MN, Na IS, Baek SW, et al. Rice yield estimation based on K-means
clustering with graph-cut segmentation using low-altitude UAV images.
Biosystems Engineering. 2019;177:109–121.
[13] Puletti N, Chianucci F, Castaldi C, et al. Use of Sentinel-2 for forest clas-
sification in Mediterranean environments. Annals of Silvicultural Research.
2018;42(1):32–38.
[14] Sathiaraj D, Huang X, and Chen J. Predicting climate types for the
Continental United States using unsupervised clustering techniques.
Environmetrics. 2019;30(4):e2524.
[15] Lizundia-Loiola J, Otón G, Ramo R, et al. A spatio-temporal active-fire
clustering approach for global burned area mapping at 250 m from MODIS
data. Remote Sensing of Environment. 2020;236:111493.
Introduction 15
[16] Sharma B, Kumar M, Denis DM, et al. Appraisal of river water quality using
open-access earth observation data set: a study of river Ganga at Allahabad
(India). Sustainable Water Resources Management. 2019;5(2):755–765.
[17] Firozjaei MK, Sedighi A, Argany M, et al. A geographical direction-based
approach for capturing the local variation of urban expansion in the appli-
cation of CA-Markov model. Cities. 2019;93:120–135.
[18] Wang L, Dong Q, Yang L, et al. Crop classification based on a novel feature
filtering and enhancement method. Remote Sensing. 2019;11(4):455.
[19] Foody GM, Ling F, Boyd DS, et al. Earth observation and machine learning
to meet Sustainable Development Goal 8.7: mapping sites associated with
slavery from space. Remote Sensing. 2019;11(3):266.
[20] Newcomers Earth Observation Guide. Available from: https://business.esa.
int/newcomers-earth-observation-guide.
[21] Earth observation: Data, processing and applications – CRC for spatial ...
Australia and New Zealand CRC for Spatial Information. Available from:
https://www.crcsi.com.au/assets/Consultancy-Reports-and-Case-Studies/Earth-
Observation-reports-updated-Feb-2019/Vol1-Introduction-low-res-0.8MB.pdf.
[22] Vasefi F, MacKinnon N, and Farkas D. Hyperspectral and multispectral
imaging in dermatology. In: Imaging in Dermatology. New York, NY:
Elsevier; 2016. p. 187–201.
[23] Principles of Remote Sensing and Processing. Available from: https://crisp.
nus.edu.sg/ research/tutorial/opt_int.htm.
[24] Satellite imagery and archive, 2021. Available from: https://www.planet.
com/products/planet-imagery/.
[25] Wei L, Wang K, Lu Q, et al. Crops fine classification in airborne hyper-
spectral imagery based on multi-feature fusion and deep learning. Remote
Sensing. 2021;13(15):2917.
[26] What is Thermal Infrared Imaging, 2022. Available from: https://movitherm.
com/knowledgebase/thermal-infrared-imaging.
[27] Teena M and Manickavasagan A. Thermal infrared imaging. In: Imaging with
Electromagnetic Spectrum. New York, NY: Springer; 2014. p. 147–173.
[28] OpenSnow. Infrared Satellite – Explained. Available from: https://open-
snow.com/news/post/infrared-satellite-explained.
[29] Rap Real-Time Weather. Available from: https://weather.rap.ucar.edu/.
[30] Agarwal T. Radar – Basics, Types, Working, Range Equation & Its
Applications, 2020. Available from: https://www.elprocus.com/radar-basics-
types-and-applications/.
[31] Radar Course 2. Available from: https://earth.esa.int/eogateway/missions/
ers/radar-courses/radar-course-2#: :text=Real
[32] Remote Sensing GIS Applications 3(2+1). Available from: http://ecourse-
sonline.iasri.res.in/mod/page/view.php?id=2068.
[33] Sentinel Datset. Available from: https://scihub.copernicus.eu/dhus/#/home.
[34] Narain A. Data Analytics to Shape the Future of Earth Observation, 2019.
Available from: https://www.geospatialworld.net/blogs/future-of-earth-
observation/.
16 Earth observation data analytics using machine and deep learning
[35] Pesaresi M, Syrris V, and Julea A. A new method for earth observation data
analytics based on symbolic machine learning. Remote Sensing. 2016;8
(5):399.
[36] Yang C, Yu M, Li Y, et al. Big Earth data analytics: a survey. Big Earth
Data. 2019;3(2):83–107.
[37] Big Earth Data Observation. Available from: https://ati.ec.europa.eu/sites/
default/files/2020-06/Big%20Data
[38] Haase D, Jänicke C, and Wellmann T. Front and back yard green analysis
with subpixel vegetation fractions from earth observation data in a city.
Landscape and Urban Planning. 2019;182:44–54.
[39] Dogan T and Uysal KA. The impact of feature selection on urban land cover
classification. International Journal of Intelligent Systems and Applications
in Engineering. 2018;6:59–64.
[40] Marmanis D, Datcu M, Esch T, et al. Deep learning earth observation clas-
sification using ImageNet pretrained networks. IEEE Geoscience and
Remote Sensing Letters. 2015;13(1):105–109.
[41] Yi Y and Zhang W. A new deep-learning-based approach for earthquake-
triggered landslide detection from single-temporal RapidEye satellite
imagery. IEEE Journal of Selected Topics in Applied Earth Observations
and Remote Sensing. 2020;13:6166–6176.
[42] Hoeser T and Kuenzer C. Object detection and image segmentation with
deep learning on earth observation data: a review – Part I: evolution and
recent trends. Remote Sensing. 2020;12(10):1667.
[43] Som-ard J, Immitzer M, Vuolo F, et al. Mapping of crop types in 1989, 1999,
2009 and 2019 to assess major land cover trends of the Udon Thani Province,
Thailand. Computers and Electronics in Agriculture. 2022;198:107083.
[44] Kuchler PC, Simões M, Ferraz R, et al. Monitoring complex integrated
crop–livestock systems at regional scale in Brazil: a big Earth Observation
data approach. Remote Sensing. 2022;14(7):1648.
[45] Memon N, Parikh H, Patel SB, et al. Automatic land cover classification of
multi-resolution dualpol data using convolutional neural network (CNN).
Remote Sensing Applications: Society and Environment. 2021;22:100491.
[46] Löw F, Stieglitz K, and Diemar O. Terrestrial oil spill mapping using
satellite earth observation and machine learning: a case study in South
Sudan. Journal of Environmental Management. 2021;298:113424.
[47] Seydi ST, Hasanlou M, Amani M, et al. Oil spill detection based on multi-
scale multidimensional residual CNN for optical remote sensing imagery.
IEEE Journal of Selected Topics in Applied Earth Observations and Remote
Sensing. 2021;14:10941–10952.
[48] Hong Z, Yang T, Tong X, et al. Multi-scale ship detection from SAR and
optical imagery via a more accurate YOLOv3. IEEE Journal of Selected
Topics in Applied Earth Observations and Remote Sensing. 2021;14:6083–
6101.
[49] El Naqa I and Murphy MJ. What is machine learning? In: Machine Learning
in Radiation Oncology. New York, NY: Springer; 2015. p. 3–11.
Introduction 17
[50] Qiu J, Wu Q, Ding G, et al. A survey of machine learning for big data
processing. EURASIP Journal on Advances in Signal Processing. 2016;2016
(1):1–16.
[51] Picoli MCA, Camara G, Sanches I, et al. Big earth observation time series
analysis for monitoring Brazilian agriculture. ISPRS Journal of
Photogrammetry and Remote Sensing. 2018;145:328–339.
[52] Löw F, Michel U, Dech S, et al. Impact of feature selection on the accuracy
and spatial uncertainty of per-field crop classification using support vector
machines. ISPRS Journal of Photogrammetry and Remote Sensing.
2013;85:102–119.
[53] Foerster S, Kaden K, Foerster M, et al. Crop type mapping using spectral–
temporal profiles and phenological information. Computers and Electronics
in Agriculture. 2012;89:30–40.
[54] Collobert R and Weston J. A unified architecture for natural language
processing: deep neural networks with multitask learning. In:
Proceedings of the 25th International Conference on Machine Learning,
2008. p. 160–167.
[55] Krizhevsky A, Sutskever I, and Hinton GE. Imagenet classification with
deep convolutional neural networks. Communications of the ACM. 2017;60
(6):84–90.
[56] Schmidhuber J. Deep learning in neural networks: an overview. Neural
Networks. 2015;61:85–117.
[57] Zhong L, Hu L, and Zhou H. Deep learning based multi-temporal crop
classification. Remote Sensing of Environment. 2019;221:430–443.
[58] Hoeser T, Bachofer F, and Kuenzer C. Object detection and image seg-
mentation with deep learning on Earth observation data: a review. Part II:
applications. Remote Sensing. 2020;12(18):3053.
[59] Goyal A. Must known data visualization techniques for Data Science.
Analytics Vidhya. 2021; Available from: https://www.analyticsvidhya.com/
blog/2021/06/must-known-data-visualization-techniques-for-data-science/.
[60] Lesley. 7 techniques to visualize geospatial data – Atlan: Humans of data,
2019. Available from: https://humansofdata.atlan.com/2016/10/7-techni-
ques-to-visualize-geospatial-data/.
[61] 12 Methods for Visualizing Geospatial Data on a Map. Available from:
https://www.safegraph.com/guides/visualizing-geospatial-data.
[62] GISGeography. Cartogram Maps: Data Visualization with Exaggeration,
2022. Available from: https://gisgeography.com/cartogram-maps/.
[63] Geothermal Heat Map. Available from: https://serc.carleton.edu/details/
images/46195.html.
[64] Rice S. Healthcare Scorecard Shows Little Progress Among U.S. States.
Modern Healthcare. 2014; Available from: https://www.modernhealthcare.
com/article/20140430/NEWS/304309964/healthcare-scorecard-shows-little-
progress-among-u-s-states.
[65] Space Applications Center, Indian Space Research Organization, Government
of India. Available from: https://vedas.sac.gov.in/en/.
18 Earth observation data analytics using machine and deep learning
The Red Edge Position (REP) index plays an important role in agricultural remote
sensing applications. A wavelength from 350 nm to 990 nm is the common range
for green. In this chapter, we have focused mainly on crop classification using the
deep learning method. We have presented a study of crop classification using deep
learning methods on hyperspectral remote sensing data. Deep learning is the
evolved form of artificial neural network (ANN). It is based on a biological concept
that deals with the network of neurons in a brain. To solve problems regarding crop
classification, many machine-learning methods are used by researchers. Traditional
machine-learning algorithms, including support vector machine, decision tree-
based, and random forest, work on structured data only. Remote sensing data is
unstructured data. Hence more computational overheads are needed to organize the
unstructured data into structured ones. One of the most adaptable state-of-the-art
approaches for feature extraction and classification of unstructured and structured
data is deep learning. Thus we have focused on deep learning convolutional neural
network (CNN) for feature extraction and classification of crops.
There are multiple sources available for the collection of remote sensing data.
Sensors are mounted on either satellites or aircraft. Each data has its own spectral
and spatial resolution. Spectral resolution defines intervals between two con-
secutive wavelengths. The finer the spectral resolution, the narrower the wave-
length range for a particular channel or band. There are two types of data collection
techniques used: active and passive. In passive, reflected sunlight is used to mea-
sure radiation. These types of sensors can work in the presence of sunlight. Active
sensors are not dependent on the Sun’s electromagnetic rays. These sensors use
their electromagnetic energy and incident on the earth’s surface. Reflected energy
is collected by active sensors.
1
MIT, India
2
Jawaharlal Nehru Engineering College, MGM University, India
22 Earth observation data analytics using machine and deep learning
Spectral reflectance can also be acquired with an ASD Spectro radiometer that
provides measurements in the spectral range starting from 350 nm to 2,500 nm with
3 nm spectral resolution and a 1 nm sampling step. These experiments can be
carried out in the field or laboratory. Obtained data can be viewed and exported in
ASCII file using View Spec Pro 6.2.
10–12 bands which are not contiguous [3]. Hyperspectral data has hundreds of
contiguous spectral bands. Hyperspectral imaging is the collection of hundreds of
continuous connecting spectral bands which can be employed to represent each
pixel [4].
derivative. The easiest way to determine REP is linear interpolation, which takes
into account both the maximum and lowest shoulders of the chlorophyll reflectance
curve. For all crops and environmental conditions, the third linear extrapolation has
taken into account the fixed four spots on the chlorophyll reflectance curve. The
two approaches mentioned above are typically utilised when there are two peaks in
the first derivative and a nitrogen-affected REP.
A narrowband reflectance capability that responds to variations in chlorophyll
content is called the REP index. Increasing chlorophyll concentration broadens the
absorption characteristic and shifts the red edge to longer wavelengths. Red edge
location refers to the wavelength with the steepest slope between 690 and 740 nm.
Green vegetation typically has a wavelength between 700 and 730 nm.
10
20
30
40
0 10 20 30 40 50 60
Flattening Output
Layer
Cotton
PCA
Mulberry
Sugarcane
Rock
Hyperspectral
Image Convolutional
Convolutional Layer2 Fully
Layer1 Connected
Data have been gathered via campaigns. The chosen terrain includes of rock as
well as sugarcane, cotton, mulberry and a limited quantity of other crop fields with
different plant heights between 0.7 and 1.5 m. Using gathered ground truth data as
well as the latitude and longitude of each pixel, the labelling of hyperspectral data
corresponding to individual pixels has been completed. This method resulted in the
creation of the final labelled training and testing dataset. Experts who obtained data
in the field manually marked various crop varieties. The extracted and labelled
hyperspectral data corresponding to each pixel were then created using the com-
bined ground truth data. The final labelled dataset used for training and testing was
created through this approach. The correctly labelled dataset for crop classification
was created by running a data-gathering effort.
Python experiments have been carried out utilising the Keras library functions
and the TensorFlow environment.
been held constant at 0.0001. The batch size has been held constant at 32. CNN has
two convolutional layers in its proposed design, which are followed by fully con-
nected layers and output layers, as shown in Figure 2.2.
% Overall accuracy
CNN Convolutional autoencoder Deep NN
88 2.43 85 1.57 78 1.15
Predicted values
Actual values 0 (Cotton) 1 (Mulberry) 2 (Sugarcane) 3 (Rock)
0 (Cotton) TP 0 E 01 E 02 E 03
1 (Mulberry) E 10 TP 1 E 12 E 13
2 (Sugarcane) E 20 E 21 TP 2 E 23
3 (Rock) E 30 E 31 E 32 TP 3
28 Earth observation data analytics using machine and deep learning
continuous spectral bands in the hyperspectral dataset, the signature of each crop is
more obvious. This method of crop classification works well. Because it effectively
handles unstructured data, deep-learning CNN has been chosen. It can auto-
matically extract the information needed for crop detection or categorization. To
obtain an optimum CNN model, a fine-tuning strategy has been applied by
adjusting the values of various parameters.
DNNs and convolutional autoencoder, two distinct supervised classification
techniques, are contrasted with optimised CNN. The model’s performance can be
enhanced by employing precisely calibrated parameters. Data for training and
testing have been created through the use of significant field research. The samples
were taken using a random sampling strategy. In Aurangabad, cotton is regarded as
a significant winter crop. Sugarcane, mulberries and a few other minor crops are
additional important crops. There are rocks all across the study area as well:
Precision 0 = (TP 0) / (TP 0 + E 10 + E 20 + E 30)
Recall 0 = (TP 0) / (TP 0 + E 01 + E 02 + E 03)
F1 Score = (2 * precision * recall) / (precision + recall)
Overall accuracy = number of correctly classified samples/number of test samples
Overall accuracy = (TP 0 + TP 1 + TP 2 + TP 3)/(total test record)
Three techniques: convolutional autoencoder, DNN, and optimized CNN have been
used and compared. The improved CNN method has been found to produce
superior classification outcomes compared to the other approaches. Through the
use of flattened one-dimensional patches, convolutional autoencoders and DNNs
can also extract deep features via deep learning, albeit the performance may be
significantly lowered. Unsupervised learning is used by the convolutional auto-
encoder to acquire feature knowledge. As a result, it cannot properly process the
information on the label.
Our goal was to create a crop classification system that had already been
trained. This study concentrated on the efficient application of crop forecast pre-
dictive models. It also looks at how well the various predictive algorithms can
classify data. The overall classification accuracy on the dataset can be significantly
increased by the optimised CNN. It also functions with a minimal number of
training samples, as was seen with the dataset for the research area.
The current study may be viewed as the initial stage in creating a crop cate-
gorization model that has already been trained. Additionally, it demonstrates that
CNN’s use of the Adam optimiser enhances efficiency even with sparse input. To
improve accuracy on other crops, it is possible to build and test deep learning-based
systems for remote sensing data classification in the future.
The signature of each crop is more distinct in the hyperspectral data set since it
has more bands. Crop identification can be aided by this hyperspectral data feature.
Typically, the confusion matrix is used to test classification accuracy. The out-
comes of crop classification are compared to and examined using the confusion
matrix, which is made up of the actual pixel classification results. The proportion of
correctly classified pixels to all pixels is used to measure classification accuracy
Deep learning method for crop classification using remote sensing data 29
overall. A fully connected neural network layer follows the encoder-decoder layers
of the convolutional autoencoder. Dimensionality is reduced during the encoding
and decoding process, and it is possible that less valuable information is preserved
as well. As a result, its accuracy is lower than CNN’s.
In order to extract features, the convolution layer with an activation function
reduces the size of the input image. It is accomplished by applying a weighted area
filter on the input data. The data with the highest value within a region is accepted
by the pooling layer. The function of these layers is to choose an essential attribute
from the input for categorization. DNNs are unable to achieve this since they have
several layers with a single activation function. Consequently, it is not possible to
extract the particular feature. A fully connected neural network and an encoder–
decoder are combined to create a convolutional autoencoder. Its accuracy therefore
falls between that of the CNN and the deep neural network.
2.5 Conclusion
Hyperspectral remote sensing data has a greater number of bands; therefore, it has
also been essential to reduce dimensions. The overall classification accuracy with a
small number of training samples can remarkably improve by the CNN model.
Each crop has a more prominent signature because the hyperspectral data set has a
greater number of bands. Such type of research improves the perception of crop
classification and its role in an agricultural field. It will help the farming commu-
nity to plan suitable crops and farming practices and that will increase yield.
Extension of this work can be used for finding crop sown area and expected yield. It
also can be implemented for classifying and distinguishing other objects. Instead of
only hyperspectral imaging, we can use multisource data that is a combination of
microwave, hyperspectral and ultra-spectral sensors.
Healthy and diseased crops can monitor using REP and can be classified using
deep learning methods. Deep learning methods can work on unstructured data like
images, video and audio efficiently. It has been observed that Adam, with ReLU
activation function and a filter size of 2 2, has given better classification accu-
racy compared to the other methods.
A pre-trained architecture for crop classification was to be developed as part of
this work. DNN and convolutional autoencoder have both been compared to opti-
mised models. For crop classification, it has been feasible to implement the pre-
dictive models successfully. The ability of the various analytical models to identify
things is also looked at.
The overall classification accuracy of the dataset is seen to be significantly
improved by the optimised CNN; it has been noticed. The suggested CNN model
performs admirably on small training samples, as was noted in the case of the study
area dataset.
It has also been crucial to lowering dimensions because hyperspectral remote
sensing data contains a greater number of bands. From 155 to 36 bands could be
30 Earth observation data analytics using machine and deep learning
successfully reduced using PCA. The results of this study have demonstrated the
importance of crop identification in farming and the research community.
The future focus of this research will be on estimating crop yield and area
seeded, classifying and identifying other items, and using multisource data from a
mix of microwave, hyper-spectral and ultra-spectral sensors.
References
[1] M. Fauvel, J. Chanussot, and J.A. Benediktsson (2009). Kernel principal
component analysis for the classification of hyperspectral remote sensing
data over urban areas. Journal on Advances in Signal Processing, 2009,
Article ID 783194, 14 pages, doi:10.1155/2009/783194.
[2] W. Zhao and S. Du (2016). Spectral–spatial feature extraction for hyper-
spectral image classification: a dimension reduction and deep learning
approach. IEEE Transactions on Geoscience and Remote Sensing, 54
(8):4544–4554.
[3] C. Yang, C.P.-C. Suh, and J.K. Westbrook (2017). Early identification of
cotton fields using mosaicked aerial multispectral imagery. Journal of
Applied Remote Sensing, 11(1):016008, doi:10.1117/1.JRS.11.016008.
[4] S.E. Hosseini Aria, M. Menenti, and B.G. H. Gorte (2017). Spectral region
identification versus individual channel selection in supervised dimension-
ality reduction of hyperspectral image data. Journal of Applied Remote
Sensing, 11(4):046010, doi:10.1117/1.JRS.11.046010.
[5] A. Ozdarici Ok, O. Akar, and O. Gungor (2012). Evaluation of random forest
method for agricultural crop classification. European Journal of Remote
Sensing, 45(1):421–432, doi: 10.5721/EuJRS20124535.
[6] A. Zhang, X. Yang, L. Jia, J. Ai, and Z. Dong (2019). SAR image classifi-
cation using adaptive neighbourhood-based convolutional neural network.
European Journal of Remote Sensing, 52(1):178–193, doi: 10.1080/
22797254.2019.1579616.
[7] T. Tian, L. Gao1, W. Song, K.-K. Raymond Choo, and J. He (2017). Feature
Extraction and Classification of VHR Images with Attribute Profiles and
Convolutional Neural Networks. Springer Multimed Tools Appl, https://doi.
org/10.1007/s11042-017-5331-4.
[8] S. Yu, S. Jia, and C. Xu (2017). Convolutional neural networks for hyper-
spectral image classification. Elsevier Neurocomputing, 219:88–98.
[9] Y. Chen, H. Jiang, C. Li, X. Jia, and P. Ghamisi (2016). Deep feature
extraction and classification of hyperspectral images based on convolutional
neural networks. IEEE Transactions on Geoscience and Remote Sensing,
54:6232–6251.
[10] K. Bhosle and V. Musande (2019). Evaluation of deep learning CNN model
for land use land cover classification and crop identification using hyper-
spectral remote sensing images. Journal of the Indian Society of Remote
Deep learning method for crop classification using remote sensing data 31
This study aims at developing an end-to-end solution for deep learning-based crop
classification using synthetic aperture radar (SAR) images. SAR provides all-
weather, day, and night imaging capabilities, sensitive to dielectric properties.
Optical images are intuitive and capture static information well, like the boundary of
the field in the absence of atmospheric disturbances. In this work, the end-to-end
solution to use deep learning algorithms for the crop-type classification is done using
SAR images. The limitation of the SAR images is handled by using the boundary
information from the optical data. For the classification of different crops in the test
site, L-band ISRO L- & S-band Airborne SAR (ASAR) and Airborne Synthetic
Aperture Radar (AIRSAR) images were acquired over an agricultural site near
Bardoli and Flevoland respectively. Pre-trained model Inception v3 and Custom
VGG like model were used for crop classification. Inception V3 enabled us to better
discriminate crops, particularly banana and sugarcane, with 97% accuracy, while the
Custom VGG like model achieved 95.17% accuracy for 11 classes.
3.1 Introduction
The contribution of agriculture to India’s GDP (Gross Domestic Production) is 15.9% &
India’s 54.6% population is engaged in agriculture and allied activities. The COVID-19
pandemic and subsequent lockdowns have influenced most of the sectors of the econ-
omy. However, the agricultural sector has performed way better with a 2.9% growth
rate amid 2019–2020, as against 2.74% accomplished amid 2018–2019 [1,2].
United Nations’ Sustainable Development Goal is to eliminate global hunger,
protect the indigenous seed and crop varieties, double agriculture productivity, and
have small farmer incomes by 2030 [3]. To achieve this objective for successful
1
Institute of Technology, Nirma University, India
34 Earth observation data analytics using machine and deep learning
3.1.1 Motivation
Remote sensing has become a potent geospatial tool in agricultural systems. As it
delivers valuable, precise, and timely information about agricultural urbanization at
high spatial and spectral resolutions, SAR imaging is an effective agricultural
solution. For crop classification, a variety of techniques were used, however, deep
learning has shown to be more effective. As per the literature, many researchers
have tried machine learning and deep learning algorithms for crop classification,
but they have only worked on AIRSAR, Advanced Land Observation Satellite
(ALOS), Polarimetric Phased Array L-band Synthetic Aperture Radar (PALSAR),
Uninhabited Aerial Vehicle Synthetic Aperture Radar (UAVSAR), Experimental
Synthetic Aperture Radar (ESAR), etc., satellite datasets. In contrast, we worked on
the ASAR dataset (provided by SAC-ISRO), which will provide global coverage of
the entire landmass. This paper aims to work on the above-mentioned dataset and
compare the crop classification results with the existing dataset.
3.1.3 Organization
This paper describes a deep learning model that does crop-type classification from full
polarization L-band SAR images. The rest of the paper is organized as follows:
Section 3.2 discusses the related work, Section 3.3 describes the methodology, Section 3.4
describes the study area, Section 3.5 shows an experimental setting, Section 3.6 presents
the experimental result and analysis, and concluding remarks are in Section 3.7.
Author Year Objectives Sensor Frequency Polarimetric parameters Multi Area No. of classes
date identified
Halder et al. [9] 2011 Evaluation L band data ALOS PALSAR L Different polarization Hisar, Haryana 5
with different polariza- combination modes
tion combination for
crop classification 1. Linear polarization
2. Circular polarization
3. Hybrid polarization
Henning [10] 2011 To study crop classifica- EMISAR L&C 1. Single polarization Denmark 11
tion accuracy for dif- 2. Coherent and incoherent dual
ferent polarization polarizations
modes 3. Fully polarimetric
Ning et al. [26] 2016 Proposed an improved 1. AIRSAR L Pauli decomposition 1. Flevoland Dataset 1 1. 10
super pixel-based 2. ESAR 2. Oberpfaffenhofen 2. 2
POLSAR image classi-
fication integrating
color features
Huapeng et al. 2019 explored Capability of UAVSAR L 1. Cloude–Pottier decomposition Sacramento Valley, 11
[11] time series data for 2. Freeman-Durden decomposition California
crop classification 3. Linear polarization (VV, HV, HH)
4. Combination of all
Julien et al. [12] 2019 To identify the best SAR RADARSAT-2 L&C 1. Back scattering coefficients Mount saint 12
configuration to identify Sentinel 2. Cloude–Pottier decomposition Michel, France
winter land use ALOS-2 3. Freeman–Durden decomposition
4. SPAN (total scattered power) and
Shannon entropy
5. Dual and quad
6. Polarization (pol) mod
Hongwei et al. 2019 Proposed a differentiable 1. AIRSAR L coherence matrix [T] 1. San Francisco 1. 5
[27] architecture search 2. AIRSAR 7. Decomposition Techniques 2. Flevoland Dataset 1 2. 15
(DAS)method 3. ESAR 3. Oberpfaffenhofen 3. 3
Chu et al. [28] 2020 Proposed a POLSAR im- 1. AIRSAR L 1. Pauli 1. Flevoland Dataset 1 11
age classification based 2. AIRSAR 2. Cloude 2. San Francisco 5
on FCN and manifold 3. AIRSAR 3. Freeman 3. Flevoland Dataset 2 14
graph embedding 4. H/A/Alpha
model
5. Huynen
6. Yamaguchi
7. Krogager
Proposed – Proposed a deep learning 1. ASAR L Freeman Decomposition 1. Bardoli 1) 2
approach based crop classifica- 2. AIRSAR 2. Flevoland 2) 11
tion model
38 Earth observation data analytics using machine and deep learning
Fields extraction
The pre-trained HED model has been integrated with the deep neural network
(DNN) module of OpenCV. Hence, the entire pipeline has been developed using
OpenCV in Python. The pipeline consists of the following steps:
1. The Bilateral Filtering was applied to the input Optical/SAR image. This fil-
tering method is known for the removal of noise in the image while preserving
the edges. It compares the pixel intensity variation of the edge and the neigh-
boring pixels of the edge which would be included for blurring [32].
2. The output image is then sent to the HED model which produces an edge map
that does a good job conserving the boundaries of the crop fields. Hence, a
binary image having the crop field boundaries is obtained.
3. The retrieval of contours is done on the binary image and every contour
representing an individual crop field is extracted separately.
Segmentation of SAR image was done by super-imposing the edge detected
image on Freeman decomposed image to extract out the fields. Crop fields for
which ground truth data was available were labeled.
3.3.3.1 Inception V3
In recent years, Inception V3 model has been applied in many fields such as
Human Eye Iris Recognition [34], maize disease detection in optical image [35],
bird voice recognition [36], vehicle pollution detection [37], breast cancer
detection [38], actions of distracted drivers [39], skin disease detection [40],
flower classification [41], and remote-sensing scene classification [42]. In
Inception-V3, the number of connections is reduced without affecting the net-
work efficiency using methods such as factorization of 55 convolutional layer
into two 33 convolutional layers, factorization into asymmetric convolutions
such as 33 into 13 and 31, and Auxiliary classifier. The model uses an input
image size of 299*299 [43].
India
N
Bardoli
nd
ola
lev
F
(d)
Figure 3.5: Bardoli dataset: (a) optical image, (b) edge detected image, (c)
Freeman decomposition image, and (d) samples of extracted fields
mapped with GT
Optical images to demarcate fields in L band SAR images 43
and testing. The classification is done in two stages. In the first stage, Dataset 1
images are classified into two classes of cultivated area and non-cultivated area. In
the second stage, cultivated areas are further classified into banana and sugarcane.
3.5.2 Dataset 2
Flevoland dataset was pre-processed according to the standard procedure such as
Lee speckle filter, and Freeman decomposition (see Figure 3.6). Edges were
detected in two different SAR images, i.e. Freeman decomposition image and the
Pauli-decomposed image using the HED model. OR operation was performed on
two edges detected images to generate a resultant image. The segmentation of the
Freeman decomposition image (RGB image) is done by super-imposing the edge
detected image (the resultant edge-detected image) on the Freeman decomposition
image to extract out the fields. A total of 121 fields were extracted. Crops were
labeled with the help of Ground Truth data provided in [28].
Initially, we have 127 labeled images of the 11 different classes: beet, lucerne,
rapeseed, steambeans, peas, potato, wheat, bare soil, forest, grass, and water. Using
the augmentation techniques, we increase the size of the dataset. To validate our
results, we divide the training dataset into two subsets of training and testing.
Table 3.5 shows the number of samples for training and testing after augmentation.
(d)
performance of the model was assessed using cross-validation with the validation
set. The training dataset was further classified into training and validation at the
time of training. Inception v3 model achieved the best performance accuracy of
58.74% for Dataset 1 and 97.52% for the cultivated area. For Dataset 2, Custom
VGG like model achieved the best performing accuracy of 95.17%.
The classification results of the proposed method are described from multiple
perspectives. Confusion matrices for crop classification are shown in Figure 3.7 for
Dataset 1, cultivated area (banana/sugarcane), and Dataset 2. Here it is seen that in
the cultivated area, banana and sugarcane were often correctly classified, whereas
in Dataset 1, while classifying cultivated and non-cultivated areas, cultivated area
was misclassified. For Dataset 2, wheat is misclassified with rapeseed.
100
100
Confusion matrix
Confusion matrix
90 80
80 Banana 96 5
Cultivated area 96 106
60
70
True label
True label
60
40
82 50
Non-Cultivated area 19 Sugarcane 0 101
40
20
Cultivated area Non-Cultivated area
Predicted label 30 Banana Sugarcane
Predicted label
20 0
(a) (b)
100
Confusion matrix
Beet 0 103 0 0 0 0 0 0 0 0 0 80
Forest 0 0 101 0 0 0 0 0 0 0 0
Grass 0 0 0 102 0 0 0 0 0 0 0
60
Luceme 0 0 0 2 99 0 0 0 0 0 0
True label
Peas 0 0 0 0 0 101 0 0 0 0 0
Potato 0 0 0 0 0 0 103 0 0 0 0
40
Rapeseed 0 0 0 0 0 0 0 102 0 0 0
Steambean 0 0 0 0 0 0 0 0 102 0 0
Water 0 0 0 0 0 0 0 0 0 101 0
20
Wheat 0 0 0 0 0 0 0 52 0 0 50
Bare Soil Beet Forest Grass Luceme Peas Potato Rapeseed Steambean Water Wheat
Predicted label
(c)
Figure 3.7 Confusion matrix for crop classification: (a) Dataset 1, (b) cultivated
area, and (c) Dataset 2
46 Earth observation data analytics using machine and deep learning
The receiver operating characteristic (ROC) curve is plotted in Figure 3.8 for
Dataset 1, cultivated area and Dataset 2. The cultivated area gives good accuracy
over Dataset 1. Higher the ROC curve’s area, better the model performance. In
particular, the cultivated area was the best performing. For Dataset 2, all the values
of curve are equal or near to 1 except wheat. Overall, the results presented in
Tables 3.6 and 3.7 for cultivated area and Dataset 2 indicate that the UA and PA for
the majority of the crop classes are near 1.
100
100
Confusion matrix
Confusion matrix
90 80
80 Banana 96 5
Cultivated area 96 106
60
70
True label
True label
60
40
82 50
Non-Cultivated area 19 Sugarcane 0 101
40
20
Cultivated area Non-Cultivated area
Predicted label 30 Banana Sugarcane
Predicted label
20 0
(a) (b)
100
Confusion matrix
Beet 0 103 0 0 0 0 0 0 0 0 0 80
Forest 0 0 101 0 0 0 0 0 0 0 0
Grass 0 0 0 102 0 0 0 0 0 0 0
60
Luceme 0 0 0 2 99 0 0 0 0 0 0
True label
Peas 0 0 0 0 0 101 0 0 0 0 0
Potato 0 0 0 0 0 0 103 0 0 0 0
40
Rapeseed 0 0 0 0 0 0 0 102 0 0 0
Steambean 0 0 0 0 0 0 0 0 102 0 0
Water 0 0 0 0 0 0 0 0 0 101 0
20
Wheat 0 0 0 0 0 0 0 52 0 0 50
Bare Soil Beet Forest Grass Luceme Peas Potato Rapeseed Steambean Water Wheat
Predicted label
(c)
Figure 3.8 ROC curve: (a) Dataset 1, (b) cultivated area, and (c) Dataset 2
Optical images to demarcate fields in L band SAR images 47
Table 3.6 Best user’s (UA), producer’s (PA), and overall (OA)
accuracy and kappa coefficient for cultivated area
Class UA PA OA Kappa
Banana 100% 95.04% 97.52% 0.95
Sugarcane 95.28% 100%
Table 3.7 Best user’s (UA), producer’s (PA), and overall (OA)
accuracy and kappa coefficient for Dataset 2
Class UA PA OA Kappa
Bare soil 100% 100%
Beet 100% 100%
Forest 100% 100%
Grass 98.07% 100%
Lucerne 100% 98.01%
Peas 100% 100% 95.17% 0.94
Potato 100% 100%
Rapeseed 66.23% 100%
Steambeans 100% 100%
Water 100% 100%
Wheat 100% 49.01%
3.7 Conclusion
In this study, full polarization SAR data for crop classification in Bardoli and Flevoland
was investigated. A deep learning-based crop classification method was used. Two
types of approaches were used in this paper. One is Inception v3 and the other is
Custom VGG like a model. Datasets were created for training a model followed by
extracting fields of SAR images by super-imposing the edge detected optical/SAR
image on the Freeman decomposition image and labeled extracted SAR field images for
crop classification. The model achieved an overall accuracy of 97.52% for the cultivated
area using Inception v3 and 95.17% for Dataset 2 using Custom VGG like model.
References
[1] Atmanirbhar Krishi and Atmanirbhar Kisan are important to achieve the goal
of Atmanirbhar Bharat –Prime Minister Shri Narendra Modi, in his address
to the nation on the 74th Independence Day. Available from: https://pib.gov.
in/PressReleasePage.aspx?PRID=1646286.
[2] Ministry of Agriculture Farmers Welfare, Government of India. Kurukshetra
Magazine: Ministry of Rural Development. Kurukshetra Magazine: Ministry
48 Earth observation data analytics using machine and deep learning
of Rural Development. Govt. of India 53, no. 1 (2004): 28. P. B. Patil and T.
A. Shivare.
[3] Take action for the sustainable development goals – United Nations sus-
tainable development. United Nations. Available from: https://www.un.org/
sustainabledevelopment/sustainable-development-goals.
[4] Cué La Rosa LE, Queiroz Feitosa R, Nigri Happ P, et al. Combining deep
learning and prior knowledge for crop mapping in tropical regions from
multitemporal SAR image sequences. Remote Sensing. 2019;11(17):2029.
[5] Sun Y, Luo J, Wu T, et al. Geo-parcel based crops classification with
Sentinel-1 time series data via recurrent neural network. In: 2019 8th
International Conference on Agro-Geoinformatics (Agro-Geoinformatics).
IEEE; 2019. p. 1–5.
[6] Jia K, Li Q, Tian Y, et al. Crop classification using multi-configuration SAR
data in the North China Plain. International Journal of Remote Sensing.
2012;33(1):170–183.
[7] Kussul N, Skakun S, Shelestov A, et al. The use of satellite SAR imagery to
crop classification in Ukraine within JECAM project. In: 2014 IEEE
Geoscience and Remote Sensing Symposium. IEEE; 2014. p. 1497–1500.
[8] Tan CP, Ewe HT, and Chuah HT. Agricultural crop-type classification of
multi-polarization SAR images using a hybrid entropy decomposition and
support vector machine technique. International Journal of Remote Sensing.
2011;32(22):7057–7071.
[9] Haldar D, Das A, Mohan S, et al. Assessment of L-band SAR data at dif-
ferent polarization combinations for crop and other land use classification.
Progress in Electromagnetics Research B. 2012;36:303–321.
[10] Skriver H. Crop classification by multitemporal C-and L-band single-and
dual-polarization and fully polarimetric SAR. IEEE Transactions on
Geoscience and Remote Sensing. 2011;50(6):2138–2149.
[11] Li H, Zhang C, Zhang S, et al. Crop classification from full-year fully-
polarimetric L-band UAVSAR time-series using the Random Forest algo-
rithm. International Journal of Applied Earth Observation and
Geoinformation. 2020;87:102032.
[12] Denize J, Hubert-Moy L, and Pottier E. Polarimetric SAR time-series for
identification of winter land use. Sensors. 2019;19(24):5574.
[13] Zhao H, Chen Z, Jiang H, et al. Evaluation of three deep learning models for
early crop classification using sentinel-1A imagery time series: a case study
in Zhanjiang, China. Remote Sensing. 2019;11(22):2673.
[14] Mullissa AG, Persello C, and Tolpekin V. Fully convolutional networks for
multi-temporal SAR image classification. In: IGARSS 2018–2018 IEEE
International Geoscience and Remote Sensing Symposium. IEEE; 2018.
p. 6635–6638.
[15] Xu Z, Meng S, Zhong S, et al. Study on temporal and spatial adaptability of
crop classification models. In: 2019 8th International Conference on Agro-
Geoinformatics (Agro-Geoinformatics). IEEE; 2019. p. 1–7.
Optical images to demarcate fields in L band SAR images 49
[16] Mei X, Nie W, Liu J, et al. Polsar image crop classification based on deep
residual learning network. In: 2018 7th International Conference on Agro-
geoinformatics (Agro-geoinformatics). IEEE; 2018. p. 1–6.
[17] Castro JDB, Feitoza RQ, La Rosa LC, et al. A comparative analysis of deep
learning techniques for sub-tropical crop types recognition from multi-
temporal optical/SAR image sequences. In: 2017 30th SIBGRAPI
Conference on Graphics, Patterns and Images (SIBGRAPI). IEEE; 2017.
p. 382–389.
[18] Gu L, He F, and Yang S. Crop classification based on deep learning in
northeast China using sar and optical imagery. In: 2019 SAR in Big Data Era
(BIGSARDATA). IEEE; 2019. p. 1–4.
[19] Avolio C, Tricomi A, Mammone C, et al. A deep learning architecture for
heterogeneous and irregularly sampled remote sensing time series. In:
IGARSS 2019 – 2019 IEEE International Geoscience and Remote Sensing
Symposium. IEEE; 2019. p. 9807–9810.
[20] Kussul N, Lavreniuk M, Skakun S, et al. Deep learning classification of land
cover and crop types using remote sensing data. IEEE Geoscience and
Remote Sensing Letters. 2017;14(5):778–782.
[21] Zhou Y, Luo J, Feng L, et al. Long-short-term-memory-based crop classi-
fication using high-resolution optical images and multi-temporal SAR data.
GIScience & Remote Sensing. 2019;56(8):1170–1191.
[22] Wei S, Zhang H, Wang C, et al. Multi-temporal SAR data large-scale crop
mapping based on U-Net model. Remote Sensing. 2019;11(1):68.
[23] Chen SW and Tao CS. PolSAR image classification using polarimetric-
feature-driven deep convolutional neural network. IEEE Geoscience and
Remote Sensing Letters. 2018;15(4):627–631.
[24] Chen SW and Tao CS. Multi-temporal PolSAR crops classification using
polarimetric-feature-driven deep convolutional neural network. In: 2017
International Workshop on Remote Sensing with Intelligent Processing
(RSIP). IEEE; 2017. p. 1–4.
[25] Ndikumana E, Ho Tong Minh D, Baghdadi N, et al. Deep recurrent neural
network for agricultural classification using multitemporal SAR Sentinel-1
for Camargue, France. Remote Sensing. 2018;10(8):1217.
[26] Xing Y, Zhang Y, Li N, et al. Improved superpixel-based polarimetric syn-
thetic aperture radar image classification integrating color features. Journal
of Applied Remote Sensing. 2016;10(2):026026.
[27] Dong H, Zou B, Zhang L, et al. Automatic design of CNNs via differ-
entiable neural architecture search for PolSAR image classification.
IEEE Transactions on Geoscience and Remote Sensing. 2020;58
(9):6362–6375.
[28] He C, He B, Tu M, et al. Fully convolutional networks and a manifold graph
embedding-based algorithm for polsar image classification. Remote Sensing.
2020;12(9):1467.
[29] European Space Agency. Available from: https://earth.esa.int/documents/
653194/656796/LNAdvancedConcepts.pdf.
50 Earth observation data analytics using machine and deep learning
Information related to land use land cover (LULC) plays an instrumental role in
land management and planning. With the advent of the field of machine learning,
accurate automation of tasks has become feasible, hence this study presents a
similarity-learning approach using twin networks for LULC classification and
extended use cases on the AVIRIS sensor’s Indian Pines standard dataset. A thor-
ough comparative study is conducted for the Siamese Network backbones, with
experiments on DiCENets, ResNets, SqueezeNets, and related state-of-the-art
approaches. Embedding augmentation is also explored along with embedding
classification and dimensionality reduction algorithms for understanding hyper-
space generation and the use of similarity learning. The performance analysis of
SiameseNets on a reduced training dataset size is also shown to reinforce the utility
of SiameseNets. Thorough experiments are conducted for improving the hyper-
parameters associated with deep learning architectures to form a non-biased and
rational comparison of the given classification approaches. The proposed meth-
odologies follow a classification accuracy approaching 98% and are validated using
baselines.
4.1 Introduction
Accurately and efficiently identifying the physical characteristics of the earth’s
surface or land cover in general, as well as how we exploit the affiliated land usage,
is a difficult topic in environmental monitoring and many other related areas and
subdomains [1]. This can be accomplished by extensive field surveys or by ana-
lyzing a large number of satellite photos utilizing the remote sensing paradigm [2].
While doing field surveys is more complete and authoritative, it is a costly
1
Institute of Technology, Nirma University, India
2
School of Technology, Pandit Deendayal Energy University, India
3
Space Application Centre, ISRO, India
52 Earth observation data analytics using machine and deep learning
endeavor that often takes a long time to update. Deep learning and convolutional
neural networks have demonstrated promising results in LULC categorization with
recent breakthroughs in the space industry and the growing availability of satellite
pictures in both free and commercially accessible data sets [3]. By assessing and
understanding the current status of predictive analysis and general research related
to vision systems, application-oriented deep learning [4–6], computing, computa-
tional facilities [7], and the overall implications of robust remote sensing, this
chapter offers novel implementations and analysis in the field of machine learning-
assisted LULC predictions. For any predictive model or any algorithm related to
the machine learning methodology, an inherent need for sufficient data is
encountered [8]. It is the general thought process and observation around deep
learning algorithms that a higher number of data samples implicates a higher pos-
sible classification or prediction accuracy [8].
However, in cases concerning a multitude of domains a deficiency of significant
data samples is observed, and special contingencies should be anointed to tackle
related predicaments. Recent literature and advancements in the field of deep learn-
ing recommend two possibilities to tackle an insufficient dataset, namely, data aug-
mentation [9], and similarity learning [10] implementations. To elaborate details on
the former, there have been significant improvements in the field of data augmen-
tations, with algorithms focusing on images, Generative Adversarial Networks
(GANs) [11], or the excessively contemporary embedding-based augmentative stra-
tegies [12]. There have been many image augmentation techniques in the literature
on images, they usually promote the use of image array manipulations like rando-
mized rotations, horizontal flipping, and vertical flipping [8]. However, the compu-
tation corresponding to image manipulation would be significantly higher when
compared to embeddings or moderately dimensionalized embedding data. The latter
argumentation mechanism for scarce data is countered by the use of Siamese
Networks or similarity learning, these specialized deep neural architectures permit
the use of learning similar features from pairs, triplets, or quadruplets and traverse the
information to favorable embeddings and embedded hyperspace [13]. With the scope
of this paper related to the use of Siamese networks, an E-Mixup embedding aug-
mentation strategy [12] is presented to accurately measure and validate the use of
Siamese networks, augmentation, and its associated prominence for LULC classifi-
cation. This chapter functions on the motivations related to automating the task of
LULC classification and the data-related predicaments, and functions on a publicly
available Indian Pines dataset, which consists of AVIRIS Hyperspectral images [14].
The chapter further contains the related work, the proposed computation strategies
and methodologies, and a thorough empirical analysis to understand the relevance of
the mentioned literature followed by the concluding statements.
sensing and LULC classification and general predictive analysis of LULC. A novel
analysis presented in [15] showcases a deeper insight into the domain of medical
imaging and the relevant machine learning and deep learning algorithms which
assisted the related cause. The paper offered a similar correlation between the use of
deep learning algorithms, and the challenges, possibilities, and benchmarks corre-
sponding to remote sensing. The article [16] offered a novel analysis of the spatial
relationships between land surface temperature, commonly abbreviated as LST, and
LULC of three urban agglomerations (UAs) belonging to the eastern portion of India.
The obtained study further promoted a planning and implementation strategy for the
effective development of small and medium-sized cities. The paper [17] presented a
comparative study, where different object-based and pixel-based classification tech-
nologies were leveraged on a high spatial-resolution multi-source dataset to accu-
rately and robustly map LULC. Significant work has also been done for learning
transferable change for LULC change detection, the novel method as presented in
[18] had significant societal information and added to the multi-disciplinary utility of
an RNN. The authors of the paper [14] presented and highlighted the use and func-
tioning of a convolutional neural network (CNN) in a study area focusing on
Aurangabad, India. The presented model procured a percentage accuracy of 79.43
when validated against the previously mentioned study area. An important aspect
concerning the algorithm was the use of CNNs for an unstructured input and a sig-
nificantly scarce data pool, as the model performance turned out to be satisfactory.
The previously mentioned research also utilized the Indian Pines dataset, further
explaining and justifying its usefulness for this chapter. The paper [19] also
functioned primarily on deep neural networks, however, offered a different study
area and learning paradigm. The related deep neural architectures were trained using
the publicly available ImageNet LargeScale Visual Recognition Competition
(ILSVRC) datasets. The models worked on the fine-tuning principle and a total of
19,000 Landsat 5/7 satellite images from the Province of Manitoba in Canada
were converted to patches of image size 224 224 for the same commission. The
article [20] also worked on LULC classification on data collected by the Sentinel-2,
the dataset collection is called EuroSAT remote sensing image dataset and was used
for the underlying tests. Several CNN-based subsidiary architectures like
InceptionV3, ResNet50, and VGG19 were deployed for feature extraction and
methodologies containing Channel Squeeze & Spatial Excitation block and Twin
SVM (TWSVM) were further used as a complete pipeline. The best performing
model outputted a 94.57% accuracy.
4.3 Methodology
The chapter offers a novel study on deep architectures, Siamese or Twin networks,
and classification approaches to the Indian Pines dataset. Multiple experiments
were conducted, where the encoder was tested without any modifications, a
Siamese variant for the encoder, and the siamese variant with E-Mixup augmen-
tation. This section additionally contains the dataset description, information about
54 Earth observation data analytics using machine and deep learning
Figure 4.1 Information flow and the primary preprocessing steps for training and
testing the proposed models, each architecture, and underlying
technologies are explained thoroughly in the forthcoming sections.
Siamese Networks and the related parameters, and a thorough description of the
experimented encoders. The primary flow of information which is observed for
training and validating the proposed neural methodologies is explained in
Figure 4.1.
4.3.1 Dataset
For this study, the Indian Pines dataset was used, and the underlying preprocessing
steps are mentioned and explained in this subsection. The images follow
145 145 spatial dimensions and 224 spectral bands in the wavelength range of
400–2,500 nm. A total of 24 spectral bands concerning water absorption regions have
been discarded. The available ground truth images are designated to 16 different
classes which adequately categorizes the study area’s land cover. The images have a
significantly high input dimensionality which is unsuited for training, hence Principal
Component Analysis or PCA [21] is deployed. The resultant is 10,249 images of size
64 64 3, the totality of the time taken to perform the said conversion was
0.616 s. To avoid the skewed dataset problem, a stratified train test split [22] is used,
and an equal percentage of images are taken out of each class to have a high-quality
training setup. The standard encoders or single architectures are trained on a 60–40
train-test split and to check the SiameseNet’s performance on a reduced training test
size a 50–50 split is leveraged. The lower train–test split was added to future-proof
LULC studies, as for cases about the inclusion of a new class, data sample-based
predicaments may be observed, and having standardized and developed neural stra-
tegies which function suitably in those scenarios, is highly favorable. A tabular
description for the 16 constituent classes [23] is mentioned in Table 4.1.
Leveraging twin networks for land use land cover classification 55
4.3.2.1 E-Mixup
This section elaborates on the embedding augmentation functionality as proposed in
the aforementioned sections. The embeddings that are generated due to the use of a
Siamese Network have a significantly lower size than a standard image, hence using
an embedding augmentation technique generally outperforms standard image
manipulation techniques in terms of computational efficiency. To use the E-Mixup
augmentation pipeline, a weighted average is taken over two embeddings, and the
binary matrix representation of the class values with the weight as lambda [12],
which is calculated as a random value from a Beta distribution with the alpha values
fixed as 0.2 [12]. By using the said method, the training set is tripled and classified.
4.3.2.2 MLP
This section elaborates on the further classification of the obtained embeddings as
proposed in the aforementioned sections. The embeddings have a dimensionality of
56 Earth observation data analytics using machine and deep learning
256 and are fed into an MLP. The architecture contains four hidden layers, batch
normalization, and the concluding Softmax layer. The final activation function
promotes a resultant probabilistic distribution, which is favorable for a multiclass
classification problem. The same architecture is used for all experiments related to
Siamese Networks, for an unbiased and nonaligned testing experience.
4.3.3 Encoders
Various encoders architectures were tested, mainly DenseNets [26], DiceNets [27],
ResNets [28], and SqueezeNets [29]. For tests involving individual networks, a soft-
max layer is used for the classification. For tests about Siamese Networks, each
encoder is modified to obtain embeddings with 256 dimensions by using dense layers.
4.3.3.1 DiCENet
Standard convolutions simultaneously encode spatial and channel-wise informa-
tion, but they are computationally intensive. Therefore, separable (or depth-wise
separable) convolutions are introduced to increase the efficiency of ordinary con-
volutions by encoding spatial and channel-wise information independently using
depth-wise and point-wise convolutions, respectively. Though successful, this
factorization places a large computational burden on point-wise convolutions,
making them a computational bottleneck. The DiCENet consists of multiple DiCE
units, which can be understood as building blocks [30].
The DiCE unit was introduced to encode spatial and channel-wise information
efficiently. Dimensionwise Convolution (DimConv) and Dimension-wise Fusion are
used to factorize ordinary convolution in the DiCE unit (DimFuse). DimConv learns
local dimension-wise representations by applying light filtering across each dimen-
sion of the input tensor. DimFuse effectively mixes these representations from many
sizes while also incorporating global data. The DiCE unit can employ DimFuse
instead of computationally intensive point-wise convolutions since DimConv can
encode local spatial and channel-wise information from all dimensions.
4.3.3.2 ResNet
This item presents a synopsis of ResNets as illustrated in the original publication
[28]. One of the proposed encoders that we will utilize is Residual Networks, often
known as ResNets. The advantage that ResNets have over unadorned networks is
that they alleviate the deterioration issue that was shown when extremely deep
networks began to converge. Identity mapping was presented by ResNets, which
meant that input from a prior layer was taken and sent to another layer as a shortcut.
A deeper understanding can be obtained from Figure 4.2 [28].
Typically, the 34-layer and 18-layer ResNets are employed because they pro-
duce a substantially small error and comparably higher accuracies than their simple
rivals. The 34-layer ResNet exhibits somewhat lower training error and tackles the
deterioration problem encountered in its plain competitor, resulting in good accu-
racy from greater depths. Not only did the 18-layer ResNet outperform its simple
competitor in terms of accuracy, but it was also able to perpetrate convergence
faster and get satisfactory solutions on fewer data entities.
Leveraging twin networks for land use land cover classification 57
Weight layer
Weight layer
G(X )
G(X) + X
ReLU
4.3.3.3 SqueezeNet
SqueezeNet’s research involves an intelligent architecture as well as a quantita-
tive analysis. As a consequence, while maintaining the same degree of accuracy,
SqueezeNet can be three times quicker and 500 times smaller than AlexNet. The
fire module is a two-layer SqueezeNet building block featuring a squeeze layer
and an expansion layer. A SqueezeNet is a network made up of many fire mod-
ules and numerous pooling layers. The squeeze and expand layers keep the fea-
ture map size constant, while the former decreases the depth while the latter
increases it [29]. Another common design is to increase the depth while lowering
the breadth. Because the squeeze module includes just 1 1 filters, it acts as a
fully-connected layer that operates on feature points in the exact position. In other
words, it cannot be spatially abstracted. As the name says, one of its advantages is
that it minimizes the depth of the feature map. When the depth is lowered, the
succeeding 3 3 filters in the expansion layer will have more undersized
calculations [29]. It boosts performance since a 3 3 filter takes nine times the
calculation of a 1 1 filter. Intuitively, too much squeezing inhibits information
flow; too few 3 3 filters, on the other hand, restrict spatial resolution. These
were the primary architectural details that were obtained from the original article
[29]. The inherent functioning of this architecture and fire modules is explained
graphically in Figure 4.3 [31].
58 Earth observation data analytics using machine and deep learning
1×1
Squeeze
Expand
1×1 3×3
Softmax
4.3.3.4 DenseNet
This section presents a synopsis of the DenseNet architecture as specified in the
original paper [26]. Another type of encoder employed in the suggested design is
the Dense Convolution Network, often known as DenseNet. The benefit of
employing this sort of network is that each of its levels collects supplemental inputs
from all of the layers before it. The data is concatenated such that each layer
acquires the cumulative intelligence of all preceding levels. As a result, when each
layer acquires feature mappings from previous layers, the whole network is com-
pressed, resulting in more infrequent total channels. The constituent functionality
of the DenseNet neural architecture can be further understood by Figure 4.4 [26].
The difference between DenseNets and ResNets is that DenseNets employ the
parameters more dexterously. Outwardly, the only significant difference between
the two networks is that DenseNets concatenate the inputs, whereas ResNets sum
the inputs. Although this appears to be a little modification, it results in a
significant shift in behavior for both of them [26]. Furthermore, DenseNets require
a small number of parameters and processing resources to deliver extremely precise
and cutting-edge findings, and results with greater precision may be obtained when
the hyperparameters are tweaked with care [26].
BN-ReLU-Conv2D
Concat
BN-ReLU-Conv2D
Concat
BN-ReLU-Conv2D
Concat
BN-ReLU-Conv2D
Concat
following equations, here TP, FP, TN, FN implicates the True Positive, False
Positive, True Negative, and True Positive respectively:
2 TP
F1-Score ¼ (4.1)
2 TP þ FP þ FN
TP þ TN
Accuracy ¼ (4.2)
TP þ TN þ FP þ FN
The obtained results concerning the aforementioned performance metrics are
mentioned in Table 4.2. A substantial temporal analysis is also obtained by asses-
sing the time taken for both the training phases and the testing phases across all
experiments for understanding the deploy-ability of each considered methodology
is mentioned in Table 4.3.
From Tables 4.2 and 4.3, it can be inferred that the best performing models
were obtained during the tests concerning Siamese DenseNets, which were aug-
mented using E-Mixup and the standard ResNet. To further estimate the best
60 Earth observation data analytics using machine and deep learning
(a) (b)
Figure 4.5 The false-color composite (a) and the corresponding ground truth
image (b)
20 20 20
40 40 40
60 60 60
80 80 80
Figure 4.6 The obtained spectral graphs for the ResNet-based architectures, the
architectural names are based on the aforementioned abbreviations.
This chapter aimed to offer a novel implementation and study for possible advancements
in the field of LULC classification. The Indian Pines dataset was used for training and
validation of the mentioned methodologies, with extensive tests on DiceNets, DenseNets,
ResNets, and SqueezeNets. The individual encoders were also tested in a low train test
split with a twin network strategy, which symbolized and paved the way for further
research involving a potential new class, where data scarcity is common. An embedding
augmentation strategy called E-Mixup or Embedding-Mixup is also explored which was
emphasized as an improvement to the standard or vanilla Siamese implementations.
For future work, the authors aim to improve upon the existing datasets and also
work on LULC predictions about Indian terrains to validate the feasibility and
traversability of the Indian Pines Dataset which can be considered a limitation for
the scope of this study. The authors also aim to assess other modern architectural
strategies and newer learning techniques to present a potentially generalized study.
62 Earth observation data analytics using machine and deep learning
DiCENet DiCENet-S + MLP DiCENet-S + E-Mixup + MLP
0 0 0
20 20 20
40 40 40
60 60 60
80 80 80
Figure 4.7 The obtained spectral graphs for the DiCENet-based architectures,
the architectural names are based on the aforementioned
abbreviations.
DenseNet DenseNet-S + MLP DenseNet-S + E-Mixup + MLP
0 0 0
20 20 20
40 40 40
60 60 60
80 80 80
Figure 4.8 The obtained spectral graphs for the DenseNet-based architectures,
the architectural names are based on the aforementioned
abbreviations.
SqueezeNet SqueezeNet-S + MLP SqueezeNet-S + E-Mixup + MLP
0 0 0
20 20 20
40 40 40
60 60 60
80 80 80
Figure 4.9 The obtained spectral graphs for the SqueezeNet-based architectures,
the architectural names are based on the aforementioned
abbreviations.
References
[1] Sathyanarayanan D, Anudeep D, Das CAK, et al. A multiclass deep learning
approach for LULC classification of multispectral satellite images. In: 2020
IEEE India Geoscience and Remote Sensing Symposium (InGARSS). IEEE;
2020. Available from: http://dx.doi.org/10.1109/ingarss48198.2020.9358947.
Leveraging twin networks for land use land cover classification 63
[2] Naushad R, Kaur T, and Ghaderpour E. Deep transfer learning for land use
and land cover classification: a comparative study. Sensors 2021;21
(23):8083. Available from: http://arxiv.org/abs/2110.02580v3.
[3] Yassine H, Tout K, and Jaber M. Improving LULC classification from
satellite imagery using deep learning – EUROSAT Dataset. In: Conference:
The XXIV ISPRS Congress (TCIII-Remote Sensing), 2021, vol. 6, p. XLIII-
B3-2021:369–376. Available from: http://dx.doi.org/10.5194/isprs-archives-
xliii-b3-2021-369-2021.
[4] Gajjar P, Shah P, and Sanghvi H. E-Mixup and Siamese Networks for
musical key estimation. In: Smart Innovation, Systems and Technologies.
Singapore: Springer Nature; 2022. p. 343–350. Available from: http://dx.doi.
org/10.1007/978-981-19-2541-2_26.
[5] Gajjar P, Garg M, Shah V, et al. Applicability analysis of attention U-Nets
over vanilla variants for automated ship detection. Reports on Geodesy and
Geoinformatics. 2022;114(1):9–14.
[6] Gajjar P, Mehta N, and Shah P. Quadruplet loss and SqueezeNets for Covid-
19 detection from chest-X ray. CSJM 2022;30(2(89)):214–222. Available
from: http://dx.doi.org/10.56415/csjm.v30.12.
[7] Mehta N, Shah P, Gajjar P, et al. Ocean surface pollution detection:
applicability analysis of V-Net with data augmentation for oil spill and other
related ocean surface feature monitoring. In: Communication and Intelligent
Systems. Singapore: Springer Nature; 2022. p. 11–25. Available from: http://
dx.doi.org/10.1007/978-981-19-2130-8_2.
[8] Mehta N, Shah P, and Gajjar P. Oil spill detection over ocean surface using
deep learning: a comparative study. Marine Systems & Ocean Technology,
2021 11;16(3–4):213–220. Available from: http://dx.doi.org/10.1007/
s40868-021-00109-4.
[9] He Z, Xie L, Chen X, Zhang Y, Wang Y, and Tian Q. Data augmentation
revisited: rethinking the distribution gap between clean and augmented data.
arXiv preprint arXiv:1909.09148. 2019. Available from: http://arxiv.org/abs/
1909.09148v2.
[10] Qiu K, Ai Y, Tian B, et al. Siamese-ResNet: implementing loop closure
detection based on Siamese network. In: 2018 IEEE Intelligent Vehicles
Symposium (IV). IEEE; 2018. Available from: http://dx.doi.org/10.1109/
ivs.2018.8500465.
[11] Ghojogh B, Ghodsi A, Karray F, and Crowley M. Generative Adversarial
Networks and Adversarial Autoencoders: Tutorial and Survey. arXiv pre-
print arXiv:2111.13282. 2021. Available from: http://arxiv.org/abs/2111.
13282v1.
[12] Wolfe CR and Lundgaard KT. E-Stitchup: Data Augmentation for Pre-Trained
Embeddings. arXiv preprint arXiv:1912.00772. 2019. Available from: http://
arxiv.org/abs/1912.00772v2.
[13] Shrestha A and Mahmood A. Enhancing Siamese networks training with
importance sampling. In: Proceedings of the 11th International Conference
on Agents and Artificial Intelligence. SCITEPRESS – Science and Technology
64 Earth observation data analytics using machine and deep learning
[25] Shepanski JF. Multilayer perceptron training using optimal estimation. Neural
Networks, 1988;1:51. Available from: http://dx.doi.org/10.1016/0893-6080
(88)90093-7.
[26] Que Y and Lee HJ. Densely connected convolutional networks for multi-
exposure fusion. In: 2018 International Conference on Computational
Science and Computational Intelligence (CSCI). IEEE; 2018. Available
from: http://dx.doi.org/10.1109/csci46756.2018.00084.
[27] Pal A, Krishnan G, Moorthy MR, et al. DICENet: fine-grained recognition
via dilated iterative contextual encoding. In: 2019 International Joint
Conference on Neural Networks (IJCNN). IEEE; 2019. Available from:
http://dx.doi.org/10.1109/ijcnn.2019.8851800.
[28] He K, Zhang X, Ren S, et al. Deep residual learning for image recognition. In:
2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
IEEE; 2016. Available from: http://dx.doi.org/10.1109/cvpr.2016.90.
[29] Iandola FN, Han S, Moskewicz MW, Ashraf K, Dally WJ, and Keutzer K.
SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and
<0.5MB model size. 2016. arXiv preprint arXiv:1602.07360. Available
from: http://arxiv.org/abs/1602.07360v4.
[30] Metha S, Hajishirzi H, and Rastegari M. DiCENet: dimension-wise con-
volutions for efficient networks. IEEE Transactions on Pattern Analysis and
Machine Intelligence 2020;44:2416–2425. Available from: https://dx.doi.
org/10.1109/TPAMI.2020.3041871.
[31] Wang A, Wang M, Jiang K, Cao M, and Iwahori Y. A dual neural archi-
tecture combined SqueezeNet with OctConv for LiDAR data classification.
Sensors, 2019;19(22):4927. Available from: http://dx.doi.org/10.3390/
s19224927.
[32] Reiser P, Eberhard A, and Friederich P. Implementing graph neural networks
with TensorFlow-Keras. arXiv preprint arXiv:2103.04318. 2021. Available
from: http://arxiv.org/abs/2103.04318v1.
[33] Mishra P. Introduction to neural networks using PyTorch. In: PyTorch
Recipes. Apress; 2019. p. 111–126. Available from: http://dx.doi.org/
10.1007/978-1-4842-4258-2_4.
[34] Bittrich S, Kaden M, Leberecht C, et al. Application of an
interpretable classification model on Early Folding Residues during protein fold-
ing. BioData mining, 2019 1;12 1–16. Available from: http://dx.doi.org/10.1186/
s13040-018-0188-2.
[35] Roy SK, Krishna G, Dubey SR, et al. HybridSN: exploring 3-D-2-D CNN
feature hierarchy for hyperspectral image classification. IEEE Geoscience
and Remote Sensing Letters, 2020;17(2):277–281. Available from: http://dx.
doi.org/10.1109/lgrs.2019.2918719.
This page intentionally left blank
Chapter 5
Exploiting artificial immune networks for
enhancing RS image classification
Poonam S. Tiwari1, Hina Pande1 and
Shrushti S. Jadawala2
The use of up-to-date and detailed information about the urban land cover is stra-
tegic for urban planning and management in present times, involving issues related
to the recent massive urban sprawl and densification, climate change, and the need
for environmental protection. The remote sensing images obtained by high spatial
resolution satellite sensors are important data sources for urban classification. The
artificial immune network (AIN), a computational intelligence model based on
artificial immune systems (AIS) inspired by the vertebrate immune system, has
been widely utilized for pattern recognition and data analysis. The algorithm is
based on the principles of the behaviours of both B cells and T cells in the biolo-
gical immune system. However, due to the inherent complexity of current AIN
models, their application to multi-/hyperspectral remote sensing image classifica-
tion has been severely restricted. The study explores the accuracy gained in land
cover classification using the AIN.
The algorithm is inspired by the clonal selection theory of acquired immunity
that explains how B and T lymphocytes improve their response to antigens over
time. The region of interest (ROI) is selected. Clonal selection algorithms are most
commonly applied to optimization and pattern recognition domains. The algorithm
is initialised by randomly chosen pixels to a set of memory cells Ab{m} and to the
set of Ab{r}. Euclidian distance is calculated to define the affinity to Ab. The cells
showing the highest affinity are labelled as memory cells. The antibody cells are
cloned and mutated, and finally the affinity is calculated, and memory cells are
updated. The final classification is carried out based on affinity to the
memory cells.
High-resolution earth observation data for Chandigarh city, India, has been
utilised for the study. The urban land cover was extracted using the AIN algorithm
and maximum likelihood classifier and a comparative analysis was carried out.
1
Department of Space, Indian Institute of Remote Sensing, Indian Space Research Organization,
Government of India, India
2
Department of Earth Sciences, School of Science, Gujarat University, India
68 Earth observation data analytics using machine and deep learning
Data was classified into 15 urban landcover classes such as roads, roof types and
vegetation. It was observed that for most of the classes, an improvement in overall
accuracy and kappa statistics was observed by using the approach based on an AIN.
Overall kappa statistics for MLC was found to be 0.72, and for AIN, it was cal-
culated to be 0.86. The study demonstrates the potential of AIN-based algorithms
for the classification of high-resolution remote sensing images.
5.1 Introduction
Due to an immense population, rapid urbanization and industrialization, land
resources worldwide have been facing unprecedented pressure in the past decades,
highlighted by the rapid loss of quality agricultural land and excessive sprawl of
urban boundaries. The situation in countries with large populations, e.g. India is
quite alarming [1–3]. Therefore there is an urgent need to create a balance between
the land-use supply and demand in a region and promote the sustainable utilisation
of land resources. This will lead to better land-use planning and optimisation of
resources [1,4].
Image classification is an important issue in remote sensing and other appli-
cations. The accurate classification of remote sensing images has a wide range of
uses, including reconnaissance, assessment of environmental damage, land use
monitoring, urban planning and growth regulation [5–7]. A significant distinction
in image classification separates supervised from unsupervised classification
methods. In remote sensing image classification, a key issue is to improve classi-
fication accuracy. For many years, a conventional statistical classifier, such as
maximum-likelihood (ML), has been applied for remote sensing image classifica-
tion. However, these conventional multivariate statistical methods require non-
singular and class-specific covariance matrices for all classes. Because of the
complexity of ground matters and the diversity of disturbance, these traditional
classification methods often have the drawback of low precision.
Machine-learning algorithms can generally model complex class signatures,
accept a variety of input predictor data, and do not make assumptions about the data
distribution. A wide range of algorithms are being used, such as SVMs, single DTs,
RFs, boosted DTs, an AIS based, and k-nearest neighbour (k-NN). Selecting a
machine-learning classifier for a particular task is challenging, not only because of
the wide range of available machine-learning methods but also because the litera-
ture appears contradictory, making it difficult to generalize the relative classifica-
tion accuracy of individual machine-learning algorithms [6,8,9].
In recent years, a new intelligence theory – AISs has also been applied to
classify remote-sensing images. The AIS is inspired by its natural counterparts and
has exhibited many benefits over traditional classifiers [7,10,11]. Compared to the
conventional statistical classifier, the AIS classifier has the capacity for self-
learning and robustness. AIS are data-driven self-adaptive methods that can adjust
themselves to the data without any explicit specification of functional or distribu-
tional form for the underlying model. Also, AIS are nonlinear models, which makes
Exploiting artificial immune networks for enhancing RS image classification 69
epitope
antigen
paratope B-Cell 2
Id2
idiotope P2
B-Cell 1 antibody 2
Id1
P1
antibody 1
suppression
B-Cell 3
stimulation
Id3
P3
antibody 3
(a) (b)
Figure 5.2 (a) High-resolution image of study area. (b) Multispectral image of
study area.
The study area should be best suited to the adopted methodology for achieving
the objectives of our research work, or simply it should be helpful for implementing
our methodology to achieve the desired goal (or it will work in an opposite manner
also as we opt for that methodology which is best suited to our predefined study
area). So, the selection of a study area is a very important part of any research work.
Keeping all these things in mind, we take Chandigarh City as a study area for our
research work. As Chandigarh is a properly planned city in India, urban feature
classification can be very easy due to the numerous urban features with proper
spacing.
The study is carried out in a part of Chandigarh city, India. The study area is
geographically located between 76 45’32.44” E to 76 56’11.82” E longitudes and
30 47’ 52.77” N to 30 38’38.4” N latitudes. The average height of the underlying
terrain is 330.77m above mean sea level. Chandigarh is located on the foothills of
the Himalayas. The study area consists of a dense urban area including buildings,
schools, hospitals, industries, vegetation cover surrounding the school and open
spaces, bare land, and road networks. The WorldView-2 images of the study area
are shown in Figure 5.2(a) and (b).
are obtained by selecting a ROI. The overall methodology is shown in Figure 5.3.
The training procedure is as follows.
5.3.1 Initialization
Available Ab repertoire that can be decomposed into several different subsets. Let
Ab{m} represent the set of memory cells. Ab{r} represent the set of remaining Ab.
Ab = Ab{m} + Ab{r} (r + m = N). This is done by randomly choosing training
Select ROI
Random selection of
antigen
Replication of selected
by cloning antibodies
Mutation
Replication of selected
by cloning antibodies
Calculation of affinity
antigens to be added to the set of memory cells Ab{m} and to the set of Ab{r}. For
each antigen (Ag) in the training set perform, the following steps.
aff i ¼ dj
where b is the a multiplying factor, N is the total number of Ab’s round, and () is
the operator that rounds its argument toward the closest integer.
smaller the mutation rate. Where mutate procedure and function mutate(x) are
defined in the equation below. The function Irandom() returns a random value in
the range [0,1] and Lrandom returns a random value in the range [1,1]. Function
D(t, y) is defined in (5.3) as follows:
l
Dðt; yÞ ¼ y 1 r ð 1Tt Þ
(5.3)
5.3.8 Decide
Whether the mccandidate replaces mcmatch that was previously identified. If mccandidate
has more affinity by the training antigen, Ag. The candidate memory cell is added to
the set of memory cells Ab {m} b and replaced with mcmatch.
S
30°42'55''N
30°42'55''N
30°42'30''N
30°42'30''N
5.3.9 Replace
Replace the d lowest affinity Ab’s from Ab {r}.
N
AIN classified image of Chandigarh W E
S
30°42'55''N
30°42'55''N
30°42'30''N
30°42'30''N
Class_Names
Bare land Young vegetation Polycarbonated roof
this one antigen stops. If the stopping criterion has not been met, repeat, beginning
at step 3.
After training is done, the evolved memory cells are available for use in
classification. Each memory cell is presented with a data item. By calculating the
closeness or affinity between the memory cell and image data, the image is clas-
sified into the class that has the maximum closeness. Based on the training data,
image was classified into 15 urban landcover classes such as different types of
roads, roof types and vegetation. The classified results for the maximum-likelihood
classifier (Figure 5.4) and AIN (Figure 5.5) are shown below.
5.4 Result
To validate the classification results, an accuracy assessment has been carried out
on both classified results. The accuracy was calculated for individual classes to
properly compare the classification algorithms. Table 5.2 gives the class-wise
achieved accuracy for both classification algorithms.
It was observed that for most of the classes, an improvement in overall accu-
racy and kappa statistics was observed by using the approach based on an AIN.
Table 5.2 shows that the AIN approach produces better classification results than
the ML method. Overall kappa statistics for MLC was found to be 0.72, and for
AIN, it was calculated to be 0.86. As shown in Table 5.2, the AIS approach
improved overall classification accuracy for each class; classes such as lawn,
footpath, glazed roof, polycarbonate roof, and concrete roof exhibit the most sig-
nificant improvement in accuracy, followed by the concrete roof, construction site,
road, tin roof, etc. For a few classes, both classifiers exhibited similar accuracy
levels, such as young and mature vegetation and bushes. The accuracy levels
decreased for metalled and unmetalled roads. This is because the ML approach
works well only when the underlying assumptions are satisfied and poor perfor-
mance may be obtained if the actual probability density functions are different from
those assumed by the model. At the same time, AIS are nonlinear models, which
make them flexible in modelling real-world complex relationship.
5.5 Conclusion
In this paper, we synthesise the advantages of the AIS and proposed a new remote
sensing image classification algorithm using clonal selection algorithm which is a basis
of the immune system. A quantitative comparison between the conventional ML sta-
tistical classifier and our algorithm was demonstrated that the ML statistical classifier is
less capable of discriminating roof types and urban features than AIS classifier. The
results also show concurrence with field observations and classification map derived
through visually interpreted reference. Experimental results show that the proposed
classification algorithm has high classification precision. It is a good and efficient
classification algorithm and can be applied to remote sensing image classification. AIS
will provide an alternative approach for accurate pattern recognition in remote sensing
data. It will not only be an effective learning algorithm and classify multi-remote
sensing images but also a very competent classifier for processing high volumes of data
e.g. in hyperspectral images. As a future scope of the work, it is planned to investigate
the approach for decreasing the number of unclassified antigens which are not recog-
nized by any artificial antibody in the trained network. In addition, it is also planned to
enhance the classifiers by considering feature selection or extraction using other AIS
models in high-dimensional feature space.
References
[20] Zhang, L., Zhong, Y., and LI, P. (2004). Application of artificial immune
systems in remote sensing image classification. In: The International
Archives of Photogrammetry, Remote Sensing and Spatial Information
Sciences.
[21] Campelo, F., Guimaraes, F.G., and Igarashi, H. (2008). Multiobjective
optimization using compromise programming and an immune algorithm.
IEEE Transactions on Magnetics, 44(6), pp. 982–985.
[22] Grazziela. F., Nelson, E., Helio, J.C.B., et al. (2007). The SUPRAIC algo-
rithm: a suppression immune-based mechanism to find a representative
training set in data classification tasks. In: ICARIS 2007, LNCS 4628, pp.
50–70.
[23] Yin, G. (2003). The multi-sensor fusion: image registration using artificial
immune algorithm. In: Soft Computing Techniques in Instrument,
Measurement and Related Applications 2003, pp. 32–36.
Chapter 6
Detection and segmentation of aircrafts in UAV
images with a deep learning-based approach
Hina Pande1, Poonam Seth Tiwari1, Parul Dhingra1 and
Shefali Agarwal1
6.1 Introduction
UAVs [1] are utilized in civil and military arenas for several purposes such as
surveillance, security, recreational, educational, rescuing, and monitoring. Due to
1
Indian Institute of Remote Sensing, ISRO, India
82 Earth observation data analytics using machine and deep learning
Various DL algorithms like R-CNN, Fast R-CNN, YOLO, Faster R-CNN, SSD,
R-FCN, etc. have been developed for various object detection tasks.
Traditionally convolutional networks were used for classification tasks [7].
However, there are several segmentation applications that require assignment of
class labels to each pixel. Ciresan et al. [8] proposed a network for such require-
ments, where the local region around the pixel was used to predict the class label.
This algorithm was quite slow as it had to run separately for each local region/
patch, and there was a trade-off between use of context in terms of size of patches
and accuracy of localization. To overcome these limitations networks were devel-
oped where features from multiple layers were taken into account for classifier
output. Further, Olaf Ronneberger et al. [7] developed a more advanced archi-
tecture where even with few training images precise segmentations could be
achieved. Pathak et al. [14] applied DL for object detection. Alganci et al. [15]
compared various DL approaches such as Faster R-CNN, Single Shot Multi-box
Detector (SSD), and You Only Look Once-v3 (YOLO-v3) for airplane detection
from very high-resolution satellite images. They concluded that Faster R-CNN
architecture provided the highest accuracy according to the F1 scores, average
precision (AP) metrics and visual inspection of the results. Zhaoa et al. [30] pro-
posed a heterogeneous model to transfer CNNs to remote-sensing scene classifi-
cation to correct input feature differences between target and source datasets. Ji
et al. [31] detected aircraft in high spatial resolution RS images by combining
multi-angle features and majority voting CNN.
The study aims to develop a supervised learning framework for detecting air-
planes in UHSR images acquired with UAV’s using MobileNet-deep neural net-
work. Since the target of interest is likely to have multiple orientations in the
image, a multi-angle feature extraction is enabled. The airport images are manually
labelled and segmented using U-Net architecture. Performance analysis of deep
neural networks is assessed in the study. The paper is organized as follows.
Section 6.1 discusses the basic technical concepts underlying our research work.
The method for detection and segmentation of airplanes in UHSR images is pre-
sented in Section 6.3. Section 6.4 discusses training and testing process, limitations,
and objective analysis of the trained models. Section 6.5 states conclusions.
6.2 Background
UAV’s has recently become popular across the fields of CV and RS due to their
comprehensive and flexible data acquisition, Inspired by recent success of DL,
many advanced object detection and tracking approaches have been widely applied
to various UAV-related tasks, such as environmental monitoring, precision agri-
culture, and traffic management.
The following section explains characteristics of high resolution images,
and further, provides the background of the neural networks and convolutional
neural networks (CNN) specifically for automated object detection and
segmentation.
84 Earth observation data analytics using machine and deep learning
6.2.3 CNNs
CNN are widely used neural networks for extracting information from 2D-image
data, where inputs are grid-structured, and there are spatial dependencies within the
local regions [12]. The pixels in the neighbourhood of an individual pixel often
have similar values; hence, image data exhibits strong spatial dependencies, which
makes it highly suitable for CNNs. The CNNs can be used for spatial, temporal, and
spatiotemporal input data. The image data exhibits translation invariance, where an
object has the same interpretation irrespective of its location in the image. In
CNNs, similar feature values are created from local regions that have a similar
pattern. The basic operation executed in CNNs is mathematical convolution. A
convolution operation is a sliding dot-product carried out between the convolution
filters and grid-structured inputs. The operation is beneficial for data that exhibits a
high level of spatial locality. CNNs are the neural networks in which at least one
86 Earth observation data analytics using machine and deep learning
layer is the convolution layer. There can be one or multiple convolution layers in a
neural network. As every feature value in the current layer is dependent on the
small local region in the previous layer, the spatial relationships get inherited from
one layer to the next layer. A three-dimensional grid structure with height, width,
and depth define each convolution layer of a CNN. The depth refers to the number
of feature maps in a convolutional layer. The basic building blocks of a typical
feed-forward convolutional neural network are convolution layer, pooling layer,
rectified linear unit (ReLU) layer, fully connected layer, and loss layer. The con-
volution layer overlaps the kernel at every location in an image and performs a
sliding dot product. The pooling layer basically performs the down sampling of the
feature maps in a non-linear manner. Max pooling is one most commonly used non-
linear function in the pooling layers. A new feature map is produced as the pooling
layer acts independently on every depth slice of the feature map. The input image is
partitioned into non-overlapping regions in a feature map, and the pooling function
obtains the maximum value in the particular region to generate a new feature
map. The pooling layer reduces the size of the feature map and the parameters
required to train the network; hence, the computational complexities within the
convolutional neural network are reduced. The commonly used activation function
in CNNs is ReLU activation functions. The ReLU function squashes the negative
values to zero and, therefore, does not permit negative values to propagate in the
network. The dimensions of a layer remain the same when an activation function is
applied as it only maps the values in the feature map corresponding to the activa-
tion function. After the implementation of convolution and max-pooling layers,
lastly, the outputs are generated by implementing a fully connected neural layer.
The loss layer which is the final layer of the convolutional neural network deter-
mines the deviation between the expected ideal outcome and the predicted out-
come. Softmax loss and sigmoid cross-entropy loss are examples of such loss
functions in the loss layer. The CNNs are used to perform object detection, clas-
sification, and segmentation tasks in CV. We have implemented CNNs to automate
the task of target extraction from UHSR images.
6.3 Methodology
6.3.1 Dataset
The UHSR image-dataset utilized in the project is captured by two eBee classic
drones [13], flying at a height of 393.7 ft. The ground resolution of images is
3.14 cm/px. The data has been acquired over the Le Bourget airport in Paris. The
dimension of the images is 4,608 3,456 pixels. The 2D-images are captured in the
visible spectral range: comprising the red, green, and blue wavelengths. The images
contain one or multiple parked-airplanes, along with several other objects like
buildings, runways, automobiles, etc. Figure 6.1 shows images from airport dataset.
The following section describes the experimental approach adopted for
extracting airplanes from the UHSR images in an automated way. Figure 6.2
depicts the broad methodology followed for the study.
Detection and segmentation of aircrafts in UAV images 87
(a)
(b)
Figure 6.1 (a) SenseFly eBee drone [1]. (b) Image samples of airport area.
GENERATING GROUND-
IMAGE ANNOTATION
TRUTH MASK IMAGES
RESULT ANALYSIS:
Mean Average Precision RESULT ANALYSIS:
(MAP) Dice coefficient
Object detection
Semantic segmentation
the image, name of the object annotated (airplane) and location of the manually
annotated bounding boxes in the image. For training, the neural network 13 images
are annotated manually which generated 13 .xml files.
Further, these 13 individual .xml annotation files are combined and converted
to Comma Separated Values (CSV) .csv file. The .csv file and image data are
converted and stored in TensorFlow Record (TFRecord) format. The TFRecord
format stores the data in binary format and significantly reduces the training time,
and occupies less space in the storage disk. The training data in TFRecord file
format is fed to the neural network.
Detection and segmentation of aircrafts in UAV images 89
path [7] are as follows. First, it performs two 3 3 convolutions with 64 filters.
After each convolution, the outputs are subjected to the ReLU activation function
and are downsampled by using a 2 2 max-pooling operation with stride 2. With
each down sampling step, the number of feature channels is doubled. The archi-
tecture for the expansion path is as follows. First, the expansion path upsamples the
feature map. Subsequently, to reduce the number of feature channels to half, a
2 2 up-convolution is implemented, followed by a concatenation step. Further,
two 3 3 convolutions are implemented. The outputs from both the convolutions
are subjected to the ReLU activation function. The final layer implements 1 1
convolutions, which relates a feature vector consisting of 64 components to the
required number of labels.
(a) (b)
(c) (d)
Figure 6.6 Test cases: RGB images (a), (b), (c), and (d)
(a) (b)
(c) (d)
6.4.1.4 Limitations
SSD-MobileNet v2 is trained to enclose the airplane feature in rectangular
bounding boxes. When the trained network is tested with nine images, it is
observed that if an image consists of an object whose spatial features are similar to
that of an airplane, then it captures that object too as the target. Figure 6.8 shows
the two limitation cases A and B, where the objects other than airplane feature are
enclosed in rectangular bounding boxes.
6.4.2.1 Training
The network [22] is trained by utilizing neural network libraries provided by Keras
API. The training data is stored in portable network graphics (PNG) format. Before
initiating the training process, the data augmentation step is carried out, as we are
aiming to train the network with a lesser number of images. The data augmentation
process helps in minimizing the risk of overfitting while training the neural net-
work. To execute the data augmentation, few of the operations executed are flip-
ping, zooming, shearing, etc. The initial weights of the network are set-up using
transfer learning. Further, we train the U-Net with our training data to update the
weights with several epochs. The model’s weights get updated after every epoch if
the loss reduces. The Adam-optimiser is used to update the model’s weights after
each iteration. The loss function implemented while training the network is binary
cross entropy loss [26]:
P
s
Loss ¼ 1s a i þ ð1 ai Þlogð1 b
ai logb aiÞ (6.5)
i¼1
where ba i is the ith value in the model output, s is the output size, and ai is the
target value.
Figure 6.9 Outputs from trained U-Net neural network. First row: original RGB images
for testing. Second row: corresponding grayscale images fed to network for
testing. Third row: outputs from the trained U-Net neural network.
96 Earth observation data analytics using machine and deep learning
regarding real-time processing using UAV datasets may also be explored. This study
mainly focuses on aircraft detection; however, the method proposed is theoretically
applicable to other kinds of targets e.g. ships, vehicles, etc. in RS images. In future,
the algorithm could consider more hyperparameters to optimise the network structure
and the training process and further improve the accuracy of detection.
References
[1] H. Shakhatreh, A.H. Sawalmeh, A. Al-Fuqaha, et al. Unmanned aerial vehi-
cles (UAVs): A survey on civil applications and key research challenges. IEEE
Access, 7, pp. 48572–48634, 2019, doi:10.1109/ACCESS.2019.2909530.
[2] S. Wang, Y. Han, J. Chen, Z. Zhang, G. Wang, and N. Du. A deep-learning-
based sea search and rescue algorithm by UAV remote sensing. In: 2018 IEEE
CSAA Guidance, Navigation and Control Conference (CGNCC), pp. 1–5, 2018.
[3] A.M. Jensen, B.T. Neilson, M. McKee, and Y. Chen. Thermal remote sen-
sing with an autonomous unmanned aerial remote sensing platform for sur-
face stream temperatures. In: 2012 IEEE International Geoscience and
Remote Sensing Symposium, pp. 5049–5052, 2012.
[4] H. Shakhatreh, A.H. Sawalmeh, A. Al-Fuqaha, et al. Unmanned aerial
vehicles (UAVs): a survey on civil applications and key research challenges.
IEEE Access, 7, pp. 48572–48634, 2019.
[5] S. Hwang, J. Lee, H. Shin, S. Cho, and D.H. Shim. Aircraft detection using
deep convolutional neural network for small unmanned aircraft systems.
AIAA Information Systems – AIAA Infotech @ Aerospace, 2018.
[6] Comparing Deep Neural Networks and Traditional Vision Algorithms in
Mobile Robotics by Andy Lee.
[7] O. Ronneberger, P. Fischer, and T. Brox. U-net: convolutional networks for
biomedical image segmentation. In: N. Navab, J. Hornegger, W.M. Wells,
and A.F. Frangi (eds.), Medical Image Computing and Computer-Assisted
Intervention – MICCAI 2015, Cham: Springer International Publishing,
2015, pp. 234–241.
[8] D.C. Ciresan, L.M. Gambardella, A. Giusti, and J. Schmidhuber. Deep
neural networks segment neuronal membranes in electron microscopy
images. In: NIPS, 2012, pp. 2852–2860.
[9] A. Parkin. Computing Colour Image Processing. Cham: Springer, 2018.
[10] S.K. Ghosh. Digital Image Processing. Oxford: Alpha Science Intl Ltd, 2012.
[11] S.S. Haykin. Neural Networks and Learning Machines, 3rd ed. Upper Saddle
River, NJ: Pearson Education, 2009.
[12] C.C. Aggarwal. Neural Networks and Deep Learning. New York, NY:
Springer International Publishing Company, 2018.
[13] Airport Dataset, SenseFly Parrot Group.
[14] A.J. Pathak, M. Pandey, and S. Rautaray. Application of deep learning for
object detection. Procedia Computer Science, 132, pp. 1706–1717, 2018.
International Conference on Computational Intelligence and Data Science.
98 Earth observation data analytics using machine and deep learning
1
Knight Foundation School of Computing and Information Sciences, Florida International University,
USA
2
Applied Research Center, Florida International University, USA
102 Earth observation data analytics using machine and deep learning
7.1 Introduction
Assessment of damage after a hurricane disaster is progressively more imperative
for emergency administration. The existing method relies on disaster response crew
members to drive into the affected area to analyse the damage using the windshield
analysis approach. This method is expensive and time-intensive. Numerous
experimental studies have been performed to support image analysis and further
decrease the time for collecting data to speed up the assessment execution time.
Learning-based algorithms are one of the noteworthy trends to distinguish whether
a building is impaired or safe, after a hurricane landfall, on metrics such as accu-
racy, precision, and F1 score using satellite imagery [1]. Other research applied
such learning algorithms to analyse panchromatic imagery and the time series of
Landsat5/7 satellite imagery [2,3]. Chen et al. use synthetic aperture radar (SAR)
data and extract the texture feature of the building to perform learning-based ana-
lysis. Deep learning techniques [4,5] have shown promising results within the field
of damage assessment. Analysis of precipitation, a network of rivers [6], and usage
of topological data to analyse the depth [7] were a few of the approaches used
before deep learning-based CNN algorithm development. To detect impairment in
the car [8,9], concrete structures [10–12], or detecting the change in the region after
the impact of a hurricane [13], CNN is used extensively. Nevertheless, such
methods require labelled datasets of good quality and quantity, which might be
expensive in a few cases. Transfer Learning is the key to mitigating this issue to a
reasonable extent. It is a method to develop and train the CNN model on a good
dataset and use it for other related datasets. This improves training time and effi-
ciency. The rest of the chapter is summarized as follows. Section 7.2 provides the
literature review. Section 7.3 details the image processing technique and an in-
depth overview of a deep learning-based CNN. Section 7.4 explains transfer
learning with the various pre-trained model. Section 7.5 describes the practical use
cases of the VGG-16 for damage classification. Finally, we conclude in Section 7.7.
et al. [22] used an in-house dataset from Hurricane Dorian to train a stacked CNN.
Hao and Wang [23] analyse the images of social networking platforms to detect the
type of damage and level of severity by training five different machine learning
algorithms. Recently, a transfer learning-based approach has been explored for uti-
lizing the pre-trained model to assess hurricane damage. Many studies use aerial
images [24–26] and apply transfer learning on a pre-trained CNN. This chapter
proposes a transfer learning-based framework to assess hurricane damage using the
open-source satellite image dataset.
Supervised learning
Input data with a label is used for training purposes. Based on the target label value,
such a learning algorithm is divided into classification and regression.
Classification is when the output label is categorical, and the problem is a regres-
sion when the output label is a real value. Decision Trees, Random Forest, K
Nearest Neighbour, Support vector machine, and linear regression are supervised
learning algorithms.
Unsupervised learning
Input data without the target label is used for training. Such algorithms learn the
hidden patterns from the unlabelled dataset. Clustering is one such type. It clusters
the data into different groups. K-Means, Agglomerative, and Divisive are some of
the unsupervised algorithms.
Semi-supervised learning
It contains a small part of the labelled dataset, with most data being unlabelled. The
algorithms learn from the labelled data to predict unknown samples.
With the increase in computational power and rapid amount of data generation,
deep learning algorithms are used heavily in almost every application area. One
such deep learning-based algorithm is a CNN as shown in Figure 7.2. For image
analysis such as object segmentation, object recognition and video processing,
CNNs are widely used.
Let us, deep dive, into the working mechanism of CNN.
CNNs consist of four layers:
(i) Convolutional layer: This layer extracts the meaningful information or fea-
tures from the image by applying a convolution operation when an input
image is fed to this layer. Numerous such convolution operations extract the
diverse distinct features from the image.
(ii) Pooling layer: The second layer after the convolutional layer. It takes the
max or min or any statistical value to further reduce the dimension of the
image. It is then passed to activation functions such as rectifier linear unit
(ReLU) to have some nonlinearity.
Image
(iii) Flatten layer: This is the third layer in the architecture. The main objective
of this layer is to flatten the input coming from the pooling layer.
(iv) Neural network: Finally, the last layer is a fully connected dense layer. The
output of the flatten layer is processed by the neural network for classifica-
tion, image segmentation, or object detection purposes.
Hyper-parameters of CNN
Every learning-based algorithm has hyper-parameters that need to be tuned to
achieve optimal results. The values for the hyperparameters have to be set before
training the model, and it is one of the active research areas. Let us discuss a few of
the crucial hyperparameters for the CNN algorithm.
1. Epochs: Number of times the entire dataset is processed by the model during
the training phase.
2. Learning rate: This parameter controls the rate of change in gradient des-
cent, thus optimizing the weights. It can either be increased or decreased
gradually or be kept fixed throughout the training time. It depends on popular
optimizers such as Adaptive Delta, Adaptive momentum, RMSprop, and
Stochastic Gradient Descent.
3. Activation function: They are used during the training of the neural network to
apply nonlinearity. There are various activation functions, and the choice depends
on the task that needs to be solved. ReLU is popularly used in CNN networks.
Tangent hyperbolic and sigmoid are other activation functions that can be used.
4. Batch size: It controls the update of the weight during each epoch. It is
usually in the range of 32–256 rows of the dataset, which can vary. It means
the update of weights will occur after the processing of each batch of rows
within the epoch.
5. The number of hidden layers and units: The total number of neurons in
each hidden layer with the total number of hidden layers parameters is deci-
ded based on how well the model performs on the dataset. There is an
underfitting issue with a small subset of it, whereas the model tends to be
overfitted when such parameters have high values. Regularization techniques
can be used to find the trade-offs.
6. Dropout for regularization: Dropout is used to prevent the overfitting of the
model. If the dropout value is 0.3, 30% of the neurons in that layer will be
deactivated during model training.
7. Weight initialization: A small set of random numbers can initialize the
weights. Such random numbers should be uniformly distributed.
8. Kernel size: It is the filter’s size used to extract features. Different kinds of
features can be extracted from different sizes of the kernel.
9. Stride: It indicates how many pixel values the kernel should be moved while
moving the kernel through the image in the convolution layer.
10. Pooling: It reduces the dimensionality of the features [28,29]; max-pooling
will use the max value to reduce the feature, whereas min-pooling will use
the min value to reduce the features.
106 Earth observation data analytics using machine and deep learning
Task1
Top Bottom
Data1 Model Predictions1
Layer Layer
Transfer Learning
Task 2
Access to a large dataset of 1.2 million training images, 50k validation images,
and 100k test images was provided to various teams in a competition for The
ImageNet Large Scale Visual Recognition Challenge (ILSVRC) [30] to develop
and train the model for classifying the images. There were a total of 1,000 cate-
gories. This open-source benchmark dataset is used to validate the models devel-
oped for transfer learning purposes.
Let us discuss various pre-trained models.
7.4.1 AlexNet
AlexNet [31] as shown in Figure 7.4 was developed in 2012 and was the winner of
the ILSVRC competition, where it has decreased the error rate to half. It was the
breakthrough towards the advancement of CNN networks. They utilize GPUs for
the ReLU activation function. AlexNet architecture contains five convolutional
layers and three fully connected layers with 60,000 parameters. Furthermore, the
images were augmented by flipping, scaling, adding noise, etc.
It leads to the stride of length 4 in the pooling layer to reduce the error rate.
Input
Conv 1 (96), stride 4, 11 × 11
MaxPooling, stride 2, 3 × 3
MaxPooling, stride 2, 3 × 3
MaxPooling, stride 2, 3 × 3
7.5 Implementation
This section discusses the practical implementation of using AlexNet for damage
classification. The following libraries are used for experimentation purposes.
Tensorflow: Tensorflow is an open-source library from google for deep learning
and machine learning [32].
Keras: Keras is a wrapper on Tensorflow used by many societies worldwide. The
code written in Keras is internally converted to TensorFlow for further execution. It
has functional API and Sequential API [33].
Scikit-learn: Scikit-learn deals with a wide variety of learning-based algorithms
(both supervised and unsupervised) [34].
Figure 7.5 shows the high-level framework for the implementation. It has three
sections: Data Collection, Data Preprocessing, and Training Algorithm with final
testing.
Stage 1: Data collection
This dataset is obtained from Kaggle. It contains satellite images of Hurricane
Harvey in the Texas region. The training, validation, and test data are depicted in
Table 7.1. Figures 7.6 and 7.7 show the sample images of No_Damage and Damage
types, respectively.
Stage-3: AlexNet Algorithm
Stage-1: Data Collection Model Training Phase
Test Data
Final
Trained Model
1
0.9
0.8
Accuracy Value
0.7
0.6
0.5
0.4
0.3
0.2
0.1
0
1 2 3 4 5 6 7 8 9 10 111213141516171819 20 21222324 252627 282930
Epoch
2
1.8
1.6
1.4
Loss value
1.2
1
0.8
0.6
0.4
0.2
0
1 2 3 4 5 6 7 8 9 10 111213141516171819 20 21222324 252627 282930
Epoch
For faster and optimal training, the ReLU activation function is used in all the
convolution layers. Table 7.2 shows the evaluated metrics. Thus, we can conclude
that a transfer learning algorithm can be applied to solve the problem where the
dataset is limited and there is a limited computational resource. The model can
further be deployed for real-time use cases. We performed all our experiments on
Google Colaboratory, where we used Keras with TensorFlow at the backend to
train the model and seaborn for visualization purposes.
7.6 Conclusion
Satellite images can prominently enable recovery and rescue efforts after the landfall
of a hurricane. Numerous machine learning-based supervised classification is being
applied to assess the damage to the building, and their use remains challenging. Since
they require a good amount of labelled datasets, thus this approach is time-intensive.
In this chapter, we provide an overview of CNN, a deep learning-based algorithm
heavily used to solve the problem of image datasets. Every learning-based algorithm
has many parameters that need to be tuned, and there is no precise formula to choose
these values. Therefore, we presented a transfer learning-based approach where
various pre-trained CNN-based algorithms can be applied directly to the new pro-
blem domain. We provide architectural details of several transfer learning-based
algorithms and, finally, we presented a high-level implementation of building
damage classification using an Alexnet pre-trained network.
This work can be extended by employing an ensemble approach by training
multiple pre-trained algorithms to increase the model’s accuracy. Additionally, we
can utilize geological data to enhance the study of building damages.
References
[1] Q.D. Cao, and Y. Choe. Building damage annotation on post-hurricane
satellite imagery based on convolutional neural networks. Natural Hazards,
vol. 103, no. 3, pp. 3357–3376, 2020.
[2] C. Van der Sande, S. De Jong, and A. De Roo. A segmentation and classi-
fication approach of ikonos-2 imagery for land cover mapping to assist flood
risk and flood damage assessment. International Journal of Applied Earth
Observation and Geoinformation, vol. 4, no. 3, pp. 217–229, 2003.
[3] S. Skakun, N. Kussul, A. Shelestov, and O. Kussul. Flood hazard and flood
risk assessment using a time series of satellite images: a case study in
Namibia. Risk Analysis, vol. 34, no. 8, pp. 1521–1537, 2014.
112 Earth observation data analytics using machine and deep learning
[17] R. Gupta, R. Hosfelt, S. Sajeev, et al. xbd: a dataset for assessing building
damage from satellite imagery, 2019. arXiv:1911.09296.
[18] E. Weber and H. Kané. Building disaster damage assessment in satellite
imagery with multi-temporal fusion, 2020. arXiv:2004.05525.
[19] K. He, G. Gkioxari, P. Dollár, and R. Girshick. Mask r-cnn. In Proceedings
of the IEEE International Conference on Computer Vision, Venice, Italy,
27–29 October 2017, pp. 2961–2969.
[20] H. Hao, S. Baireddy, E.R. Bartusiak, et al. An attention-based system for
damage assessment using satellite imagery, 2020. arXiv:2004.06643.
[21] R. Gupta, B. Goodman, N. Patel, et al. Creating xBD: a dataset for assessing
building damage from satellite imagery. In Proceedings of the IEEE/CVF
Conference on Computer Vision and Pattern Recognition Workshops, Long
Beach, CA, USA, 16–17 June 2019, pp. 10–17.
[22] C.S. Cheng, A.H. Behzadan, and A. Noshadravan. Deep learning for post-
hurricane aerial damage assessment of buildings. Computer-Aided Civil and
Infrastructure Engineering,, vol. 36, pp. 695–710, 2021.
[23] H. Hao and Y. Wang. Hurricane damage assessment with multi-, crowd-
sourced image data: a case study of Hurricane Irma in the city of Miami. In
Proceedings of the 17th International Conference on Information System for
Crisis Response and Management (ISCRAM), Valencia, Spain, 19–22 May
2019.
[24] Y. Li, W. Hu, H. Dong, and X. Zhang. Building damage detection from post-
event aerial imagery using single shot multibox detector. Applied Science,
vol. 9, p. 1128, 2019.
[25] M. Presa-Reyes and S.C. Chen. Assessing building damage by learning the
deep feature correspondence of before and after aerial images. In
Proceedings of the 2020 IEEE Conference on Multimedia Information
Processing and Retrieval (MIPR), Shenzhen, China, 6–8 August 2020,
pp. 43–48.
[26] Y. Pi, N.D. Nath, and A.H. Behzadan. Convolutional neural networks for
object detection in aerial imagery for disaster response and recovery.
Advanced Engineering Informatics, vol. 43, p. 101009, 2020.
[27] J. Soni and N. Prabakar. Effective machine learning approach to detect
groups of fake reviewers. In Proceedings of the 14th International
Conference on Data Science (ICDATA’18), Las Vegas, NV, 2018, pp. 3–9.
[28] J. Soni, N. Prabakar, and H. Upadhyay. Visualizing high-dimensional data
using t-distributed stochastic neighbor embedding algorithm. In Principles of
Data Science. Cham: Springer, 2020, pp. 189–206.
[29] J. Soni, N. Prabakar, and H. Upadhyay. Feature extraction through deep-
walk on weighted graph. In Proceedings of the 15th International
Conference on Data Science (ICDATA’19), Las Vegas, NV, 2019.
[30] O. Russakovsky, J. Deng, H. Su, et al. Imagenet large scale visual recogni-
tion challenge. International Journal of Computer Vision, vol. 115, no. 3,
pp. 211–252, 2015.
114 Earth observation data analytics using machine and deep learning
[31] Md.Z. Alom, T.M. Taha, C. Yakopcic, et al. The history began from alexnet:
a comprehensive survey on deep learning approaches, 2018. preprint
arXiv:1803.01164.
[32] J.V. Dillon, I. Langmore, D. Tran, et al. Tensorflow distributions, 2017.
preprint arXiv:1711.10604.
[33] N. Ketkar. Introduction to keras. In Deep Learning with Python. Berkeley,
CA: Apress, 2017, pp. 97–111.
[34] F. Pedregosa, G. Varoquaux, A. Gramfort, et al. Scikit-learn: machine
learning in Python. Journal of Machine Learning Research, vol. 12,
pp. 2825–2830, 2011.
[35] S. Kaur, S. Gupta, S. Singh, et al. Transfer learning-based automatic hurri-
cane damage detection using satellite images. Electronics,; vol. 11, no. 9,
p. 1448, 2022.
Chapter 8
Wildfires, volcanoes and climate change
monitoring from satellite images using deep
neural networks
Yash Kumar Shrivastava1 and Kavita Jhajharia1
Climate change refers to the long-term change in weather patterns and tempera-
tures, driven by multiple factors. Since 1880, human activities have been the pri-
mary accelerators of climate change. Climate change has broad, far-reaching
impacts on the environment. Sometimes these impacts result in catastrophic dis-
asters, causing huge economic (infrastructure and property loss, poverty, etc.) and
social (human lives, animal lives, diseases, malnutrition, displacement of families,
mental health issues, job losses, food, and water shortage, etc.) losses. These
impacts include Wildfires, Volcanic Eruptions, Landslides, etc. In 2020, in the
United States alone, there were 58,950 wildfires, burning about 10.12 million acres
of land. In India, in 2019 alone, landslides were responsible for the loss of
264 lives. The current global warming situation only accelerates these disasters
even more.
Monitoring the impacts of climate change will be crucial in tackling it, and in
reducing the consequences and losses caused due to disasters. Active and efficient
monitoring of disasters can help in early warning, which can ensure a faster and
more effective response, saving countless lives and reducing losses. Satellite ima-
gery plays a crucial role in monitoring climate change and observing its impacts
around the world. Wildfires, landslides, volcanoes, etc., can all be identified and
monitored with satellite imagery. Satellite imagery has historically been used to
monitor the long-term effects of climate change, like changes in the ice cap, forest
cover, etc. Satellite images of volcanic eruptions, wildfires, landslides, etc. help in
mapping the data resulting in more efficient rescue operations.
Today, methods like deep learning (DL) and machine learning (ML) can play a
crucial role in improving the efficiency of monitoring climate change and natural
disasters. ML is a branch of Artificial Intelligence (AI) and computer science that
involves the study of computer algorithms that can improve themselves through
experience, observations, and the use of data. DL is a part of a broader family of
ML methods that use multiple layers to progressively extract higher-level features
1
Manipal University Jaipur, India
116 Earth observation data analytics using machine and deep learning
from raw input. Applying DL and convolutional neural network (CNN) techniques
to satellite images can help in automating the disaster identification process,
thereby making the process faster and much more efficient. Image segmentation
can be of great help in monitoring environmental changes and natural disasters.
Image segmentation is a process that involves the division of a digital image into
various parts. These parts are called ‘Image segments’. This process helps in
reducing the complexity of the image and it also makes further processing and
analysis of the image much more simplified. Implementation of image segmenta-
tion techniques on very high-resolution (VHR) satellite images can substantially
help in monitoring climate change and natural disasters like wildfires, landslides,
volcanic eruptions, etc.
Wildfires are spontaneous destructive fires that spread quickly over forests,
woodland, and other combustible vegetation. CNNs can be used to help automate
the detection process of wildfires using VHR satellite imagery after training the
model with data. Image segmentation can be performed over satellite images for
automated detection of wildfires based on training data. The model can be trained
to divide the image into segments of fire and non-fire, burnt area and unburnt area.
Data augmentation techniques can also be used to enlarge the training data set.
CNNs like U-Net, Inception-v3, etc. have been used in the past and be very
effective in classifying images. Synthetic aperture radar (SAR) can also play a
critical role in wildfire detection and monitoring, since it can penetrate through
clouds and smoke, and is also capable of imaging day and night. A CNNs-based DL
framework can be used to identify the burnt area and differentiate it from the
unburnt area.
Wildfires can also trigger many other disastrous hazards, one such particular is
the occurrence of landslides in the burnt area. The burnt plots of land area are
highly susceptible to debris flows. Satellite imagery has made it much easier to
identify the areas affected by wildfires, and it has also enabled efficient assessment
of burnt areas after the wildfire. Assessment of post-wildfire burnt areas is a very
challenging task and poses a lot of risks. Satellite imagery makes this much easier
and risk-free. Manual monitoring of landslides too is a very labour-intensive and
time-consuming task. This problem can also be dealt with through DL methods, by
automating the task of landslide classification through satellite imagery for more
robust classification results. CNN’s like U-Net can be very useful for classification
purposes and landslide mapping can be done using VHR satellite images. These DL
techniques can automate the task of landslide detection, hence reducing the time
and human effort.
Volcano deformation is another phenomenon that can be monitored using DL
techniques. Satellite data can help in large-scale global monitoring and identifica-
tion of volcanic activity and provide the first indications of volcanic eruptions,
instability, etc. Volcanic deformation is a very good indicator of volcanic eruptions
as it usually happens before a volcanic eruption. Here, interferometric SAR
(InSAR) can be very useful in monitoring volcanic activity, as it can help in plot-
ting surface deformation. The images generated by InSAR are known as inter-
ferograms and they contain contributions from both deformation of volcanoes and
Wildfires, volcanoes and climate change monitoring 117
radar paths through the atmosphere. ML models can identify useful signals from a
large collection of InSAR data. Since the number of deforming volcanoes is very
low, implementing ML-based monitoring solutions becomes even more difficult.
CNNs like AlexNet can be very useful in automating this task. AlexNet can be
trained using synthetic interferograms for this task. The synthetic interferograms
contain deformation patterns based on multiple probability selection, effects caused
by stratified atmospheres derived from weather models, and other atmospheric
effects caused by irregularities in air motion. Such efficient DL models can help
identify volcanic activity faster and give warnings before volcanic unrest.
8.1 Introduction
Climate change has increasingly become a cause of concern among experts and
common citizens around the globe alike. But it must be noted that climate change is
not a new concern and has been discussed for a long time. The earliest concerns
were raised in 1938 by a British engineer who went by the name ‘Guy Callendar’
[1]. He compiled the records of 147 weather stations located around the globe and
used the data to show that the temperatures had risen over the preceding century.
He also noted that the concentration of carbon dioxide had increased significantly
over the same period. He suggested that this increment in CO2 levels was the
primary reason for the rise in the temperatures. This theory was widely known as
the ‘Callendar Effect’. It was largely dismissed by experts at that time [2].
Today, several governments around the world have started taking stringent
measures to keep a check on climate change and reduce its impact as much as
possible, and as soon as possible. Several countries have set respective goals to
reach the so-called ‘Net-Zero Emissions’ within a span of a few years. Countries
around the globe have come together on different occasions and summits to set
different goals for curbing climate change and reducing its impacts effectively and
quickly. One such summit is the COP26 [3], hosted in March 2021 [4], in Glasgow,
United Kingdom. COP – an abbreviation for ‘Conference of Parties’ – is the latest
yearly climate change conference held by the United Nations. It was attended by all
the countries that had signed the United Nations Framework Convention on
Climate Change (UNFCCC). This framework came into effect in 1994. COP26 was
the 26th summit of its kind. This summit saw many countries pledging the attain-
ment of ‘Net-Zero Emissions’ within a fixed period. In simple terms, ‘Net-Zero’
emissions are achieved when the amount of carbon present in the atmosphere is
equal to the amount of carbon that is being removed from the atmosphere. And
‘Net-Positive’ emissions are achieved when the amount of carbon present in the
atmosphere is less than the amount of carbon that is being taken away from the
atmosphere. Most countries pledged to attain Net Zero emissions before 2050.
Some examples are Canada (2050), Germany (2045), and Nepal (2045), some
countries have set longer deadlines, like India (2070) and China (2060). Some
countries have already achieved Net Zero emissions (Mauritius, Bhutan, etc.).
Some countries, including Japan, Canada, and the EU countries, have legally
118 Earth observation data analytics using machine and deep learning
the prevention of such CIDs. Monitoring wildfires has historically been a complex
challenge. But with the latest advancements in technology, monitoring wildfires
has become much easier.
The availability of data and resources like satellite imagery can be of immense
help in monitoring and predicting the occurrence and the spread of wildfires [10],
which in turn can be of great use to the authorities carrying out rescue operations,
thus making the rescue process much smoother and easier. Modern DL methods
can be immensely useful in analysing the data from satellite images and monitor-
ing/predicting the occurrence or spread of wildfires. Today, methods like DL and
ML can play a crucial role in improving the efficiency of monitoring climate
change and natural disasters. ML is a branch of -AI and computer science, that
involves the study of computer algorithms that can improve themselves through
experience, observations, and the use of data. DL is a part of a broader family of
ML methods that use multiple layers to progressively extract higher-level features
from raw input [11]. Applying DL and CNN techniques to satellite images can help
in automating the disaster identification process, thereby making the process faster
and much more efficient. Image segmentation can be a great way of monitoring
environmental changes and natural disasters. Image segmentation is a process that
involves the division of a digital image into various parts. These parts are called
‘Image segments’. This process helps in reducing the complexity of the image and
it also makes further processing and analysis of the image much more simplified.
Implementation of image segmentation techniques on VHR satellite images can
substantially help in monitoring climate change and natural disasters like wildfires,
landslides, volcanic eruptions, etc. CNNs can be used for the automatic detection of
wildfires using VHR satellite imagery after training the model with data. Image
segmentation can be performed over satellite images for automated detection of
wildfires based on training data. The model can be trained to divide the image into
segments of fire and non-fire, burnt area, and unburnt area. Data augmentation
techniques can also be used to enlarge the training data set [12]. SAR can also play
a critical role in wildfire detection and monitoring since it can penetrate through
clouds and smoke, and is capable of imaging day and night. A CNNs-based DL
framework can be used to identify the burnt area and differentiate it from the
unburnt area. Wildfires are only one of the many CIDs that occur in different parts
of the world. One such CID is landslide. Landslides occur when a mass of rock,
earth, or debris slides down a slope of the land. Landslides cause catastrophic
economic loss and claim many lives every year.
According to the World Health Organization (WHO), between the years 1998
and 2017, approximately 4.8 million people were affected by landslides. Landslides
claimed around 18,000 lives during that same year. It has become extremely
important to monitor landslides. Landslides can be triggered by wildfires [13]. The
burnt plots of land area are highly susceptible to debris flows. Thus, preventing
wildfires from occurring can hugely reduce the number of landslides occurring
every year. Satellite imagery has made it much easier to identify the areas affected
by wildfires, and it has also enabled efficient assessment of burnt areas after the
wildfire. Assessment of post-wildfire burnt areas is a very challenging task and
120 Earth observation data analytics using machine and deep learning
poses a lot of risks. Satellite imagery makes this much easier and risk-free. CNN
can be trained to classify the parts of the land into segments of burnt and unburnt
areas, thus making it far easier to predict the occurrence of landslides, and thereby
making rescue and evacuation operations much faster. Since manual monitoring of
landslides is very challenging and labour intensive which poses a lot of risk to the
personnel, such DL methods can be greatly useful in making the task of monitoring
much easier. Assessment of post-wildfire burnt areas is a very challenging task and
poses a lot of risks. Satellite imagery makes this much easier and risk-free and this
problem can also be dealt with through DL methods, by automating the task of
landslide classification through satellite imagery for more robust classification
results. CNN like U-Net can be very useful for classification purposes and landslide
mapping can be done using VHR satellite images. These DL techniques can auto-
mate the task of Landslide detection, hence reducing the time and human effort.
Volcano deformation is another phenomenon caused by climate change. Volcano
deformation is when the surface of a volcano and the land around a change shape.
This is a very important indicator of volcanic eruptions [14].
Volcanic eruptions are extremely disastrous and cause a lot of economic and
human resource loss. Monitoring volcano deformation is extremely important as it
enables researchers and scientists to understand the situation of the volcano and
estimate what is happening within it. This data can be very helpful in determining
the possibility of a volcanic eruption. This can be very helpful in the case of an
eruption as the people living around such an area can be informed in advance and
hence the rescue and evacuation operations can be carried out in a much smoother
and more efficient manner, hence, it will lead to many more lives being saved [15].
Modern DL methods can be very helpful in monitoring such volcanic activity with
the help of satellite imagery and technologies like InSAR [16]. The images gen-
erated by InSAR are known as interferograms and they contain a contribution from
both deformation of volcanoes and radar path through the atmosphere. DNNs like
AlexNet can be trained to monitor InSAR data and warn about the occurrence of
such deformations, so that action can be taken in time by the rescue authorities.
AlexNet can be trained using synthetic interferograms. The synthetic inter-
ferograms contain deformation patterns based on a multiple probability selection,
effects caused by stratified atmosphere derived from weather models, and other
atmospheric effects caused by irregularities in air motion. Thus, an efficient DL
model can identify volcanic activity and give warnings before volcanic unrest.
Often, volcanic eruptions may also lead to wildfires, and wildfires can also lead to
landslides. Hence, having monitoring systems in place can be crucial and can
hugely reduce the impact caused by these CIDs, saving countless lives [17].
Cyclones are yet another major cause of concern around the globe. Cyclones
cause a lot of economic damage and take many lives every year. In the last 50 years,
around 14,942 calamities and disasters have been attributed to tropical cyclones. An
estimated 779,324 people have lost their lives due to cyclone-related disasters. This
number averages out to be around 43 deaths every day. These cyclones have caused
1.4 trillion dollars in economic losses, which is an average of 78 million dollars in
losses every single day for those 50 years. Hence, monitoring cyclones and coastal
Wildfires, volcanoes and climate change monitoring 121
water bodies is crucial. CNN-based models can be trained to identify and predict the
possibility of a cyclone approaching the coast and can also predict its path and speed.
This can be crucial in evacuation and rescue missions. Any time saved in such
missions can save many more lives. Thus, the existence of such monitoring systems
can be of great benefit. It will make the task of evacuation authorities a lot easier and
will make the whole evacuation process a lot faster and smoother. If people living in
coastal areas are alerted about the possibility of a cyclone hitting the area, the eva-
cuation can start much before the actual calamity is about to occur, thus giving a lot
more time for people to move themselves as well as their belongings to safer places.
Monitoring cyclones has been attempted with UAV, but this method is unsustainable
and impractical for a variety of reasons. First, while the cost of UAVs themselves is
relatively low, their imaging elevation height is very low, and they need to get
dangerously close to the cyclone. This may lead to accidents and the loss of drones.
Second, since the drones need to be so close to the cyclone, they do not stay stabi-
lized, and the film comes out to be very shaky and at times unusable. Another issue
with drones is their flight time, satellites have a practically unlimited flying time,
compared to the minutes of flight time drones have. This makes them very imprac-
tical to use. And drones are unusable for 24/7 monitoring of water bodies. Also, they
have a very limited range, and hence, they are unviable for this purpose. Hence,
satellites prove to be the best medium for monitoring water bodies and cyclones [18].
CNN-based models can be very useful in extracting deep features from live
image feeds and can be used to predict the intensity of the cyclone and its path. A
once crucial way of tracking a cyclone is by detecting its ‘eye’. The eye of a
cyclone is the area of the cyclone where the wind flow is extremely slow and is
generally very calm. This ‘eye’ usually lies in the centre of the cyclone, and I am
usually 30–35 km in diameter. The eye is surrounded by the ‘eye wall’ of the
cyclone. This is the region where the severity of the cyclone is the highest and very
fast winds blow in this region. DL models can be trained with images to identify
these features in the images and estimate their severity, speed and path, based on
data from prior cyclones.
8.3.1 U-Net
U-Net is a new neural network and was first used in 2015 for biomedical image
segmentation. U-Net’s primary use case is image segmentation. U-Net takes in an
image as an input and it outputs a label. U-Net does its classification on every
pixel, and hence the size of the input and out in U-Net is the same in every case.
This property of U-Net is very useful in bio-medical image segmentation and
anomaly detection and it enables both detection and localization of the anomaly.
U-Net has many use cases in climate change monitoring, and especially the
availability of satellite imagery makes the case even stronger for U-Net applica-
tions. UNet is a u-shaped neural network, and it contains an encoder and a
decoder. These are the two main components of U-Net. This is a very useful
neural network as it can achieve great accuracy in results with very little training
data. U-Net finds its use in wildfire segmentation. A model can be trained to
identify burnt and un-burnt land plots and hence makes the task of authorities
carrying out rescue operations much easier.
8.3.2 AlexNet
AlexNet is a CNN model that solves a problem faced by many traditional neural
networks – they are difficult to be applied to high-resolution images. Since the
monitoring of climate change requires the usage of VHR satellite images, it makes
perfect sense to apply AlexNet in this case. AlexNet is an incredibly powerful CNN
model that can achieve high accuracies on very challenging datasets. The AlexNet
architecture consists of eight convolutional layers. Five of these layers are con-
volutional layers while three are fully connected layers. One of the many unique
AlexNet features that are not present in traditional CNNs is the ability to work with
multiple graphical processing units (GPUs). Half of the model’s neurons can be put
on 1 GPU and half on another. This allows much bigger models to be trained and
cuts down on the time required for training. AlexNet can be very useful in mon-
itoring volcanic deformation using VHR satellite imagery, primarily because
AlexNet’s application on high-resolution images is much easier and more con-
venient than traditional CNNs [21].
8.3.3 Inception-v3
Inception-v3 in a CNN model that is used for image segmentation. It has achieved
an accuracy rate of 78.1% on the ImageNet database. The Inception-v3 is also less
power/resource-consuming in comparison with the older versions. It has a total of
42 layers and the error rate is much lower than that of its earlier iterations/versions.
Inception-v3 is used in the classification of VHR satellite images using image
segmentation [22].
harmful gases hence this approach will also help keep many workers healthier and
make their jobs easier on them.
CNN-based models can be trained to keep a check on deforestation and illegal
cutting of trees. Deforestation is a major contributor to global warming and green
and climate change. Deforestation not only contributes in a major way to climate
change, but at the same time, it affects the animal life of the forest area and takes
away the natural habitat of many animals. This forces the animals to leave the
forests and travel into cities, causing much havoc and accidents. Several animals
die and species go extinct only because of deforestation. It is estimated that on earth
137 species of animals go extinct every year because of deforestation alone. DL
models can be trained to monitor forest caps and look for changes over a short and
long time. This can give us crucial information on illegal deforestation activities
and changes in forest caps in the long term. Action can be taken if the model shows
changes in forest cap over a small time and the respective authorities in charge can
be informed and action can be taken before the severity of the situation increases.
In the long term, DL models can give us a new perspective on the changes in
forest cap over time. This can also be later on compared to the no of deforestation
incidents that occurred during that time, again using DL models, and a relation can
be established. DL models can also be used in monitoring ice caps, glaciers, and
rivers. Models can be trained to observe long-term changes in ice caps and pre-
dictions can be done for the future. As glacier and ice cap melting is directly
correlated to rising sea levels, which is yet another major sign of trouble that is
about to come, the data regarding those is highly crucial. The melting of glaciers
can be a very harmful thing. Some of the biggest cities around the globe are coastal
cities that house millions of people and contribute significantly to global econo-
mies. Global trade relies on these coastal countries, and they are the ones which
will face the worst impact due to the melting of glaciers and ice caps. Especially
island cities that have a lot of coastal areas will be heavily impacted by this glacial
melting. For example, the Maldives, an Island nation and a popular vacation des-
tination for many travellers, located in the Indian Ocean, is predicted to lose 80% of
its habitable land to rising sea levels by 2050. This is an alarming situation.
DL models can be trained to observe glacial melting patterns and make pre-
dictions. They can be trained to look for concerning signs and in general the overall
changes in the ice caps. They can also be used to look at glacier movements in real-
time and alert if something unusual is observed. They can also be trained to keep
track of the ice cap and melting of the same. It can give us crucial data about the
melting of ice caps and we can then predict its impact and prepare for the same.
Another effect of climate change is the drying of rivers, caused due to global
warming. Many rivers dry out in different parts of the world, causing water
shortages. In some cases, rivers have over time changed their course due to climatic
changes. DL models can be implemented to keep a check on the drying of rivers
and the general long-term change in the pattern and the path rivers take. Satellite
imagery can also be used to look after the cleanliness of the rivers. Models can be
trained to identify unclean sections of the rivers, and the respective authorities can
be alerted to act before the situation becomes worse. In many cases, many factories
Wildfires, volcanoes and climate change monitoring 127
people remain left to be evacuated from flood sites. If we have already predicted
the occurrence of a flood well in advance, evacuation can start much early on and
hence everyone can be safely evacuated. This also solves the problem of chaos that
occurs at places where the evacuated people are taken. Such places can be prepared
well in advance and hence much of the mismanagement that happens due to lack of
time can be prevented.
With the proper implementation of DL models, such floods can be a lot less
damaging and much more lives can be saved. Another application of DL models
could be to monitor the desertification of land areas in the long-term using satellite
imagery. Now, this approach is slightly different because of one major reason –
desertification, unlike natural calamities, does not occur instantaneously, it occurs
over a while. This means that a 24/7 feed on live satellite data is not needed,
although it can certainly give a lot more data than it is needed to monitor deserti-
fication, this approach only requires a few snapshots every day. Previous research
titled ‘Status of Desertification in the Patagonian Region: Assessment and Mapping
from Satellite Imagery’, published by Argentinean researchers H.F. Del Valle, N.O.
Elissalde, D.A. Gagliardini, and J. Milovich, explores the usage of satellite imagery
in monitoring desertification of an area called ‘Patagonian Region’. They found
that of the total region studied (78.5 million ha), 93.6% (73.5 million ha) showed
different degrees of desertification. Categories of desertification for the whole
region were slight (9.3%), moderate (17.1%), moderate to severe (35.4%), severe
(23.3%), and very severe (8.5%). This paper was published in 1997 and the mea-
surements were done manually. And while no DL/ML algorithms were used here, it
just demonstrates the ability of satellite imagery and its potential just by itself.
If DL models are trained to identify areas that are going through desertification
and also train the models to determine the severity of desertification and give them
a rating, then this opens up a lot of possibilities. Areas with controllable severity
levels can be saved. Local authorities can be informed, and action can be taken.
Moreover, models can be trained to predict the speed of desertification and its
spread, and potential areas where desertification can start to happen. This can be
used to alert the stakeholders to take action. This data can also be very useful to
corporations and government projects, where teams look for land for different
projects. This can help projects where a very specific type of land is required.
Droughts have also become a significant challenge in many countries around the
globe, and droughts can lead to the desertification of lands. Observing the growth
and spread patterns of droughts using satellite images can also help us immensely.
Models can be trained with images of pre-drought land so that the possibility of the
occurrence of droughts can be predicted, and hence the required action can
be taken.
Droughts take countless lives every year around the globe in many different
countries. They cause a lack of water, any kind of water, be it drinking or utilitarian
water. This leads to countless deaths. Droughts primarily affect relatively poorer
people, and it is the poor section of society that faces the worst. Predicting droughts
can enable authorities to act accordingly to prevent such dire situations from
happening.
Wildfires, volcanoes and climate change monitoring 129
8.8 Conclusion
Climate change has become a majorly discussed issue around the globe. It has
become one of the main causes of deaths and economic losses worldwide. CIDs
have increased in frequency in the past few years and action needs to be taken to
reverse the issue of climate change. The first step in curbing climate change is to
effectively monitor it and take instant action at the time of a calamity. Monitoring
the impacts of climate change will be crucial in tackling it, and reducing the con-
sequences and losses caused due to disasters. Active and efficient monitoring of
disasters can help in early warning, which can help in a faster and more effective
response, saving countless lives and reducing losses.
Satellite imagery plays a crucial role in monitoring climate change and
observing its impacts around the world. Wildfires, landslides, volcanoes, etc. can
be identified and monitored with satellite imagery. Satellite imagery has histori-
cally been used to monitor the long-term effects of climate change, like changes in
the ice cap, forest cover, etc. Satellite images of volcanic eruptions, wildfires,
130 Earth observation data analytics using machine and deep learning
landslides, etc. help in mapping the data resulting in more efficient rescue opera-
tions. Modern DL methods can be extremely useful in assisting the monitoring of
climate change using satellite imagery.
The idea of applying DL methods to satellite images for monitoring climate
change and its impacts has been proven to be highly effective and efficient in both
identifying and predicting the spread of natural calamities and can be of great use to
the authorities responsible for rescue and evacuation missions. The DL methods
discussed here have the ability to not only identify the happening of the calamity
but also to localize and find out the exact region that is being affected. Over time,
some models can learn and begin to predict the probability of the occurrence of a
climate-induced disaster. The usage of U-Net, a DNN model initially introduced as
a bio-medical image segmentation model, in land segmentation for detecting pat-
ches of land burnt after a wildfire, shows the possibilities that DL provides.
The use of modern DNNs is much more efficient than manual monitoring done by
humans. It cuts down on labour costs, reduces the time taken, increases the efficiency
of disaster detection, and is even capable of predicting the occurrence of such events.
This ends up saving much more lives and reduces economic losses. It also helps the
rescue authorities in preparing for such an event in advance. Hence, this is an extre-
mely useful approach and if implemented this will save many lives and reduce eco-
nomic losses when calamities happen. The implementation of these DL methods and
CNN models requires far less capital investment and requires a lot less personnel. It
also keeps improving itself as time goes by and requires very little change.
This approach has also been proven to be extremely accurate and reliable and
is much faster than conventional methods. Hence, this approach makes a very
strong case for itself and can be extremely useful not just in climate change mon-
itoring, but at the same time, it can also in helping save countless lives at the time
of CIDs and natural calamities.
References
[1] Callendar, G.S. ‘The artificial production of carbon dioxide and its influence
on temperature’, Q. J. R. Meteorol. Soc., 1938, 64(275), pp. 223–240.
[2] ‘A brief history of climate change’, BBC News, 2013.
[3] ‘COP26: Together for our planet | United Nations’, https://www.un.org/en/
climatechange/cop26, accessed March 2022.
[4] ‘UN Climate Change Conference (COP26) at the SEC – Glasgow 2021’,
https://ukcop26.org/, accessed March 2022.
[5] ‘Evidence | Facts – Climate Change: Vital Signs of the Planet’, https://
climate.nasa.gov/evidence/, accessed March 2022.
[6] Forcada, J. and Trathan, P.N. ‘Penguin responses to climate change in the
Southern Ocean’, Glob. Change Biol., 2009, 15(7), pp. 1618–1630.
[7] Veytia, D. and Corney, S. ‘Climate change threatens Antarctic krill and the
sea life that depends on it’, http://theconversation.com/climate-change-
threatens-antarctic-krill-and-the-sea-life-that-depends-on-it-138436, accessed
March 2022.
Wildfires, volcanoes and climate change monitoring 131
[23] Iandola, F.N., Han, S., Moskewicz, M.W., Ashraf, K., Dally, W.J., and
Keutzer, K. ‘SqueezeNet: AlexNet-level accuracy with 50x fewer para-
meters and <0.5 MB model size’, 2016, ArXiv160207360 Cs.
[24] He, K., Zhang, X., Ren, S., and Sun, J. ‘Deep residual learning for image
recognition’, 2015, ArXiv151203385 Cs.
Chapter 9
A comparative study on torrential slide
shortcoming zones and causative factors using
machine learning techniques: a case study of an
Indian state
G. Bhargavi1 and J. Arunnehru1
9.1 Introduction
Landslide susceptibility is heavily influenced by topography. Failures can happen on
the coastline, in the highlands, in the Midlands, and near shore. The plain landscape has
low gravitational forces, which makes it the ideal factor for disasters [1]. Kerala is a
tropical climatic region in India’s southwest. Landslides are common in this state due
to the torrential rains and the persistent clay soil. It covers an area of 38,863 square
1
Department of Computer Science and Engineering, SRM Institute of Science and Technology, India
134 Earth observation data analytics using machine and deep learning
INDIA
Kerala
(a)
(b) (c)
Figure 9.1 (a) Map of India highlighting the state Kerala. (b) Highlights the
landslide zones and (c) highlights the 2021 annual rainfall data.
research. This sensing and monitoring can assess the danger of a disaster, and an
alarm system can be created as a result. Even though we can predict landslide
frequency, the Google Earth engine datasets will define the landslide vulnerability
zone [9]. As a result, monitoring and forecast zones will almost certainly be
established. The goal is to improve the centralised public database by employing
satellite photographs and accompanying data to aid disaster response by pinpoint-
ing specific regions – the ability to develop accurate landslide risk zonings for large
areas using verifiable approaches. Furthermore, as confirmed by a survey done with
the relevant research publications, machine learning approaches such as regression
models and support vector machines (SVM) were employed to develop accurate
landslide predicting models noting that the Western Ghats of Kerala are prone to
repeated landslides as a result of complex tectonic geomorphology and significant
precipitation penetration, resulting in increased pore pressure and landslides [10].
Deforestation has also made it easier for rainwater to enter directly. The preceding
investigations corroborate our hypothesis that landslides are caused by pore pres-
sure. Thus, avalanche peril planning is an initial step that will incorporate the
information we wanted to settle on choices about the affected region and distin-
guish inclinations for proper relief arrangements [11]. However, these are practi-
cally equivalent, they are unique concerning the mark of cycles and the ideas. Since
the current review has zeroed in additional on geomorphology in the avalanche-
inclined regions, the wording ‘inclination’ has been utilized. Figure 9.2 shows the
Figure 9.2 Viewpoint on Idukki torrential slide and adjoining areas. The dim square
shapes put aside in the figure show minor torrential slides occurred in a
comparable whirlwind event close by, and the yellow square shape on the
upper right of the figure exhibits the space of the breaks.
Landslide identification using machine learning – a case study 137
aerial view of a landslide that occurred in the year 2018 in the Idukki district of
Kerala. The landslide expanded up to 1,200 m from the peak.
effectively utilized. The data was assembled by the as of late dispatched Sentinel-
2 satellite [14]. The most prominent trait of the Sentinel-2 satellite is its 5-day
return to the Equator. In sans cloud circumstances, the spatial goal of a satellite
fluctuates up to 10 m for various recurrence groups or stations, for example, Band 2
– blue, Band 3 – green, Band 4 – red, and Band 8 – infrared. We can join at least
two groups for our execution because each band has exceptional properties [15].
The system that utilises many groups produces comparative discoveries that
demonstrate relative patterns. These outcomes were made by broadening the pic-
ture into mathematical geological representation utilizing the DEM. DEM was
made by the IRS satellite. DEMs are utilised to make geographic qualities like
slants at some random position, tendency, and look. A DEM is a three-dimensional
image of a landscape surface. Printing shape lines are changed into DEMs (active
form approaches in DIP). Polygons are made utilising specific computerized limits,
and every polygon follows the height information from the limit layout before
being utilised as a raster or vector information design [16]. DEMs are fundamen-
tally liable for GIS appropriateness. Changes to the DEM can be made to gauge
disintegration, precipice disappointment, and decide avalanche volume.
Precipitation and the timetable time frame for the locale repository are other
key setting-off factors for the avalanche. Precipitation, as per [17], is an occasional
impacting factor with a stage-like edge bend. Other affecting components, for
example, tremors and anthropogenic exercises produce an undulation shape in the
edge bend.
Avalanche identification is portrayed [18]. This distinguishing proof is fun-
damental and addresses a stage forward during the time spent fast danger evalua-
tion and relief. The avalanche weakness region is determined utilising numerous
open-source methods utilising avalanche planning. The ghastly properties of the
avalanche regions are utilised to group them. Picture division is utilised to distin-
guish the change discovery, trailed by the expulsion of non-helpless areas utilising
an item-based methodology. At last, utilising the solo strategy to bunch the picture
protests, the yield is recovered and characterized [19]. The splendour of the crea-
tion of the post-avalanche picture recognises the pre and post-avalanche satellite
photos. From the avalanche planning, the post-avalanche edge esteem shows the
avalanche-harmed region. Likewise, avalanche separating limitations, for example,
varieties in the DEM, slant, principal component analysis (PCA), and Green
Normalized Difference Vegetation Index (GNDVI) evaluations were utilized to
diminish invigorating sham expectations. Utilising the k-implies bunching calcu-
lation, the things that stay after counterfeit asset decrease are isolated into two
gatherings. By setting up the principal characteristics, the topographical attributes
included with a model can be approximated. Figure 9.3 portrays the rainfall range
that occurred in various districts in Kerala for researching avalanche subtleties. The
parkway street between Madurai and Munnar was enlarged because of an ava-
lanche that occurred in 2021. The avalanche stock is made utilising high-goal
satellite information from before and then afterwards the avalanche [20]. For ava-
lanche planning, an assortment of uninhibitedly accessible satellite pictures is
utilised.
Landslide identification using machine learning – a case study 139
Rainfall in mm
800
600
400
200
0
..
LL ...
I D U ...
.
A R
E
K ...
U .
PA LA ..
W ISS .
M D
R
K UZ.
Z H Y ..
M
..
..
SA I
TH D
K RA
A O.
H
K KK
U
A
LA PP
TH AN
U
TH VA
PA KA
A
A
ER NN
K
K TA
N
M IK
PP
A
T
A
N
LA
Y
R
IR
O
A
A
O
K
A
Sub Division/Districts
Figure 9.3 2021 Kerala rainfall data from Indian Metrological Department,
Thiruvananthapuram
in scantily vegetated regions. Utilising a SAR satellite might work on the current
technique for noticing incline movement at a geological split. For avalanche
observing or unexpected tendencies, earthbound or ground-based SAR inter-
ferometry framework strategies are an urgent instrument. The joined presentation
of the sensors and territory activities, while thinking about projected headways, can
fundamentally assist with a few pieces of information and observing limitations
related to the avalanche pointer.
9.10 Conclusion
This audit paper plans to understand the viewpoints that impact avalanches in India’s
meteorological conditions. Avalanches are a typical normal disaster, especially in the
Himalayan and Western Ghats areas. As per this review, water and stressed water after
precipitation from an adjoining state assume a basic part of Indian avalanche mishaps.
We discussed the course of avalanche occasions since planning avalanches is the initial
step. Then, at that point, it is separated into steps like estimation, checking, demon-
strating, and at last danger appraisal and alleviation, which is alluded to as the board. AI
calculations and remote detecting information can propel this work and give an authentic
contribution as a powerful occasion or catastrophe caution. We can likewise make an
early admonition model to forestall the annihilation. We can download an assortment of
time-series picture information to comprehend avalanche highlights in a specific region
better. Outrageous environmental changes will be knowledgeable about the future,
inciting the advancement of information put together innovation based on past infor-
mation utilising profound learning. The public authority can give assets for compelling
alleviation activities because of the phenomenal advancement in satellite photographs,
which permits them to detect the lowered zones and sort out salvage tasks all the more
rapidly. Besides, the exploration exertion will be useful to an overseer, engineers in
arranging, and specialised specialists in making financial exercises in such an area.
References
[1] P.P. Sun, M.S. Zhang, and L.F. Zhu, “Review of ‘workshop on landslide in
Southeast Asia: management of a prominent geohazard’ and its enlight-
enment,” Northwest. Geol., 2013.
Landslide identification using machine learning – a case study 145
[14] T.R. Martha, R. Priyom, K. Kirti, et al., “Landslides mapped using satellite
data in the Western Ghats of India after excess rainfall during August 2018,”
Curr. Sci., vol. 117(5), pp. 804–812.
[15] L. Ayalew and H. Yamagishi, “The application of GIS-based logistic
regression for landslide susceptibility mapping in the Kakuda-Yahiko
Mountains, Central Japan,” Geomorphology, vol. 65, pp. 15–31, 2004, doi:
10.1016/j.geomorph.2004.06.010.
[16] K. Khosravi, B.T. Pham, K. Chapi, et al., “A comparative assessment of
decision trees algorithms for flash flood susceptibility modeling at Haraz
Watershed, Northern Iran,” Sci. Total Environ., vol. 627, pp. 744–755, 2018,
doi: 10.1016/j.scitotenv.2018.01.266.
[17] Y. Song, R. Niu, S. Xu, et al., “Landslide susceptibility mapping based on
weighted gradient boosting decision tree in Wanzhou section of the Three
Gorges Reservoir Area (China),” mdpi.com, vol. 8, no. 4, p. 4, 2019, doi:
10.3390/ijgi8010004.
[18] B. Guru, R. Veerappan, and F. Sangma, “Comparison of probabilistic and
expert-based models in landslide susceptibility zonation mapping in part of
Nilgiri District, Tamil Nadu, India,” Spatial Inf. Res., vol. 25, pp. 757–768, 2017.
[19] K.M.R. Hunt and A. Menon, “The 2018 Kerala floods: a climate change
perspective,” Clim. Dyn., vol. 54, no. 3–4, pp. 2433–2446, 2020, doi:
10.1007/s00382-020-05123-7.
[20] E. Thennavan, G.P. Ganapathy, S.S. Chandra Sekaran, and A.S. Rajawat,
“Use of GIS in assessing building vulnerability for landslide hazard in The
Nilgiris, Western Ghats, India,” Nat. Hazards, vol. 82, no. 2, pp. 1031–1050,
2016, doi: 10.1007/S11069-016-2232-1.
[21] K.S. Sajinkumar, S. Anbazhagan, A.P. Pradeepkumar, and V.R. Rani, “Weathering
and landslide occurrences in parts of Western Ghats, Kerala,” J. Geol. Soc. India,
vol. 78, no. 3, pp. 249–257, 2011, doi: 10.1007/s12594-011-0089-1.
[22] V. Mishra and Harsh L. Shah., “Hydroclimatological perspective of the
Kerala flood of 2018,.” J. Geol. Soc. India, vol. 92, pp. 645–650, 2018.
[23] G. Bhargavi, and J. Arunnehru, “Identification of landslide vulnerability
zones and triggering factors using deep neural networks – an experimental
analysis,” . Advances in Computing and Data Sciences: 6th International
Conference, ICACDS, Kurnool, India, April 22–23, 2022, Revised Selected
Papers, Part I. Cham: Springer International Publishing, 2022.
[24] S. Rahamana, S. Arychmay, and R. Jegankumar, “Geospatial approach on
landslide hazard zonation mapping using multicriteria decision analysis: a
study on Coonoor and Ooty, part of Kallar watershed, The Nilgiris, Tamil
Nadu,” Int. Arch. Photogramm. Remote Sens. Spatial Inf. Sci., vol. XL-8,
pp. 1417–1422, 2014, doi: 10.5194/isprsarchives-XL-8-1417-2014.
[25] F. Guzzetti, P. Reichenbach, M. Cardinali, M. Galli, and F. Ardizzone,
“Probabilistic landslide hazard assessment at the basin scale,” Geomorphology,
vol. 72, no. 1–4, pp. 272–299, 2005, doi: 10.1016/j.geomorph.2005.06.002.
[26] L. Felipe, P. Sarmiento, M.G. Trujillo-Vela, and A.C. Santos, “Linear dis-
criminant analysis to describe the relationship between rainfall and
Landslide identification using machine learning – a case study 147
10.1 Introduction
Increasing demand of fossil fuels makes reservoir characterization (RC) [1] more and
more crucial and at the same time challenging. Reservoir here means petroleum
reservoir and characterizing it as the process of modeling the reservoir to understand
the distribution of different petrophysical properties to decide a location to drill a
production well that can enhance the production of an oil field. ML models [2–4] had
a great contribution in modeling reservoir with geophysical data including seismic
and well logs. However, the success rate varies depending on the characteristics of
1
Rajiv Gandhi Institute of Petroleum Technology, India
2
Indian Institute of Technology Guwahati, India
150 Earth observation data analytics using machine and deep learning
the reservoirs [5]. Seismic data is acquired through a seismic survey, which is
recorded with respect to time. Well logs represent some of the important reservoir
characteristics, acquired using well logging, which is recorded with respect to depth.
Petrophysical properties are the reservoir characteristics that can indicate the pre-
sence of petroleum or oil. The important petrophysical properties of reservoir include
porosity, permeability, fluid content, grain composition, and structure of the sub-
surface rocks. When seismic and well logs data of an oil field are properly processed
and correlated, a good understanding of the reservoir could be obtained and hence
can help in deciding the optimal position to place a production well in the field. One
of the very initial works of reservoir characterization with the data-driven integration
of seismic and well log was proposed by Schultz in 1994 [6]. After that, many works
have been reported in this field that integrated these data sources with different ML
[6–14] approaches for the prediction of porosity [15–21], permeability [15,18], water
saturation [19,21,22], and many more other characteristics [18,19,21]. Although
petrophysical modeling for reservoir characterisation has been widely explored till
date, it still remains a challenging problem [23] due to the varying characteristics
across reservoirs that include subsurface heterogeneity, spatial variability, and the
presence of different complex geological features like fractures and faults in reser-
voirs. Hence, geoscientists still rely on other sources and disciplines [5,24] of science
and engineering like Geology, Geophysics, Petrophysics, Geochemistry, Reservoir
Engineering, etc. for a more certain conclusion of a reservoir. Remarkable research
efforts on effectively applying of ML modeling in this field have been made in past
three decades [5,24]; however, ML literature [18] in RC is comparatively weak
compared to other fields of ML like vision and speech and it lacks uniformity of
applied models and lacks explicit consideration and analysis of data characteristics
that ultimately define the success of ML model in a particular field.
In this work, we considered a case study on the prediction of a petrophysical
property, porosity. Porosity [16] indicates the fraction of total void volume present
within the reservoir rock. It is an intrinsic characteristic of geological rock and one of
the primary factors to identify hydrocarbon prospects in a reservoir, hence, it is a
suitable petrophysical parameter to characterize a reservoir. The case study is per-
formed on a prospect survey area, in a motive of in-depth analysis of reservoir
characterization problem using ML. The ML approach followed is a framework of
preprocessing and modeling techniques that are based on different concepts of signal
processing, ML, and data analytics. Preprocessing techniques include well tie, seis-
mic signal reconstruction, smoothing of well log property, and outlier removal to
prepare the data well for the modeling. Different state-of-the-art regression techni-
ques are applied for modeling and a comparative analysis is performed on the con-
sidered case study. To improve the generalization of neural network modeling, deep
neural network with the regularization approaches is also investigated. At the end,
postprocessing is performed to effectively visualize the distribution of porosity var-
iation over the reservoir. Moreover, every phase of the framework have been ana-
lyzed in a data-driven manner to improve the modeling of the reservoir. Through this
case study, we also identified different prominent challenges and opportunities that
can open up future avenues of ML research in this field.
ML paradigm for predicting reservoir property 151
Seismic waves
Layer boundaries
of seismic wave generating sources like thumper. The produced acoustic signals
flow through the earth crust and gets reflected back by different rock boundaries at
different speeds. Geophones are placed on the earth surface that detects and records
the reflected seismic signals from the earth to illustrate the subsurface structure. As
different rocks transmit the signals differently, the measurement of the transmitted
signals can reveal different properties and location of each rock layer. With the
advancement of sensor technology, it is possible to sense and scan the earth crust
with apparently good resolution seismic data and hence it is possible to explore
even the reservoirs present deep beneath the earth crust.
characteristics. Some of the challenges that are posed while developing ML for
predicting reservoir characteristics from seismic and well logs are listed below:
● Reservoir characteristics that are recorded in the form of well logs include
porosity, permeability, water saturation, grain and sand fractions and share a
complex and nonlinear relationship with the seismic data. For modeling a
complex and nonlinear relation, ML models demand a huge set of data sam-
ples. In newly developed oil fields, collecting such data is costly and also time
consuming. This type of scenarios limits the ability of ML methods to effec-
tively model the reservoir.
● Due to the heterogeneous nature of earth subsurface, the characteristics of
reservoirs vary away from the well locations [34]. Predicting the petro-
physical properties of a well (test well) that is farther from the training well
locations can be difficult if the test well location is not in their neighbourhood.
The test well is called as blind well as the trained model has not seen any data
characteristic of it during its training. Hence, the blind well prediction can be
challenging if the blind well distribution largely varies from the trained wells
distribution.
Our work focuses on the estimation of petro-physical properties to effectively
characterize a given reservoir using seismic data and well logs. The objective of
this research is to address the afore-mentioned challenges by investigating and
developing different ML paradigm for the estimation of petro-physical properties.
We provided an initial analysis of basic ML models of the regression for the pre-
diction of petro-physical properties in the considered oil field to analyze how
effectively they can predict on our considered dataset, and on which scenarios we
can expect a good performance. To improve the performance of modeling, we
effectively pre-processed the dataset using signal processing approaches. We
compared different standard preprocessing approaches for preparing the seismic
and well logs for modeling. Preprocessing techniques are applied basically to
calibrate the data sources that are different in characteristics. Through the case
study of ML models on petrophysical property modeling, we investigated the
reservoir characterization problem from data science perspective and highlighted
potential research directions incorporating few challenges and opportunities.
12,150 12,150
Well 1
Well 5
11,675 11,675
Well 2 Well 4
11,200 11,200
2,800 2,975 3,150 3,325 3,500
Inlines
(a) (b)
Figure 10.2 Geological setting of survey area [35]. (a) Survey Area, (b) Location
of Krishna Godavari Basin.
250
Amplitude
Porosity
0.3
0 0.2
–250 0.1
2,000 2,100 2,200 2,300 2,400 2,500 2,600 2,700 1,600 1,800 2,000 2,200 2,400
Time (ms) Depth (m)
(a) (b)
Figure 10.3 Representative plots of seismic and well log; seismic: timewise
measurement, low sampling rate, contains low-frequency, low resolution,
large area coverage; well log: depthwise measurement, high sampling
rate, contains high-frequency, high resolution, small area coverage. (a)
Seismic Signal recording w.r.t. time, (b) Well log property (Porosity)
recording w.r.t. depth.
10.4.1 Geological background of the survey area
The base-map of the survey area is provided in Figure 10.2a, which is located in the
region of Krishna–Godavari basin oil field, having seven exploratory and production
wells. The Krishna–Godavari Basin is located in the South-Eastern part of India. The
basin lies across the Bay of Bengal and the blue rectangle in the figure is the
approximate location of the Basin, east of Hyderabad [35]. The sediment contents of
the basin are of thick sequences formed with several cycles of deposition. The basin
has several sub-basins located in on-land as well as in offshore and has the potential
reservoirs of oil and gas. To the southeast, the basin extends into the deep water of the
Bay of Bengal. The area is drained by two major rivers: Krishna and Godavari. A large
stratigraphic section of the basin from oldest Permo-Triassic Mandapeta Sandstone in
on-land to the youngest Pleistocene channel levee complexes in deep water offshore
has the great potential to host large volumes of hydrocarbon reserves [36].
Preprocessing Modelling
Well tie
Seismic reconstruction ML models for
Well log smoothing regression
Feature selection
Outlier removal
signal reconstruction, so that we need not compromise with the samples of the
wells. The reconstruction of the seismic signal is performed by cubic spline inter-
polation. Spline interpolation [37] is applied as the seismic signal has the char-
acteristics of local and abrupt changes. Using higher order polynomial to
interpolate the data may oscillate largely between the data points that it may
increase the reconstruction error of the data points between the interval. Whereas
cubic spline interpolation fit a series of unique cubic polynomial functions between
each data point to obtain a continuous and smooth curve that can be helpful for
approximating the actual graph of the seismic signal. Figure 10.5 provides the
graph of the seismic traces corresponding to respective well before and after
interpolation. The dots represent the original values, and the curves represent
interpolated seismic traces using cubic spline interpolation. We can observe that
cubic spline interpolation smoothly approximating the seismic signal, which very
closely satisfies the smoothness characteristics of seismic signal, along with pre-
serving the local and abrupt changes.
250
Seismic
0
–250
Original points Interpolated Curve
2,000 2,100 2,200 2,300 2,400 2,500 2,600 2,700
(a)
500
Seismic
Seismic
0
0
–100
2,000 2,200 2,400 2,600 2,800 2,000 2,200 2,400 2,600 2,800
(b) (c)
200 100
Seismic
Seismic
0 0
–100
2,200 2,300 2,400 2,500 2,600 2,700 2,800 2,200 2,300 2,400 2,500 2,600 2,700 2,800 2,900
(d) (e)
200
50
Seismic
Seismic
0 0
–50
–200
2,100 2,200 2,300 2,400 2,500 2,600 2,200 2,300 2,400 2,500 2,600 2,700 2,800
(f) Time
(g)
Figure 10.5 Interpolated seismic traces on the original sample points for Well 1
to Well 7 (from a to g)
ML paradigm for predicting reservoir property 157
actual actual
0.3 mean 0.3 median
Porosity
Porosity
0.2 0.2
0.1 0.1
2,000 2,200 2,400 2,600 2,000 2,200 2,400 2,600
Time Time
(a) (b)
actual actual
0.3 mmf 0.3 FT
Porosity
Porosity
0.2 0.2
0.1 0.1
2,000 2,200 2,400 2,600 2,000 2,200 2,400 2,600
Time Time
(c) (d)
Figure 10.6 Smoothing porosity log using different filtering techniques. (a) Mean
filtering, (b) median filtering, (c) mean median filtering, and (d)
fourier regularised.
filtered logs, Fourier regularization seems to follow better trend of the original trace
compared to the other filtering approaches, i.e., mean filter, median filter, and
mean–median filter. Fourier transform could more smoothly approximate, even the
abrupt peaks of the signal (as shown by the red circle), and from the domain of
geoscience, these changes have some inevitable information that needs to be con-
sidered. Fourier regularization is based on frequency selective filtering technique
performed by analyzing a signal in the frequency domain. With this technique, we
selectively remove the frequency components from the signal and hence it helped
to match the frequency contents of both seismic and well log signals. Also, removal
of high frequency components from a signal, sometimes help to remove high fre-
quency noise. When we analyzed the seismic and well log from all the available
wells, the seismic signal contains around lower 10% of the total frequency com-
ponents as in well log. After removing the rest high frequency components using
Fourier regularization, the regularized plot is shown in Figure 10.6(d), that is
smoothly approximating the variation of the original porosity log. However, the
Fourier regularization imposed some ill samples in the endpoints of the well log.
Hence, we remove those samples to maintain the quality of the data.
resolution as the well log sees. Hence, we considered few analytically derived
seismic attributes [7,43]: amplitude envelope (A1), amplitude-weighted cosine
phase (A2), amplitude-weighted frequency (A3), amplitude-weighted phase (A4),
derivative (A5), dominant frequency (A6), instantaneous frequency (A7), instan-
taneous phase (A8), integrate (A9), integrated absolute amplitude (A10), quad-
rature trace (A12), and seismic amplitude (A13), along with two more attributes
generated by seismic inversion [21], P-impedance (A11) and VpVs (A14). All the
attributes derived here not necessarily contain relevant information about the por-
osity prediction. To understand the importance of these derived attributes about the
porosity, we used different statistical methods [44,45] of feature importance that
includes Pearson correlation (P_Corr), F Test (F_Test), and Mutual Information
(M_I) and the comparison results are provided in Figure 10.7 with its normalised
values. P_Corr and F_Test could capture only linear dependency whereas M_I [46]
can capture any kind of dependencies including nonlinear too. It can be observed
from Figure 10.7 that all the methods marked that attributes A6, A11, and A14
captured the dependencies better than the other attributes including main seismic
signal (A13) about the porosity. Better the dependency, the better can be the por-
osity prediction using these attributes. These higher dependency attributes together
seems to carry high information compared to other attributes. Hence, computa-
tionally efficient model can be build using these selected important features.
Another important consideration while selecting the attributes is that the feature
analysis methods as listed above do not provide information about multi-
collinearity. Having multicolinear attributes can affect the performance of training
models. So, we must remove the multicolinearity of features well before training
the models for data. The colinearity of the attributes can be identified by calcu-
lating the correlation coefficients (CC) between the attributes. The CC for the
attributes comprising the input and target is provided in Figure 10.8. Larger the box
in the corresponding location higher the correlation. White box means positively
correlated and darker box means negatively correlated. It is observed that the
amplitude-weighted cosine phase (A2) and seismic (A13) is highly correlated with
each other. Consideration of these kinds of attributes together in prediction should
be avoided. Considering the above two aspects of feature importance and colli-
nearity, we considered A6 (dominant frequency), A11 (P-impedance), and A14
(VpVs) as the final selected features for the ML modeling.
1.0
P_Corr
f_Test
Value
0.5 mutual_info
0.0
A1
A2
A3
A4
A5
A6
A7
A8
A9
A10
A11
A12
A13
A14
Seismic attributes
1
A1
A11 A8
A8 A9
A9 A4
A4 2
A1
A12 A5
A5 0
A1
A10 A1
A1 A3
A3 A2
A2 3
A1
A13 A7
A7 PT
PT A6
A6 4
A1
A14
Figure 10.8 Correlation coefficients (CC) among the seismic attributes and target
The outliers detected are presented in Figure 10.9 with red samples. In
Figure 10.9(b), ED detected samples with the consideration of Euclidian distance,
160 Earth observation data analytics using machine and deep learning
4 Outlier 4 Outlier
2 2
PC2
PC2
0 0
–2 –2
–5 0 5 10 –5 0 5 10
PC1 PC1
(a) (b)
so it may consider some valid samples as outlier if distance measure is low and in
case of high distance measure, it can even leave the outlier samples as clean sam-
ples. Comparatively, MD provided better detection of outlier as it considers cov-
ariance to declare a outlier. Hence, we removed the detected outliers using MD to
keep our data clean.
3
40
2
VpVs
DF
20
1
0 0
5,000
0.3
P-Imp
4,000
PT
0.2
2,000
0.1
0 0
Wells Wells
6 4
5
3
4
Density
Density
3 2
2
1
1
0 0
0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0
Value (DF) Value (VpVs)
5
5
4
4
3
Density
Density
2 2
1 1
0 0
0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0
Value (P-Imp) Value (PT)
building. Generally, predictive models expect the training and test data to come
from similar distribution. On combining all the wells, the data distribution of the
attributes looks like in Figure 10.12. The attributes behave like following the nor-
mal distribution, with the lack of enough data samples in some regions. The dis-
tribution can be more complete with the increase number of wells and the samples.
Hence, considering the distribution of input attributes, we normalized the attributes
with z-score normalization, which corresponds with the normalization parameter
mean (m) and standard deviation (s). However, the target attribute, porosity, is
normalized within the range of 0–1, so that it should be suitable for the modeling
methods including neural network with sigmoid activation function, that demands
target in the range of 0–1. The considered field contains 20–40% porosity varia-
tions with very less samples going out of this bound. Hence the output variables
mostly lies between 0.2 and 0.4, with very less samples going out of this bound. So,
to make all the possible values to normalize between 0 and 1, we kept the nor-
malized parameters with minimum 0.1 and maximum 0.9 with keeping some offset
to handle the values of test data which comes out of range of the minimum and
maximum values of the train data.
ML paradigm for predicting reservoir property 163
DF VpVs P-Imp PT
3.0
2.5
2.0
Density
1.5
1.0
0.5
0
0.0 0.2 0.4 0.6 0.8 1.0
Value of features
Figure 10.12 Kernel density estimation for attributes with all wells combined
1 1
0.75 0.75
0.5
PT
0.5
PT
0.25 0.25
0 0
–120 –60 0 60 120 –120 –15 –12 –9
X_embed X_embed
(a) (b)
Figure 10.13 Porosity (PT) plot in reduced dimension Xembed . (a) All samples
considered. (b) First 1,000 samples.
We have also analyze the patterns of the data to prepare modeling hypothesis.
With a motive to visualize the behavior of the data we employed t-SNE
(T-distributed stochastic neighbor embedding) [52] plot as provided in
Figure 10.13. It is the representation of target property in reduced dimension denoted
as xembed. Two points closer on the X-axis indicates that the two samples are closer
to each other in the feature hyperspace. Figure 10.13(b) is of first 1,000 samples of
Figure 10.13a, which is the t-SNE plot of all the samples considered. This is done to
zoom in the samples to clearly understand how the patterns vary according to its
neighboring samples. it is observed that changing of porosity is of smooth changing
curve but with very complex pattern. This suggests that there exists a complex
nonlinear relationship between input attributes and the target property. Nonlinear ML
model capable to handle nonlinear complexity seems to be useful with this data.
subsequently. The training dataset for ML models is obtained with 80% of the total
samples from the combined data from all the wells. It was generated by shuffling
the data first, to remove any possible trend along the depth, rest 20% is considered
for testing. The most cumbersome task with machine learning models is to deal
with fiddling with the models, tuning their parameters. So, from the training data,
around 20% data are taken out for validation set to tune the hyperparameters. The
generalization capability of the trained model is evaluated during testing phase. The
model performances are evaluated in terms of root mean square error (RMSE),
coefficient of determination (R-square), and correlation coefficient (CC) measures.
Several runs of training, testing, and validation phases have been carried out in
order to decide the best set of hyperparameters and parameters for every ML
model. The input attributes for ML models constitutes the selected features as
mentioned in Section 10.5.4, and the main seismic signal. Different ML regression
models [18,26,53–57] are explored on our dataset. We have considered artificial
neural network (ANN), k-nearest neighbors regressor (KNN), support vector
regression (SVR), decision tree (DT), and linear models from Scikit Learn Library
[58]. Even though it is clear from the statistical analysis of Section 10.6.1 that the
underlying relationship among data is nonlinear, we have also explored linear
models as in some literature linear models have also been effectively used for the
prediction of petrophysical properties [18], and linear model can provide the
information about the percentage of linear information the data contains. Results of
the linear methods that include stochastic gradient descent (SGD) regressor, Ridge,
Lasso, and Elastic Net are provided in Table 10.2 with their comparisons.
All the linear models provided almost comparable performance. It is apparent
from the result that the data is having very less fraction of linear dependency
between input and output (around 22%), as visible from the R-square score.
However, R-square score in nonlinear model is not a good practice as it sometimes
draw false conclusions about nonlinear model performance [59]. Hence, we discard
this metric for nonlinear models evaluations. Elastic Net [60], which has the cap-
ability to outperform other linear models in large dataset (as it combines the
advantages of both ridge and lasso regression that take care of both L1 and L2
regularization, respectively), is considered to present the prediction curve as pro-
vided in Figure 10.14a. The non-linearity in the data has not been captured by the
model as can be seen in predicted vs. actual plot. The patterns that could not be
Residual plot
1.00 0.50
0.75 0.25
Residuals
PT
0.50 0.00
0.25 Actual –0.25
Predicted
0.00 –0.50
0 10,000 20,000 30,000 0.4 0.6 0.8
Samples Measured values
(a) (b)
Figure 10.14 Analysis plots of linear model (ElasticNet). (a) Prediction plot and
(b) Residual plot.
understood by the linear models are provided in Figure 10.14(b), which is the error
plot corresponding to the fitted values. It can be observed that the residual error
variance is not constant over the measured values. This indicates
Heteroskedasticity [61] issue that indicates the presence of nonlinearity in the data
which has not been captured by the model. Hence, nonlinearity modeling is
essential with this kind of data as the linear models could not explore the input–
output relation properly. Experiments are performed with the popular nonlinear
regression models, and the results are presented in Table 10.3. The k-nearest
neighbors regression surprisingly outperformed other algorithms, SVR, DT, and
ANN models. The model achieved best result with RMSE of 0.0035 and CC of 0.99
on the test data, even in very little time (of 1.2 s) compared to other models. Two
variations of KNN based on the weighting of its neighbor samples are applied, one
with uniformly weighting the considered k-neighbours and the other with weight-
ing the neighbours based on its distance from the considered data point. Distance-
based weighting of neighbors is comparatively better approach with our dataset as
can be observed from Table 10.3.
When compared both the approaches with the variation of its hyperparameter,
i.e., number of neighbours, it is observed that the distance-based weighing provided
better results in every hyper parameter configurations as can be observed in
Figure 10.15. This observation indicates that the prediction can get better if the test
samples are from those regions of the sample space, more numbers of nearby
166 Earth observation data analytics using machine and deep learning
Error (RMSE)
0.010 Distance
Uniform
0.005
1 2 3 4 5 6 7 8 9
Number of neighbors
ML models RMSE CC
Linear 0.055 0.42
ANN 0.05 0.50
KNN 0.052 0.34
SVR 0.056 0.004
DT 0.064 0.152
0.10 0.75
CC
0.50
0.05 0.25
0.00 0.00
Linear ANN KNN SVR DT Linear ANN KNN SVR DT
training patterns are available for the model. This motivates us to investigate for the
prediction results for blind location/well, where the data is expected to comes from
a location further from the wells considered during training. Due to geological
characteristics heterogeneity, it is assumed that the characteristics of data can vary
from one location to another, hence the prediction can be difficult in this scenario.
Separating Well 1 samples from all the other wells, we define Well 1 as blind
well as its samples are not considered in training dataset. The result of blind well
prediction is provided in Table 10.4, and its comparison to prediction on shuffled
data (as provided in Table 10.3) is provided in Figure 10.16.
It clearly indicates that KNN that performed best could not perform that good
in blind well prediction as its performance is completely based on the availability
of nearby samples in its samples space, which seems is not properly present here
due to the consideration of test well from different location having different
ML paradigm for predicting reservoir property 167
Normalised input
Input layer
1 2 3 4
Hidden layer 1
0.08
Batch normalisation 0.06
RMSE
Dropout 0.04
0
5 10 20 30 40 50 100
Output layer
Neurons
Output
(a) (b)
Figure 10.17 Realization of DNN for blind well prediction: (a) DNN architecture
and (b) results on varying layers
168 Earth observation data analytics using machine and deep learning
The ultimate goal of the modeling is to generate the spatial distribution of the
petrophysical property to visually interpret the underlying reservoir. In order to
generate the spatial distribution of porosity, the best model obtained above (DNN)
is considered and the porosity values are predicted for each spatial point corre-
sponding to available seismic traces. The 2D visualization of generated porosity for
Inline no. XX78 is provided in Figure 10.18. However, when we visualize seismic
data as shown in Figure 10.19, we see that seismic attribute varies smoothly across
the subsurface. Hence, we understand that the spatial variation of petrophysical
property must be smooth, it cannot vary so abruptly as in Figure 10.18. Similarly
the volumetric visualization of porosity along with other petrophysical properties
can be made to detect the potential zones of hydrocarbon presence. Hence, we
applied this rationale with a smoothing technique, median filtering, and the result
of filtering as shown in Figure 10.20 that enhanced the continuity of the generated
subsurface.
At Inline: XX78
1,000 0.18
Time (ms)
1,100 0.16
1,200 0.14
0.12
1,300 0.10
1,400 0.08
13,000 12,800 12,600 12,400 12,200 12,000 11,800 11,600 11,400 11,200
Cross Line
At Inline: XX78
1,000 400
Time (ms)
1,100 200
1,200 0
1,300 –200
1,400 –400
13,000 12,800 12,600 12,400 12,200 12,000 11,800 11,600 11,400 11,200
Cross Line
At Inline: XX78
1,000 0.18
0.16
Time (ms)
1,100
0.14
1,200
0.12
1,300 0.10
1,400 0.08
13,000 12,800 12,600 12,400 12,200 12,000 11,800 11,600 11,400 11,200
Cross Line
10.8 Conclusion
Acknowledgment
This study was supported by Geodata Processing & Interpretation Centre
(GEOPIC), Oil and Natural Gas Corporation (ONGC) Limited, Dehradun, India (M
984563). The authors are grateful to Mr Sanjai Kumar Singh and Mr P. K.
Chaudhury of GEOPIC, ONGC for their valuable suggestions.
References
[1] Lines LR and Newrick RT. Fundamentals of Geophysical Interpretation.
Society of Exploration Geophysicists; 2004.
[2] Mishra S and Datta-Gupta A. Applied Statistical Modeling and Data
Analytics: A Practical Guide for the Petroleum Geosciences. Elsevier; 2017.
[3] Bhattacharya S and Mishra S. Applications of machine learning for facies
and fracture prediction using Bayesian Network Theory and Random Forest:
Case studies from the Appalachian basin, USA. Journal of Petroleum
Science and Engineering. 2018;170:1005–1017.
[4] Sebtosheikh MA and Salehi A. Lithology prediction by support vector
classifiers using inverted seismic attributes data and petrophysical logs as a
new approach and investigation of training data set size effect on its per-
formance in a heterogeneous carbonate reservoir. Journal of Petroleum
Science and Engineering. 2015;134:143–149.
ML paradigm for predicting reservoir property 171
[52] van der Maaten L and Hinton G. Visualizing data using t-SNE. Journal of
Machine Learning Research. 2008;9:2579–2605.
[53] Abhishek K, Singh M, Ghosh S, et al. Weather forecasting model using
artificial neural network. Procedia Technology. 2012;4:311–318.
[54] Anifowose FA, Labadin J, and Abdulraheem A. Ensemble model of non-
linear feature selection-based Extreme Learning Machine for improved
natural gas reservoir characterization. Journal of Natural Gas Science and
Engineering. 2015;26:1561–1572.
[55] Ahmadi MA and Chen Z. Analysis of gas production data via an intelligent
model: application to natural gas production. First Break. 2018;36(12):91–98.
[56] Wong KW, Fung CC, Ong YS, et al. Reservoir characterization using sup-
port vector machines. In: International Conference on Computational
Intelligence for Modelling, Control and Automation and International
Conference on Intelligent Agents, Web Technologies and Internet
Commerce. 2005, vol. 2. pp. 354–359.
[57] Nikravesh M and Aminzadeh F. Past, present and future intelligent reservoir
characterization trends. Journal of Petroleum Science and Engineering.
2001;31(2–4):67–79.
[58] Pedregosa F, Varoquaux G, Gramfort A, et al. Scikit-learn: Machine learn-
ing in Python. Journal of Machine Learning Research. 2011;12(Oct):2825–
2830.
[59] Spiess AN and Neumeyer N. An evaluation of R2 as an inadequate measure
for nonlinear models in pharmacological and biochemical research: a Monte
Carlo approach. BMC Pharmacology. 2010;10(1):6.
[60] Zou H and Hastie T. Regularization and variable selection via the elastic net.
Journal of the Royal statistical Society: Series B (Statistical Methodology).
2005;67(2):301–320.
[61] Long JS and Trivedi PK. Some specification tests for the linear regression
model. Sociological Methods & Research. 1992;21(2):161–204.
[62] Bengio Y. Learning deep architectures AI. Foundations and Trends in
Machine Learning. 2009;2(1):1–127.
[63] Mhaskar H, Liao Q, and Poggio T. When and why are deep networks better
than shallow ones? In: Thirty-First AAAI Conference on Artificial
Intelligence. 2017. pp. 2343–2349.
[64] Liang S and Srikant R. Why deep neural networks for function approxima-
tion? In: ICLR; 2017.
[65] Srivastava N, Hinton G, Krizhevsky A, et al. Dropout: a simple way to
prevent neural networks from overfitting. The Journal of Machine Learning
Research. 2014;15(1):1929–1958.
[66] Ioffe S and Szegedy C. Batch normalization: accelerating deep network
training by reducing internal covariate shift. In: Proceedings of the 32nd
International Conference on Machine Learning. 2015, vol. 37. pp. 448–456.
[67] He H and Garcia EA. Learning from imbalanced data. IEEE Transactions on
Knowledge & Data Engineering. 2008;21(9):1263–1284.
Part III
Tools and technologies for Earth Observation
data
This page intentionally left blank
Chapter 11
The application of R software in water science
Nasrin Fathollahzadeh Attar1 and
Mohammad Taghi Sattari2
Nowadays, dealing with data in all sciences is very critical. Data science and data
engineering are exciting disciplines to turn data into bright understanding. Finding
statistical characteristics and communicating data by some techniques such as visua-
lization. R is a rapidly growing, statistical, open-source software with many libraries
and packages. A core team of R developers and the R Foundation for Statistical
Computing support it. Recently, many scholars in different fields have used R daily
and produced publication-quality graphics. Hydrologists use this software to tidy,
transform, visualize, and model their hydrometric data. There are many packages
available in hydrology science. Hydroinformatics is a branch of informatics that deals
with water purposes. They are starting from downloading hydrological data in special
packages, cleaning up hydrological and climate data, managing data such as aggre-
gating, dealing with missing data, extracting indicators, analyzing extreme events such
as floods, droughts, bushfires, dealing with data scales, spatial and temporal dataset
tools, dealing with surface and groundwater, hydrographs, rainfall, snowfall water
quality, reservoirs packages, watershed modelling, soil water systems, evaporation, and
multiple water-related packages. This chapter aims to analyze these packages by
explaining the resources and their use in water and hydrology science, finding the gaps
in existing packages, and suggesting researchers develop new packages.
11.1 Introduction
Earth and environmental science, derived data technologies provide comprehensive
information about the earth system. Research on environmental science has a long
tradition because of continuously changing phenomena. Recent theoretical devel-
opments have revealed that data-driven methods such as machine learning (ML)
and deep learning can be applied to deal with extensive environmental observa-
tions. The latter results in efficient and accurate modelling boosted by statistical,
advanced cloud computing techniques. There are growing appeals for machine and
1
Department of Statistics, University of Padova, Padua, Italy
2
Department of Water Engineering, University of Tabriz and Ankara University, Iran
178 Earth observation data analytics using machine and deep learning
rich set of graphs with more creative potential than Microsoft Excel), reproduci-
bility (works with big data, building reports), advanced modelling (advanced
growing statistical models), automation (model computations, cloud environ-
ments), generating reports, dashboards, and web applications. R software can be
downloaded from the CRAN website [12]. One of the essential things about R,
because of its open access and a significant number of users in all fields, is that
there are lots of websites and information on the web available about R. But
sometimes, it is hard to find them using Google search. Therefore, a search engine
named ‘Rseek’ [13] makes it easy to search for R. Sasha Goodman and some
volunteers have maintained this engine since 2007. Another similar search engine
searches R in multi-sites [13].
Some integrated development environments (IDEs) software facilitates using
R. IDEs have different windows simultaneously, which helps the users code. They
consist of a code script window, code compiler or console, debugging options, data
environment, and a single graphical user interface (GUI) for visualization of plots,
help window, package installation, updating the installed packages, and documents
and vignettes. The most popular IDE for R is RStudio, founded by Joseph J. Allaire
in 2009, and in 2011 RStudio IDE for R was launched. The RStudio is extremely
user-friendly by providing a helpful environment for all researchers. RStudio is
available in an open-source edition and can be downloaded freely for use by
desktops for all platforms such as Windows, Linux, or MacOS. Utilizing RStudio
can help us conduct various tasks such as creating different code scripts, version
control, creating presentations in Rmarkdown in different formats (HTML, Latex,
PowerPoint), creating fruitful graphs, and creating web applications in shiny. More
information about RStudio can be found at [14].
events, and conferences for R-related topics by providing ample and rich blogs,
forums, and websites. R-bloggers is a good blog updated daily and has many
authors [20]—on Twitter, using the hashtag #rstats and R tip of the day [21] to get a
handful and widely used R-related topics. The Revolution Analytics blog has
contributed to R development topics [22]. Another online community that helps
beginners to experts in question and answering platforms is StackOverflow [23]; to
comment, one should be signed into the website. Another similar site to
StackOverflow is a cross-validated site that helps R users with statistics, statistical
models, and data mining techniques [24]. RStudio also has a huge community of
R-users and developers [25]. Their community also has some mailing lists for
getting help using R [26]. These mailing lists have been active since the 1990s.
Weekly also contains learning R materials, blog posts, podcasts, and the newest
updates [27]. R meet-up groups are also overgrowing [28], with 38 countries,
92 groups, and approximately 72,101 members until the 28th of February 2022.
Another global community that promotes gender diversity is the R users commu-
nity, which has many active branches from all over the world, with 140 groups and
79,251 members to the date of this study (2022.02.28) [29]. R girls school is also a
network that promotes the R for secondary schools (11–16 years old) and is cur-
rently developing lesson plans for use in the classrooms [30]. Lastly, R commu-
nities, developers, users, distributors, maintainers, and R-related conferences are
supported by R Consortium [31].
JSM 2022 (August 6–11), one of the largest statistical events in the world, will
likely have several R-related talks in its program.
R Medicine (August 23–26) promotes using R-based tools to improve clinical
research and practice.
EARL Conference (September 6–8) is a cross-sector conference focusing on
the commercial use of the R programming language.
R Government (December, dates TBA)
Past events
Evidence synthesis and meta-analysis in R—The talks from this workshop
series aim to develop and promote open software tools for evidence synthesis.
5 – Model building
6 – Reporting the
results
Hydrological
models
Table 11.3 List of packages for physically based and deterministic models
References
[1] Hydroshare. Available from: https://www.hydroshare.org/
[2] Hydroclient. Available from: https://data.cuahsi.org/
The application of R software in water science 193
[3] Essawy BT, Goodall JL, Zell W, et al. Integrating scientific cyberinfras-
tructures to improve reproducibility in computational hydrology: example
for HydroShare and GeoTrust. Environ Model Softw. 2018;105:217–229.
Available from: https://doi.org/10.1016/j.envsoft.2018.03.025
[4] hydro server. Available from: https://hydroserver.cuahsi.org/
[5] Tarboton DG and Idaszak R. HydroShare: advancing hydrology through
collaborative data and model sharing. iRODS User Gr Meet. 2015. Available
from: https://irods.org/documentation/articles/irods-user-group-meeting-2015/,
https://irods.org/wp-content/uploads/2015/06/Tarboton-HydroShare.pdf.
[6] Abbott MB. Introducing hydroinformatics. J Hydroinformatics. 1999;1
(1):3–19.
[7] Mosaffa H, Sadeghi M, Mallakpour I, Naghdyzadegan Jahromi M, and
Pourghasemi HR. Application of machine learning algorithms in hydrology.
In: Pourghasemi HR, (eds.) Computers in Earth and Environmental
Sciences. Elsevier. 2022. pp. 585–591 (Chapter 43). Available from: https://
www.sciencedirect.com/science/article/pii/B9780323898614000270
[8] Rossum G. Python programming language. In: USENIX Annual Technical
Conference. 2007. p. 36.
[9] AGU. Available from: https://agu-h3s.org/2021/03/29/resources-for-pro-
gramming-in-hydrology/
[10] Python-Hydrology-Tools. Available from: https://github.com/raoulcollen-
teur/Python-Hydrology-Tools
[11] Dabbish L, Stuart C, Tsay J, and Herbsleb J. Social coding in GitHub:
transparency and collaboration in an open software repository. In: CSCW
’12: Proceedings of the ACM 2012 Conference on Computer Supported
Cooperative Work; 2012; pp. 1277–1286.
[12] CRAN. Available from: https://cran.r-project.org/
[13] Search Engine. Available from: http://www.dangoldstein.com/search_r.html
[14] Rstudio. Available from: https://www.rstudio.com/
[15] RPackage site. Available from: https://www.r-pkg.org/
[16] ropensci. Available from: https://ropensci.org/
[17] r packages. Available from: http://www.datasciencemeta.com/rpackages
[18] Package downloads. Available from: https://cran.r-project.org/web/packa-
ges/dlstats/vignettes/dlstats.html
[19] Cheat sheet. Available from: https://www.rstudio.com/resources/
cheatsheets/
[20] r-bloggers. Available from: http://www.r-bloggers.com/
[21] R tip of day. Available from: https://twitter.com/rlangtip
[22] R blog. Available from: https://blog.revolutionanalytics.com/
[23] Stack over flow. Available from: https://stackoverflow.com/
[24] stackexchange. Available from: https://stats.stackexchange.com/questions/
tagged/r
[25] R studio community. Available from: https://community.rstudio.com/
[26] stat.ethz. Available from: https://stat.ethz.ch/mailman/listinfo/r-help
[27] R weekly. Available from: https://rweekly.org/
194 Earth observation data analytics using machine and deep learning
[48] Shears BS and Nick TS. New Zealand’s Climate Data in R – An Introduction
to clifro. 2015. Available from: http://stattech.wordpress.fos.auckland.ac.nz/
2015-02-new-zealands-climate-data-in-r-an-introduction-to-clifro/
[49] Sommerlot A, Fuka D, and Easton Z. getMet: Get Meteorological Data for
Hydrologic Models. 2016. Available from: https://cran.r-project.org/pack-
age=getMet
[50] Sparks AH, Hengl T, and Nelson A. GSODR: global summary daily weather
data in R. J Open Source Softw. 2017;2(10):177.
[51] Tuck SL, Phillips HR, Hintzen RE, Scharlemann JP, Purvis A, and Hudson
LN. MODISTools – Downloading and Processing MODIS Remotely
Sensed Data in R. Ecol Evol. 2014;4(24):4658–4668.
[52] Sparks AH. nasapower: a NASA POWER global meteorology, surface solar
energy and climatology data client for R. J Open Source Softw. 2018;3(30):1035.
[53] Campitelli E. metR: Tools for Easier Analysis of Meteorological Fields.
2021.
[54] Hart EM, Bell K, and Butler A. prism: Download Data from the Oregon
Prism Project. 2015. R package version 0.0, 6(10.5281).
[55] Boessenkool B. rdwd: Select and Download Climate Data from “DWD”
(German Weather Service). 2021. Available from: https://cran.r-project.org/
package=rdwd
[56] Kemp MU, Loon EEV, Baranes JS, and Bouten W. RNCEP: Obtain,
Organize, and Visualize NCEP Weather Data, 2020. Available from https://
cran.r-project.org/web/packages/RNCEP/index.html.
[57] Chamberlain S. rnoaa: “NOAA” Weather Data from R. 2021. Available
from: https://cran.r-project.org/package=rnoaa
[58] Thorley J, Mantua N, and Hare SR. rpdo: Pacific Decadal Oscillation Index
Data. 2020. Available from: https://cran.r-project.org/package=rpdo
[59] Shum A. rwunderground: R Interface to Weather Underground API. 2018.
Available from: https://cran.r-project.org/package=rwunderground
[60] Joseph M, Oakley M, and Schira Z. smapr: Acquisition and Processing of
NASA Soil Moisture Active-Passive (SMAP) Data. 2019. Available from:
https://cran.r-project.org/package=smapr
[61] Iannone R. stationaRy: Detailed Meteorological Data from Stations All Over
the World. 2020. Available from: https://cran.r-project.org/package=stationaRy
[62] Carslaw D. worldmet: Import Surface Meteorological Data from NOAA
Integrated Surface Database (ISD). 2021. Available from: https://cran.r-
project.org/package=worldmet
[63] Lisic LC and Chen L. Tools to Download and Work with USDA Cropscape
Data. 2018. Available from: https://cran.r-project.org/package=cdlTools
[64] Michael CJK, Markus G, and Filippo G. FAOSTAT: Download Data from
the FAOSTAT Database. 2022. Available from: https://cran.r-project.org/
package=FAOSTAT
[65] Van Den Broeck J, Cunningham SA, Eeckels R, and Herbst K. Data
cleaning: detecting, diagnosing, and editing data abnormalities. PLoS
Medicine, 2005;2(10):966–971.
196 Earth observation data analytics using machine and deep learning
[118] Tsay RS, Wood D, and Lachmann J. MTS: All-Purpose Toolkit for
Analyzing Multivariate Time Series (MTS) and Estimating Multivariate
Volatility Models. 2021. Available from: https://cran.r-project.org/
package=MTS
[119] Laurinec P. TSrepr R package: Time series representations. J Open Source
Softw, 2018:3(23):577. Available from: https://doi.org/10.21105/joss.00577
[120] Aldrich E. wavelets: Functions for Computing Wavelet Filters, Wavelet
Transforms and Multiresolution Analyses. 2020. Available from: https://
cran.r-project.org/package=wavelets
[121] Machine Learning Cran. Available from: https://cran.r-project.org/web/
views/MachineLearning.html
[122] Venables WN and Ripley BD. Modern Applied Statistics with S. 2002.
Available from: https://www.stats.ox.ac.uk/pub/MASS4/
[123] Rong X. deepnet: Deep Learning Toolkit in R. 2014. Available from:
https://cran.r-project.org/package=deepnet
[124] Therneau T, Atkinson B, and Ripley B. rpart: Recursive Partitioning and
Regression Trees. 2019. Available from: https://cran.r-project.org/
package=rpart
[125] Witten IH, Frank E, Hall MA, and Pal CJ. Practical machine learning tools
and techniques. In Data Mining 2005, vol. 2, no. 4. Elsevier.
[126] Liaw A and Wiener M. Classification and regression by randomforest. R.
News, 2002;2:18–22. Available from: https://cran.r-project.org/doc/Rnews/
[127] Chen T, Tong He T, Benesty M, et al. xgboost: Extreme Gradient Boosting.
2022. Available from: https://cran.r-project.org/package=xgboost
[128] Sekhon JS and Mebane WR. Genetic optimization using derivatives. Polit
Anal. 1998;7:187–210.
[129] Riza LS, Bergmeir C, Herrera F, and Benı́tez JM. frbs: Fuzzy rule-based
systems for classification and regression in R. J Stat Softw, 2015;65(6),
1–30. Available from: http://www.jstatsoft.org/v65/i06/
[130] Campbell JB and Wynne RH. Introduction to Remote Sensing (5th ed.).
New York, NY: Guilford Press, 2011, pp. 117–118.
[131] Strong JA and Elliott M. The value of remote sensing techniques in sup-
porting effective extrapolation across multiple marine spatial scales. Mar
Pollut Bull. 2017;116(1–2):405–419. Available from: http://dx.doi.org/10.
1016/j.marpolbul.2017.01.028
[132] Schwalb-willmann J, Benjamin A, Horning N, Hijmans RJ, and Leutner
MB. R: RStoolbox: Tools for Remote Sensing Data Analysis. 2018.
Available from: https://cran.r-project.org/package=RStoolbox.
[133] Goslee SC. Analyzing remote sensing data in R: the landsat package. J Stat
Softw, 2011;43(4):1–25. Available from: https://doi.org/10.18637/jss.v043.i04
[134] Lehnert LW, Meyer H, Obermeier WA, et al. Hyperspectral data analysis
in R: the hsdar package. J Stat Softw. 2019;89:1–23.
[135] Lamigueiro OP and Hijmans R. rasterVis: Visualization Methods for Raster
Data”, 2023. Available from: https://oscarperpinan.github.io/rastervis/
This page intentionally left blank
Chapter 12
Geospatial big data analysis using neural
networks
Hemi Patel1, Jai Prakash Verma1, Sapan Mankad1,
Sanjay Garg2, Pankaj Bodani3 and Ghansham Sangar3
Geospatial information system (GIS) produces large and complex data. Geo-spatial
big data analysis is a critical field nowadays because a large amount of data is
generated every day by various space mission programs running through space
agencies all over the globe. It requires robust data storage and retrieval systems for
decision-making in various GIS-based systems. This paper introduces a method for
data analysis by adding a fog layer in the cloud-based GIS. The Fog environment is
integrated mainly for data pre-processing and data cleaning for GIS systems. The
load on the cloud environment will reduce when the pre-processing tasks are exe-
cuted on the fog layer. The Weather dataset is used for weather prediction using an
artificial neural network in a cloud environment.
12.1 Introduction
Big Data Analysis is a significant concern nowadays because the amount of data
generated in a day is around 2.5 quintillion bytes [1]. Therefore, we need technolo-
gies that pre-process these data for various applications. Furthermore, we must
develop a mechanism that handles extensive geospatial data for the data generated by
satellites or geographic information system (GIS). This chapter discusses methods,
tools, technologies, and platforms for handling big geospatial data efficiently. (i)
Google BigQuery GIS is used for data generated from GIS and deals with a large
dataset for finding the information. Therefore, it is the most powerful tool for sig-
nificant dataset analysis from google. (ii) Open-source libraries and binaries like
Python GDAL binding are used to process the data. First, it will divide the large data
set into small chunks and find the results using libraries. Then it will merge all the
results from these small chunks and generate the final result. (iii) SpatialHadoop –
1
Institute of Technology, Nirma University, India
2
CSE Department, Jaypee University of Engineering and Technology, India
3
Space Applications Centre (SAC), Indian Space Research Organisation (ISRO), India
202 Earth observation data analytics using machine and deep learning
Apache Hadoop deals with big data analysis, and it has a spatial version for geos-
patial big data analysis, known as Spatial Hadoop. (iv) Google Earth Engine – this
tool contains a large data set, and we can use and manipulate these real-time spatial
data sets without downloading. (v) AWS Athena – this gives a simple interface to
perform a query on storage using SQL. (vi) NoSQL and graph databases – databases
like MongoDB and Elastic search also support some functionalities for geospatial
data sets. Following are some terminologies related to geospatial data analysis.
Fog layer
Data processing
Data mining
Data filtering
Data compression
Fog node Fog node
Data collection
12.1.5 Contribution
This chapter addresses the future enhancement identified for the method mentioned in
the paper City Geospatial Dashboard, IoT and Big Data Analytics for Geospatial
Solution Providers in Disaster Management [6]. A dashboard is created for data
visualization. It is used for data collection, data sharing, and data visualization. These
data are collected from satellites, IoT devices, and other big data sources. The paper’s
objective is to improve the method by adding one layer to this system and sharing the
load between these layers. The basic idea is to add a Fog layer and perform some data
204 Earth observation data analytics using machine and deep learning
pre-processing to reduce the cloud layer’s burden. Figure 12.1 contains three layers: the
first layer is the data collection layer. In this layer, data is collected from different
devices and sensors, and it sends all the collected data to the upper layer, known as Fog
layer. Here, it will perform data pre-processing and data filtering. The basic definition is
mentioned in the introduction. After applying these methods, data is now clean to per-
form data analysis on the cloud layer.
There are many methods and technologies for big data analysis on geospatial data of
the existing methods, some methods are discussed below in Tables 12.1–12.4. The
first subsection contains techniques for big data analysis on geospatial data, and the
second subsection contains different data processing techniques for fog environment.
Paper name GIS cloud computing-based government big data analysis platform
Approach GIS – big data analysis in cloud computing
Objective Analyse the data
Methodology First analyses the spatial association analysis method of government big
data, then proposes the architecture and function of GIS cloud
computing-based government big data platform, and finally explores
the application case based on traffic accident data
Pros Business function customizer, report query customizer, workflow
customizer
Cons Need to select the attributes that may be analysed in the vector data as
mining items
Data GIS cloud-computing-based government big data analysis platform
big data analysis is mist computing which will apply the algorithms for finding all the
unused microcomputers in the computing environment [6]. The last big data analysis
technique is cloud computing [14,15]. Tasks like data storage, data analysis, data pre-
processing, data mining, and data management are performed on the cloud. Another
modified version of cloud computing is hybrid cloud computing [16]. This architecture
facilitates geospatial big data processing in hybrid cloud environments by leveraging
and extending standards released by the Open Geospatial Consortium (OGC). Many
more big data analysis techniques are used for geospatial data [17].
Fog layer
Geo-spatial Collection, clean, Pass the data to the Find the hidden patterns Apply big data analysis
big data preprocess the data cloud layer using neural network techniques
Cloud layer
Data Input Output
warehouse Hidden
For weather prediction, ANN is used. A neural network contains three layers:
input layer, output layer, and hidden layer. Each layer contains weights, bias, and
activation functions. For training the data, there are three ways: first is supervised
learning in which input and output are provided. The second method is unsu-
pervised learning, in which the input is provided, and the model will find the hidden
pattern. The last method is reinforcement learning; input and output are provided,
and the task is to find the hidden pattern. NN is used for tabular data, image data,
and text data. In this paper, the geo-spatial data set is used for training which
consists of longitude, latitude, date, and time information.
h1 w7
w1
w4 w8
x1 y1
w2
w9
w3
h2
Input output
w5 w10
x2 y2
w11
h3
w6
w12 Output Layer
Input Layer
Hidden Layer
Feed Forward
Back Propagation
activation function is not used, it will only use a linear function for calculation,
which is not a good idea. So, the activation function is like a powerhouse of the
ANN. So, one neuron or perceptron is one logistic regression, and ANN is a group
of neurons.
Here, weights are numeric values. The first step is to give the input, find the output
value, and find the loss function based on predicted and actual values. This process is
known as feed-forward. If the weights are updated of the ANN and repeat the process
until getting good accuracy, it is known as backward propagation. But two basic pro-
blems occur in ANN: the first is the vanishing gradient and exploding gradient. A
vanishing gradient occurs when the weight updating is less or does not change after
some time or epoch. Exploding gradient occurs when the change value is large. The
next section will discuss the implementation part and compare the results.
Data set information: Weather prediction dataset selected from Kaggle (source:
https://www.kaggle.com/alihanurumov/weather-prediction-network/data). The dataset
contains seven files for temperature, pressure, wind direction, wind speed, city attri-
bute, humidity and weather description. Pre-processed data and convert it into the
Geospatial big data analysis using neural networks 209
proper data frame by removing null values (by putting mean values). Then apply ANN
with three layers and the activation function used is sigmoid. The proposed model gets
an accuracy of 70% for 1,800 records and 3,600 records.
Data pre-processing on fog environment: platform configuration: As
shown in Table 12.5, the implementation is done on Google colab, local pc is
considered as the fog layer. So, data pre-processing is done on the fog layer. As the
results mention, almost 6 min are saved by pre-processing data on the fog layer for
1.8k records and 13 mins for 3.6k records. By using the fog layer, the CPU’s power
consumption and processing power are reduced on the cloud layer.
Prediction on cloud environment (comparison of algorithm on cloud layer –
execution steps):
● Train/test split: 70% train and 30 % split
● Label encoding: one hot encoding
● Three layers:
* Input layer:
* Activation function: Sigmoid
* Input features/independent variables: 5 (temperature, pressure, wind
speed, wind direction, humidity)
* Number of neurons and input shape: 5
* Kernel initialization: he_normal
* Hidden layer (ANN-1 layer):
* Activation function: Sigmoid
* Kernel initialization: he_normal
* Number of neurons: 16
* Output layer:
* Activation function: Sigmoid
* Kernel initialization: he_normal
* Dependent variable: 1 (Weather description)
● Optimizer: adam
● Loss function: binary_crossentropy
As shown in Table 12.6, the algorithms have an accuracy of 70%. At first, all
the algorithms were applied for nearly 1.8k records for different dates and times in
a particular area. The same process was applied for 3.6k records and got an accu-
racy of almost 71%. So, all the files contain the date, time, longitude, latitude, and
weather prediction attributes (temperature, pressure, humidity, wind speed, and
wind direction).
12.6 Conclusion
This paper discussed the fog layer and neural networks on the cloud layer. A fog layer
is introduced in cloud computing to reduce the load on the cloud layer. On this layer, a
task like data pre-processing is performed. A local machine is used as a fog layer and
performs data pre-processing on the weather dataset. As shown above, a time of about
6 min is saved for 1.8k records. Time will increase as the number of records increases.
Other than time, the processing power is reduced on the cloud layer. And for weather
prediction, an ANN is used. Here, nearly 71% accuracy is achieved for 3.6k data
records. The results are compared with algorithms like decision trees, KNN, Gaussian
NB, and SVM. We can use CNN, RNN, etc., for better results.
References
[1] C. Dobre and F. Xhafa, “Intelligent services for big data science,” Future
Generation Computer Systems, vol. 37, 2014, pp. 267–281, ISSN 0167-
739X.
[2] D. Li, Y. Gong, G. Tang, and Q. Huang, “Research and design of mineral
resource management system based on big data,” in 5th IEEE International
Conference on Big Data Analytics, 2020.
[3] L. Geng1 and P. Xue, “Research on the protection and utilization evaluation
of historic and cultural streets based on big data technology,” in 2nd
International Conference on Machine Learning, Big Data and Business
Intelligence (MLBDBI), 2020.
[4] S.K. Siledar, B. Deogaonkar, N. Panpatte, and J. Pagare, “Map reduce over-
view and functionality,” in Proceedings of the 6th International Conference on
Communication and Electronics Systems, 2021.
Geospatial big data analysis using neural networks 211
[5] D. Li, Y. Gong, G. Tang, and Q. Huang, “GeoBD2: geospatial big data
deduplication scheme in fog assisted cloud computing environment,” in 8th
International Conference on Computing for Sustainable Global Development,
2020.
[6] K.K., Sekimoto, Y., Takeuchi, W., et al., “City geospatial dashboard: IoT and
big data analytics for geospatial solutions provider in disaster management,”
in 2019 International Conference on Information and Communication
Technologies for Disaster Management (ICT-DM). IEEE, 2019.
[7] P.R. Merla and Y. Liang, “Data analysis using Hadoop MapReduce envir-
onment,” in IEEE International Conference on Big Data (BIGDATA), 2017.
[8] M. Dick, J.G. Ji, and Y. Kwon, “Practical difficulties and anomalies in
running Hadoop,” in International Conference on Computational Science
and Computational Intelligence, 2017.
[9] V. Sontakke and R.B. Dayanand, “Optimization of Hadoop MapReduce
model in cloud computing environment,” in Second International Conference
on Smart Systems and Inventive Technology, 2019.
[10] P. Praveen, C.J. Babu, and B. Rama, “Big data environment for geospatial
data analysis,” in 2016 International Conference on Communication and
Electronics Systems (ICCES). IEEE, 2016.
[11] K. Rattanaopas and S. Kaewkeeree, “Improving hadoop mapreduce perfor-
mance with data compression: a study using wordcount job,” in 14th
International Conference on Electrical Engineering/Electronics, Computer,
Telecommunications, 2017.
[12] D. Sik, K. Csorba, and P. Ekler, “Implementation of a geographic informa-
tion system with big data environment on common data model,” in IEEE,
2017.
[13] Krämer, M. and I. Senner, “A modular software architecture for processing
of big geospatial data in the cloud,” Computers & Graphics, vol. 49, 2015,
pp. 69–81.
[14] Q. Wang and Y. Jiang, “GIS cloud computing based government Big Data
analysis platform,” in IEEE 2nd International Conference on Big Data,
Artificial Intelligence and Internet of Things Engineering, 2021.
[15] Huang, Y., P. Gao, Y. Zhang, and J. Zhang, “A cloud computing solution for
big imagery data analytics,” in 2018 International Workshop on Big
Geospatial Data and Data Science (BGDDS), IEEE, 2018, pp. 1–4.
[16] S. Ingo, “Geospatial Big Data processing in hybrid cloud environments,” in
IGARSS 2018–2018 IEEE International Geoscience and Remote Sensing
Symposium, IEEE, 2018, pp. 419–421.
[17] R.K. Barik, S.S. Patra, P. Kumari, S.N. Mohanty, and A.A. Hamad, “New
energy-aware task consolidation scheme for geospatial big data application
in mist computing environment,” in IEEE, 2019.
[18] H. Sulimani, W.Y. Alghamdi, T. Jan, G. Bharathy, and M. Prasad,
“Sustainability of load balancing techniques in fog computing environment:
review,” Procedia Computer Science, vol. 191, 2021, pp. 93–101, ISSN
1877-0509.
212 Earth observation data analytics using machine and deep learning
13.1 Introduction
Space and time are the basic dimensions of our existence and thus needed to explore
and analyze to reveal hidden knowledge in various fields, from the micro level of
cells in our physical body to the macro level of our planet Earth and its geography.
In the past few years, many software tools have been developed for spatio-
temporal data analysis. Many of them are application specific, which means they
are developed and designed to be meant for analysis in the specific domain with
limited functionalities. Some of them are only meant for spatial analysis; some are
performing well enough to answer spatiotemporal queries but are incapable to
resolved complex queries.
1
Symbiosis Institute of Technology, Symbiosis International University, India
2
U.V. Patel College of Engineering, Ganpat University, India
3
Space Application Centre, ISRO, India
4
Institute of Technology, Nirma University, India
214 Earth observation data analytics using machine and deep learning
13.3 Challenges
There are several issues with the existing spatiotemporal data analysis software that
prompted us to think of designing and developing the proposed framework. They
are as follows:
● Existing open-source GIS tools do not provide sophisticated data mining
facilities.
● Multi-level granularities in space and time should be handled effectively.
● Many tools lack effective visualization components.
● Analysis software should be able to support spatiotemporal data in various
possible formats.
● Software should provide all the possible functionalities for analysing spatio-
temporal data inside a single workbench.
216 Earth observation data analytics using machine and deep learning
Visualization
Layer
Functional
Layer
Data Layer
1. Input module
2. Pre-processing module
3. Data mining module
4. Multidimensional analysis module
5. Application generation module
The framework consists of the above modules, for accomplishing the follow-
ing specific task or functionality:
Preprocessing
Data mining
I/P data in machine and
various forms deep learning
technique
ST-DAME
features
Domain Multi-
specific dimensional
application analysis
Vector format
data
Raster
format
Input
module
Database:
spatial
non-spatial
Nose
removal Classification
Data Output
Input conversion Clustering
textual, graphs,
Input module maps
considered while applying data mining on such data. For the above-mentioned
reasons, separate data mining techniques are needed for spatiotemporal data.
The data is needed to be processed through several modules as per the require-
ment of the user. Typical data flow for data mining is illustrated in Figure 13.5.
The data mining module should offer the features mentioned below:
● Integration of new techniques: it should provide the facility to add new tech-
niques to the workbench or modify existing techniques.
● Application of existing techniques: all the techniques would take input data
in the predefined format only, to fulfil this requirement data conversion is
done to change the input format into the acceptable format by data mining
technique(s).
ST-OLAP server
Logical-
physical
schema
mapping in
XML
SQL queries
Spatiotemporal databases
Domain-
Input specific Output
application
Query Maps, graphs,
(Space+time+measure) histograms
Output
Input Preprocessing Datamining Map
Graph
module module module Report
Machine and deep
Data input file learning technique
Output
Map
Spatio temporal Multi-dimensional Graph
data-warehouse data analysis Report
Output
Datamining Map
Spatio temporal Graph
data-warehouse module
Report
Machine and deep
learning technique
Output
Datamining Multi-dimensional Map
Spatio temporal Graph
data-warehouse module data analysis
Report
Machine and deep
learning technique
Output
Multi-dimensional Datamining Map
Spatio temporal data analysis module Graph
data-warehouse Report
Machine and deep
learning technique
Automated
system
Customized
system
13.5 Result
In our proposed system, we have designed and developed two separate systems for
analysis which are named as:
1. Automated system
2. Customized system
Software framework for spatiotemporal data analysis and mining 223
At first, the start window as shown in Figure 13.13 is splashed over the screen,
asking for an option to be selected between the two. Depending on the choice of the
user, a user interface for the selected subsystem is displayed next.
Figure 13.15 NDVI and rainfall status for selected spatial and temporal
granularities
Software framework for spatiotemporal data analysis and mining 225
13.6 Conclusion
The proposed framework has offered generic functionalities by integrating three
approaches for spatiotemporal data analysis that are data mining, multi-
dimensional analysis (MDX queries) and visualization i.e., generation of graphs
and maps after a rigorous survey of existing tools that are meant for such analysis.
This resulted in overcoming issues of existing tools mentioned earlier and ease of
analysis of spatiotemporal data. The framework also proposes an automated and
customized spatiotemporal multidimensional data analysis environment where
depending on end-user requirements, the result of MDX queries can be obtained
which is evident from the result section. To deal with a huge amount of data, deep
learning techniques (CNN/LSTM/RNN) can be used in future.
References
[1] Alvares LO, Palma A, Oliveira G, et al. Weka-STPM: from trajectory
samples to semantic trajectories. In: Proceedings of the XI Workshop de
Software Livre, WSL. Citeseer, 2010, vol. 10, pp. 164–169.
[2] Rao KV, Govardhan A, and Rao KC. An architecture framework for spa-
tiotemporal datamining system. International Journal of Software
Engineering & Applications. 2012;3(5):125.
[3] Karssenberg D, Schmitz O, Salamon P, et al. A software framework for
construction of process-based stochastic spatio-temporal models and data
assimilation. Environmental Modelling & Software. 2010;25(4):489–502.
[4] Andrienko N and Andrienko G. A visual analytics framework for spatio-
temporal analysis and modelling. Data Mining and Knowledge Discovery.
2013;27(1):55–83.
226 Earth observation data analytics using machine and deep learning
[5] Patel S and Garg S. Issues and challenges in data mining architecture. In:
National Conference on Emerging Technology in C.S., 2007.
[6] Andrienko G, Andrienko N, Demsar U, et al. Space, time and visual analy-
tics. International Journal of Geographical Information Science. 2010;24
(10):1577–1600.
[7] Cavalcanti VM, Schiel U, and de Souza Baptista C. Querying spatio-
temporal databases using a visual environment. In: Proceedings of the
Working Conference on Advanced Visual Interfaces; 2006, pp. 412–419.
[8] Shekhar S, Lu C, Tan X, et al. A visualization tool for spatial data ware-
houses. Geographic Data Mining and Knowledge Discovery. 2001;73:16–72.
[9] Shekhar S, Lu CT, Liu R, et al. CubeView: a system for traffic data visua-
lization. In: Proceedings. The IEEE 5th International Conference on
Intelligent Transportation Systems. IEEE, 2002, pp. 674–678.
[10] Körting TS, Fonseca LMG, and Câmara G. GeoDMA—geographic data
mining analyst. Computers & Geosciences. 2013;57:133–145.
[11] de Oliveira MG and de Souza Baptista C. An approach to visualization and
clustering-based analysis on spatiotemporal data. Journal of Information and
Data Management. 2013;4(2):134–134.
[12] Eibe F, Hall MA, and Witten IH. The WEKA workbench. In: Online
Appendix for Data Mining: Practical Machine Learning Tools and
Techniques. Morgan Kaufmann, 2016.
Chapter 14
Conclusion
Sanjay Garg1, Kimee Joshi2 and Nebu Varghese3
This book covers a wide range of topics about applications for Earth Observation
data (EO) i.e. variety in applications, variety of data formats used, and variety of
machine learning (ML) tools and techniques. This text covers applications for
weather forecasting, crop monitoring, crop classification, LULC classification,
hurricane, climate change, reservoir characterization, etc. Studies of various types
of data such as optical data, metrological data, microwave data, spectral data, and
hyperspectral data are considered for these applications. Various ML techniques
like ANN, CNN, autoencoders, and auto-regression are used and also chapter on
the R tool is included for comprehensive knowledge of the relevant tool.
1
CSE Department, Jaypee University of Engineering and Technology, India
2
Institute of Technology, Nirma University, India
3
Dholera Industrial City Development Limited, India
228 Earth observation data analytics using machine and deep learning
computing to reduce the load on the cloud layer and for weather prediction, an
artificial neural network is used and finally accuracy of weather prediction is
increased.
Chapter 13 fabricated a framework to offer generic functionalities by inte-
grating three approaches for spatiotemporal data analysis that are data mining,
multi-dimensional analysis (MDX Queries), and visualization. This resulted in
overcoming issues of existing tools mentioned earlier and ease of analysis of spa-
tiotemporal data. The framework also encompasses an automated and customized
spatiotemporal multidimensional data analysis environment.
observation are optical and radar images [1]. Understanding the characteristics and
features of satellite imagery is essential to know how to use it. The requirements for
data derived from satellite imagery can vary depending on a particular application
[2]. The level of resolution, both temporal and spatial, is increasing as EO is
applied to more local requirements. Aside from cost and computing requirements,
the suitability of imagery is being questioned. This is especially true in humid areas
where long-term cloud cover or critical periods (e.g., flooding) make optical ima-
gery difficult to use [3]. For measuring crop health and doing vegetation analysis,
the multispectral bands are commonly used [4]. The monitoring of the weather,
agriculture harvest monitoring, field segmentation of icebergs, and natural or man-
made disasters like tsunamis and oil spills are just a few of the applications of SAR
satellite data [2].
14.2.5 Standardization
It is clear from the preceding that global data sets from EO are generated by a
variety of sensors, processed by numerous organizations, and available from var-
ious sources. This means that the data have different resolutions and formats, so
integration is required to make the best use of the data, and standardization would
help with this process [5].
References
[1] Sudmanns M, Tiede D, Lang S, et al. Big Earth data: disruptive changes in
Earth observation data management and analysis? International Journal of
Digital Earth. 2020;13(7):832–850.
[2] ICEYE. Satellite Data: How to use Satellite Data for Better Decision
Making. Available from: https://www.iceye.com/satellite-data.
232 Earth observation data analytics using machine and deep learning
[3] Addressing the Challenges of using Earth Observation Data for SDG
Attainment: Evidence from the Global South and West Africa Region.
Available from: https://static1.squarespace.com/static/5b4f63e14eddec
374f416232/t/624334541bf6aa586f04d166/1648571478139/Earth+Observation-
3.29.pdf.
[4] Murray R. Choosing the right type of imagery for your project. 2021.
Available from: https://www.l3harrisgeospatial.com/Learn/Blogs/Blog-
Details/ArtMID/10198/ArticleID/15204/Choosing-the-right-type-of-ima-
gery-for-your-project.
[5] Dowman I and Reuter HI. Global geospatial data from Earth observation:
status and issues. International Journal of Digital Earth. 2017;10(4):328–341.
[6] Haklay M, Mazumdar S, and Wardlaw J. Citizen Science for Observing and
Understanding the Earth. Springer; 2018.
[7] Baumann P, Rossi AP, Bell B, et al. Fostering cross-disciplinary earth sci-
ence through datacube analytics. In: Earth Observation Open Science and
Innovation. Springer, Cham; 2018. p. 91–119.
[8] Giuliani G, Chatenoux B, De Bono A, et al. Building an earth observations
data cube: lessons learned from the swiss data cube (sdc) on generating
analysis ready data (ard). Big Earth Data. 2017;1(1–2):100–117.
[9] Purss MB, Lewis A, Oliver S, et al. Unlocking the Australian landsat
archive – from dark data to high performance data infrastructures. GeoResJ.
2015;6:135–140.
[10] Ground Truth Data Collection – egyankosh.ac.in. Available from: https://
egyankosh.ac.in/bitstream/123456789/39538/1/Unit-9.pdf.
[11] Agarwal C. Challenges of using machine learning on Earth Observation
Data; Geospatial World, 2021. Available from: https://www.geospatial-
world.net/top-stories/challenges-of-using-machine-learning-on-earth-obser-
vation-data/.
[12] Cerbaro M, Morse S, Murphy R, et al. Challenges in using Earth
Observation (EO) data to support environmental management in Brazil.
Sustainability. 2020;12(24):10411.
Index
choropleth map 9 decision tree (DT) 23, 104, 143, 152, 164
climate change 1, 117, 125 deep learning (DL) 8, 23, 52, 102, 142
climate-induced disasters (CIDs) 118 classification using 39
Cloud computing 206 custom VGG-like model 40
coefficient of determination 164 inception V3 model 39–40
Cohen’s kappa coefficient 58 deep neural network (DNN) 26, 39
Comma Separated Values (CSV) 88 background and related work
computational hydrology 178 121–2
Conference of Parties 117 benefits of 124
confidence score (CS) 89 long-term climate change monitoring
confusion matrix 28 using DL methods 125–7
convolutional autoencoder 26 modern DL methods 122
convolutional layers 25–6, 86, 104 AlexNet 123
convolutional neural networks (CNNs) 8, inception-v3 123
21, 35, 52, 53, 83, 206 other neural networks 123–4
architecture of 25 U-Net 123
confusion matrix for 27 other applications 127–8
hyper-parameters of 105–6 possible problems 129
core sampling 152 DeepUNet 8
correlation coefficients (CC) 158–9, 164 defence and security 13
Cuckoo Search algorithm 206 deforestation 126, 136
customized system 224–5 DenseNets 58, 60, 228
custom VGG-like model 40 DiceNets 56, 228
cyclones 120 dice similarity coefficient (DSC) 96
digital elevation model (DEM) 137
data analytics 150 digital images 84
data analytics methodology 6 digital surface model 82
deep learning 8 dropout 167
machine learning 7–8 dynamic data 202
data collection 108
data layer 216 Earth observation (EO) data 1
data mining 203, 214 applications of 10
data mining module 218–19 categories of 2
data pre-processing 109, 203 active imaging system 5–6
data science 179 passive imaging system 3–4
dataset 54–5, 86–7 data analytics methodology 6
dataset collections 137 deep learning 8
data storage 230 machine learning 7–8
data visualization techniques 8 data visualization techniques 8
cartogram map 8–9 cartogram map 8–9
choropleth map 9 choropleth map 9
heat map 9 heat map 9
Index 235
Earth Observation Data Analytics Using Machine and Deep Learning: Modern tools,
applications and challenges covers the basic properties, features and models for Earth
observation (EO) recorded by very high-resolution (VHR) multispectral, hyperspectral,
synthetic aperture radar (SAR), and multi-temporal observations.
Approaches for applying pre-processing methods and deep learning techniques to satellite
images for various applications – such as identifying land cover features, object detection,
crop classification, target recognition, and the monitoring of earth resources – are described.
Cost-efficient resource allocation solutions are provided, which are robust against common
uncertainties that occur in annotating and extracting features on satellite images.
This book is a valuable resource for engineers and researchers in academia and industry
working on AI, machine and deep learning, data science, remote sensing, GIS, SAR, satellite
communications, space science, image processing and computer vision. It will also be of
interest to staff at research agencies, lecturers and advanced students in related fields.
Readers will need a basic understanding of computing, remote sensing, GIS and image
interpretation.
Swati Jain is an associate professor in the Computer Science and Engineering Department at
the Institute of Technology, Nirma University, India.
Nitant Dube is Group Director of Planetary and Meteorology Data Processing Group at Space
Applications Centre (SAC), ISRO, India.
Nebu Varghese is an assistant manager (GIS) in the Land and Municipal Service at Dholera
Industrial City Development Limited (DICDL), India.