Flood Forecasting
Flood Forecasting
Flood Forecasting
Introduction
With the evolution of the Internet, the ways how businesses, economies, stock
markets, and even the governments function and operate have also evolved, big
time. It has also changed the way people live. With all of this happening, there
has been an observable rise in all the information floating around these days;
it’s more than ever before. This outburst of data is relatively new. Before the
past couple of years, most of the data was stored on paper, film, or any other
analog media; only one-quarter of all the world’s stored information was digital.
But with the exponential increase in data, the idea of storing it manually just
does not hold appeal anymore. You will learn more about applications and
examples of big data in this big data analytics tutorial.
The first trace of big data was evident way back in 1663. It was during
the bubonic plague that John Graunt dealt with overwhelming amounts of
information during his study of the disease. He was the first person ever to
make use of statistical data analysis. The field of statistics expanded later to data
collection and analysis in the early 1800s.The US Census Bureau estimated that
it would take eight years to handle and process the data collected during the
census program in 1880, which was the first overwhelming collection of raw
data. The Hollerith Tabulating Machine was invented to reduce the calculation
work in the subsequent 1890 census.
After that, data evolved at an unprecedented rate throughout the 20th century.
There were machines that stored information magnetically. Scanning patterns in
messages and computers were also prominent during that time. In 1965, the first
data center was built with the aim to store millions of fingerprint sets and tax
returns.
Everyone knows that customers are the most important asset of any
business. However, even with a solid customer base, it is foolish to disregard
competition. A business should be aware of what customers are looking for.
This is where big data comes in.Applying big data allows businesses to identify
and monitor customer-related trends and patterns. This contributes toward
gaining loyalty. More data collection allows for more patterns and trends to be
identified.With a proper customer data analytics mechanism in order, critical
behavioral insights can be derived to act on and retain the customer base. This is
the most basic step to retain customers.Big data analytics is strongly behind
customer retention at Coca-Cola. In 2015, Coca-Cola strengthened its data
strategy by building a digital-led loyalty program.
A good example of a brand that uses big data for targeted advertisements is
Netflix. It uses big data analytics for targeted advertising. The data gives
insights into what interests the subscribers the most.
Risk Management
Big data analytics has contributed immensely toward the development of risk
management solutions. Tools allow businesses to quantify and model regular
risks. The rising availability and diversity of statistics have made it possible for
big data analytics to enhance the quality of risk management models, thus
achieving better risk mitigation strategies and decisions.UOB in Singapore uses
big data for risk management. The risk management system allows the bank to
reduce the calculation time of the value at risk.
Big data has become a smart way for creating additional revenue streams
through innovations and product improvement. Organizations are first correct as
much data as possible before moving on to designing new product lines and
redesigning existing ones.The design processes have to encompass the
requirements and needs of customers. Various channels are available to help
study these customer needs. Big data analytics helps a business to identify the
best ways to capitalize on those needs.Amazon Fresh and Whole Foods are the
perfect examples of how big data can help improve innovation and product
development. Data-driven logistics provides companies with the required
knowledge and information to help achieve greater value.
Supply Chain Management
Big data offers improved clarity, accuracy, and insights to supplier networks.
Through big data analytics, it is possible to achieve contextual intelligence
across supply chains. Suppliers are now able to avoid the constraints and
challenges that they faced earlier.Suppliers incurred huge losses and were prone
to making errors when they were using traditional enterprise and supply chain
management systems. However, approaches based on big data made it possible
for suppliers to achieve success with higher levels of contextual
intelligence.PepsiCo depends on enormous amounts of data for the efficient
supply chain management. The company tries to ensure that it replenishes the
retailers’ shelves with appropriate numbers and types of products. Data is used
to reconcile and forecast the production and shipment needs.
Structured Data
Any data that can be stored, accessed, and processed in a fixed format is known
as structured data. Businesses can get the most out of this type of data by
performing analysis. Advanced technologies help generate data-driven insights
to make better decisions from structured data.
Unstructured Data
Data that has an unknown structure or form is unstructured data. Processing and
analyzing this type of data for data-driven insights can be a difficult and
challenging task as they are under different categories and putting them together
in a box will not be of any value. A combination of simple text files, images,
videos, etc., is an example of unstructured data.
Semi-structured data
Semi-structured data, as you may have already guessed, has both structured and
unstructured data. Semi-structured data may seem structured in form, but it is
not exactly well-defined with table definition in relational DBMS. Web
applications have unstructured data such as transaction history files, log files,
etc.
Every time one opens an application on his/her phone, visits a web page, signs
up online on a platform, or even types into a search engine, a piece of data is
gathered.So, whenever we turn to our search engines for answers a lot of data is
created and gathered.
But as users, we are usually more focused on the outcomes of what we are
performing on the web. We don’t dwell on what happens behind the scenes. For
example, we might have opened up our browser and looked up for ‘big data,’
then visited this link to read this blog. That alone has contributed to the vast
amount of big data. Now imagine the number of people spending time on the
Internet visiting different web pages, uploading pictures, and whatnot.
There are some terms associated with big data that actually help make
things even clearer about big data. These are essentially called the
characteristics of big data and are termed as volume, velocity, and variety,
giving rise to the popular name 3Vs of big data, which I am sure we must have
heard before. But, if it feels new to you, do not worry. We are going to discuss
them in detail here. As people are understanding more and more about the ever-
evolving technological term, big data, it shouldn’t come as a shock if more
characteristics are added to the list of the 3Vs. These are called veracity and
value.
Characteristics of Details
Big Data
Volume Organizations have to constantly scale their storage
solutions since big data requires a large amount of space to
be stored.
Velocity Since big data is being generated every second,
organisations need to respond in real time to deal with it.
Variety Big data comes in a variety of forms. It could be structured
or unstructured, or even in different formats such as text
format, videos, images, and more.
Veracity Big data, as large as it is, can contain wrong data too.
Uncertainty of data is something organisations have to
consider while dealing with big data.
Value Just collecting big data and storing it is of no consequence
unless the data is analyzed and a useful output is produced.
It must be pretty clear by now that while talking about big data one can’t ignore
the fact that there are some obvious big data challenges associated with it. So
moving forward in this blog, let’s address some of those challenges.
Data growing at such a quick rate is making it a challenge to find insights from
it. There is more and more data generated every second from which the data that
is actually relevant and useful has to be picked up for further analysis.
Storage
Security
Large amounts of data in organizations can easily become a target for advanced
persistent threats, so here lies another challenge for organizations to keep their
data secure by proper authentication, data encryption, etc.
Unreliable Data
We can’t deny the fact that big data can’t be 100 percent accurate. It might
contain redundant or incomplete data, along with contradictions.
Miscellaneous Challenges
These are some other challenges that come forward while dealing with big data,
like the integration of data, skill and talent availability, solution expenses, and
processing a large amount of data in time and with accuracy so that the data is
available for data consumers whenever they need it.
Before we go further into getting to know technologies that can help manage
big data, we should first get familiar with a very popular programming
paradigm called MapReduce.
What it does is, allows performing computations on huge data sets on multiple
systems in a parallel fashion.
MapReduce mainly consists of two parts: the Map and the Reduce. It’s kind of
obvious! Anyway, let’s see what these two parts are used for:
Map: It sorts and filters and then categorizes the data so that it’s easy to
analyze it.
Reduce: It merges all data together and provides the summary.
These are some of the many technologies that are used to handle and manage
big data. Hadoop is the most widely used among them. If you wish to learn
more about Big Data and Hadoop, along with a structured training program,
visit HERE.
There are many real-life Big Data applications in various industries. Let’s find
out some of them in brief.
Fraud Detection
Big data helps in risk analysis, management, fraud detection, and abnormal
trading analysis.
Advertising and Marketing
Big data helps advertising agencies understand the patterns of user behavior and
then gather information about consumers’ motivations.
Agriculture
Big data can be used to sensor data to increase crop efficiency. This can be done
by planting test crops to record and store the data about how crops react to
various environmental changes and then using that data for planning crop
plantation, accordingly.
Knowledge about big data is one of the most important skills required for some
of the hottest job profiles which are in high demand right now and the demand
in these profiles won’t be dropping down any time sooner, because, honestly,
the accumulation of data is only going to increase over time, increasing the
number of talents required in this field, thus opening up multiple doors of
opportunities for us. Some of the hot job profiles are given below:
Data analysts analyze and interpret data, visualize it, and build reports to
help make better business decisions.
Data scientists mine data by assessing data sources and using algorithms
and machine learning techniques.
Data architects design database systems and tools.
Database managers control database system performance, perform
troubleshooting and upgrade hardware and software.
Big data engineers design, maintain and support big data solutions.
Once we learn about big data and understand its use, we will come to know that
there are many analytics problems we can solve, which were not possible earlier
due to technological limitations. Organizations are now relying more and more
on this cost-effective and robust method for easy data processing and storage.
Identify the problem
Data lifecycles
For example, social media data may be combined with internal customer
data to uncover new insights. This new data could be discarded once it has
served its immediate use, or it could be retained for a period. If an organisation
decides to retain the data, it needs to consider not only the regulatory
implications, but also the technical question of how and where to store that data:
There are no definitive answers to these questions. They will vary depending on
the contents of the data, an organisation’s policies and any legal or regulatory
restrictions. However, these considerations underline the need for robust data
lifecycles when adopting Big Data solutions. In effect, data is created (arrives),
is maintained (exists), and then is deleted (disappears) at some point in the
future. That data needs to be managed in a way that ensures it is kept for the
right length of time and no longer. But Big Data’s potential to find new insights
and generate value from old data prompts another question — shouldn’t
organisations be keeping it forever, as long as they can do so securely, cost-
effectively and legally?
When selecting tools for Big Data analysis, organisations face a number of
considerations:
The chart above has been designed to help readers of this book identify the
key Big Data priorities for their businesses. At the heart of business is the
drive for increased profit, represented here in the centre of the target.
Working outwards, businesses can either increase profit by increasing
revenue or reducing costs. Both of these methods can be achieved through
either direct or indirect actions, but by combining the two we move outwards
towards the appropriate segment. The outer circle shows the various actions
a business can take. The left hemisphere contains revenue-increasing actions,
while the right side contains cost-reducing actions. The diagram also splits
horizontally to show direct actions (in the top half) and indirect actions
(bottom half). From this it is easy to see the possible actions a business can
take to increase revenues or reduce costs, either directly or indirectly. These
are also listed below with examples:
By shading the outer grey segments of the diagram on the previous page
— green (for high value), yellow (for medium value), blue (for low
value) or leaving them blank (for no value) — it would be straightforward
for organisations to identify their business priorities. They can then use
the table on the pages overleaf to identify the actions that need to be
taken. Ensuring success for a Big Data project In common with any
business change initiative, a Big Data project needs to be business-led
and (ideally) executive-sponsored — it will never work as an isolated IT
project. Even more importantly, Big Data is a complex area that requires
a wide range of skills — it spans the whole organisation and the entire
executive team needs to work together (not just the CIO). In addition,
there is a dearth of data scientists, and it may be necessary to fill gaps
with cross training. The next two chapters of this book look at the
changing role of the executive team and the rise of the data scientist
1. Supervised Learning
2. Unsupervised Learning
How it works: In this algorithm, we do not have any target or outcome variable
to predict / estimate. It is used for clustering population in different groups,
which is widely used for segmenting customers in different groups for specific
intervention. Examples of Unsupervised Learning: Apriori algorithm, K-means.
3. Reinforcement Learning:
How it works: Using this algorithm, the machine is trained to make specific
decisions. It works this way: the machine is exposed to an environment where it
trains itself continually using trial and error. This machine learns from past
experience and tries to capture the best possible knowledge to make accurate
business decisions. Example of Reinforcement Learning: Markov Decision
Process
1. Linear Regression
2. Logistic Regression
3. Decision Tree
4. SVM
5. Naive Bayes
6. kNN
7. K-Means
8. Random Forest
9. Dimensionality Reduction Algorithms
10.Gradient Boosting algorithms
1. GBM
2. XGBoost
3. LightGBM
4. CatBoost
It is used to estimate real values (cost of houses, number of calls, total sales etc.)
based on continuous variable(s). Here, we establish relationship between
independent and dependent variables by fitting a best line. This best fit line is
known as regression line and represented by a linear equation Y= a *X + b.
In this equation:
Y – Dependent Variable
a – Slope
X – Independent variable
b – Intercept
These coefficients a and b are derived based on minimizing the sum of squared
difference of distance between data points and regression line.
Look at the below example. Here we have identified the best fit line having
linear equation y=0.2811x+13.9. Now using this equation, we can find the
weight, knowing the height of a person.
Here’s a coding window to try out your hand and build your own linear
regression model in Python:
2. Logistic Regression
Let’s say your friend gives you a puzzle to solve. There are only 2 outcome
scenarios – either you solve it or you don’t. Now imagine, that you are being
given wide range of puzzles / quizzes in an attempt to understand which
subjects you are good at. The outcome to this study would be something like
this – if you are given a trignometry based tenth grade problem, you are 70%
likely to solve it. On the other hand, if it is grade fifth history question, the
probability of getting an answer is only 30%. This is what Logistic Regression
provides you.
Coming to the math, the log odds of the outcome is modeled as a linear
combination of the predictor variables.
Now, you may ask, why take a log? For the sake of simplicity, let’s just say that
this is one of the best mathematical way to replicate a step function. I can go in
more details, but that will beat the purpose of this article.
Build your own logistic regression model in Python here and check the
accuracy:
There are many different steps that could be tried in order to improve the model:
3. Decision Tree
The best way to understand how decision tree works, is to play Jezzball – a
classic game from Microsoft (image below). Essentially, you have a room with
moving walls and you need to create walls such that maximum area gets cleared
off with out the balls.
So, every time you split the room with a wall, you are trying to create 2
different populations with in the same room. Decision trees work in very similar
fashion by dividing a population in as different groups as possible.More:
Simplified Version of Decision Tree Algorithms
Let’s get our hands dirty and code our own decision tree in Python!
For example, if we only had two features like Height and Hair length of an
individual, we’d first plot these two variables in two dimensional space where
each point has two co-ordinates (these co-ordinates are known as Support
Vectors)Now, we will find some line that splits the data between the two
differently classified groups of data. This will be the line such that the distances
from the closest point in each of the two groups will be farthest away.
In the example shown above, the line which splits the data into two differently
classified groups is the black line, since the two closest points are the farthest
apart from the line. This line is our classifier. Then, depending on where the
testing data lands on either side of the line, that’s what class we can classify the
new data as.
You can draw lines/planes at any angles (rather than just horizontal or
vertical as in the classic game)
The objective of the game is to segregate balls of different colors in
different rooms.
And the balls are not moving.
Try your hand and design an SVM model in Python through this coding
window:
Naive Bayes
Naive Bayesian model is easy to build and particularly useful for very large data
sets. Along with simplicity, Naive Bayes is known to outperform even highly
sophisticated classification methods.Bayes theorem provides a way of
calculating posterior probability P(c|x) from P(c), P(x) and P(x|c). Look at the
equation below:
Here,
Step 3: Now, use Naive Bayesian equation to calculate the posterior probability
for each class. The class with the highest posterior probability is the outcome of
prediction.
Problem: Players will pay if weather is sunny, is this statement is correct?
Here we have P (Sunny |Yes) = 3/9 = 0.33, P(Sunny) = 5/14 = 0.36, P( Yes)=
9/14 = 0.64
Now, P (Yes | Sunny) = 0.33 * 0.64 / 0.36 = 0.60, which has higher probability.
Naive Bayes uses a similar method to predict the probability of different class
based on various attributes. This algorithm is mostly used in text classification
and with problems having multiple classes.
KNN can easily be mapped to our real lives. If you want to learn about a person,
of whom you have no information, you might like to find out about his close
friends and the circles he moves in and gain access to his/her information!
K-Means
Remember figuring out shapes from ink blots? k means is somewhat similar this
activity. You look at the shape and spread to decipher how many different
clusters / population are present!
In K-means, we have clusters and each cluster has its own centroid. Sum of
square of difference between centroid and the data points within a cluster
constitutes within sum of square value for that cluster. Also, when the sum of
square values for all the clusters are added, it becomes total within sum of
square value for the cluster solution.
We know that as the number of cluster increases, this value keeps on decreasing
but if you plot the result you may see that the sum of squared distance decreases
sharply up to some value of k, and then much more slowly after that. Here, we
can find the optimum number of cluster.
Random Forest
For more details on this algorithm, comparing with decision tree and tuning
model parameters, I would suggest you to read these articles:
In the last 4-5 years, there has been an exponential increase in data capturing at
every possible stages. Corporates/ Government Agencies/ Research
organisations are not only coming with new sources but also they are capturing
data in great detail.
As a data scientist, the data we are offered also consist of many features, this
sounds good for building good robust model but there is a challenge. How’d
you identify highly significant variable(s) out 1000 or 2000? In such cases,
dimensionality reduction algorithm helps us along with various other algorithms
like Decision Tree, Random Forest, PCA, Factor Analysis, Identify based on
correlation matrix, missing value ratio and others.
To know more about this algorithms, you can read “Beginners Guide To Learn
Dimension Reduction Techniques“.
10.1. GBM
GBM is a boosting algorithm used when we deal with plenty of data to make a
prediction with high prediction power. Boosting is actually an ensemble of
learning algorithms which combines the prediction of several base estimators in
order to improve robustness over a single estimator. It combines multiple weak
or average predictors to a build strong predictor. These boosting algorithms
always work well in data science competitions like Kaggle, AV Hackathon,
CrowdAnalytix.
. XGBoost
The XGBoost has an immensely high predictive power which makes it the best
choice for accuracy in events as it possesses both linear model and the tree
learning algorithm, making the algorithm almost 10x faster than existing
gradient booster techniques.
One of the most interesting things about the XGBoost is that it is also called a
regularized boosting technique. This helps to reduce overfit modelling and has a
massive support for a range of languages such as Scala, Java, R, Python, Julia
and C++.
LightGBM
LightGBM is a gradient boosting framework that uses tree based learning
algorithms. It is designed to be distributed and efficient with the following
advantages
Since the LightGBM is based on decision tree algorithms, it splits the tree leaf
wise with the best fit whereas other boosting algorithms split the tree depth wise
or level wise rather than leaf-wise. So when growing on the same leaf in Light
GBM, the leaf-wise algorithm can reduce more loss than the level-wise
algorithm and hence results in much better accuracy which can rarely be
achieved by any of the existing boosting algorithms.
3. LITERATURE SURVEY
“Characteristics of the 2011 Chao Phraya River flood in central Thailand,” for
A massive flood, the maximum ever recorded in Thailand, struck the Chao
Phraya River in 2011. The total rainfall during the 2011 rainy season was 1,439
mm, which was 143% of the average rainy season rainfall during the period
1982–2002. Although the gigantic Bhumipol and Sirikit dams stored
approximately 10 billion m³ by early October, the total flood volume was
estimated to be 15 billion m³. This flood caused tremendous damage, including
813 dead nationwide, seven industrial estates, and 804 companies with
inundation damage, and total losses estimated at 1.36 trillion baht
(approximately 3.5 trillion yen). The Chao Phraya River watershed has
experienced many floods in the past, and floods on the same scale as the 2011
flood are expected to occur in the future. Therefore, to prepare of the next flood
disaster, it is essential to understand the characteristics of the 2011 Chao Phraya
River Flood. This paper proposes countermeasures for preventing major flood
damage in the future[3]. “The 2011 Great Flood in Thailand: Climate Diagnostics
and Implications from Climate Change” for Severe flooding occurred in Thailand
during the 2011 summer season, which resulted in more than 800 deaths and
affected 13.6 million people. The unprecedented nature of this flood in the Chao
Phraya River basin (CPRB) was examined and compared with historical flood
years. Climate diagnostics were conducted to understand the meteorological
conditions and climate forcing that led to the magnitude and duration of this
flood. Neither the monsoon rainfall nor the tropical cyclone frequency
anomalies alone was sufficient to cause the 2011 flooding event. Instead, a
series of abnormal conditions collectively contributed to the intensity of the
2011 flood: anomalously high rainfall in the premonsoon season, especially
during March; record-high soil moisture content throughout the year; elevated
sea level height in the Gulf of Thailand, which constrained drainage; and other
water management factors. In the context of climate change, the substantially
increased premonsoon rainfall in CPRB after 1980 and the continual sea level
rise in the river outlet have both played a role. The rainfall increase is associated
with a strengthening of the premonsoon northeasterly winds that come from
East Asia. Attribution analysis using phase 5 of the Coupled Model
Intercomparison Project historical experiments pointed to anthropogenic
greenhouse gases as the main external climate forcing leading to the rainfall
increase. Together, these findings suggest increasing odds for potential flooding
of similar intensity to that of the 2011 flood[4]. “Adaptive hydrological flow
field modeling based on water body extraction and surface information” for
Hydrological flow characteristic is one of the prime indicators for assessing
flood. It plays a major part in determining drainage capability of the affected
basin and also in the subsequent simulation and rainfall-runoff prediction. Thus
far, flow directions were typically derived from terrain data which for flat
landscapes are obscured by other man-made structures, hence undermining the
practical potential. In the absence (or diminutive) of terrain slopes, water
passages have a more pronounced effect on flow directions than elevations. This
paper, therefore, presents detailed analyses and implementation of hydrological
flow modeling from satellite and topographic images. Herein, gradual
assignment based on support vector machine was applied to modified
normalized difference water index and a digital surface model, in order to
ensure reliable water labeling while suppressing modality-inherited artifacts and
noise. Gradient vector flow was subsequently employed to reconstruct the flow
field. Experiments comparing the proposed scheme with conventional water
boundary delineation and flow reconstruction were presented. Respective
assessments revealed its advantage over the generic stream burning.
Specifically, it could extract water body from studied areas with 98.70%
precision, 99.83% recall, 98.76% accuracy, and 99.26% F-measure. The
correlations between resultant flows and those obtained from the stream burning
• System : I3 PROCESSOR
• Hard Disk : 300 GB.
• Floppy Drive : 1.44 Mb.
• Monitor : 15 VGA Colour.
• Mouse : Logitech.
• Ram : 4 gb.
SOFTWARE REQUIREMENTS:
CLASS DIAGRAM
SEQUENCE DIAGRAM
The sequence diagrams are an easy and intuitive way of describing the
system’s behavior, which focuses on the interaction between the system and the
environment. This notational diagram shows the interaction arranged in a time
sequence. The sequence diagram has two dimensions: the vertical dimension
represents the time; the horizontal dimension represents different objects. The
vertical line also called the object’s lifeline represents the object’s existence
during the interaction.
ACTIVITY DIAGRAM
6.CONCLUSION:
This paper proposed a unique allotted flood forecasting device, primarily based
totally on integrating meteorological, hydrological, geospatial, and crowdsource
facts. Big facts made to be had via way of distinguished groups had been
received by using diverse move platform APIs. Forecasting became done
primarily based totally on those facts found out via way of present day ML
strategies. They had been choice tree, RF, Naïve Bayes, MLP and RBF ANN,
SVM, and fuzzy logics. Evaluation consequences on studied regions indicated
that the device may want to forecasted flood occasions notably accurately.
Three first-class acting MLs had been MLP ANN, SVM, and RF, respectively.
It became elucidated empirically that the advanced device might be used to alert
the general public and government alike of now no longer handiest a cutting-
edge flood however additionally destiny ones. This device additionally superior
person revel in thru responsive graphical interfaces, interoperable on one of a
kind computing gadgets inclusive of mobiles. This benefit correctly
recommended extra contribution of crowdsource facts from the general public,
enriching facts aggregation and as a result growing device accuracy and
reliability. As such, the advanced device is adaptive, in a feel that because the
device have become extra “experienced” (i.e., via learnings), the forecasting
receives extra realistic In prospects, the device may be effortlessly hired in
current floods control schemes, e.g., the ones led via way of authorities groups
or non-earnings organizations. Moreover, way to allotted architecture, the
device can attain wider public, and consequently serves as an powerful manner
of speaking with them (and specifically the flood victims), concerning cutting-
edge repute and improvement of the disaster. Future upgrades of the device
encompass preliminary flood illustration and its quantity being tailored to the
cutting-edge region of the device, in order that they may be immediately made
privy to via way of its person. Moreover, flooded region pined via way of an
icon can be augmented with color-coded regions, in order that the conditions
(e.g., stages and extents) of affected regions can be higher comprehended.Some
facts taken into consideration on this study, including GLOFAS, had been now
no longer of intrinsically excessive spatial decision, However, they had
beenaccurate. The accrued precipitations and their chance at one of a kind
stages, for instance, corresponded to the real event. Their API device became
additionally reliable. These traits had been desired via way of the proposed
device. Their decision short comings had been remedied via way of
incorporating different extra targeted layers in addition to crowdsource elements
into the ML framework. Possible upgrades for this problem consist of involving
the Internet of Things (IoT) in measuring real meteorological facts with desired
coverage.
7.REFERENCE:
[1] Asian Disaster Reduction Center, “Natural Disaster Data Book,” Japan,
Asian Disaster Reduction Center )ADRC(, 2012.
[2] Asian Disaster Reduction Center, “Natural Disaster Data Book,” Japan,
Asian Disaster Reduction Center )ADRC(, 2015
[3] D. Komori, S. Nakamura, M. Kiguchi, A. Nishijima, D. Yamazaki, S.
Suzuki and T. Oki, “Characteristics of the 2011 Chao Phraya River flood in
central Thailand,” Hydrological Research Letters, Vol. 6, pp. 41-46, 2012.
[4] P., Promchote, S. Y. Simon Wang and P. G. Johnson, “The 2011 great flood
in Thailand: Climate diagnostics and Implications from climate change,”
Journal of Climate, Vol. 29, no. 1, pp. 367-379, Jan. 2016.
[5] S. Puttinaovarat, P. Horkaew, K. Khaimook and W. Polnigongit, “Adaptive
hydrological flow field modeling based on water body extraction and surface
information,” Journal of Applied Remote Sensing, Vol. 9, no. 1, pp. 095041,
Jan. 2015.
[6] S. K. Jain, P. Mani, S. K. Jain, P. Prakash, V. P. Singh, D. Tullos and A. P.
Dimri, “A Brief review of flood forecasting techniques and their applications,”
International journal of river basin management, Vol. 16, no. 3, pp. 329-344,
Jan. 2018
[7] N. Belabid, F. Zhao, L. Brocca, Y. Huang and Y. Tan, “Near-real time flood
forecasting based on satellite precipitation products,” Remote Sensing, vol. 11,
no. 3, pp. 252, Jan. 2019.
[8] A. K. Kar, A. K. Lohani, N. K. Goel and G. P. Roy, “Rain gauge network
design for flood forecasting using multi-criteria decision analysis and clustering
techniques in lower Mahanadi river basin, India,” Journal of Hydrology:
Regional Studies, vol. 4, pp. 313-332, Sep. 2015.
[9] M. Dembélé and S. J. Zwart, “Evaluation and comparison of satellite based
rainfall products in Burkina Faso, West Africa,” International Journal of
Remote Sensing, Vol. 37, no. 17, pp. 3995-4014, Jul. 2016.
[10] A. Foehn, J. G. Hernández, B. Schaefli and G. De Cesare, “Spatial
interpolation of precipitation from multiple rain gauge networks and weather
radar data for operational applications in Alpine catchments,” Journal of
hydrology, Vol. 563, pp. 1092-1110, Jul. 2018.
[11] F. Cecinati, A. Moreno-Ródenas, M. Rico-Ramirez, M. C. Veldhuis and J.
Langeveld, “Considering Rain Gauge Uncertainty Using Kriging for Uncertain
Data,” Atmosphere, Vol. 9, no. 11, pp. 446, Nov. 2018.
[12] S. Stisen and M. Tumbo, “Interpolation of daily raingauge data for
hydrological modelling in data sparse regions using pattern information from
satellite data,” Hydrological Sciences Journal, Vol. 60, no. 11, pp. 1911-1926,
Sep. 2015.
[13] Q. Hu, Z. Li, L. Wang, Y. Huang, Y. Wang and L. Li, “Rainfall Spatial
Estimations: A Review from Spatial Interpolation to MultiSource Data
Merging,” Water, Vol. 11, no. 3, pp. 579, May. 2019.
[14] C. Berndt and U. Haberlandt, “Spatial interpolation of climate variables in
Northern Germany—Influence of temporal resolution and network density,”
Journal of Hydrology: Regional Studies, Vol. 15, pp. 184-202, Feb. 2018.
[15] S. Zhang, J. Zhang, Y. Liu and Y. Liu, “A mathematical spatial
interpolation method for the estimation of convective rainfall distribution over
small watersheds,” Environmental Engineering Research, Vol. 21, no. 3, pp.
226-232, Sep. 2016.
[16] A. Mosavi, P. Ozturk and K. W. Chau, “Flood Prediction Using Machine
Learning Models: Literature Review,” Water, Vol. 10, no. 11, pp. 20734441,
Nov. 2018.
[17] H. L. Cloke and F. Pappenberger, “Ensemble flood forecasting: A review,”
Journal of hydrology, Vol. 375, no. 3, pp. 613-626, Sep. 2009.
[18] T. E. Adams, and T. C. Pagano, “Flood Forecasting: A Global
Perspective,” Academic Press, 2016.
[19] Q. Liang, Y. Xing, X. Ming, X. Xia, H. Chen, X. Tong and G. Wang, “An
open-Source Modelling and Data System for near real-time flood,” In proc. 37th
IAHR World Congress, Newcastle University, 2017, pp. 1-10.
[20] L .C. Chang, F .J. Chang, S .N. Yang, I. Kao, Y .Y. Ku, C .L. Kuo and
M .Z. Bin Mat, “Building an Intelligent Hydroinformatics Integration Platform
for Regional Flood Inundation Warning Systems,” Water, Vol. 11, no. 1, pp.
20734441, Jan. 2019.
[21] S. Puttinaovarat, P. Horkaew and K. Khaimook, “Conguring ANN for
Inundation Areas Identication based on Relevant Thematic Layers,” ECTI
Transactions on Computer and Information Technology (ECTI-CIT), Vol. 8, no.
1, pp. 56-66, May. 2014.
[22] G. Tayfur, V. Singh, T. Moramarco and S. Barbetta, “Flood hydrograph
prediction using machine learning methods,” Water, Vol. 10, no. 8, pp. 968, Jul.
2018.
[23] C. Choi, J. Kim, J. Kim, D. Kim, Y. Bae and H. S. Kim, “Development of
heavy rain damage prediction model using machine learning based on big data,”
Advances in Meteorology, Vol.2018, pp. 1-11, May. 2018.
[24] J. A. Pollard, T. Spencer and S. Jude, “Big Data Approaches for coastal
flood risk assessment and emergency response,” Wiley Interdisciplinary
Reviews: Climate Change, Vol. 9, no. 5, pp. e543, Jul. 2018.
[25] N. Tkachenko, R. Procter and S. Jarvis, “Predicting the impact of urban
flooding using open data,” Royal Society open science, Vol. 3, no. 5, pp.
160013, May. 2016.
[26] K. Sene, “Flash floods: forecasting and warning,” Springer Science &
Business Media, 2012.
[27] World Meteorological Organization, “Manual on flood forecasting and
warning,” World Meteorological Organization (WMO), 2011.
[28] X. Chen, L. Zhang, C. J. Gippel, L. Shan, S. Chen and W. Yang,
“Uncertainty of flood forecasting based on radar rainfall data assimilation,”
Advances in Meteorology, Vol. 2016, pp. 1-12, Aug. 2016.
[29] L. Alfieri, M. Berenguer, V. Knechtl, K. Liechti, D. Sempere-Torres and
M. Zappa, “Flash flood forecasting based on rainfall thresholds,” Handbook of
Hydrometeorological Ensemble Forecasting, pp. 1223-1260, 2019.
[30] J. Ye, Y. Shao, and Z. Li, “Flood forecasting based on TIGGE
precipitation ensemble forecast,” Advances in Meteorology, Vol. 2016, pp. 1-9,
Nov. 2016.
[31] M. Santos and M. Fragoso, “Precipitation Thresholds for Triggering
Floods in the Corgo Basin, Portugal,” Water, Vol. 8, no. 9, pp. 376, Aug. 2016.
[32] F. Dottori, M. Kalas, P. Salamon, A. Bianchi, L. Alfieri and L. Feyen, “An
operational procedure for rapid flood risk assessment in Europe,” Natural
Hazards and Earth System Sciences, Vol. 17, no. 7, pp. 1111-1126, Jul. 2017.
[33] C. Li, X. Cheng, N. Li, X. Du, Q. Yu and G. Kan, “A framework for flood
risk analysis and benefit assessment of flood control measures in urban areas,”
International journal of environmental research and public health, Vol. 13, no.
8, pp. 787, Aug. 2016.
[34] D. Cane, S. Ghigo, D. Rabuffetti and M. Milelli, “Real-time flood
forecasting coupling different postprocessing techniques of precipitation
forecast ensembles with a distributed hydrological model, The case study of
may 2008 flood in western Piemonte, Italy,” Natural Hazards and Earth System
Sciences, Vol. 13, no. 2, pp. 211-220, Feb. 2013.
[35] Thailand Meteorological Department. TMD Big Data. Retrieved from
http://www.rnd.tmd.go.th/bigdata.php, 2019.
[36] Globalfloods. GLOFAS. Retrieved from http://www.globalfloods.eu
/accounts/login/?next=/glofas-forecasting/, 2019.
[37] D. D. Konadu and C. Fosu, “Digital elevation models and GIS for
watershed modelling and flood prediction–a case study of Accra Ghana.,” In
proc. Appropriate Technologies for Environmental Protection in the Developing
World Springer, Dordrecht, 2009, pp. 325- 332.
[38] J. García-Pintado, D. C. Mason, S. L. Dance, H. L. Cloke, J. C. Neal, J.
Freer, and P. D. Bates, “Satellite-supported flood forecasting in river networks:
A real case study,” Journal of Hydrology, Vol. 523, pp. 706-724, Mar. 2015.
[39] M. Shafapour Tehrany, F. Shabani, M. Neamah Jebur, H. Hong, W. Chen
and X. Xie, “GIS-based spatial prediction of flood prone areas using standalone
frequency ratio, logistic regression, weight of evidence and their ensemble
techniques. Geomatics,” Natural Hazards and Risk, Vol. 8, no. 2, pp. 1538-
1561, Aug. 2017.
[40] S. Ramly, W. Tahir and S. N. H. S. Yahya, “Enhanced Flood Forecasting
Based on Land-Use Change Model and Radar-Based Quantitative Precipitation
Estimation,” In proc. ISFRAM 2014, Springer, Singapore, 2015, pp. 305-317.
[41] R. E. Emerton, E. M. Stephens, F. Pappenberger, T. C. Pagano, A. H.
Weerts, A. W. Wood and C. A. Baugh, “Continental and global scale flood
forecasting systems,” Wiley Interdisciplinary Reviews: Water, Vol. 3, no. 3, pp.
391-418, Feb. 2016.
[42] Google, Google Map API. Retrieved from
https://developers.google .com/maps /documentation/javascript/tutorial, 2019.
[43] A. Pejic, S. Pletl and B. Pejic, “An expert system for tourists using Google
Maps API,” In proc. 2009 7th International Symposium on Intelligent Systems
and Informatics, IEEE, 2009, pp. 317-322.
[44] Y. Wang, G. Huynh and C. Williamson, “Integration of Google
Maps/Earth with microscale meteorology models and data visualization,”
Computers & Geosciences, Vol. 61, pp. 23-31, Dec. 2013.
[45] N. Mount, G. Harvey, P., Aplin and G. Priestnall, “Representing,
modeling, and visualizing the natural environment,” CRC Press, 2008.
[46] S. Nedkov and S. Zlatanova, “Google maps for crowdsourced emergency
routing,” In proc. XXII ISPRS Congress, Commission IV, Melbourne,
International Society for Photogrammetry and Remote Sensing (ISPRS), 2012,
pp. IAPRS XXXIX-B4.
[47] J .P. De Albuquerque, M. Eckle, B., Herfort and A .Zipf, “Crowdsourcing
geographic information for disaster management and improving urban resilience
:an overview of recent developmentsand lessons learned,” European Handbook
of Crowdsourced Geographic Information, 2016, pp. 309.
[48] J. Fohringer, D. Dransch, H. Kreibich and K.Schröter, “Social media as an
information source for rapid flood inundation mapping,” Natural Hazards and
Earth System Sciences, Vol. 15, no. 12, pp. 2725- 2738, Jul. 2015.
[49] G. Schimak, D. Havlik and J .Pielorz, “Crowdsourcing in crisis and
disaster management–challenges and considerations,” In proc. International
Symposium on Environmental Software Systems, Springer International
Publishing, 2015 .pp .56-70.
[50] T. Simon, A. Goldberg, and B .Adini, “Socializing in emergencies— A
review of the use of social media in emergency situations,” International Journal
of Information Management, Vol. 35, no. 5, pp. 609- 619, Oct. 2015.
[51] F .E .A. Horita, L .C. Degrossi, L .F .G. de Assis, A. Zipf and J .P. De
Albuquerque, “The use of volunteered geographic information )VGI (and
crowdsourcing in disaster management :a systematic literature review,” In proc.
the Nineteenth Americas Conference on Information Systems, 2013, pp .1-10.
[52] M. Zook, M. Graham, T. Shelton and S. Gorman, “Volunteered geographic
information and crowdsourcing disaster relief :a case study of the Haitian
earthquake,” World Medical & Health Policy, Vol. 2, no. 2, pp. 7-33, 2010.
[53] L. See, P. Mooney, G. Foody, L. Bastin, A. Comber, J. Estima and
H .Y .Liu, “Crowdsourcing, citizen science or volunteered geographic
information? The current state of crowdsourced geographic information,”
ISPRS International Journal of Geo-Information, Vol. 5, no. 5, pp. 55, May.
2016.
[54] P.-S. Yu, T.-C. Yang, S.-Y. Chen, C.-M. Kuo and H.-W. Tseng,
“Comparison of random forests and support vector machine for realtime radar-
derived rainfall forecasting,” J. Hydrol. Vol. 552, pp. 92– 104, Sep. 2017.
[55] M. Dehghani, B. Saghafian, F. Nasiri Saleh, A. Farokhnia and R. Noori,
“Uncertainty analysis of streamflow drought forecast using artificial neural
networks and Monte-Carlo simulation,” Int. J. Climatol, Vol. 34, pp. 1169–
1180, Aug. 2014.
[56] A.K. Lohani, N. Goel and K. Bhatia, “Improving real time flood
forecasting using fuzzy inference system,” J. Hydrol, Vol. 509, pp. 25– 41, Feb.
2014.
[57] M. Azam, H. San Kim and S. J. Maeng, "Development of flood alert
application in Mushim stream watershed Korea," International journal of
disaster risk reduction, Vol. 21, pp. 11-26, 2017.
[58] J. Xiaoming, H. Xiaoyan, D. Liuqian, L. Jiren, L. Hui, C. Fuxin and R.
Minglei, "Real-Time Flood Forecasting and Regulation System of Poyanghu
Lake Basin in China," EPiC Series in Engineering, Vol. 3, pp. 2368-2374, 2018.
[59] N. A. Sulaiman, N. F. Ab Aziz, N. M. Tarmizi, A. M. Samad and W. Z. W.
Jaafar, W. Z. W. "Integration of geographic information system (GIS) and
hydraulic modelling to simulate floodplain inundation level for bandar
segamat," In 2014 IEEE 5th Control and System Graduate Research
Colloquium, pp. 114-119, IEEE, 2014.
[60] S. Ghosh, S. Karmakar, A. Saha, M. P. Mohanty, S. Ali, S. K. Raju and P.
L. N. Murty, "Development of India's first integrated expert urban flood
forecasting system for Chennai," CURRENT SCIENCE, 117(5), 741-745, 2019.
[61] Q. Ran, W. Fu, Y. Liu, T. Li, K. Shi and B. Sivakumar, "Evaluation of
Quantitative Precipitation Predictions by ECMWF, CMA, and UKMO for
Flood Forecasting: Application to Two Basins in China," Natural Hazards
Review, Vol. 19, No. 2, pp. 05018003, 2018.
[62] E. Mayoraz, and E. Alpaydin, "Support vector machines for multiclass
classification," In International Work-Conference on Artificial Neural
Networks, pp. 833-842, Springer, Berlin, Heidelberg, 1999.
[63] E. R. Davies, "Computer vision: principles, algorithms, applications,
learning," Academic Press, 2017