Model Design and Simulation Analysis
Model Design and Simulation Analysis
Model Design
and Simulation Analysis
15th International Conference, AsiaSim 2015
Jeju, Korea, November 4–7, 2015
Revised Selected Papers
123
Communications
in Computer and Information Science 603
Commenced Publication in 2007
Founding and Former Series Editors:
Alfredo Cuzzocrea, Dominik Ślęzak, and Xiaokang Yang
Editorial Board
Simone Diniz Junqueira Barbosa
Pontifical Catholic University of Rio de Janeiro (PUC-Rio),
Rio de Janeiro, Brazil
Phoebe Chen
La Trobe University, Melbourne, Australia
Xiaoyong Du
Renmin University of China, Beijing, China
Joaquim Filipe
Polytechnic Institute of Setúbal, Setúbal, Portugal
Orhun Kara
TÜBİTAK BİLGEM and Middle East Technical University, Ankara, Turkey
Igor Kotenko
St. Petersburg Institute for Informatics and Automation of the Russian
Academy of Sciences, St. Petersburg, Russia
Ting Liu
Harbin Institute of Technology (HIT), Harbin, China
Krishna M. Sivalingam
Indian Institute of Technology Madras, Chennai, India
Takashi Washio
Osaka University, Osaka, Japan
More information about this series at http://www.springer.com/series/7899
Syng Yup Ohn Sung Do Chi (Eds.)
•
Model Design
and Simulation Analysis
15th International Conference, AsiaSim 2015
Jeju, Korea, November 4–7, 2015
Revised Selected Papers
123
Editors
Syng Yup Ohn Sung Do Chi
Korea Aerospace University Korea Aerospace University
Goyang-si Goyang-si
Korea (Republic of) Korea (Republic of)
Honorary Chairs
Jong-Hyun Kim Yonsei University, Korea
Axel Lehmann Universität der Bundeswehr München, Germany
Bo Hu Li Beihang University, China
Koji Koyamada Kyoto University, Japan
General Chair
Yun Bae Kim Sungkyunkwan University, Korea
General Co-chairs
Zhao Qinping Beihang University, President of the CASS, China
Satoshi Tanaka Ritsumeikan University, President of the ASIASIM and
the JSST, Japan
Organizing Committee
Seong-Yong Jang (Chair) Seoul National University of Science and Technology,
Korea
Byeong-Yun Chang (Chair) Ajou University, Korea
Jinsoo Park Yong In University, Korea
Chul-Jin Park Hanyang University, Korea
Publication Committee
Dong-Won Seo Kyunghee University, Korea
(Proceedings Chair)
Ja-Hee Kim Seoul National University of Science and Technology,
Korea
Sung-Do Chi Korea Aerospace University, Korea
(Special Issue Chair)
Yun-Ho Seo Korea University, Korea
(Special Issue Chair)
Jang-Se Lee Korea Maritime and Ocean University, Korea
Dong-Hyun Choi Korea Aerospace University, Korea
Sol Ha Seoul National University, Korea
Industrial Committee
Dug-Hee Moon (Chair) Changwon National University, Korea
Byung Hee Kim VMS SOLUTIONS Co., Ltd., Korea
Seung Whan Kim Siemens Industry Software Ltd, Korea
Young Suk Park ATWORTH CO., LTD., Korea
Ku-kil Chang Dassault Systems Korea, Korea
Young Gyo Chung SimTech Systems, Inc., Korea
Seong-Hoon Choi Sangmyung University, Korea
Award Committee
Hyung-Jong Kim (Chair) Seoul Women’s University, Korea
1 Introduction
moisture dynamics are very complex, and appropriate mathematical and/or numerical
models are needed for a practical comprehension of the system.
Moisture dynamics in porous media can be effectively described with Porous
Medium Equations (PMEs), an important class of degenerate nonlinear diffusion
equations based on the mass conservation principle and constitutive laws [2]. Similar
differential equations arise in other applied study areas, such as plasma physics [3] and
astrophysics [4]. Solving PMEs involves mathematical and computational difficulties
due to degeneracy and nonlinearity of the coefficients. Practical numerical methods for
solving PMEs should have sufficiently high accuracy, stability, and versatility. How-
ever, such methods are still not common and developing a simple numerical method
that is sufficient for practical use is an important research topic.
The main purpose of this article is to present a new numerical model of an extended
PME governing longitudinally one-dimensional (1D) moisture dynamics occurring in
non-woven fibrous sheets based on regularization, finite volume, and a differential
equation-based adaptive re-meshing technique. The proposed numerical model is
extensively verified through test cases with exactly or partially known solutions.
2 Mathematical Model
~t ¼ ðm pÞKs2 D1 2
x ¼ ðm pÞKs D1
s hs t; ~
1 ~ 1 2
s hs x; and Es ¼ ðm pÞ Ks Ds hs Es
2
ð2Þ
A Regularized Finite Volume Numerical Method 5
Fig. 1. The 1D domain of the extended PME (1) indicating the inclination a of the fibrous sheet.
In deriving (1), the pressure head wðuÞ and the permeability K ðuÞ have been
assumed to be physically parameterized as
Note that the pressure head reduces to the conventional case as p ! þ 0 [2, 6, 7].
The pressure head wðuÞ is continuous and increasing in 0\u 1, vanishes at u ¼ 1,
and diverges as u ! þ 0. The permeability K ðuÞ is continuous and increasing in
0 u 1 and vanishes at u ¼ 0. Figure 2(a) and (b) plot the pressure head wðuÞ and
the permeability K ðuÞ for several p and m, respectively where the coefficients Ks1 Ds hs
for wðuÞ and Ks for K ðuÞ are set to be 1 for the sake of simplicity. The angle a in (1)
may be spatially distributed, but is assumed to be constant in this article. There exist
two contrasting cases on arrangement of the sheets, which are the horizontal (a ¼ 0)
and vertical cases (a ¼ 0:5p). The advection term of (1) vanishes in the horizontal
case (sin a ¼ 0). The standard theory for moisture evaporation from the surface of a soil
body involves a near constant evaporation rate for most values of u. It is only as u
approaches zero that the evaporation rate rapidly decreases to zero [8]. This behavior
has been observed in our preliminary laboratory experiments using thin non-woven
fibrous sheets, which are not discussed here but will be addressed in a succeeding
paper. In the extended PME, this phenomenon occurs with q 1, which has actually
been qualitatively validated through preliminary laboratorial experiments. In addition,
the order of the parameters m and p have experimentally been estimated as O ¼ ð100 Þ
and O ¼ ð101 Þ, respectively, meaning that the relation p m 1 is satisfied for
moisture dynamics in non-woven fibrous sheets. Hereafter, “~” representing
non-dimensional variables are omitted for the sake of brevity.
Fig. 2. Plots of (a) the pressure wðuÞ and (b) permeability K ðuÞ. (Color figure online)
@u @ mp1 @u @
¼ ð m pÞ ½ he ð uÞ þ sin a ½he ðuÞm1 u Es uq ð4Þ
@t @x @x @x
pffiffiffiffiffiffiffiffiffiffiffiffiffiffi
using the regularization kernel he ðuÞ ¼ u2 þ e2 e with a small positive parameter
eð¼ 1010 Þ, which mathematically as well as numerically rules out the degeneration of
the advection and diffusion terms. Application of another conventional regularization
method to (1) remains as an option; however, such methods typically rely on trans-
formations of the solution u to a new dependent variable, which in general cannot
numerically realize mass conservation [7].
A numerical method for solving the regularized PME (4), which is referred to as the
Dual-Finite Volume Method (DFVM) [10], is briefly explained in this section.
The DFVM was originally developed for numerically solving Kolmogorov’s forward
equations, which are linear and conservative advection-diffusion-decay equations
(ADDEs), defined on connected graphs [11]. The DFVM concurrently uses primal and
dual computational meshes and evaluates the numerical fluxes based on analytical
A Regularized Finite Volume Numerical Method 7
and
PE u ¼ Es uq ; ð6Þ
where uðkÞ is the solution at the k th time step. Temporal integration in the evaporation
sub-steps is performed with local analytical solutions to the temporal ordinary differ-
ential equation at each node [12]. Temporal integration in the advection-diffusion
sub-step is performed with the conventional backward Euler method and the DFVM
spatial discretization, which is coupled with a Picard algorithm for handling the non-
linear advection and diffusion terms. Although the DFVM is theoretically uncondi-
tionally stable for linear problems, preliminary computation demonstrated that it is not
the case for the extended PME with advection term when Dt is sufficiently large. The
reason for this phenomenon would be the loss of ellipticity of the discretized system
with large Dt as implied in Pop et al. [9].
The present DFVM can be optionally equipped with an adaptive re-meshing
technique based on the Moving Mesh Partial Differential Equation (MMPDE) method
for accurately resolving sharp transitions of solution profiles [13]. The MMPDE
method solves an additional parabolic PDE that governs nodal positions; this is distinct
from the primary PDE to be solved, which in this article is the extended PME. An
advantageous point of the MMPDE method is that it does not alter the topology of the
mesh, which is not common to the other types of adaptive re-meshing methods.
The MMPDE method has successfully been used for numerically approximating
solutions to the Hamilton-Jacobi-Bellman equations governing upstream fish migration
in 1D river flows [14] and Kolmogorov’s forward equations in unbounded domains
governing probability density functions of stochastically-excited systems [15].
8 H. Yoshioka and D. Triadis
The present DFVM is extensively applied to test cases for its verification of accuracy,
stability, and versatility. Comparing performance with other published numerical
models is beyond the scope of this article, and will be discussed elsewhere. The test
cases considered in this article are (Sect. 4.1) Barenblatt problems, (Sect. 4.2) hori-
zontal infiltration subject to evaporation, (Sect. 4.3) inclined infiltration subject to
evaporation, (Sect. 4.4) support divorce problem in a horizontal sheet, and (Sect. 4.5)
downward infiltration in a vertical sheet.
Fig. 3. Comparisons of Barenblatt numerical and analytical solutions (a) over the entire domain
and (b) around the right front. (Color figure online)
8 sffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi !mq
2 9
< Es =
uð xÞ ¼ max 1 ðm qÞ x ;0 : ð8Þ
: 2mðm þ qÞ ;
The solution has bounded support for m [ q. As can be directly seen from (8), this
solution is also a weak solution since it is continuous but not differentiable at an edge of
the support. The focus here is whether the DFVM can capture the solution profiles and
the front position. The sheet is identified with the domain X ¼ ð0; 1Þ, which is uni-
formly discretized into 100 cells. The time increment is set as Dt ¼ 0:001. Steady
numerical solutions are computed starting from the initial guess u ¼ 0 over X. The
parameters Es , m, and q can be chosen such that the support of (8) becomes ½0; 0:5: We
have fixed the parameter q at 0:01, and shown solutions for m ¼ 2; 4; and 6, so that
Es ¼ 8:12; 8:06; and 8:04, respectively. The terminal time of computation is empir-
ically set as t ¼ 20 at which the numerical solutions are observed to be sufficiently
close to steady state.
Figure 4(a) compares numerical and analytical solutions for different values of m
without the MMPDE. The numerical solutions successfully approximate analytical
solutions for all m examined. Figure 4(b) provides an enlarged view of numerical and
10 H. Yoshioka and D. Triadis
Fig. 4. Horizontal infiltration subject to evaporation (a) without and (b) with the MMPDE.
(Color figure online)
analytical solutions around the front for m ¼ 4. Numerical solutions without the
MMPDE appear to perform better than those with it, indicating that using the MMPDE
with the DFVM may not be a good option for the problems with highly singular
evaporation term with small q. Qualitatively similar computational results have been
obtained for the other values of m ¼ Oð100 Þ.
where W1 ½ y denotes the second branch of Lambert W-function, which is valid for the
range expð1Þ y\1. To the authors’ knowledge, this implicit analytical solution has
not been derived elsewhere. The solution (9) can be verified by differentiating twice
with respect to x, using the well known result
dW1 ½ y W1 ½ y
¼ : ð10Þ
dy yð1 þ W1 ½ yÞ
The solution (9) has the bounded support ½0; z where z is the value of the right
hand-side of (9) with u ¼ 0. The domain X is specified as ð0; 2Þ. Parameter values are
specified as a ¼ 0:5p and p ¼ q ¼ 0:01, with four cases of ðEs ; mÞ examined;
ð3:01; 1Þ, ð3:01; 4Þ, ð5:01; 1Þ, and ð5:01; 4Þ, to see the effect that varying nonlinearity
of the advection-diffusion terms, and varying strength of the evaporation term have on
solution profiles. The domain is uniformly discretized into 200 cells and the time
increment is set as Dt ¼ 0:001. The terminal time of computation is empirically set as
t ¼ 20 at which the numerical solutions are observed to be sufficiently close to steady
state. The MMPDE method is not used since the previous problem showed that it does
not to successfully work for the problem with nonlinear evaporation.
Figure 5(a) and (b) compare the numerical and the analytical solution (9) for
stronger (Es ¼ 1) and weaker evaporative environments (Es ¼ 4). The agreement
between two is satisfactory in all cases. The computational results indicate that the
DFVM can handle steady infiltration problems in inclined sheets subject to nonlinear
evaporation. From the form of the exact solution it follows that the support of each
solution is not significantly dependent on the parameter m ¼ Oð100 Þ if 0\q 1, this
is demonstrated in Fig. 5.
Fig. 5. Comparisons of numerical and analytical solutions for inclined infiltration subject to
evaporation for Es ¼ 1 (a) and (b) Es ¼ 4.
and Wu [17]. The obtained results are comparable to those in Zhang and Wu [17] who
used a local DG method, which we consider to be a more complicated numerical
algorithm than the DFVM because a given computational mesh will have a higher
degree of freedom. Hence the computation results show that the DFVM can handle
transient problems having an evaporative source term without numerical instability.
Fig. 6. Numerical solution of the support divorce problem (a) in the x t phase space and (b) at
the time t ¼ 0:788. (Color figure online)
Es ¼ 0. The focus here is assessing whether the DFVM can capture infiltration fronts
where diffusion and advection terms would degenerate, potentially serving as severe
computational difficulties. The sheet is identified with the domain X ¼ ð0; 5Þ with the
boundary condition F ¼ 0 at x ¼ 0 and 5 where F is the flux associated with (4).
A delta-function initial condition is adopted: uð0; xÞ ¼ dð xÞ. The deltaic condition is
approximated as h1 where h is the size of the leftmost cell. The domain X is uniformly
divided into N regular cells and the time increment is set as N 1 . The values of N
examined are 100 2k for k ¼ 0; 1; 2; 3; 4. The MMPDE method is not used here
because preliminary computation indicated instability of numerical solutions just after
the time t ¼ 0, due to the deltaic initial condition.
Figure 7(a) and (b) compare numerical (N ¼ 200) and analytical solutions with
m ¼ 2 and m ¼ 3 at selected time steps at which the boundary effects at x ¼ 5 are
considered to be sufficiently small. The analytical solutions with m ¼ 2 and m ¼ 3
have unbounded and bounded supports, respectively, and present profiles with different
curvatures. They therefore have qualitatively different mathematical properties.
14 H. Yoshioka and D. Triadis
Fig. 7. Comparisons of numerical and analytical solutions to the downward infiltration problem
with (a) m ¼ 2 and (b) m ¼ 3. (Color figure online)
The figures show that the numerical solutions compare well with the analytical solu-
tions where the maximum values and the tails of the solutions are quite accurately
reproduced at each time step. The computational results with other values of N, not
presented in this article, implied almost first-order convergence of the numerical
solutions in both cases. This observation is consistent with the fact that the DFVM is
equivalent to a fully-upwind FVM for advection-dominant ADDEs [11].
5 Conclusions
A numerical model of the extended PME with a regularization method for the nonlinear
coefficients was proposed and applied to test cases whose solution behaviors are
known. The obtained computational results demonstrated its satisfactory accuracy,
stability, and versatility, indicating that it can potentially serve as an effective tool for
simulating complex moisture dynamics in non-woven fibrous sheets. There were
indications that the MMPDE method may not be a good option for problems with
highly singular evaporation term with small q. A new exact solution for extended
PMEs was derived in this paper, and was accurately reproduced with the DFVM.
A Regularized Finite Volume Numerical Method 15
The solution may be useful for verification of numerical methods and also for
parameter identification of non-woven fibrous sheets and porous media. Spatially 2-D
extension of the present mathematical and numerical models does not encounter
additional technical difficulties. Future research will address mathematical analysis of
the extended PME, on its regularity and analytical solutions in particular following
[25]. In addition, assessing consistency error between the original and regularized
extended PMEs will be addressed; we expect that the error can be a decreasing function
of the parameter e under an appropriate normed space as in Atkinson and Han [26].
Detailed numerical comparisons of a simplified extended PME with experimental
observations have been performed in Yoshioka et al. [1]. Their approach can be further
sophisticated with the help of the present numerical model for more realistic and
reliable assessment of moisture dynamics in non-woven fibrous sheets. Finally, the
presented regularization method and the DFVM can be applied to other degenerate
nonlinear diffusion equations arising in applied problems, such as Keller-Segel systems
for chemotaxis [27] and population dynamics models [28].
Acknowledgements. We thank Dr. Ichiro Kita and Dr. Kotaro Fukada in Faculty of Life and
Environmental Science, Shimane University, Japan for providing helpful comments and sug-
gestions on this article.
References
1. Yoshioka, H., Ito, Y., Kita, I., Fukada, K.: A stable finite volume method for extended
porous medium equations and its application to identifying physical properties of a thin
non-woven fibrous sheet. In: Proceedings of JSST2015, pp. 398–401 (2015)
2. Landeryou, M., Eames, I., Cottenden, A.: Infiltration into inclined fibrous sheets. J. Fluid
Mech. 529, 173–193 (2005)
3. Wilhelmsson, H.: Simultaneous diffusion and reaction processes in plasma dynamics. Phys.
Rev. A 38, 1482–1489 (1988)
4. Haerns, J., Van Gorder, R.A.: Classical implicit travelling wave solutions for a quasilinear
convection-diffusion equation. New Astron. 17, 705–710 (2012)
5. Broadbridge, P., White, I.: Constant rate rainfall infiltration: a versatile nonlinear model.
Analytic solution. Water Resour. Res. 24, 145–154 (1988)
6. Lockington, D.A., Parlange, J.Y., Lenkopane, M.: Capillary absorption in porous sheets and
surfaces subject to evaporation. Transport Porous Med. 68, 29–36 (2007)
7. Brooks, R.H., Corey, A.T.: Hydraulic properties of porous media. Colorado State
University, Hydrology Papers, Fort Collins, Colorado (1964)
8. Stewart, J.M., Broadbridge, P.: Calculation of humidity during evaporation from soil. Adv.
Water Resour. 22, 495–505 (1999)
9. Pop, I.S., Radu, F., Knabner, P.: Mixed finite elements for the Richards’ equation:
linearization procedure. J. Comput. Appl. Math. 168, 365–373 (2004)
10. Yoshioka, H.: On dual-finite volume methods for extended porous medium equations, arXiv
preprint. arXiv:1507.05281 (2015)
11. Yoshioka, H., Unami, K.: A cell-vertex finite volume scheme for solute transport equations
in open channel networks. Prob. Eng. Mech. 31, 30–38 (2013)
16 H. Yoshioka and D. Triadis
12. Li, Y., Lee, G., Jeong, D., Kim, J.: An unconditionally stable hybrid numerical method for
solving the Allen-Cahn equation. Comput. Math Appl. 60, 1591–1606 (2010)
13. Huang, W., Russel, R.D.: Adaptive Moving Mesh Methods, pp. 27–133. Springer,
Heidelberg (2011)
14. Yoshioka, H., Unami, K., Fujihara, M.: A Petrov-Galerkin finite element scheme for 1-D
tome-independent Hamilton-Jacobi-Bellman equations. J. JSCE. Ser. A2, 71 (in press)
15. Yaegashi, Y., Yoshioka, H., Unami, K., Fujihara, M.: An adaptive finite volume scheme for
Kolmogorov’s forward equations in 1-D unbounded domains. J. JSCE. Ser. A2, 71 (in press)
16. Li, H., Farthing, M.W., Dawson, C.N., Miller, C.T.: Local discontinuous Galerkin
approximations to Richards’ equation. Adv. Water Resour. 30, 555–575 (2007)
17. Zhang, Q., Wu, Z.L.: Numerical simulation for porous medium equation by local
discontinuous Galerkin finite element method. J. Sci. Comput. 38, 127–148 (2009)
18. Aronson, D.G., Caffarelli, L.A.: The initial trace of a solution of the porous medium
equation. Trans. Am. Math. Soc. 280, 351–366 (1983)
19. Fila, M., Vázquez, J.L., Winkler, M., Yanagida, E.: Rate of convergence to Barenblatt
profiles for the fast diffusion equation. Arch. Ration. Mech. An. 204, 599–625 (2012)
20. Rosenau, P., Kamin, S.: Thermal waves in an absorbing and convecting medium. Physica D
8, 273–283 (1983)
21. Nakaki, T., Tomoeda, K.: A finite difference scheme for some nonlinear diffusion equations
in an absorbing medium: support splitting phenomena. SIAM J. Numer. Anal. 40, 945–954
(2002)
22. Tomoeda, K.: Numerically repeated support splitting and merging phenomena in a porous
media equation with strong absorption. J. Math-for-Ind. 3, 61–68 (2011)
23. Tomoeda, K.: Numerical and mathematical approach to support splitting and merging
phenomena in the behaviour of non-stationary seepage. Theor. Appl. Mech. Jpn. 63, 15–23
(2015)
24. Hayek, M.: Water pulse migration through semi-infinite vertical unsaturated porous column
with special relative-permeability functions: exact solutions. J. Hydrol. 517, 668–676 (2014)
25. Vazquez, J.L.: The Porous Medium Equation. Oxford University Press, Oxford (2007)
26. Atkinson, K., Han, W.: Theoretical Numerical Analysis, pp. 449–451. Springer, Heidelberg
(2009)
27. Bian, S., Liu, J.G.: Dynamic and steady states for multi-dimensional Keller-Segel model
with diffusion exponent m > 0. Commun. Math. Phys. 323, 1017–1070 (2013)
28. Naldi, G., Cavalli, F., Perugia, I.: Discontinuous Galerkin approximation of porous
Fisher-Kolmogorov equations. Commun. Appl. Indust. Math. 4 (2013). doi:10.1685/journal.
caim.446
An Integrated Network Modeling for Road Maps
Zhichao Song1 ✉ , Kai Sheng2, Peng Zhang1, Zhen Li1, Bin Chen1, and Xiaogang Qiu1
( )
1
College of Information System and Management, National University of Defense Technology,
Changsha 410073, Hunan, People’s Republic of China
[email protected]
2
Electronic Engineering College, Naval University of Engineering, Wuhan 430033, Hubei,
People’s Republic of China
1 Introduction
Road network is an important part of environment models in both social simulation and
military simulation. Identifying the critical junctions or roads on the whole road map
and pay more attention to them is very helpful to assign traffic resources reasonably in
traffic simulations. As the traffic resources are always limited, more reasonable the
assignment is means that more important parts of the road network can be protected
better, so these important parts can be restored very soon once they are invalidated and
traffic on the road network can be more stable and more smooth.
Maybe it is because the roads linking with each other can be regarded as a graph or
network naturally. The issue that identifying key parts on road maps is usually dealt with
by graph theory or complex network theory. Urška et al. [1] researched how to identify
critical locations in a spatial network and extracted the key nodes in the street network
of Helsinki Metropolitan Area. Porta et al. [2] studied how centrality works in cities
based on the road graphs. Wu et al. [3] analyzed the topological bottleneck for traffic
networks.
Currently, there are three broadly used road network modeling methods: “Name
Street” method [4], primal graph approach [5] and dual graph approach [6]. In the street
network modeled by “Name Street” method, nodes represent named streets and edges
describe the junctions among these streets. The named-street-centred network reflects a
higher level of abstraction topological connections in a given street network and is very
suitable to analyze the topological character of the street network. Based on the network,
Jiang and Claramunt [4] gave their conclusions that the topology of street networks
reveals a small-world property but has no scale-free property. Whereas, this network
model does not fit the high-resolution critical-location identification in a road map. On
the one hand, the critical-location identification needs to consider geometrical informa‐
tion but lengths of the streets are not described in the model. On the other hand, a named
street is usually composed of many road segments and functions of these segments in a
whole road network should be different, so the critical location analysis in the network
can only give a large-scale location. In the dual graph of a road map, nodes represent
roads and edges represent junctions. This graph can provide a road-segment-level crit‐
ical-location analysis. However the length of each road segment still cannot be consid‐
ered. The critical-location analysis based on a dual graph is just a topological step-
distance concept. The primal graph of a road map is more comprehensive, objective and
realistic for the network analysis of centrality. In the graph, nodes are junctions and
edges are roads. This graph is usually a weighted graph because lengths of roads are
assigned to the edges. Therefore, the primal graph of a road map is the most suitable of
the three for the critical-location analysis on a road map.
Centrality indices, which are usually used to measure the importance of the node
in a network, are also used to identify the critical levels of the locations on a road
map. However, even based on the primal graph method, it still does not satisfy
enough to simultaneously calculate the critical levels of roads and junctions of a road
network by a measure of centrality once. Mostly, centrality scores of nodes are
calculated firstly and then the critical levels of edges are estimated by the scores of
nodes they connect directly. For example each edge is assigned a centrality score
equal to the average score of the pair of nodes on the both sides of it in literature [5]
and several kinds of “edge degree” are defined in literature [7]. In this method, crit‐
ical levels of edges which are estimated by nodes’ scores are totally depended on the
nodes on their both sides. It is in such a local scale that actual effect on the whole
network an edge has is not considered accurately.
To settle this simultaneous calculation for the centrality of nodes and edges, we
propose a new road network modeling method named “Integrated Graph Method” in
this paper. In the integrated graph of a road map, junctions and roads are all abstracted
as nodes and the connective relations are regarded as edges. The remainder of this paper
is organized as follows. In Sect. 2, the integrated graph method is introduced. In Sect. 3,
we present the key parts identification for a small road map. In Sect. 4, a real large-scale
case is given. Our conclusions are given in Sect. 5.
2 Method
We consider that no matter it is a road or a junction the critical locations on the road
map should be identified by one measure of centrality once. As we have introduced, the
road network modeling methods such as primal graph method, dual graph method or
“Named Street” method cannot meet our demands. Among the three methods, primal
graph is much more suitable for discovering critical elements of the network. However,
An Integrated Network Modeling for Road Maps 19
in primal graph network the importance of nodes and edges still cannot be simultane‐
ously calculated by one measure of centrality. Always the edges’ importance is estimated
by the centrality sores of nodes connecting the edge directly. In this paper, we integrate
the ideas of node presentation in primal graph and dual graph that both roads and junc‐
tions on a road map are represented by nodes. That is why the network modeling method
we proposed is named “Integrated Graph Method”.
Specifically, both roads and junctions on a road map are represented by nodes when
the road network is constructed. If a road is directly connected with a junction, there is
an edge between them and the weight of every edge is half of the road’s length. Based
on this weighted graph or network, the importance of roads and junctions can be
computed simultaneously by one measure of centrality. Figure 1 gives the different
(a) (b)
(c) (d)
Fig. 1. The different network models of a simple road map: (a) is the original road map; (b) is
the primal graph of the road map, in which nodes are junctions and lines are roads; (c) is the dual
graph of the road map, in which nodes mean roads and lines mean junctions; (d) is the network
modeled by integrated graph method, in which red nodes represent junctions and green nodes
represent roads (Color figure online)
20 Z. Song et al.
network models of a simple road map based on primal graph method, dual graph method
and the integrated graph method.
Being the importance measure methods, centrality indices are often used to identify key
parts on road maps. So far, there have been various measures of structural centrality,
which can be divided into three classes. The first-class measures are based on the idea
that the a central node in a network is near to the other nodes. The second-class measures
are based on the idea that central nodes stand between other node pairs on their paths of
connection. The third-class measure are the ones which combine the two main ideas of
centrality mentioned above. The four measures of centrality: degree centrality [8],
closeness centrality [9], betweenness centrality [10], and information centrality [11] are
four classic ones among the various measures of centrality. The four measures of
centrality cover the three classes mentioned above. Degree centrality is a local-scale
first-class measure, closeness centrality is the global-scale first-class measure, betwe‐
enness centrality is a representative one of the second-class measure, and information
centrality belongs to the third-class measure.
Here we apply the four classic measures of centrality namely degree centrality,
closeness centrality, betweenness centrality, and information centrality to calculate the
importance of nodes and edges in a simple hypothetic road network. To show the differ‐
ences among the four measures, we used a hypothetic road map. Figure 2 gives the
primal graph and the integrated graph of it. From the primal graph we can see that the
simple road map consists of 19 junctions that are labeled by “v1, v2, v3 …, v19” and
18 roads which are labeled by “e1, e2, e3 …, e18”.
The importance of roads and junctions on this road map is calculated by four meas‐
ures of centrality (degree centrality, closeness centrality, betweenness centrality, and
information centrality) based on the primal graph and the integrated graph. Since the
importance of an edge in the primal graph cannot be directly calculated by the node
centrality indices, it is quantified by the average centrality scores of the pair of nodes
on the both sides of it. As the centrality expressions we use are the normalized ones, the
results are limited between 0 and 1.
Figure 3 shows the centrality scores of junctions and roads calculated based on the
primal graph and the integrated graph of the road map. The x-axis is marked by the
sequence numbers of junctions and roads, in which 1–19 represent the nodes “v1, v2,
v3 …, v19” and 20–37 represent the edges “e1, e2, e3 …, e18”. The y-axis is marked
by the centrality scores. The scores are calculated by the four different measures of
centrality. In the integrated graph of road maps, importance scores of the roads are
evaluated by their impacts on the whole network and not only depending on the pair of
junctions on their both ends, so the scores of road nodes are much more reasonable than
the averages.
The centrality scores calculated in the integrated graph are universally lower than
those in the primal graph except the closeness. This is because the primal graph does
not consider the roads as nodes and the normalization factor is related to the node
An Integrated Network Modeling for Road Maps 21
(a)
(b)
Fig. 2. A simple hypothetic road network: (a) is the primal graph and (b) is the integrated graph
number. If the normalization factors applied to the two network graphs are the same,
the degree centrality sores of junctions in the two networks will be the same. Whereas
the importance of roads are not exact when degree centrality is used in the integrated
graph. We should not ascribe this to the road network model, and it is because the
shortage of degree centrality itself that degree centrality can only give a so local scale
evaluated result. For example, the importance of “v4” and “v5”, of which scores are
very high when closeness centrality, betweenness centrality and information centrality
are used, is not very obvious in Fig. 3(a). For the other three measures of centrality,
effects come from the normalization factor are not obvious. It is counteracted by the
decrease of node pairs’ distances for closeness centrality and the increase of the shortest
path for betweenness centrality. Moreover, the increase of the shortest path makes an
extra contribution to heighten the closeness scores. The normalization factor of infor‐
mation centrality is actually not affected by the node number. The decrease of informa‐
tion centrality scores in integrated graph is only due to the increase of network efficiency
caused by the new road nodes. The trend of the four centrality scores in the two graphs
is similar and the consistency of the closeness centrality scores is the best. The relative
importance of junctions does not change very much in the integrated graph. Especially
when betweenness centrality is used, importance of junctions in these two networks is
22 Z. Song et al.
0.5 0.5
Primal Graph Primal Graph
0.4 Integrated Graph 0.45 Integrated Graph
Centrality Scores
Centrality Scores
0.3 0.4
0.2 0.35
0.1 0.3
0 0.25
0 10 20 30 40 0 10 20 30 40
Sequence Number Sequence Number
(a) (b)
0.8 0.7
0.5
0.4
0.4
0.3
0.3
0.2
0.2
0.1 0.1
0 0
0 10 20 30 40 0 10 20 30 40
Sequence Number Sequence Number
(c) (d)
Fig. 3. Importance of junctions and roads calculated based on the two road networks by different
measures of centrality: (a) is calculated by degree centrality; (b) is calculated by closeness
centrality; (c) is calculated by betweenness centrality; (d) is calculated by information centrality
(Color figure online)
nearly the same (Fig. 3(c)). That is because the measure of betweenness centrality is the
least affected by the new nodes in integrated graphs.
From the comparison of the roads’ scores in the two network graphs, we can conclude
that the evaluation of junctions’ importance is hardly affected by the road nodes and the
calculation of road’s importance can be more exact when integrated graph method is
used, because the scores are calculated based on the contributions of the road nodes in
the whole network.
An Integrated Network Modeling for Road Maps 23
In this section, we will use the integrated graph method to model the Beijing’s main-
road map which is a large-scale road network and then simultaneously compute the
importance of junctions and road segments by the three classic global-scale centrality
namely closeness centrality, betweenness centrality and information centrality. As
shown in Fig. 4, there are 4192 junctions and 5166 road segments in the map. The line
only depicts the shape, location and length of a main road between two junctions. As
we have discussed, the road network should be modeled using the integrated graph
method before the importance of roads and junctions is computed. The network which
was built using the integrated graph method has 9358 vertices and 10332 edges and is
shown in Fig. 5, in which red vertices represent junctions and green vertices are roads.
Fig. 5. The integrated graph of the main road map of Beijing (Color figure online)
Based on the integrated network model, the importance of roads and junctions are
computed by closeness centrality, betweenness centrality and information centrality.
Degree centrality is not considered here, because it is a local-scale evaluated method
and is not suitable to evaluate the importance of road nodes in the integrated graph.
Actually, when the information centrality is used, the junctions’ importance is evaluated
by the group information centrality [11]. Because we consider that if a junction is inva‐
lidated the roads that directly connect it are always useless too.
The cumulative distributions of the three measures of centrality are shown in Fig. 6.
The cumulative distribution probability is defined by , where
N(C) is the number of nodes whose centrality scores equal to C and N is the number of
nodes in the whole network. The centrality scores in the figure are all renormalized by
dividing the maximum ones, so the maximum scores are 1. From it we can have conclu‐
sions as follows:
• Closeness centrality is mainly linear and betweenness centrality shows an obvious
exponential distribution. Information centrality which is the third-class centrality
measure also has an exponential character.
• Importance levels computed by closeness centrality are more uniform than the other
two but they do not cover the whole range 0–1. Betweenness centrality and
An Integrated Network Modeling for Road Maps 25
information centrality have a better resolving power for the nodes’ importance and
are both cover the whole range.
• We can see from the figure that the distribution curve of information centrality scores
locates between the closeness centrality and betweenness centrality. Maybe it is
because information centrality is the third-class measure of centrality which
combines the two ideas of closeness centrality and betweenness centrality.
1
Closeness
0.9 Betweenness
Information
0.8
0.7
0.6
P(C)
0.5
0.4
0.3
0.2
0.1
0
0 0.2 0.4 0.6 0.8 1
Importance Level
Fig. 6. The cumulative distributions of the three classic global-scale measures of centrality in
the integrated graph of Beijing’s main road map. (Color figure online)
Importance Level
0 0.5 1
Fig. 7. The geographic distribution of critical roads and junctions in Beijing’s main road map
(Color figure online)
5 Conclusions
This paper proposes a road network modeling method called integrated graph to settle
the simultaneous computation of the importance of roads and junctions on a road map
in traffic simulations. This method that both roads and junctions on a road map are
represented by nodes integrates the ideas of node presentation in primal graph and dual
graph. Based on this, the importance of junctions and roads can be evaluated by a
measure of centrality once.
The method has been validated by analyzing the importance of locations in a small
road network and Beijing’s main road network. The results show that our modeling
An Integrated Network Modeling for Road Maps 27
method of road networks is feasible and efficient. The calculation of a road’s importance
in the integrated network is based on its contribution to the whole network, and is more
reasonable than the evaluation based on the junctions on its both sides. Furthermore, the
integrated network modeling method makes the importance of roads and junctions
measured by the same standard, so the importance of roads and junctions can be
compared more accurately and clearly. These improvements are very helpful to identify
the critical junctions or roads on the whole road map in traffic simulations and realize a
more reasonable traffic resources assignment in traffic simulations.
Acknowledgements. The authors would like to thank National Nature and Science Foundation
of China under Grant Nos. 91024030, 91224008, 61503402, and 71303252.
References
1. Demšar, U., Špatenková, O., Virrantaus, K.: Identifying critical locations in a spatial network
with graph theory. Trans. GIS 12, 61–82 (2008)
2. Porta, S., Latora, V.: Centrality and cities: multiple centrality assessment as a tool for urban
analysis and design. In: New Urbanism and Beyond: Designing Cities for the Future, pp. 140–
145 (2008)
3. Wu, J., Gao, Z., Sun, H.: Topological-based bottleneck analysis and improvement strategies
for traffic networks. Sci. China Ser. E Technol. Sci. 52, 2814–2822 (2009)
4. Jiang, B., Claramunt, C.: Topological analysis of urban street networks. Environ. Plann. B
31, 151–162 (2004)
5. Porta, S., Crucitti, P., Latora, V.: The network analysis of urban streets: a primal approach.
Environ. Plann. B Plann. Des. 33, 705–725 (2005)
6. Porta, S., Crucitti, P., Latora, V.: The network analysis of urban streets: a dual approach.
Phys. A Stat. Mech. Appl. 369, 853–866 (2006)
7. Nieminen, J.: On the centrality in a graph. Scand. J. Psychol. 15, 332–336 (1974)
8. Holme, P., Kim, B.J.: Attack vulnerability of complex networks. Phys. Rev. E 65, 056109
(2002)
9. Latora, V., Marchiori, M.: Efficient behavior of small-world networks. Phys. Rev. Lett. 87,
198701 (2001)
10. Freeman, L.C.: Centrality in social networks conceptual clarification. Soc. Netw. 1, 215–239
(1979)
11. Latora, V., Marchiori, M.: A measure of centrality based on network efficiency. New J. Phys.
9, 188 (2007)
A Novel Flexible Experiment Design Method
1 Introduction
A critical property is the space-filling property, i.e., how the design points
distributed in the experiment space. Among various DOE methods, the Latin
Hypercube Design (LHD), which was proposed by McKay [3], is most used in
simulations [4]. It is always the concern in different variants of LHD design to
obtain better space-filling property. Park developed a row-wise element exchange
algorithm to obtain the optimal LHD [5]. Morris and Mitchell applied simulated
annealing algorithm to optimize the design [6]. Bates et al. employed the Genetic
algorithm to optimize LHD [7]. Although these variants present good perfor-
mance, the involved number of experiment trails is large. Felipe Viana et al.
proposed a fast optimal Latin Hypercube Design using Translational Propaga-
tion algorithm (TPLHD) [8]. TPLHD generates nearly optimal design instead of
the global optimal design. However, the number of experiment trials of TPLHD
is fixed for discrete experiment space, which is often insufficient to fully explore
the experiment space.
In this paper, an Extended TPLHD method (ETPLHD) is proposed to gen-
erate the design points flexibly. ETPLHD can generate more points in the design
space than traditional LHDs, which will be helpful in study of the target system.
It was shown in the comparison with other DOE methods that ETPLHD is effi-
cient to produce the experiment points, and achieved better evaluation results
about the studied system.
The rest of this paper is organized as follows. Section 2 gives a brief intro-
duction of TPLHD. Then the proposed method ETPLHD is described in detail.
Section 3 introduces the linear regression model for system approximation and
predication. Section 3.2 presents the comparative experiment between ETPLHD
and the other two DOE methods, and the results analysis is given. Finally, the
remarkable features about ETPLHD approach is concluded in Sect. 4.
2 Method
2.1 Review of Latin Hypercube via Translational
Propagation (TPLHD)
TPLHD method works in real time at the cost of finding the near optimal design
instead of the globally optimal design. Suppose a design space with nv variables,
each variable has np levels. The first step in TPLHD is to create a seed design that
contains ns points. This seed design is used as a pattern to fill the experiment
space iteratively. To fill the space, the experiment space is partitioned into nb
blocks firstly:
nb = np /ns (1)
The number of levels contained by each block is determined as follows:
The second step is to fill the seed design into each block. An example is
illustrated in Fig. 1. A seed design containing only 1 point is created in a 9 × 2
30 G. Zhai et al.
(two variables and each has 9 levels) design space. Then the original space is
divided into 9/1 = 9 blocks, and each block has 91/2 = 3 levels. As Fig. 1(a)
shows, the seed design is placed in the left-bottom block firstly. Then the seed
design is iteratively shifted by np /nd = 3 levels along one dimension until this
dimension is filled with the seed design, as Fig. 1(b) and (c) shows. Next, the
design along this dimension is adopted as a new seed design to fill along other
dimension until all blocks are filled, as shown in Fig. 1(d).
As all the key operations are to translate the design points, The TPLHD
method requires much less computation compared to other optimal methods
which normally need to search the best design among all (np !)nv LHD designs.
A criterion parameter φp , is used to measure the space-filling quality of the
experiment design. For an experiment space, a smaller φp indicates that the
created experiment points are better in distribution to fill up the space.
np −1
np
φp = [ d−p
ij ]
1/p
(3)
i=1 j=i+1
A Novel Flexible Experiment Design Method 31
where p is a pre-selected integer value and dij is the distance between any two
design points xi , xj :
nv
dij = d(xi , xj ) = [ |xik − xjk |t ]1/t (4)
k=1
(a) A TPLHD design (b) Translate within 1 (c) 4 layer of 16×2 space
layer
and then perform the experiment design in a lower dimensional subspace, until
the TPLHD method is applied.
Figure 2 illustrates this process taking a design in 16 × 3 experiment space.
First, one dimension (denoted as the 1d ) of the experiment space is chosen and
partitioned into 4 parts according to Eq. (5). 4 intervals along this dimension are
determined, as shown in Fig. 2(a). Next, a 16 × 2 TPLHD designs (the dimen-
sions are denoted as 2d and 3d ) is conducted in each interval. By this way, the
ETPLHD generates 64 points. For comparison, TPLHD obtains a design with
16 points.
In this example, the experiment points within low-dimensional subspace need
to be expanded to the higher dimension. Generally, if the points set xim has been
obtained within a subspace of m dimensions, the design points can be expanded
to the higher dimension as follows:
where k(1 ≤ k ≤ dm+1 ) is the index of the interval along dimension m + 1. The
expansion is repeated until all dimensions are included. As for the example in
Fig. 2, the experiment points form the slanted layer along the dimension 1d , as
shown in Fig. 2(c).
After the expansion to higher dimension, some points may stay in a small
region of the experiment space. Consider the example in Fig. 2, an experi-
ment point [x0 , y0 ] designed in the 2d × 3d subspace will generates a series
experiment points along dimension 1d by taking the expansion operation:
[x0 , y0 , z0 ], [x0 , y0 , z3 ], [x0 , y0 , z7 ], [x0 , y0 , z11 ]. From the direction of the dimen-
sion 1d , they are lines. This problem will be worse when the dimension grows.
As a result, it is necessary to adjust the distribution of the experiment points
before expansion to the higher dimension. Assuming the current dimension is m,
then the points are shifted p(0 ≤ p ≤ dm ) unit along the 2 ∼ m dimension, where
dm is the number of the intervals along dimension m. Equation (7) is changed as
follows:
xki(m+1) = (xim + p)mod d + (k − 1) · d (8)
A Novel Flexible Experiment Design Method 33
There are two issues should be noted. First, the value of p is normally deter-
mined empirically. Second, a control parameter s can be defined to specify how
many dimensions need to apply the ETPLHD method. For a n × m space, s = 2
means only two dimensions apply the ETPLHD method and the rest m−2 dimen-
sions will apply the TPLHD method. By the control parameters, the number of
experiment points determined by ETPLHD is:
s
+1
n∗ = np · ds = np2 (9)
Figure 3 demonstrates the shift within the 2d × 3d subspace before expansion
to dimension 1d .
The goal of DOE method is to evaluate the target system with less experi-
ment trails. Many approximation methods such as the linear regression, Kriging
model, neural nets, support vector regression and so on are often employed to
evaluate and then predict the target system. The linear regression is the most
used approach among them10 .
A first-order polynomial can be given by
y = Xβ + e (10)
where y = (y1 , . . . , yn ) is the predicted value of the target system with n experi-
ments. X = (xij )(i = 1, . . . , n, j = 1, . . . , q) is the experiment data recorded in n
trails, where i in the index of experiment trail, and j is the index of data within
each trail. β = (β1 , . . . , βq ) is the regression coefficients; and e = (e1 , . . . , en )
denotes the residuals in each experiment.
A Novel Flexible Experiment Design Method 35
Let w = (w1 , . . . , wn ) be the true output of the simulation system, the Sum
of Squared Residuals(SSR) is given by Eq. (11)
β = (X X)−1 X w (12)
The accuracy of the model is normally measured by the mean square error
(MSE ), which is the accurate estimate of the true error of the prediction model.
Five parameters of SAM missile are adopted as the design variables and each
has 4 levels, as shown in Table 2.
TPLHD and ETPLHD are tested in this scenario for comparison. In
ETPLHD, s = 3 produces 32 experiment points. Each experiment point is tested
for 5 times to get the mean value of the output for further analysis. The linear
regression approach is employed to obtain an approximated system model that
can be used to predict the output of the combat. Table 3 demonstrates a part
of the results and the predictions of the approximated models based on the two
DOE methods.
A Novel Flexible Experiment Design Method 37
(a)TPLHD
No X1 X2 X3 X4 X5 y ypred
1 1 1 1 1 1 8.2 10.2520
2 3 1 1 3 2 17.2 18.1212
3 4 2 2 2 3 14.4 13.6219
4 2 2 4 4 4 27 28.7731
· · · · ··
32 4 4 4 4 4 26.2 25.2967
(b)ETPLHD
No X1 X2 X3 X4 X5 y ypred
1 3 2 1 1 1 8 8.2456
2 3 3 4 2 1 15.6 17.7121
3 2 1 1 3 2 20.4 19.7148
4 4 1 3 2 3 17.2 16.0479
· · · · ··
32 4 3 1 1 3 8 6.7885
Given the data in Table 3, the MSE of the two methods can be calculated
according to Eq. (13). The MSE of ETPLHD is 0.6141, which is much less than
that (1.0533) of TPLHD. This result indicates that the approximated system
model using the ETPLHD method performs better than the model using TPLHD
method.
20 random points are selected from the experiment space to verify the effec-
tiveness of the approximated models. Table 4 demonstrates a part of the results.
ya is the prediction by the approximated model that uses the TPLHD method,
while yb is the predication by the model that uses the ETPLHD method.
The plots of the simulation values and two prediction values are shown in
Fig. 5. The MSE of the two predication is 4.0147 and 3.217 respectively, indi-
cating that the prediction model with ETPLHD is more precise.
38 G. Zhai et al.
No X1 X2 X3 X4 X5 y ya yb
1 3 1 3 3 4 24.2 22.1132 22.0739
2 2 3 2 2 1 17 14.1160 15.4181
3 4 3 1 2 1 13.2 10.4750 11.6806
4 1 3 1 3 1 21.2 18.4052 19.7464
· · · · ··
20 4 4 2 1 2 8 7.0334 7.9351
4 Conclusion
In order to break the limitation of the fixed number of design points in tra-
ditional LHD method, a new flexible DOE method ETPLHD is proposed in
this paper. The comparison shows that ETPLHD method performs better than
TPLHD for no more than 6 variables. The ETPLHD method is also applied
in a complex combat simulation to help to find proper approximation model
of the combat system, which is used to predict the outcomes of the combat
A Novel Flexible Experiment Design Method 39
with different parameter configurations of the SAM missiles. The results show
that the approximation model using the ETPLHD method has better predicting
accuracy than the model using the TPLHD method.
References
1. Sanchez, S.M., Wan, H.: Work smarter, not harder: a tutorial on designing and
conducting simulation experiments. In: Proceedings of the 2012 Winter Simulation
Conference, pp. 1–15 (2012)
2. Ghosh, S.P., Tuel, W.G.: A design of an experiment to model data base system
performance. IEEE Trans. Softw. Eng. SE-2(2), 97–106 (1976)
3. McKay, M.D., Beckman, R.J.: A comparison of three methods for selecting values
of input variables from a computer code. Technometrics 21(2), 239–245 (1979)
4. Kelton, W.D.: Designing simulation experiments. In: Proceedings of the 1999 Win-
ter Conference, pp. 33–38 (1999)
5. Park, J.S.: Optimal Latin-hypercube designs for computer experiments. J. Stat.
Plann. Inference 30(1), 95–111 (1994)
6. Morris, M.D., Mitchell, T.J.: Exploratory designs for computational experiments.
J. Stat. Plann. Inference 43(3), 381–402 (1995)
7. Bates, S.J., Sienz, J., Toropov, V.V.: Formulation of the optimal Latin hyper-
cube design of experiments using a permutation genetic algorithm. In: 45th
AIAA/ASME/ASCE/AHS/ASC Structures, Structural Dynamics and Materials
Conference, AIAA-2004-2011, Palm Springs, California, pp. 19–22 (2004)
8. Viana, F.A.C., Venter, G., Balabanov, V.: An algorithm for fast optimal latin
hypercube design of experiments. Int. J. Numerical Methods Eng. 82(2), 135–156
(2009)
9. Ye, K.Q., Li, W., Sudjianto, A.: Algorithmic construction of optimal symmetric
latin hypercube designs. J. Stat. Plann. Inference 90, 145–159 (2000)
10. Ma, Y., Gong, G.: A research on CGF reasoning system based on Fuzzy Petri
Net. In: Proceedings of System Simulation and Scientific Computing, Beijing, pp.
406–411 (2005)
Modeling Behavior of Mobile Application
Using Discrete Event System Formalism
Yun Jong Kim, Ji Yong Yang, Young Min Kim, Joongsup Lee,
and Changbeom Choi ✉
( )
1 Introduction
According to a new market research report published by ‘VisonMobile’, the total global
mobile applications market is expected to be worth US$58.0 billion by the end of 2016.
The market of the mobile application is growing fast and user’s requirement is also
becoming diversified. To further, there have been many studies and attempts to find
solutions that reflects the user’s needs efficiently.
One of the biggest challenges in mobile application development is that the require‐
ments for mobile application change frequently and rapidly [1]. Many developers have
difficulties to keep the code up to date and it takes a lot of time. To deal with this problem,
rapid prototyping has been used [2]. In this methodology, research and development
stages are conducted in parallel to another when creating the prototype [3]. This enables
developer and designer to acquire precise feedback and analyze user requirements by
providing high-fidelity prototypes.
In order to develop a high-fidelity prototype, the developer should consider the
behavior of mobile application. The typical behavior of mobile application can be
in turn categorized into temporal behavior and non-temporal behavior. The temporal
behavior does not require users’ input, but instead it is triggered by external sources.
Display of animation on mobile screen and pop-up message from mobile applica‐
tions are such examples. On the other hand, non-temporal behavior is the event that
requires user’s input. Screen touch, swipe and long touch are some examples of this
behavior. We have applied discrete event system specification formalism to monitor
these behaviors.
This paper contains five sections. Section 2 describes the related works concerning
the making of high-fidelity mobile application prototype. Section 3 introduces how to
model the temporal behavior and non-temporal behavior using DEVS formalism, which
sets the theoretical background of our study. Furthermore, we explain the overall struc‐
ture of the simulation using DEVS formalism. Section 4 introduces the case study with
the proposed approach and we conclude our study in Sect. 5.
2 Related Works
Mobile application can be grouped by the three categories: native, web-based, and hybrid
[4, 5]. All three types of mobile applications needs prototyping process under the devel‐
opment in common. To support the process, many prototyping tools currently exist in
the market.
The table below shows the limits of existing tools. We compared four well-known
prototyping tools with our proposed approach. There are several prototyping tools that
support making of high-fidelity prototype. These tools simulate the mobile application
by using the expected behavior information embedded in the pre-designed image. Note
the existing tools only support the non-temporal behavior. Also, it is not possible to
discrete each event with formal representation for the user to obtain additional infor‐
mation about their prototype (Table 1).
Table 1. The comparison between existing prototyping tools and proposed system
Marvel POP Fluid Axure Proposed system
Modeling non- O O O O O
temporal behavior
Modeling temporal X X X X O
behavior
Formal representation X X X X O
The DEVS formalism supports the structure of a system in modular and hierarchical
point of view. There are two kinds of DEVS model: one is atomic model and the other
is coupled model [6]. While a part of system can be represented as an atomic model, a
system represented in DEVS coupled model shows how a system can be coupled
together and also how they interact with each other.
42 Y.J. Kim et al.
The coupled model is composed of various models. This may express larger system
by taking an atomic model or a subordinated coupled model as a child model. A coupled
model ‘CM’ is defined as:
The Atomic model is the most basic module in the hierarchical structure. ‘Mi’ in the
CM corresponds to the atomic model. It describes the system behavior. An atomic model
‘M’ is defined as follows:
Mobile application can be viewed as the collection of various components and this
can be expressed by using DEVS formalism models. To explain in detail, we use the
term ‘Activity’ in Android, which is an application component that provides a screen
for the users to interact. As shown in Fig. 1, one activity consists of many mobile
Fig. 1. The composition of mobile application activity with DEVS formalism model
Modeling Behavior of Mobile Application 43
shown in the animation. After TR seconds, the state goes back to the initial state and the
cycle is repeated.
The execution model sends output message to implement specific event after given time,
Tn. Note the amount of time taken is noted as Tn. Tn corresponds to the number that
the event data includes. Figure 4 demonstrates the details of Execution model.
There are two states: WAIT, EVENT. The initial state is WAIT. Once the message
is arrived from the behavior atomic model, the state changes to EVENT. Consequently,
the model turns itself back into WAIT state through the internal transition and sends
message allowing the mobile OS to operate.
Modeling Behavior of Mobile Application 45
4 Case Study
In this section, we introduce two specific events which can be expressed by using the
behavior models we have proposed earlier. Note various features of simulation using
the behavior model can be seen from the case study. The events that we simulate are
twofold: (1) Activity conversion by button touch and (2) Image conversion with the
lapse of time.
Activity conversion by touch button is one of the events that is triggered by the user’s
input. Figure 6 describes the event with DEVS formalism models. This event can be
expressed as the coupled model, which comprises the non-temporal behavior model and
46 Y.J. Kim et al.
the execution model. The execution model of non-temporal behavior enables the conver‐
sion to occur and as the result, the image displayed on screen changes into the upper-
left image shown in Fig. 6.
The image conversion event in Fig. 7 requires current activity as an input source.
Because the direct user input is unnecessary for this event, the event in Fig. 7 can be
expressed as the coupled model, which comprises the temporal behavior model and the
execution model. The execution model takes some time as temporal behavior model,
and then changes the screen image into the upper-right image shown in Fig. 7.
Modeling Behavior of Mobile Application 47
5 Conclusion
Mobile application system is a highly hierarchical and time-based process [7]. Because
of these characteristics, DEVS formalism is one of the appropriate ways to model mobile
application system. However, there has been no attempt to express mobile application
workings through DEVS.
This study simulates the mobile application by stating the mobile app behavior in
DEVS models and introduces how to model the mobile app behavior. We classified by
the type of the mobile app behavior; and made models of such behavior using the DEVS
atomic model. By combining the behavior atomic model and coupled model we can state
the various mobile application events.
The approach that we proposed represents the temporal behavior which is not
possible to state in existing prototyping tools. This helps to implement the prototype
with high fidelity. In addition, the mobile app simulation using DEVS model enables to
accumulate formal data. This means that the simulation environment that we are
proposing can in turn be used to test usability by analyzing the accumulated data. Many
other industries utilize the data from the prototype to analyze the task-based usability
48 Y.J. Kim et al.
test [8]. Moreover, the formal method, which is ensured by DEVS based simulation
environment, has a mathematical basis [9]; thus with the formal method, the user can
specify, develop, and verify the prototype in systematic manner.
However, the implementation framework still does not exist; hence the user must
develop the entire simulator model from the bottom up. Also, simulator heavily depends
on platform because of the absence of simulation engine. For further research topics we
may look into simulation engine and finding ways to implement the device-independent
framework.
References
1. Joorabchi, M.E., Mesbah, A., Kruchten, P.: Real Challenges in Mobile App Development. In:
International Symposium on Empirical Software Engineering and Measurement, IEEE
computer society, pp. 15–24 (2013)
2. Adelmann, R., Langheinrich, M.: SPARK rapid prototyping environment – mobile phone
development made easy. In: Tavangarian, D., Kirste, T., Timmermann, D., Lucke, U., Versick,
D. (eds.) IMC 2009. CCIS, vol. 53, pp. 225–237. Springer, Heidelberg (2009)
3. Tripp, S.D., Bichelmeyer, B.: Rapid prototyping: an alternative instructional design strategy.
ETR&D 38(1), 31–44 (1990). ISSN 1042-1629
4. Masi, E., Cantone, G., Mastrofini, M., Calavaro, G., Subiaco, P.: Mobile apps development: a
framework for technology decision making. In: Uhler, D., Mehta, K., Wong, J.L. (eds.)
MobiCASE 2012. LNICST, vol. 110, pp. 64–79. Springer, Heidelberg (2013)
5. “Native, web or hybrid mobile-app development,” IBM Software, Thought Leadership
WhitePaper. http://www.computerworld.com.au/whitepaper/371126/native-web-or-hybrid-
mobile-app-development/download/
6. Zeigler, B.P.: Multi-facetted Modeling And Discrete Event Simulation. Academic Press, San
Diego (1984)
7. Zeigler, B.P., Chow, A.C.H.: Parallel DEVS, hierarchical, modular, modeling formalism
(1994)
8. Stone, B., Wang, Y.: AirportLogic: usability testing, prototyping, and analysis of an airport
wayfinding application. In: Stephanidis, C. (ed.) Posters, Part II, HCII 2011. CCIS, vol. 174,
pp. 81–84. Springer, Heidelberg (2011)
9. Seo, K.-M., Choi, C., Kim, T.G., Kim, J.H.: DEVS-based combat modeling for engagement-
level simulation. Simul. Trans. Soc. Model. Simul. Int. 90(7), 759–781 (2014)
Multi-modal Transit Station Planning Method
Using Discrete Event System Formalism
1 Introduction
resources. Moreover, it should spend a huge budget and time to fix such problems after
the station has been built. Modeling and simulation is one of the alternatives to help the
planner to design multi-modal transit station before the construction. By using modeling
and simulation, an urban planner may consider the traffic flow and volume of the
passenger flow before building the multi-modal transit station. Also, the planner may
consider the geographical layout of access roads to the transit station and the behavior
of the individual passenger to predict the traffic flow and volume of the passenger flow.
Modeling and simulation of urban planning is nothing new to urban planning domain.
There were several studies that utilize the modeling traffic and the layout of the road in
the cities and simulate them to acquire data to find the congestion point. Especially, a
traffic simulation model is widely used to analyze the alternatives of transportation
system without directly changing the real world, such as constructing additional road or
modulating traffic signals [1, 2]. In order to capture such behaviors of traffics, passengers,
and pedestrian, two modeling and simulation methods have been proposed: macro
simulation model and micro simulation model. Both models were used to capture the
characteristics of the traffic of a transportation system. However, the previous studies
were used to predicting the traffic situations. Also, to the best of the authorsʼ knowledge,
there are no studies that confirm whether the design of multi-modal transit station satis‐
fies the requirements or not.
In this paper, we propose Discrete Event System (DEVS) Formalism based modeling
method to capture the behavior of the vehicle and passenger. Also, we propose the
simulation environment to check the transit station model against to the requirement.
The DEVS Formalism is set theoretical formalism which has hierarchical structure and
modular feature. First, we use model of the DEVS Formalism, the DEVS model, to
model spatial layout of the multi-modal transit station in order to capture the behavior
of the vehicles and passengersʼ movements. Especially, we model the spatial layout by
connecting DEVS model consecutively. Also, we model temporal and non-temporal
behaviors of the vehicle and passenger as discrete events from previous DEVS model
to next DEVS model. Also, we model the interaction between vehicle and passenger by
exchanging discrete event from DEVS model to DEVS model.
Second, we propose simulation environment that supports various requirement
specification methods to check the model against to its requirements. In this paper, we
mainly utilize requirement specification method based on the propositional logic.
However, in order to support various requirement specification methods, we unify the
modeling method and the requirement specification method using DEVS model. The
model for requirement specification continuously monitors the event sequences from
model of the multi-modal transit station, and checks the temporal and non-temporal
requirements. A preliminary version of this paper appeared in AsiaSim ʼ15. In the
previous paper, there are no details in modeling method and simulation environment.
This paper is an extended and improved version of the preliminary version. The rest of
this paper is organized as follows. Section 2 presents related works. In Sect. 3 introduces
a modeling method for multi-modal transit station, and Sect. 4 presents the simulation
environment for multi-modal transit station. Finally, Sect. 5 shows the case study of our
proposed method and we conclude this paper.
Multi-modal Transit Station Planning Method Using DEVS Formalism 51
In this section, we introduce the previous researches in modeling and simulation in urban
planning domain and modeling and simulation theories. The previous researches can be
categorized by macro simulation and micro simulation. The macro simulation model
captures the characteristics of transportation system in macroscopic point of view rather
than modeling individual elements. On the other hand, the micro simulation model
models each individual transportation element to analyze the interaction among various
transportation elements.
The macro simulation model abstracts the type of the vehicles into the differential equa‐
tions and captures the macroscopic characteristics, such as aggregated traffic flow
dynamics and density of the traffic [3, 4].
A macro simulator has advantage on understanding traffic flows, but hard to predict
behavior of each object. For example, a macro simulation model DYNEMO was
designed for the development, evaluation and optimization of traffic control systems for
motorway networks [4]. Traffic flow model included with the simulation package
combines the advantages of a macroscopic model (computational simplicity) with the
advantages of a microscopic model (output statistics relating to individual vehicles).
On the other hand, most studies about micro simulations were concentrated on indi‐
vidual vehicles, crossroads or traffic signal based simulations. For example, Errampalli
et al. has provided a basis for evaluating the public transport policies such as public
transport priority systems by developing microscopic traffic simulation model [2]. They
have considered two types of public transport policies for evaluation. One is bus lanes
and another is public transport priority systems at traffic signal. Fuzzy logic reasoning
has been incorporated in route choice analysis while choosing the route based on the
level of satisfaction of the available routes. The developed simulation model has been
applied on the part of Gifu city network and it is able to predict the vehicular movements
with a fair amount of accuracy. From this study, it can be concluded that the developed
microscopic simulation model can be applied to evaluate public transport polices partic‐
ularly bus lane and public transport priority systems at intersections with fair amount of
accuracy.
Among various modeling and simulation theories in micro simulation, cellular
automata and agent based simulation are the representative theory that model the trans‐
portation system and simulate them [5]. Especially, Zamith et al. has introduced a flex‐
ible and robust traffic cellular automata model [5]. They consider the motion expectation
of vehicles that are in front of each driver. Then, they define how a specific vehicle
decides to get around, considering the foreground traffic configuration. The model
utilizes stochastic rules for both situations, using the probability density function of the
beta distribution to model three drivers behavior, adjusting different parameters of the
beta distribution for each one. The three drivers are as follows. First is Hunter and
Tailgater that describes the behavior of a driver that intends to stay close to the upfront
vehicle and wants to go fast. Second is Ultraconservative that describes the driver with
52 J. Choi et al.
the opposite profile of the Hunter and Tailgater. This driver prefers to keep a great
distance to the upfront vehicle and drives slower. Third is Flow Conformist that
describes a driver that intends to go according with the flow and is not characterized by
none of the previous profiles. It models only driversʼ behavior and it means they focus
on cars.
Above mentioned previous studies have solid theoretical background to model the
transportation network or the transportation systems. Our approach is similar to the
microscopic simulation. Especially, we utilize the cellular automata to model the spatial
layout and agents to denote the elements of the transportation system. However, our
approach is differentiated by utilizing cellular automata and agents to model the multi-
modal transit station. In our approach we focus to the spatial layout of the multi-transit
station and we model the vehicles and the passengers as agents. Also, we model the
boarding in the transit station by modeling the interaction among the vehicles and the
passengers.
where
X: a set of external input event types,
Y : an output set,
S: a sequential state set,
δext : Q × X → S, an external transition function
where Q is the total state set of
M = {(s, e)|s ∈ S and 0 ≤ e ≤ ta(s)},
δint : S → S, an internal transition function,
λ: S → Y , an output function,
ta: S → R+ 0,∞ , a time advance function, where the R0,∞
+
internal events, the output specified by the output function λ is produced based on the
state s, which means the output function is processed before the internal transition func‐
tion. Then, as before, a new ta(sʹ) is computed, and the elapsed time e is set to zero.
A Coupled Model is defined as follow:
where
D: a set of component names,
For each i in D,
Mi : a component basic model
(an atomic or coupled model),
Ii : a set of influences of i,
and for each j in Ii ,
Zi,j : Yi → Xj , an i-to-j output translation,
SELECT : 2M − φ → M , a tie-breaking selector.
In this section we introduce modeling method for multi-modal transit station. A multi-
modal transit station is complex system consist of different types of area. For example,
a modeler should consider the transportation region and passenger region. Following
Fig. 5 shows compositions of general multi-modal transit station. As shown in the
Fig. 5, when a group of vehicles arrives to the station it is categorized by transportation
type and it flows in to the waiting lane. After that, it waits on the waiting lane until the
vehicle picks up a passenger, then it leaves the transit station through the exit roads. On
the other hand, when a passenger arrives to the multi-modal transit station, a passenger
54 J. Choi et al.
decides the transportation method and move to the right platform. Therefore, a passenger
uses the crosswalk to change the platform and a vehicle should wait for the passenger
to cross the road (Fig. 1).
A spatial layout of multi-modal transit station is one of the reason that causes traffic
congestion, so it is one of the important consideration for designing multi-modal transit
station. In order to model a spatial layout of the station, we divide a region into cells and
each cell represents the characteristics of the region. For instance, a single lane which
can hold 10 vehicles can be modeled as 10 cells. Then, each cell represents one vehicle
and a modeler can model the behavior of the single lane by array of cells that models
the characteristics of the single lane. Modeling a spatial layout of the multi-modal transit
station using cells have two merits. First, a modeler can decide the resolution of the
simulation by choosing the number of cells. When a modeler uses massive cells to model
the transit station, the model can simulate the movements of passengers and vehicles in
micro scale with high computation power. On the other hand, when a modeler wants to
see the result quickly with the proper resolution, the modeler can reduce the number of
cells. Therefore, a modeler first select a region in the multi-modal transit station, and
Multi-modal Transit Station Planning Method Using DEVS Formalism 55
selects the proper simulation resolution with respect to the simulation objective. Finally,
each sub-region is mapped to a cell to model the region. Following Fig. 2 shows the
representation of a road using cells. As shown in the Fig. 2, if a modeler wants to model
a road, the modeler should decide the resolution of the simulation. In this case, five cells
represent a sub-region of the given road with 10 m.
To model the region by dividing smaller region to model the behaviors of the sub-
region, we adopted the cellular automata theory [13]. Therefore, we can model the road
for a vehicle and a passenger by cells that interact with each other.
In the real world, a passenger and a vehicle in the multi-modal transit station show certain
tendency to move from one place to other place. They are under controlled by rules
which are given from the society. For example, if we want to model a multi-modal transit
station with taxi, a passenger and a vehicle should follow following rules: “A taxi cannot
overtake the preceding taxi in the single lane.” and “When there is a passenger waiting
in front line, a passenger from backside of the line should not take a taxi”. Such rules
are the propositional statements that the simulation models of the multi-modal transit
station must be followed. A proposition is a declarative sentence that is either true or
false, but not both [14]. Also, we can develop a proposition by connecting various prop‐
ositions using operators, such as conjunction, disjunction, and negation. As a result, we
can develop a complex proposition by combining propositions to give requirement
specification to a given system. Such propositions should always be true during the
simulations. As a result, an urban planner can use propositional logic to check the viola‐
tion of the requirement specifications by monitoring the events during the simulation,
since some events sequence may lead the system into violation status, i.e., some circum‐
stance that the given propositions are not true.
56 J. Choi et al.
Also, the behaviors of the vehicle and the passenger can be captured by automata-
theoretic approach with respect to the propositional logic [15]. Following Fig. 3 shows
the simple single lane road system.
The number of cells in the figure denotes the length of the road. There are two cells
in this model, so the length of the transportation system is two. Each cell can have two
states: filled or empty. Each state in the automata shows the circumstance of the trans‐
portation system, for example, state denotes there is one agent is in the cell A, and
A|B denotes there are two agents in the system. Also, four events are defined: ϵ, In,
and Each event denotes, no events are occurred, an agent has arrived to
the system, an agent have departed from cell A, and an agent have departed from cell
B, respectively.
Since we can consider the passenger line as a queue, we can use automata to represent
the behavior of the passengers. Also, since there are two lanes, one for the taxis and
other for the passengers, we should cross product two automata to represent the
combined behavior between taxis and passengers. Figure 4 shows the cross product of
the two single lane transportation system. As shown in the figure, all transitions are the
valid behaviors in the multi-modal transit station. Therefore, a user can detect the viola‐
tion of the requirement specifications by monitoring the transitions of the transportation
system.
Building a simulation model for multi-modal transit station is not an easy task, especially
to the urban planners. In this section we introduce the simulation environment of the
multi-modal transit station. First, we explain the key simulation model that models the
spatial layout of the multi-modal transit station. Then, we introduce the verification
method of the simulation environment.
road, it moves forward in their lane. In order to capture such semantics, we adopted one-
dimensional cellular automata using DEVS formalism. Figure 7 shows the layout of
each road model.
Each agent in the road model checks the state of the next cell and decides whether
to move or stop. After the cells agent moves, the state of the cell is changed. This infor‐
mation is delivered to previous cell, and previous cell decides whether to go or stop.
Therefore, the information propagates backward. The DEVS formalism is suitable to
adopt this backward propagation mechanism.
The Fig. 8 shows the Coupled Model and Atomic Model which each cell checks the
state of the next cell and decides whether to send agent to next cell or not. The difference
between vehicle road model and human road model is the agent which enters to the
model.
Figure 9 shows the Traffic Manager Model of the road model. The Traffic Manager
Model generates the agents to the road, and urban planner may control the traffic flow
by changing parameters in the model. The Traffic Manager Model has two model the
Poll Manager model and Agent Generator model.
The Agent Generator Model generates agent objects and sends it to first road cell.
Each road cell has Poll Manager Model and Transfer Manager Model as shown in
Fig. 8. When a Poll Manager Model sends state information of the cell to corresponding
Transfer Manager Model when the state is changed. If a Poll manager gets this acknowl‐
edgement message, the road cell delivers agent object to next cell or matches the taxi
and human. On the other hand, it will be blocked because other agent is existing in the
front cell.
The verification model has two models, the Validity Check Model and Invalidity
Check Model. The Validity Check Model continuously monitors the event sequences
from the transit station model to check the transition model that generates valid events.
On the other hand, the Invalidity Check Model only check the invalid events that cannot
be existed through the simulation. For example, the Invalidity Check Model check the
events that lead to the abnormal states, which is matching has been occurred at the back
of the line despite empty vehicle exist at the front or customer has ridden to the vehicle
despite the customer at the front did not ride a vehicle.
This section introduce a case study of taxi station of Pohang KTX (Korean Train
eXpress) Station in South Korea. The Pohang KTX station has opened in 2015, and
many of the passenger had suffered terrible traffic jam while using transit station. One
of the reason for traffic jam was the Taxi station, so we have built an abstracted simu‐
lation model to model the taxi station of Pohang KTX station. The main objective of the
simulator is to check the fairness of the taxi station system. In order to check the fairness
of the system, we first specify the requirement specifications for the taxi and the passen‐
gers. Following propositions are the requirements of the tax station.
Proposition 1. A taxi cannot overtake the preceding taxi in the single lane.
Proposition 2. A passenger must be served in a First Come First Serve (FCFS) fashion.
Based on the propositions, a passenger should be served as FCFS fashion and the
Taxi should leave the transit station one by one. By simulating the model, we have found
the following situations: (1) ideal case and (2) abnormal case. The ideal case, every
passenger and the taxi are matched in FCFS fashion, as shown in the left part of the
Fig. 12.
62 J. Choi et al.
In Fig. 12, the white circle denotes the passenger and the black circle denotes the
taxi. However, in the abnormal case as shown in the right part of the Fig. 12, a passenger
can be matched to a taxi, although there exists a passenger in the line. This abnormal
case happens when the passenger does not recognize the line. Since the traffic light
prevents the taxi to enter the single lane, a passenger, who arrived to the taxi station,
may think he/she is the first passenger in the line, so he/she will take a taxi although
there exists a passenger in the line. We have also analyzed the alternative solution of
the taxi station. Figure 13 shows the alternative solution, which is installing fence. In
order to prevent random matching at the taxi station, a fence can be installed to the transit
station. Fortunately, the decision maker of the Pohang KTX station has installed a fence
to prevent a passenger to take a taxi anywhere. Figure 14 shows the initial phase of the
taxi station and current taxi station of Pohang KTX station.
6 Conclusion
References
1. Hernndez, J., Ossowski, S., Garca-Serrano, A.: Multiagent architectures for intelligent traffic
management systems. Transp. Res. Part C Emerg. Technol. 10(56), 473–506 (2002)
2. Errampalli, M., Okushima, M., Akiyama, T.: Microsopic simulation model considering public
transport policy. J. Eastern Asia Soc. Transp. Stud. 6, 2718–2733 (2005)
3. Papageorgiou, M.: Concise encyclopedia of traffic & transportation system: simulation
program, pp. 491–502 (1991)
4. Schwerdtfeger, T.: Dynemo: a model for the simulation of traffic flow in motorway networks.
In: The 9th International Symposium on Transportation and Traffic Theory, pp. 65–87. VNU
Science Press (1984)
5. Zamith, M., et al.: A new stochastic cellular automata model for traffic flow simulation with
driversʼ behavior prediction. J. Comput. Sci. 9, 51–56 (2015)
6. OBrien, E.J., Lipari, A., Caprani, C.C.: Micro-simulation of single-lane traffic to identify
critical loading conditions for long-span bridges. Eng. Struct. 94, 137–148 (2015)
7. Zeigler, B.P.: Multifacetted Modeling and Discrete Event Simulation. Academic Press,
London (1984)
8. Zeigler, B.P.: Object-Oriented Simulation with Hierarchical, Modular Models. Academic
Press, London (1990)
9. Zeigler, B.P., Kim, T.G., Praehofer, H.: Theory of Modeling and Simulation. Academic Press,
London (2000)
10. Song, H.S., Kim, T.G.: Application of real-time DEVS to analysis of safety critical embedded
control systems: railroad-crossing control example. Simul. Trans. Soc. Model. Simul. Int.
81(2), 119–136 (2005)
11. Hong, J., Seo, K.M., Kim, T.G.: Simulation-based optimization for design parameter
exploration in hybrid system: a defense system example. Simul. Trans. Soc. Model. Simul.
Int. 89(3), 362–380 (2013)
12. Seo, K.M., et al.: DEVS-based combat modeling for engagement-level defense simulation.
Simul. Trans. Soc. Model. Simul. Int. 90(7), 759–781 (2014)
64 J. Choi et al.
13. Wolfram, S.: Theory and Applications of Cellular Automata, vol. 1. World Scientific,
Singapore (1986)
14. Rosen, K.: Discrete Mathematics and Its Applications, 7th edn. McGraw-Hill Education, New
York (2012)
15. Vardi, M.: An automata-theoretic approach to linear temporal logic. In: Moller, F., Birtwistle,
G. (eds.) Logics for Concurrency. LNCS, vol. 1043, pp. 238–266. Springer, Heidelberg
(1996)
16. Tag Gon Kim, DEVSim++ Manual. http://smslab.kaist.ac.kr
Reliability Analysis of Software Defined Wireless
Sensor Networks
1 Introduction
Nowadays, as an important technique in the next generation wireless and mobile
networks, the wireless sensor network (WSN) is becoming more and more com-
plex and demanding [6,7,17]. It has been widely used in many areas, for exam-
ple military, navigation and environmental monitoring [15]. Typical WSNs are
self-configuring networks [10,15], and their topologies are frequently changed
[11]. This feature of WSNs poses the demand for dynamic network manage-
ment. Therefore, software-defined wireless sensor network (SDWSN), which is
a combination of WSN and software-defined networking (SDN), is proposed
[5,10,11,15,18,23].
SDN is a new paradigm of networking system that provides dynamic net-
work management and configuration. Why SDN is needed? Traditional networks
consist of many devices for forwarding and processing data packets [21]. It is
extremely time-consuming to deploy and maintain traditional networks, because
the packet routing control functions are located in routers, and they need to be
updated locally in most cases [11]. To remedy this limitation, SDN is proposed.
It separates the control plane and the data plane. The data plane composed
by physical network devices is responsible to perform packet forwarding and
processing functions. The control layer is in charge of the whole network man-
agement [22]. The applications hosted in the application layer can program the
devices in data plane [14].
c Springer Science+Business Media Singapore 2016
S.Y. Ohn and S.D. Chi (Eds.): AsiaSim 2015, CCIS 603, pp. 65–78, 2016.
DOI: 10.1007/978-981-10-2158-9 6
66 N. Gong and X. Huang
Similar to SDN, SDWSN aims at separating the control plane from the
data forwarding plane in WSNs. These days, the dramatically raising applica-
tion of the wireless and sensing technology expands the requirement for the
dynamic, flexible and secure management of WSN. Hence, to satisfy these
needs, the Software-Defined Wireless Sensor Network (SDWSN) is proposed
[5,6,11,15,18,23]. Compared with the traditional WSN, SDWSN has several
advantages:
– The network becomes programmable. This allows for more dynamic network
segmentation and utilization without changes in other devices [12].
– The centralized controller maintains a global view of the whole network. Thus,
security and flexibility of SDWSN could be improved.
However, one major concern about SDWSN is its reliability. Because SDWSM
has a central controller, it may becomes the bottleneck of the whole system. If
it fails, the whole system will be down. However, so far, few researchers study
the reliability of the SDWSN which however is not a negligible issue. This paper
therefore aims to study the reliability of the SDWSN by using the probabilistic
model checking technique with the tool PRISM.
The reminder of this paper is organized as follows. In Sect. 2, SDWSN is
described. The probabilistic model checking technique is explained in Sect. 3.
Section 4 describes the modeling of the SDWSN structure. Section 5 analyzes
the results. Finally, Sect. 6 provides the conclusion.
where Xn stands for the state of the system at time n and P(Xn = xn ) is the
corresponding probability. The probability distribution of system state at time
n + 1 (the future state) depends only on the state at time n (the current state)
and does not relate with the states at time (1, n) (the past states).
A continuous-time Markov chain (CTMC) M is a four-tuple (S, s, R, L).
Assume each component of the system has n states, then S is a finite set of
states where S = {s0 , s1 , s2 , . . . , sn } and s ∈ S; s is the initial status of the
component; R is the transition rate matrix of |S| × |S|, and L is the labeling
function with atomic propositions [2]. Figure 3 is an example of CTMC:
S = {s0 , s1 , s2 , s3 , s4 };
s = s0 ;
Reliability Analysis of Software Defined Wireless Sensor Networks 69
⎡ ⎤
0 0.5 0.2 0.3 0 0
⎢ 0 0 0 0 0.7 0 ⎥
⎢ ⎥
R=⎢
⎢0 0 0 0 0 0 ⎥;
⎥
⎣ 0 0 0 0 0 0.2 ⎦
0 0 0 0 0 0.2
L(s0 ) = {OK}, L(s1 ) = L(s3 ) = L(s4 ) = {}, L(s2 ) = {Failed}, L(s5 ) = {Waiting}.
3.3 PRISM
PRISM is a popular probabilistic model checking tool. It now mainly sup-
ports four types of probabilistic model: discrete-time Markov chains (DTMCs),
continuous-time Markov chains (CTMCs), Markov decision processes (MDPs)
and Probabilistic timed automata (PTAs) [8].
The model in PRISM is specified using the PRISM modeling language. Each
System is described as modules. Each module contains its local variables which
is used to indicate the possible states. The status transition behavior in modules
is represented in command. For example:
[] x=0 -> 0.8:(x’=0) + 0.2:(x’=1);
This command claims that when the variable x has value 0, the update state
will remain at 0 with probability 0.8 and change into 1 with probability 0.2.
Additionally, the module uses cost and reward to permit reasoning about the
quantitative measure of the system such as the expected power consumption and
expected time.
PRISM’s property specification language is based on temporal logics, namely,
PCTL, CSL, probabilistic LTL and PCTL∗ [13]. It mainly has three operators:
– P operator: this is used to reason about the probability of an event occurrence,
e.g. what is the probability of y = 6:
P=? [y = 6].
– S operator: this refers to the steady-state behavior of the model, e.g. what is
the long-run probability of the queue being more than 75 % full:
S=? [q > 0.75].
– R operator: this is used for reasoning about the expected cost of the model.
For example: what is the expected number of lost requests within 15.5 time
units
R=? [C <= 15.5].
4.2 Implementation
In the formal model, two modules (controller layer and WSN layer) are realized.
The detail structures of modules are as follows:
– Module 1 - Controller layer: Two controllers are c1 and c2 . The states are:
both two controllers are working (m = 2), only one is working (m = 1) and
fail (m = 0). Each controller has the same failure rate λc . The transition rate
of the controller layer whose initial status is m = 2 to the state m−1 is m×λc .
– Module 2 - WSN layer: At least two sensors are considered in this module.
Additionally, the number of sensors can be manually changed. Each sensor
has the same failure probability λa .
To verify the dependability of this model, the following failure types are
considered:
– Controller layer fails if none of controllers is working.
– WSN fails if none of sensors is working.
– System fails if the controller layer fails.
– System fails if the number of working sensors is less than the lower bound.
– System fails if the skipping cycle is greater than the upper limit. (The con-
troller can automatically skip the current operation cycle when the application
or WSN layer is abnormal, then a Max COUNT is set to limit the skipping.)
5 Results
The main results are presented in this section. The factors of influencing the
reliability of SDWSN are also analyzed.
Fig. 4. The performance of WSN layer in 24 h with different number of sensor nodes.
The y-axis is the failure probability of the WSN layer; and the x-axis is the duration
(24 h). The number of sensor nodes is represented in the right; e.g., “5 sensors” means
that the WSN layer contains 5 sensors. The failure rate of each sensor node is once a
day.
72 N. Gong and X. Huang
Fig. 5. The performance of WSN layer in 30 days with different number of sensor nodes.
The y-axis is the failure probability of the WSN layer; and the x-axis is the duration
(30 days). The number of sensor nodes is represented in the right; e.g., “5 sensors”
means that the WSN layer contains 5 sensors. The failure rate of each sensor node is
once a day.
Fig. 6. The performance of WSN layer in 24 h with different number of sensor nodes.
The y-axis is the failure probability of the WSN layer; and the x-axis is the duration
(24 h). The number of sensor nodes is represented in the right; e.g., “5 sensors” means
that the WSN layer contains 5 sensors. The failure rate of each sensor node is once per
30 days.
In Fig. 4, the failure rate of the WSN layer is shown. Its setting is:
– Failure rate of each sensor: once per day;
– Duration: 24 h;
– The number of sensors: 5, 10, 15, and 20.
As shown in Fig. 4, increasing the number of sensors can decrease the failure
probability of the entire WSN layer. However, if we set the duration to 30 days,
this strategy will be challenged. The result is shown in Fig. 5. From this figure,
we can see that the failure probability of the WSN plane is 100 % after 4 days.
Reliability Analysis of Software Defined Wireless Sensor Networks 73
Fig. 7. The performance of WSN layer in 30 days with different number of sensor nodes.
The y-axis is the failure probability of the WSN layer; and the x-axis is the duration
(30 days). The number of sensor nodes is represented in the right; e.g., “5 sensors”
means that the WSN layer contains 5 sensors. The failure rate of each sensor node is
once per 30 days.
Fig. 8. The performance of WSN layer in 24 h with different failure rate of individual
sensor nodes. The y-axis is the failure probability of the WSN layer; and the x-axis is
the duration (24 h). The failure rate of sensor nodes is represented in the right; e.g.,
“s-1/2” means that the failure rate is once per 2 days. The number of sensor nodes is 5.
This suggests that when the failure probability of individual sensor node is high
(once per day), increasing the number sensor nodes is not a good way of enhanc-
ing the reliability of the WSN layer.
In Figs. 6 and 7, the failure rate of each sensor in the WSN plane is changed
to once per month. Compared with Figs. 4 and 5, the failure probability of the
WSN plane is significantly lower. This depicts that decreasing the failure rate of
each sensor node can decrease the failure probability of the sensor layer.
In order to study the relation between the sensor node failure rate and the
WSN layer failure rate, we did several new experiments. In Figs. 8 and 9, we can
74 N. Gong and X. Huang
Fig. 9. The performance of WSN layer in 30 days with different failure rate of individual
sensor nodes. The y-axis is the failure probability of the WSN layer; and the x-axis is
the duration (30 days). The failure rate of sensor nodes is represented in the right; e.g.,
“s-1/2” means that the failure rate is once per 2 days. The number of sensor nodes is 5.
Fig. 10. The performance of the whole system (SDWSN) in 24 h with different failure
rate of individual sensor nodes. The y-axis is the failure probability of the system; and
the x-axis is the duration (24 h). The failure rate of sensor nodes is represented in the
right; e.g., “s-1/2” means that the failure rate is once per 2 days. The number of sensor
nodes is 5. The failure rate of the controller is?
see that lower sensor node failure rate is corresponding to lower WSN failure
rate. In Figs. 10 and 11, we can see that lower sensor node failure rate can also
enhance the system (SDWSN) reliability given a fixed controller failure rate.
Except the sensor nodes, the system reliability is also related to controllers’
status. We have the following findings:
Reliability Analysis of Software Defined Wireless Sensor Networks 75
Fig. 11. The performance of the whole system (SDWSN) in 30 days with different
failure rate of individual sensor nodes. The y-axis is the failure probability of the
system; and the x-axis is the duration (30 days). The failure rate of sensor nodes is
represented in the right; e.g., “s-1/2” means that the failure rate is once per 2 days.
The number of sensor nodes is 5. The failure rate of the controller is once per 30 days
Fig. 12. The performance of the whole system (SDWSN) in 24 h with different failure
rate of one controller. The y-axis is the failure probability of the system; and the x-axis
is the duration (24 h). The failure rate of the single controller is represented in the
right; e.g., “c-1/2” means that the failure rate is once per 2 days. The number of sensor
nodes is 5. The failure rate of the sensor node is once per 30 days.
– From Figs. 12 and 13, we can see that reducing the failure rate of each con-
troller can enhance the system reliability.
– From Figs. 12 and 14, we can see that increasing the number of controllers can
also enhance the system reliability.
76 N. Gong and X. Huang
Fig. 13. The performance of the whole system (SDWSN) in 30 days with different
failure rate of one controller. The y-axis is the failure probability of the system; and the
x-axis is the duration (30 days). The failure rate of the single controller is represented
in the right; e.g., “c-1/2” means that the failure rate is once per 2 days. The number
of sensor nodes is 5. The failure rate of the sensor node is once per 30 days.
Fig. 14. The performance of the whole system (SDWSN) in 24 h with different failure
rate of the two controllers. The y-axis is the failure probability of the system; and the
x-axis is the duration (24 h). The failure rate of the two controllers is represented in
the right; e.g., “c1c2-1/2” means that the failure rate is once per 2 days. The number
of sensor nodes is 5. The failure rate of the sensor node is once per 30 days.
6 Conclusion
In this paper, to verify the reliability of SDWSN, a probabilistic model has
been realized based on CTMC using PRISM. After experiments, the results
illustrate the impacts of controllers and sensors on the network performance.
Reliability Analysis of Software Defined Wireless Sensor Networks 77
Firstly, increasing the number of controllers and sensors can enhance the system
reliability. Secondly, the lower failure rate of controllers and sensors makes the
network more dependable. Moreover, these two factors are more cognizable in
long term.
Nevertheless, the limitation of the assumptions inevitably results in obstacles
to find more potential correlation among components in SDWSN. For example,
the relationship between the controller and sensor is omitted and this will be
considered in the further research.
Acknowledgment. This work has been supported in part by the XJTLU RDF140243,
by the Natural Science Foundation of Jiangsu Province under Grant BK20150376, and
by the Suzhou Science and Technology Development Plan under Grant SYG201516.
References
1. Ali, S.T., Sivaraman, V., Radford, A., Jha, S.: A survey of securing networks using
software defined networking. IEEE Trans. Reliab. 64(3), 1086–1097 (2015)
2. Aziz, A., Sanwal, K., Singhal, V., Brayton, R.: Model-checking continuous-time
Markov chains. ACM Trans. Comput. Logic (TOCL) 1(1), 162–170 (2000)
3. De Mil, P., Jooris, B., Tytgat, L., Hoebeke, J., Moerman, I., Demeester, P.: snap-
Mac: a generic MAC/PHY architecture enabling flexible MAC design. Ad Hoc
Netw. 17, 37–59 (2014)
4. Ding, A.Y., Crowcroft, J., Tarkoma, S., Flinck, H.: Software defined networking
for security enhancement in wireless mobile networks. Comput. Netw. 66, 94–101
(2014)
5. Granelli, F., Gebremariam, A.A., Usman, M., Cugini, F., Stamati, V., Alitska, M.,
Chatzimisios, P.: Software defined, virtualized wireless access in future wireless
networks: scenarios and standards. IEEE Commun. Mag. 53(6), 26–34 (2015)
6. Han, Z.J., Ren, W.: A novel wireless sensor networks structure based on the SDN.
Int. J. Distrib. Sens. Netw. 2014 (2014)
7. Harrop, P., Das, R.: Wireless sensor networks 2010-2020. Networks 2010, 2020
(2010)
8. Hinton, A., Kwiatkowska, M., Norman, G., Parker, D.: PRISM: a tool for automatic
verification of probabilistic systems. In: Hermanns, H., Palsberg, J. (eds.) TACAS
2006. LNCS, vol. 3920, pp. 441–444. Springer, Heidelberg (2006)
9. Hunkeler, U., Lombriser, C., Truong, H.L., Weiss, B.: A case for centrally controlled
wireless sensor networks. Comput. Netw. 57(6), 1425–1442 (2013)
10. Jacobsson, M., Orfanidis, C.: Using software-defined networking principles for wire-
less sensor networks. In: 11th Swedish National Computer Networking Workshop
(SNCNW), 28–29 May 2015, Karlstad, Sweden (2015)
11. Jagadeesan, N.A., Krishnamachari, B.: Software-defined networking paradigms in
wireless networks: a survey. ACM Comput. Surv. (CSUR) 47(2), 27 (2014)
12. Kirkpatrick, K.: Software-defined networking. Commun. ACM 56(9), 16–19 (2013)
13. Kwiatkowska, M., Norman, G., Parker, D.: Probabilistic model checking in prac-
tice: case studies with PRISM. ACM SIGMETRICS Perform. Eval. Rev. 32(4),
16–21 (2005)
14. Lin, Y.-D., Lin, P.-C., Yeh, C.-H., Wang, Y.-C., Lai, Y.-C.: An extended SDN
architecture for network function virtualization with a case study on intrusion
prevention. IEEE Netw. 29(3), 48–53 (2015)
78 N. Gong and X. Huang
15. Luo, T., Tan, H.-P., Quek, T.Q.S.: Sensor openflow: enabling software-defined wire-
less sensor networks. IEEE Commun. Lett. 16(11), 1896–1899 (2012)
16. Menshikov, M., Petritis, D.: Explosion, implosion, and moments of passage times
for continuous-time Markov chains: a semimartingale approach. Stoch. Process.
Appl. 124(7), 2388–2414 (2014)
17. Oualhaj, O.A., Kobbane, A., Sabir, E., Ben-othman, J., Erradi, M.: A ferry-assisted
solution for forwarding function in wireless sensor networks. Pervasive Mob. Com-
put. (2015)
18. Xu, R., Huang, X., Zhang, J., Lu, Y., Wu, G., Yan, Z.: Software defined intelligent
building. Int. J. Inf. Secur. Priv. (2016)
19. Steiner, R.V., Mück, T.R., Fröhlich, A.A.: C-MAC: a configurable medium access
control protocol for sensor networks. In: 2010 IEEE Sensors, pp. 845–848. IEEE
(2010)
20. Van Dam, T., Langendoen, K.: An adaptive energy-efficient MAC protocol for
wireless sensor networks. In: Proceedings of the 1st International Conference on
Embedded Networked Sensor Systems, pp. 171–180. ACM (2003)
21. Wood, T., Ramakrishnan, K.K., Hwang, J., Liu, G., Zhang, W.: Toward a software-
based network: integrating software defined networking and network function vir-
tualization. IEEE Netw. 29(3), 36–41 (2015)
22. Xie, J., Guo, D., Zhiyao, H., Ting, Q., Lv, P.: Control plane of software defined
networks: a survey. Comput. Commun. 67, 1–10 (2015)
23. Lu, Y., Huang, X., Huang, B., Xu, W., Zhang, Q., Xu, R., Liu, D.: A study on
the reliability of software defined wireless sensor network. In: 2015 IEEE Interna-
tional Conference on Smart City (IEEE Smart City 2015), Chengdu, China, 19–21
December 2015
A Path-Line-Based Modeling for Fishing
Prediction
1 Introductions
Global warming is causing climate change, which increases the variance in environ-
mental variables, such as temperature, salinity and velocity. Climate change results in a
change of locations of the habitats of fish because the habitats of fish are always looking
for a favorable environment. For fishery grounds modeling, such an environment has
been represented as a function of environmental variables. If the number of the habitats
and the environmental variables can be measured, we can construct a predictive model
for the population.
There is a strong need for the development of a predictive model for fishing to
maintain steady fish catches. Fishing is under greater risk due to global warming
because the fisherman’s tacit knowledge may be useless. To avoid wasting fuel oil in
fishing, the fishermen understand that they should have a fishery prediction technique.
The fishermen asked the academic community to develop a highly accurate predictive
model for their fishery. They recorded fish catches and made the results open to the
community.
Recently, fishermen have recorded their fish catches when they are on their fish
boats for future reference. The catches are represented using a unit named the catch per
unit effort (CPUE). In fisheries, the CPUE is an indirect measure of the abundance of a
target species. Changes in the catch per unit effort are inferred to signify changes to the
true abundance of the target species. The CPUE is recorded with the location at which
the fishing is conducted. The location includes the latitude and longitude of the fishing
boat but not the depth at which the fish are caught.
In fishery ground prediction, the CPUE is generally represented as a function of
environment variables. It is difficult to measure sufficient variable data. High perfor-
mance computing facilities enable us to calculate a highly accurate environmental
model that includes the temperature, salinity and velocity. With down-scaling and data
assimilation techniques, we can obtain highly accurate environmental variables with
sufficient resolution around possible fishing grounds.
Current modeling techniques consider an ocean simulation dataset only on the
fishing day. With such a technique, which does not use time-varying datasets, it is
difficult to explore the causal relationship between the CPUE and environment vari-
ables, even if their fitness has been maximized. The food chain is important when we
estimate fishery prediction. It originates from nutrient salts that precipitate at the bottom
of the sea. When the salts move upwards where the sunlight arrives along an upwelling
current, phytoplankton can reproduce. Then, zooplankton and fish eat the phyto-
plankton and the zooplankton, respectively. It is natural that we assume that these food
chain components follow the sea water current. The phytoplankton and the zooplankton
follow the ocean current passively in their moving.
In our research, we employ a standard least squares method with the L0 norm
regularization technique [3] to sparsely model the causal relationship because we have
a large number of possible explanatory variables considering a time-varying ocean
simulation dataset. The technique is a type of signal processing technique for efficiently
acquiring and reconstructing a signal by finding solutions to underdetermined linear
systems based on the principle that, through optimization, the sparsity of a signal can
be exploited to recover it from far fewer samples than required by the Shannon-Nyquist
sampling theorem.
The remainder of the paper is organized as the follows. In Sect. 2, we describe the
related work on fishery grounds predictive modeling. In Sect. 3, we explain our visual
analytics system, which supports the discovery of the food chains along a path-line in
the ocean flow. In Sect. 4, we describe the L0 norm regularization technique for sparse
modeling. In Sect. 5, we apply the technique to construct a predictive fishery model for
neon-flying fishes to confirm the effectiveness by comparing the proposed technique
with the previous one.
2 Related Work
Tian et al. developed two types of HSI models for predicting neon squid fish catches
and compared the models [1]. The HSI model represents an index that indicates
whether an animal feels comfortable by using a value from 0.0 to 1.0. The index is
defined by the United States Environmental Preservation Agency (EPA). In this HSI
A Path-Line-Based Modeling for Fishing Prediction 81
model, a Suitability Index (SI) is first constructed so that the CPUE can be described
using a single environmental variable, such as temperature, salinity or velocity. We call
this method Tian’s method. For example, when we consider a sea animal, we select one
of the temperature, salinity and velocity variables and construct a function of the CPUE
with the selected variable. To develop an HSI model, we integrate several SI indexes
into a single HSI model using a mathematical method, such as the geometric mean.
Xinjun et al. proposed an HSI model in which the CPUE is explained using a set of
weighted summations of the environmental variables and evaluated the effect caused by
the weighting methods [2]. We call this method Xinjun’s method. These previous
models considered only the correlation between the CPUE and the environmental
variables but did not consider any causal relationship in the food chain.
Koyamada et al. proposed a modeling technique that employs environmental
variables, such as water temperature, velocity and salinity, sampled on flow path-lines
that terminate at fishing locations [6]. Although the results indicated that the accuracy
of the model was improved when it considered the variables sampled along the
path-lines, they did not determine how many sampling points along the path-line are
required for effective modeling.
For the exploration of the causal relationship in the food chain, we focus on a path-line
that represents a curved line whose tangent is parallel to a velocity vector at a
spatio-temporal point. In this research, we assume that the food chain components,
such as plankton and fish, move along the ocean current. Although the number of the
target fish that the fishermen try to catch may eat small fishes can be measured by the
CPUE, it is difficult to measure those of the other components. It is naturally inferred
that each can have a set of environmental variables and specific value ranges.
Along a path-line that passes through a fishing location, we sample environmental
variables, such as the water temperature, velocity and salinity. We regard these vari-
ables as the factors that can be regarded as the occurrences of the food chain com-
ponents. In other words, these variables can be explained by the latent variables of the
occurrences.
Fig. 1. Fishing ground points and the corresponding path-lines. (Color figure online)
82 K. Koyamada et al.
4 Method
We develop a fishing ground model using an ocean simulation dataset and a fish catch
dataset called CPUE. The simulation dataset is generated from a data assimilation
system for the Northwest Pacific Ocean. The values of the environmental variables
were obtained from an ocean simulation based on the 3D-VAR data assimilation
product MOVE (MRI Multivariate Ocean Variational Estimation). The simulation
generated values of temperature (T), salinity (S), and current velocity (U, V, W). Thus,
the number of variable types is five. Because the grid size is 0.1° between 43° N and
Fig. 2. Path-lines that terminate at a single CPUE point with varying depth values. (Color figure
online)
A Path-Line-Based Modeling for Fishing Prediction 83
35° S and 141° W and 147° E in the North-West Pacific Ocean, the grid resolution is
90 70 in the horizontal plane. We calculate the path-lines whose terminating points
are located at the CPUE points. In the path-line, environmental variables, such as water
temperature, velocity and salinity, are sampled [5]. Figure 2 shows the path-lines (left)
and the sampled variables on the lines (right). In the lower right of the figure, several
path-lines for a single CPUE point are displayed with the depth values changed.
In the fishing ground model, we assume that the fish catch is explained using
environmental variables sampled along path-lines that are terminated at the CPUE
points. We regard the depths of the CPUE points as unknown variables because there is
no information on the depths where the fish were caught. Thus, we have 16,200
(¼ 5 54 60) variables at each CPUE point because the depth and time resolutions
are 54 and 60, respectively. In our research, we consider two methods, Tian’s and
Xinjun’s methods, to confirm the effectiveness in which time series data sets are
employed for constructing the HSI model.
X
60 X
54
logCPUE ¼ hTk;j logTk;j þ hSk;j logSk;j þ hUk;j logUk;j þ hVk;j logVk;j þ hWk;j logWk;j ;
k¼1 j¼1
ð1Þ
where Tk;j , Sk;j , Uk;j , V and Wk;j denote the temperature, salinity and velocity compo-
nents of sea water at depth j and time step k, respectively, and hk;j ¼ hTk;j hSk;j
hUk;j hVk;j hWk;j T are design parameters. To simplify the problem, the value of log CPUE
is roughly quantized to be in f1; 2g, that is, in (1), we replace the value with
qa ðlog CPUEÞ, where qa is a quantization function defined by
1 if x\a
qa ðxÞ ¼ :
2 if x [ a
60 X
X 54
qa ðlogCPUE Þ ¼ hTk;j logTk;j þ hSk;j logSk;j þ hUk;j logUk;j þ hVk;j logVk;j þ hWk;j logWk;j
k¼1 j¼1
ð2Þ
In the numerical experiments of the next section, we use the average of log CPUE as
the threshold a.
Next, we focus on a scheme to estimate the parameters hk;j . Let us consider N
ðiÞ ðiÞ ðiÞ ðiÞ ðiÞ
CPUE points CPUEðiÞ and their environmental variables Tk;j , Sk;j ,Vk;j ,Uk;j and Wk;j for
84 K. Koyamada et al.
where kk0 denotes the L0 norm of a vector, that is, the number of non-zero entries of a
vector, y 2 RN is a constant vector whose i th entries are CPUE ðiÞ , A 2 RN16200 is a
ðiÞ ðiÞ ðiÞ ðiÞ ðiÞ
constant matrix consisting of log Tk;j ; log Sk;j ; log Vk;j ; log Uk;j and log Wk;j as follows,
h i
ðiÞ ðiÞ ðiÞ ðiÞ ðiÞ
aik;j ¼ log Tk;j log Sk;j log Vk;j log Uk;j log Wk;j ;
2 3
a11;1 a11;2 ... a11;54 a12;1 . . . a160;54
6 a2 a12 ... a21;54 a22;1 . . . a260;54 7
6 1;2 7
A¼6 .. 7
1;2
6 .. .. .. .. 7;
4 . . . . . 5
aN1;2 aN1;2 ... aN1;54 aN2;1 . . . aN60;54
and h 2 R16200 is a design variable vector. To solve this problem, we apply the iterative
reweighted least squares (IRLS) algorithm [4]. In IRLS, the solution at the ðt þ 1Þ th
iteration is obtained as
2
hðt þ 1Þ ¼ arg minky Ahk2 þ kW ðtÞ h ; ð4Þ
h
ðtÞ
where W ðtÞ is a diagonal matrix whose diagonal elements Wii are calculated using the
optimal solution of the previous iteration as follows,
ðtÞ 1
Wii ¼ ; ð5Þ
ðkÞ
hi þ e
and e is a small constant. The empirical results show that we can obtain the optimal
solution of (3) by iterating (4) and (5) until convergence.
IRLS algorithm for (3)
Step 1. Set W ð0Þ to be the identity matrix for i ¼ 1; 2; ; 16200:
ii
Step 2. Set t 1:
Step 3. Repeat Step 3-1 to Step 3-3 until
hðt þ 1Þ hðtÞ \tol
Step 3-1. Calculate hðt þ 1Þ using (4).
Step 3-2. ðt þ 1Þ
Calculate Wii for i ¼ 1; 2; ; 16200 using (5).
Step 3-3. t tþ1
A Path-Line-Based Modeling for Fishing Prediction 85
60 X
X 54
SIX ¼ logCPUE ¼ hXk;j logXk;j ; ð6Þ
k¼1 j¼1
Next, we focus on a scheme to estimate the parameters hXk;j . Let us consider N CPUE
ðiÞ
points CPUEðiÞ and their environmental variables Xk;j for i ¼ 1; 2; . . .; N, where X
denotes S, T, V, U or W. The parameters can be obtained using Eq. 3, in which
ðiÞ
A 2 RN3240 is a constant matrix consisting of zik;j ¼ log Xk;j as follows:
2 ð1Þ ð1Þ ð1Þ ð1Þ ð1Þ 3
z1;1 z1;2 ... z1;54 z2;1 ... z60;54
6 ð2Þ ð2Þ ð2Þ ð2Þ ð2Þ 7
6 z1;1 z1;2 ... z1;54 z2;1 ... z60;54 7
A¼6
6 . .. .. .. .. 7
7
4 .. . . . . 5
ðNÞ ðNÞ ðNÞ ðNÞ ðNÞ
z1;1 z1;2 ... z1;54 z2;1 ... z60;54
and h 2 R3240 is a design variable vector. To solve this problem, we apply the IRLS
algorithm described in the previous section for each variable S, T, V, U or W.
In this research, an important point is how sparsely the CPUE can be modeled, that is,
how much the number of variables is minimized. To examine the method proposed in
the previous section, we perform some numerical experiments. In all experiments, we
use e ¼ 108 and tol ¼ 105 . Because the optimal solution of (3) depends on the value
of k, we increase it from 0.01 to 1 by 0.1 and select the best k.
To show the efficiency of the proposed method, we compare the proposed method
with the previous method, in which a time-varying dataset is not employed. In Xinjun’s
method, the CPUE can be described by leaving terms in which k equals 1 in Eq. 6.
X
54
logCPUE ¼ hTi logT1;i þ hSi logS1;i þ hTi logU1;i þ hTi logV1;i þ hTi logW1;i ;
j¼1
86 K. Koyamada et al.
In Tian’s method, we assume that the log CPUE, that is, the SI model, can be
described in the same way.
X
54
SIX ¼ logCPUE ¼ hXk;j logXk;j
j¼1
where X denotes S, T, V, U or W.
Here, the parameters are estimated by a least squares method. We examine these
methods using fish cash data of January 2006. The parameters are estimated using 45
CPUE points of January 1 to 10, and we examine the accuracy rate of the estimated 137
CPUE points from January 11 to 31. Because the values of logCPUE should be 1 or 2,
we round them into 1 or 2 and calculate the percentage of correct answers.
Table 1 shows the results that indicate that the accuracy of the model is improved
when it considers the variables that are sampled along the path-lines.
The purpose of this research is to demonstrate the advantage that the environmental
variables sampled along path-lines can explain the CPUE points that are measured at
the terminating points. Numerical experiments indicate that the value of CPUE depends
highly on the velocity of water. We showed the effectiveness by comparing the pro-
posed model with the previous one, which does not consider the path-line sampling and
the efficiency of the L0 norm regularization to select important environmental variables.
Fig. 3. Model accuracy distribution along the number of days. (Color figure online)
A Path-Line-Based Modeling for Fishing Prediction 87
6 Conclusion
References
1. Tian, S., Chen, X., Chen, Y., Xu, L., Dai, X.: Evaluating habitat suitability indices derived
from CPUE and fishing effort data for Ommatrephes bratramii in the northwestern Pacific
Ocean. Fish. Res. 95(2–3), 181–188 (2009)
2. Xinjun, C., Siquan, T., Bilin, L., Yong, C.: Modeling a habitat suitability index for the eastern
fall cohort of Ommastrephes bartramii in the central North Pacific Ocean. Fish. Res. 29(3),
493–504 (2011)
3. Daubechies, I., Devore, R., Fornasier, M., Gntrk, C.S.: Iteratively reweighted least squares
minimization for sparse recovery. Comm. Pure Appl. Math 63(1), 1–38 (2010)
88 K. Koyamada et al.
4. Futami, K., Onoue, Y., Sakamoto, N., Koyamada, K.: Visualization of ocean current effects
for fishing ground formation. J. Jpn Soc. Fluid Mech. 34(2), 83–86 (2015)
5. Uenaka, T., Sakamoto, N., Koyamada, K.: Visual analysis of habitat suitability index model
for predicting the locations of fishing grounds. In: Proceedings of IEEE Pacific Visualization
2014 (Visualization Notes), pp. 306–310 (2014)
6. Koyamada, K., Konishi, K., Sakamoto, N., Takami, M.: A path-line-based approach for
developing a fishing ground model. In: Proceedings of AsiaSim 2015 (2015)
Simulation and Analysis
Numerical Simulation
of a Hamilton-Jacobi-Bellman Equation
for Optimal Management Strategy of Released
Plecoglossus Altivelis in River Systems
1 Introduction
Plecoglossus altivelis (Ayu) is annual and diadromous fish species in Japan (Photo 1:
left panel) [1]. P. altivelis accounts for 7.5 % (2,353 t) of total fish catch (31,264 t) of
Japanese inland fisheries in 2013 [2]. Life history of P. altivelis is unique as reviewed
in Tanaka et al. [3]. During autumn, adult fishes spawn eggs in the downstream reaches
of their living river and die soon afterwards. Hatched larval fishes descend to coastal
areas of the downstream water body of the river, which is a sea or a brackish lake in
general, and grow up to juveniles with feeding on zooplanktons till the next spring. The
grown fishes ascend the river toward its midstream and upstream reach where
rock-attached algae, which are staple foods of P. altivelis, are available in riverbed.
They feed on the algae to mature till the coming autumn when they descend the river.
Photo. 1. Adult P. altivelis (left panel) and adult P. carbo (right panel). (The photo of P. carbo
is taken from Photo by (c)Tomo.Yun http://www.yunphoto.net).
profit by harvesting the fish and the cost of countermeasure to prevent the feeding
damage is then presented. Finding the optimal management strategy of the population
of P. altivelis reduces to solving a Hamilton-Jacobi-Bellman equation (HJBE), which is
a time-backward degenerate parabolic nonlinear partial differential equation (PDE).
The HJBE is numerically solved with a stable finite element scheme.
The rest of this paper is organized as follows. Section 2 presents the mathematical
model used in this paper. Section 3 applies the mathematical model to optimal pop-
ulation management of released P. altivelis in Hii River, Shimane Prefecture, Japan.
Section 4 concludes this paper and presents future perspectives of our research.
2 Mathematical Model
with the Verhulst-type coefficients arising from the growth curve of fishes as
aðt; xÞ ¼ r 1 K 1 x x ð2Þ
and
bðt; xÞ ¼ rx ð3Þ
where Bt (day1/2) is the 1-D standard Brownian motion, K ¼ mX0 (kg), m (−), r
(1/day), and r (1/day1/2) are positive parameters, R (1/day) is the natural mortality rate
(1/day), k ðuÞð 0Þ (1/day) is the predation pressure of P. carbo as a function of the
control variable ð0 Þuð 1Þ (−), which is effort to prevent the predation,
ð0 Þcð cM Þ (1/day) is the fishing pressure, cM is the maximum fishing pressure,
ð0\ÞTc ð\T Þ (day) is the opening time of catching P. altivelis, and vS is the
94 Y. Yaegashi et al.
characteristic function for generic set S. The term bðt; Xt ÞdBt describes temporal
fluctuation of the population of P. altivelis due to natural and artificial environmental
changes in the habitat [11, 12]. A major difference with the present and conventional
stochastic Verhulst models [9] is that the parameter K of the former depends on X0 but
that of the latter does not. Population dynamics of P. carbo is not directly considered in
the SDE (1), but its influences on the population dynamics of P. altivelis are considered
in the term k ðut ÞXt to formulate a minimum mathematical model.
The generator Au;c for the coupled process Yt ¼ ðt; Xt Þ conditioned on Ys ¼ ðs; xÞ
with s\t for sufficiently regular u ¼ uðs; xÞ is expressed as [10]
@u @u 1 @ 2 u
Au;c u ¼ þ a Rx k ðuÞx vfs Tc g cx þ b2 2 : ð4Þ
@s @x 2 @x
U ¼ fuj0 u 1g ð5Þ
and
C ¼ fcj0 c cM g; ð6Þ
respectively. The objective function to be maximized, which represents the total profit
and is denoted as v ¼ vðs; T; X; u; cÞ (kg), is proposed as
Z T Z T
vðs; T; X; u; cÞ ¼ f ðut Þdt þ vft Tc g ct Xt dt ð7Þ
s s
where f ð 0Þ (kg/day) with f ð0Þ ¼ 0 is an increasing function. The first and second
terms in the right hand-side of (7) represent the minus of the total cost of operating a
countermeasure to prevent the feeding damage by P. carbo and the total amount of
harvested P. altivelis by human during the period ðs; T Þ, respectively. The coefficients k
and f are set as
and
respectively where k0 (1/day) is the predation pressure from the bird without any
countermeasures, ð0\Það 1Þ modulates the effectiveness to decrease the predation
Numerical Simulation of a Hamilton-Jacobi-Bellman Equation 95
pressure with the control u, bð [ 0Þ (−) modulates the cost to prevent feeding damage,
and x (kg/day) is a parameter serving as a weight that determines the balances on the
cost and the benefit in the objective function. The parameter a represents the effec-
tiveness of the countermeasure, while b represents its inefficiency to prevent the
feeding damage. Values of these parameters would depend on the countermeasures
chosen [6, 7].
The goal of the optimal control problem is to find the optimal controls u ¼ u and
c ¼ c that maximize the objective function v. The maximized objective function
V (kg) is referred to as the value function, which is expressed as
The optimal controls u ¼ u and c ¼ c are analytically expressed through the value
function V as
and
c ¼ vfc 1g cM ; ð13Þ
Z T
J¼E s;x
gðXt Þdt ð14Þ
s
subject to appropriate
terminal and boundary conditions, which are J ¼ 0 at the time
t ¼ T, Jx¼0 and @J ¼ 0. Hereafter, the notation J0 ¼ J ð0; X0 Þ is used for the sake of
@x x¼L
brevity. Ecologically and economically relevant gðXt Þ can be k ut Xt , vft Tc g ct Xt ,
and, f ut with which the index J0 represents the total weight of predated P. altivelis
by P. carbo, the total weight of harvested P. altivelis as fish catches, and the total cost
to prevent the feeding damage from P. carbo, respectively. The indices with the
above-mentioned three gðXt Þ are denoted in order as J0;pr , J0;cau , and J0;cos , respectively.
3 Application
which are Lake Shinji and Lake Nakaumi from upstream, are connected to the main-
stream of Hii River. Hatched larval fishes of P. altivelis are thought to descend to Lake
Shinji; however, it is not known where actually the larval fishes survive during winter.
Constituent members and union members of Hii River cooperatives say that the
number of juvenile fishes ascending the river during spring is significantly decreasing
[8]. In this river system, juvenile fishes of P. altivelis are released in its midstream and
upstream reach during May. The fishes grow up in the system and spawn in down-
stream reaches during October to November.
The initial time t ¼ 0 for numerical computation is set on a day in May. The
terminal time T is set as 180 (day), which is in the autumn in the same year. The
parameters r and K are set as 8:7 102 (1/day) and 20X0 (kg), respectively, con-
sidering a deterministic counterpart of (1) with k ¼ c ¼ 0 [19]. The natural mortality R
is set as 4:6 103 (1/day) [20]. The range of the parameter r is assumed to be
restricted to r\0:4 (1/day1/2), which is the condition that (1) with k ¼ c ¼ 0 has a
non-trivial and non-singular PDF for sufficiently large T [21]. The value of r ¼ 0:25
(1/day1/2) is thus assumed. The other parameters are k0 ¼ 4:0 103 (1/day) [8],
cM ¼ 0:01 (1/day), and x ¼ 1 (kg/day). a ¼ 1 (−) is fixed unless otherwise specified.
The initial condition is X0 ¼ 1:5 103 (kg), which amounts to 3:0 105 juveniles of
P. altivelis [8]. The domain of the population x is set as X ¼ ð0; 6:0 104 Þ (kg), which
is discretized into a mesh with 300 elements and 301 nodes. The time increment for
temporally integrating the HJBE is 0.01 (day). Increasing spatial and temporal reso-
lution of the computation does not significantly change the obtained computational
results below.
prevent feeding damage by P. carbo after the opening time Tc except for the neigh-
borhood of the terminal time T. In these cases, it is also indicated that the intensive
countermeasure should also be performed before the opening time Tc if the population of
P. altivelis is small. For the largest b, the optimal strategy is not to perform the coun-
termeasure. On the other hand, Figs. 3(a)–(c) indicate that the optimal control c does
not significantly depend on b; the optimal strategy is to harvest P. altivelis after the
opening time Tc except when the population of P. altivelis is small.
Comparisons of a variety of countermeasures to prevent the feeding damage, which
are characterized through the non-dimensional parameters a and b, are next carried out.
The indices considered are the total weight of predated P. altivelis (J0;pr ), the total
weight of caught P. altivelis (J0;cau ), and the total cost to prevent the feeding damage by
P. carbo (J0;cos ) presented in Sect. 2.3. The values examined are a ¼ 0:1i and b ¼ j
where i and j are nonnegative integers such that 0 i; j 10. In total 121 numerical
computations are performed for each index. Figures 4(a)–(c) plot the computed indices
J0;pr , J0;cau , and J0;cos in the a b phase space. These figures can validate the coun-
termeasures in terms of the indices. The focus here is comparing the methods with high
efficiency and low effectiveness (small b and small a) and those with low efficiency and
high effectiveness (large b and large a). This is because countermeasures with low
efficiency and low effectiveness (large b and small a) should not be used in practice,
and those with high efficiency and high effectiveness (small b and large a) would
merely exist; all the existing countermeasures would have both advantages and dis-
advantages on efficiency and effectiveness. Figures 4(a) and (c) indicate that the
Fig. 4. The indices J0;pr , J0;cau , and J0;cos for a variety of countermeasures.
method with higher efficiency and lower effectiveness (small b and small a) can more
effectively reduce the total weight of predated P. altivelis and the cost to preventing the
feeding damage. In addition, Fig. 4(b) indicates that the method with higher efficiency
and lower effectiveness more effectively increases the total fish catch. The obtained
computational results thus recommend using a countermeasure with higher efficiency
and lower effectiveness for managing the population of P. altivelis.
4 Conclusions
Exploring links between the present population dynamics model and the other kind of
related mathematical models, such as the local swimming behavior model [15], is also
an important future research topic for more effectively analyzing migration of
P. altivelis in river systems.
Photo. 2. Yoshii Weir (left panel) and Hinobori Weir (right panel) installed in the mainstream
of Hii River.
Acknowledgements. We thank Dr. Kayoko Kameda of Lake Biwa Museum, and officers in
Hii-River Fishery Cooperatives, Yonago Waterbirds Sanctuary, and the Ministry of Environment
for their valuable comments and providing data. This research is supported by the River Fund in
charge of The River Foundation and JSPS Research Grant No.15H06417.
References
1. Takahashi, I., Azuma, K.: The Up-to-now Knowledge Book of Ayu. Tsukiji-shokan, Tokyo
(2006). (in Japanese)
2. MAFF: Statistics of catch and production amounts of aquatic resources in Japan during 2015
(2015). http://www.maff.go.jp/j/tokei/sokuhou/gyogyou_seisan_13/ (in Japanese)
3. Tanaka, Y., Iguchi, K., Yoshimura, J., Nakagiri, N., Tainaka, K.: Historical effect in the
territoriality of ayu fish. J. Theor. Biol. 268(1), 98–104 (2011)
4. Iguchi, K., Tsuboi, J., Tsuruta, T., Kiryu, T.: Foraging habits of Great Cormorant in relation
to released Ayu stocks as a food source. Aquacult. Sci. 56(3), 415–422 (2008)
5. Katano, O., Abe, S., Nakamura, T.: Relationships between ayu Plecoglossus altivelis and
other organisms in stream communities. Bull. Fish. Res. Agen. Suppl. 5, 203–208 (2006)
6. Yamamoto, M.: What Kind of Bird is the Great Cormorant. http://www.naisuimen.or.jp/
jigyou/kawau/01-1.pdf (in Japanese)
7. Yamamoto, M.: Stand face to face with Great Cormorant II. http://www.naisuimen.or.jp/
jigyou/kawau/03-1.pdf (in Japanese)
8. Hii River Fishery Cooperative. Personal Communication (2015)
9. Tsoularis, A., Wallace, J.: Analysis of logistic growth models. Math. Biosci. 179(1), 21–55
(2002)
10. Øksendal, B.: Stochastic Differential Equations. Springer, Berlin (2003)
11. Doyen, L., Thébaud, O., Béné, C., Martinet, V., Gourguet, S., Bertignac, M., Fifas, F.,
Blanchard, F.: A stochastic viability approach to ecosystem-based fisheries management.
Ecol. Econ. 75, 32–42 (2012)
Numerical Simulation of a Hamilton-Jacobi-Bellman Equation 101
12. Unami, K., Yangyuoru, M., Alam, A.H.M.B.: Rationalization of building micro-dams
equipped with fish passages in West African savannas. Stoch. Environ. Res. Risk Assess.
26(1), 115–126 (2012)
13. Yoshioka, H., Unami, K., Fujihara, M.: Mathematical analysis on a conforming finite
element scheme for advection-dispersion-decay equations on connected graphs. J. JSCE Ser.
A2 70(2), I_265–I_274 (2014)
14. Mincsovics, M.E.: Discrete and continuous maximum principles for parabolic and elliptic
operators. J. Appl. Math. Comput. 235(2), 470–477 (2010)
15. Yoshioka, H., Unami, K., Fujihara, M.: A Petrov-Galerkin finite element scheme for 1-D
time-independent Hamilton-Jacobi-Bellman equations. J. JSCE Ser. A2. 71, I_149–I_160
(2016). (in press)
16. MLIT (2002). http://www.mlit.go.jp/river/toukei_chousa/kasen/jiten/nihon_kawa/87072/
87072-1 (in Japanese)
17. Sato, H., Takeda, I., Somura, H.: Secular changes of statistical hydrologic data in Hii river
basin. J. JSCE Ser. B1 68(4), 1387–1392 (2012). (in Japanese with English Abstract)
18. Sato, H., Takeda, I., Somura, H.: Secular changes of statistical dry-season discharges in Hii
river basin. J. Rainwater Catchment Syst. 19(2), 51–55 (2014). (in Japanese with English
Abstract)
19. Miyaji, D., Kawanabe, H., Mizuno, N.: Encyclopedia of Freshwater Fishes in Japan.
Hoiku-sha, Tokyo (1963). (in Japanese)
20. Murayama, T.: Why population of Ayu annually changes? In: Furukawa, A., Takahashi, I.
(eds.) River Works to Grow up Ayu. Tsukiji-shokan, Tokyo (2010). (in Japanese)
21. Grigoriu, M.: Stochastic Calculus: Applications in Science and Engineering. Birkhäuser,
Boston (2002)
22. Breitburg, D.L., Rose, K.A., Cowan, J.H.: Linking water quality to larval survival: predation
mortality of fish larvae in an oxygen-stratified water column. Mar. Ecol. Prog. Ser. 178,
39–54 (1999)
23. Budy, P., Baker, M., Dahle, S.K.: Predicting fish growth potential and identifying water
quality constraints: a spatially-explicit bioenergetics approach. Environ. Manage. 48(4),
691–709 (2011)
24. Chernogorova, T., Valkov, R.: Analysis of a finite volume element method for a degenerate
parabolic equation in the zero-coupon bond pricing. Comput. Appl. Math. 34(2), 619–646
(2015)
25. Chernogorova, T., Valkov, R.: Finite volume difference scheme for a degenerate parabolic
equation in the zero-coupon bond pricing. Math. Comput. Model. 54(11), 2659–2671 (2011)
26. Crandall, M.G., Lions, P.L.: Viscosity solutions of Hamilton-Jacobi equations. Trans. Am.
Math. Soc. 277(1), 1–42 (1983)
27. Fleming, W.H., Soner, H.M.: Controlled Markov Processes and Viscosity Solutions.
Springer Science & Business Media, New York (2006)
Fault Displacement Simulation Analysis of the Kamishiro
Fault Earthquake in Nagano Prefecture Using the Parallel
Finite Element Method
1
Disaster Reduction and Environmental Engineering Deptartment,
Kozo Keikaku Engineering, Inc., 4-5-3 Chuo, Nakano-Ku, Tokyo 164-0011, Japan
[email protected]
2
Graduate School of Frontier Sciences, University of Tokyo, 5-1-5 Kashiwanoha,
Kashiwa-shi, Chiba 277-8563, Japan
Abstract. In recent years, the evaluation of fault displacement has been required
for evaluating the soundness of underground structures during an earthquake.
Fault displacement occurs as the result of the rupture of the earthquake source
fault, and studies have been conducted using the finite difference method, the
finite element method, etc. The present study used the nonlinear finite element
method to perform a dynamic rupture simulation analysis of the Kamishiro fault
earthquake in Nagano Prefecture on November 22, 2014. A model was prepared
using a solid element for the crust and a joint element for the fault surface. The
Kamishiro fault earthquake in Nagano was a reverse-fault earthquake whose fault
plane included a part of the Kamishiro fault and extended northward from there.
The total extent was 9 km, and the surface fault displacement confirmed was
approximately 1 m at maximum. Initial stress was applied to the fault to inten‐
tionally rupture the hypocenter to perform a propagation analysis of the rupture,
and the displacement and response time history obtained in the analysis were
compared with observational records. At this time, joint elements according to
Goodman et al. that had been expanded were introduced to the finite element
method code FrontISTR, which can analyze large-scale models, and the simula‐
tion analysis was performed.
1 Introduction
In recent years, the evaluation of fault displacement has been required for evaluating
the soundness of underground structures during an earthquake. Fault displacement
occurs as the result of the rupture of the earthquake source fault, and studies have been
conducted using the finite difference method (i.e. [1]) and the finite element method (i.e.
[2]). In particular, many studies centering on the finite difference method have been
performed using dynamic rupture simulation analysis, which reproduces the sponta‐
neous rupture process of a fault using a slip-weakening model (i.e. [1]).
The present study used the 3D nonlinear finite element method to perform a simu‐
lation analysis of the Kamishiro fault earthquake in Nagano Prefecture on November
22, 2014. The Kamishiro fault earthquake in Nagano (Mj 6.7, Mw 6.2) was a reverse-
fault earthquake whose fault plane included a part of the Kamishiro fault and extended
northward from there, with a length of approximately 20 km and a depth of approxi‐
mately 10 km. The total extent was 9 km, and the surface fault displacement confirmed
was approximately 1 m at maximum [3]. Also, a maximum acceleration of approxi‐
mately 600 cm/s2 was observed in the K-NET [4] observation points near the fault. A
40 km × 40 km × 20 km model that included the earthquake source fault was prepared
using a solid element for the crust and a joint element for the fault. A model was created
for the rupture process of the fault according to the nonlinear constitutive law incorpo‐
rating stress drop, and for the initial conditions, initial stress was applied to the joint
element to rupture the fault, and the ground surface response time history caused by
propagation of the rupture was compared with the observational records of K-NET. Also,
the surface fault displacement observed in studies of the actual site after the earthquake
was compared with the displacement obtained in the present analysis. This research was
conducted using the finite element method code FrontISTR [5], which can perform
parallel computation of large-scale models. This allowed a wide area of the crust to be
analyzed using relatively fine mesh.
In the present study, a model of the fault was created by the joint elements shown in
Fig. 1. Those joint elements are finite elements that easily simulate the contact/sliding/
exfoliation between two physical bodies. Several joint elements have been proposed,
such as those proposed by Goodman et al. [6] and those that have been formulated on
the basis of 3D isoparametric elements. The shape of the elements according to Goodman
et al. was hypothesized to be rectangular, and the deformation between the two
contacting surfaces is represented by the 6-mode combination shown in Fig. 2. The
authors expanded this to the desired shape of a triangle or quadrilateral, and implemented
FrontISTR. This allowed a high-precision analysis to be performed, even with distorted
mesh.
6 8 7
5
3 4 3 6
2
4 5
2
1 1
q1 q3 q5
q2 q4 q6
The relationship between the shearing stress and relative displacement of the joint
element used in this study is shown in Fig. 3(a). The joint elements have a bi-directional
degree of freedom with respect to the translation deformation in the surface (q1 and q2
shown in Fig. 2), so the shearing stress τ and the relative displacement ε are evaluated
in Eqs. (1) and (2). However, fq1 and fq2 are the shearing stress according to modes q1
and q2, and δq1 and δq2 are the displacement of modes q1 and q2.
(1)
(2)
As shown in Fig. 3(a), sliding rupture occurs when the shearing stress τ of joint elements
reaches the yield stress τy, and a stress drop to τ0 occurs. According to past studies (i.e.
[1]), the behavior of the shearing stress after sliding rupture is that it does not rapidly
drop to τ0 but has a certain degree of inclination, and with the slip-weakening model in
the finite difference method, a frequently used model is one in which the shearing stress
drops linearly to critical slip displacement Dc shown in Fig. 3(b). For convenience sake,
τ τ
τy Surface energy
ks τy
Surface energy
τi τi
Stress drop Δτ Stress drop Δτ
τ0 τ0
δ δ
Critical slip distance Dc
(a) This study (b) Slip-weakening model
the present study used a model in which the shearing stress dropped to τ0 the instant that
sliding rupture occurred. Also, according to past studies [2], the analytical results depend
on the relative quantity Δτ = τy − τ0, not on the absolute quantity τ0. Therefore, this study
used an analysis that focused on the relative value Δτ when τ0 = 0MPa. In addition,
referring to Dan et al. ([1]), the initial shearing stiffness ks of the joint elements was
specified so that the surface energy calculated from Dc = 25 cm in the slip-weakening
model was equivalent, and the vertical stiffness kv was made to be linear and a sufficiently
rigid value.
(3)
Here, x is the displacement; M, C, and K are the mass, damping, and stiffness matrix,
respectively; F is the external force vector. For damping, stiffness-proportional damping
was used, and referring to the maximum transmission frequency of the model and the
study by Mizumoto et al. [2], it was established that there would be 2 % damping at
1 Hz. Also, the effect of the reflected wave was eliminated by setting the viscous boun‐
dary to the model periphery that excluded the crust surface. According to the constitutive
law of the joint elements shown in Fig. 3(a), this problem had non-linearity, so a conver‐
gent calculation according to the Newton-Raphson method was performed.
3 Analysis Model
The model used in analysis is shown in Fig. 4, and the fault parameters are shown in
Table 1. The fault shape was established by referring to the AIST active fault database
[7] and aftershock distribution, and a strike angle of 12°, dip angle of 50°, length of
18 km, and depth of 10 km (width of 12.2 km) were established. As for generally used
physical values, ν = 0.25 was hypothesized with the share modulus of stiffness μ set to
30 GPa and the unit weight γ was set to 2.5 t/m3. Referring to the records on the hypo‐
center of this earthquake, the hypocenter was set to a position at a 5-km depth, and the
τy of the joint elements was set at a large value. The Kamishiro fault earthquake was a
reverse-fault earthquake. But the earthquake was also indicated to have been accompa‐
nied by a left-lateral slip. So we assume the direction of initial stress λ as 90° (pure
reverse-fault), 75°, 60° and 45°, and we compare their results(See Fig. 4). By setting an
initial stress, rupture would occur at the same time that the analysis started. Since ground
surface displacement was not observed on the north side of the fault, the shape of the
fault surface might not have been a simple rectangle, but for convenience sake, this study
used a rectangular fault.
106 Y. Mitsuhashi et al.
λ =75° λ =60°
Observation point North
(K-NET Hakuba)
12.2km
λ =90°
20km
18km
λ =45°
hypocenter
40km fault surface model
40km
The stress drop was uniformly set to 1 MPa, the condition in which the moment
magnitude Mw, calculated from the fault displacement Δu of each element in the final
state of the analysis by using Eqs. (4) and (5), with the fault surface as Σ, was close to
the value observed in the actual earthquake. τy was established according to Eq. (6),
referring to Andrew [8].
(4)
(5)
(6)
The maximum frequency which can be covered by the mesh fmax, that is calculated in
Eq. (7), is 0.875 Hz when n = 4. However, β is the shear wave velocity at the crust.
(7)
Fault Displacement Simulation Analysis 107
4 Analytical Results
The analytical results are shown. The simulation analysis matched well when initial
stress direction λ is 60°. So deformation diagram, fault rupture time, fault displacement
only the case is shown as a representative case.
C A
C
N
A
Cross section B
Unit: m
B Deformation
magnification: ×7000
Fig. 5. Analysis model deformation diagram and vertical displacement contour diagram.
In addition, the surface displacement of the final time was approximately 62 cm, and
the vertical displacement observed for the actual earthquake after it had occurred was
approximately 90 cm. Therefore, the simulation was believed to have simulated the
actual phenomenon well.
The response time history at the observation point location was compared with the results
actually observed by K-NET Hakuba, the response acceleration on the ground surface.
In this study, the displacement time history was compared to focus on the relatively
long-term results. So we can neglect the effect of the surface layer. The acceleration
response time history of K-NET Hakuba was corrected by using the method of Boore
et al. [10], and the displacement time history waveform that were obtained by time
integration were compared with the waveform obtained in the analysis. A comparison
of the displacement time history obtained from the observational records with the
analytical results is shown in Fig. 7. However, the time axis origin has been adjusted as
appropriate.
Because the observation points of K-NET Hakuba were relatively close to the fault,
a relatively large, permanent displacement occurred in the analysis final time. The simu‐
lation analysis matched well when initial stress direction is 60°.
5 Conclusion
Acknowledgments. This study used the K-NET strong motion seismograms from the National
Research Institute for Earth Science and Disaster Prevention.
References
1. Dan, K., Muto, M., Torita, H., Ohhashi, Y., Kase, Y.: Basic examination on consecutive fault
rupturing by dynamic rupture simulation. In: Annual Report on Active Fault and
Paleoearthquake Researches, no.7, pp. 259–271 (2007). (In Japanese)
2. Mizumoto, G., Tsuboi, T., Miura, F.: Fundamental study on fault rupture process and
earthquake motions on and near a fault by 3D-FEM. J. JSCE, no.780, I-70, 27–40 (2005). 1
(In Japanese)
3. Japan Association for Earthquake Engineering: Report about the Earthquake at the Nagano
Prefecture north in 2014 (2015). (In Japanese)
4. National Research Institute for Earth Science and Disaster Prevention: Strong-motion
Seismograph Networks. http://www.kyoshin.bosai.go.jp/. Accessed 01 Aug 2015
5. FrontISTR Workshop HP. http://www.multi.k.u-tokyo.ac.jp/FrontISTR/. Accessed 01 Aug
2015 (In Japanese)
6. Goodman, R.E.: Methods of geological engineering in discontinuous rocks, Chap. 8, pp. 300–
368. West Publishing Company (1976)
7. National Institute of Advanced Industrial Science and Technology: Active fault data base.
https://gbank.gsj.jp/activefault/index_gmap.html. Accessed 01 Aug 2015 (In Japanese)
8. Andrews, D.J.: Rupture velocity of plane strain shear cracks. J. Geophys. Res. 81, 5679–5687
(1976)
9. Earthquake Research Committee, Headquarters for Earthquake Research Promotion, Strong
motion forecasting method of the earthquake which specified an earthquake source fault
(Recipe). Revision on 21 Dec. 2009 (In Japanese)
10. Boore, D.M., Stephens, C.D., Joyner, W.B.: Comments on baseline correction of digital
strong-motion data: examples from the 1999 Hector Mine, California, Earthquake. Bull.
Seismol. Soc. Am. 92(4), 1543–1560 (2002)
Prediction of 137Cs-Contaminated Soil Losses by Erosion
Using USLE in the Abukuma River Basin, Japan
Carine J. Yi ✉
( )
Abstract. The 2011 Great East Japan Earthquake and tsunami triggered a
significant nuclear power plant accident. Subsequent measurements of the
concentration of cesium-137 (137Cs) showed that the litter and surface layers in
the forest areas near the plant were significantly contaminated. This study applied
the Universal Soil Loss Equation (USLE), which has been widely used to estimate
soil losses from erosion, in the Abukuma River Basin. The greatest soil loss was
predicted to be 1762.75 t ∙ yr−1 ∙ ha−1. To predict losses of 137Cs-contaminated
soil, a 137Cs-soil transfer factor was applied in place of a crop factor, and it
yielded an average contaminated-soil loss rate of 190.65 t ∙ yr−1 ∙ ha−1, whereas
the standard USLE calculation yielded an estimated average soil loss rate of
184.14 t ∙ yr−1 ∙ ha−1. Higher soil losses were predicted in steeper areas west of
the river. However, contaminated soil may be deposited along a comparatively
flat area, such as that on the east side of the river.
1 Introduction
Studies of nuclear events were first conducted during World War II and focused on topics
such as the life cycle, mechanisms, and environmental impacts of fallout. Cesium-137
(137Cs) is among the fission products that generate the most concern because of its beta
and gamma emissions, high radioactivity, and relatively long, 30-year half-life
(Okumura 2003), and it is only produced during nuclear fission events (Ritchie and
McHenry 1990). According to Carter and Moghissi (1977), 137Cs was first released
during early nuclear tests in 1945, and studies on the global dispersal of 137Cs into the
environment began with high-yield thermonuclear tests in November 1952. Perkins and
Thomas (1980) and Ritchie and McHenry (1990) developed the earliest studies of the
dispersal of 137Cs. 137Cs fallout is strongly influenced by local precipitation patterns
and rates (Longmore 1982). However, 137Cs concentrates on the soil surface when the
soil is left undisturbed (Gerspar 1970; Ritchie et al. 1972). Once in contact with the soil,
137Cs is firmly adsorbed to finer soil particles, such as clay, and further movement of
137Cs by natural chemical processes is limited (Tamura, 1964; Ritchie et al. 1982;
Guimarães et al. 2003). Rogowski and Tamura (1965) investigated the difference
between predicted and experimental soil losses by spraying 137Cs onto areas of bare
clipped meadow and tall meadow cover. Walton (1963) determined the vertical radio‐
activity profiles associated with weapons tests in several New Jersey soils by examining
four radionuclides: strontium 90, ruthenium 106, 137Cs, and 144Cs. McHenry and
Ritchie (1973) found that 137Cs concentrated in accumulation sites, such as floodplains,
lakes, and reservoirs. Okumura (2003) described the life cycle of radioactive 137Cs and
analyzed its material flow in the U.S. Guimarães et al. (2003) described the vertical
distribution of 137Cs in the soil profile and showed its exponential decrease with
increasing soil depth at undisturbed sites. Mizugaki et al. (2008) analyzed suspended
sediment sources using 137Cs and 210Pbex in Japan before the Fukushima Daiichi
Nuclear Power Station explosion.
The 1986 Chernobyl accident, which was classified as level 7 (major accident)
according to the International Nuclear Event Scale (Cardis et al. 2006), has been the
only nuclear accident to provide in-field data that were used to identify contamination
areas and trace radioactive materials, determine the impacts to human health from
external or internal contamination via long-term monitoring, and evaluate the impact of
radioactive fallout on the environment and agricultural activities. Subsequent to the
Chernobyl accident, the Great East Japan Earthquake and Tsunami occurred in 2011
and triggered an explosion at the Fukushima Daiichi Nuclear Power Station. Residents
within a 20-km radius of the Fukushima Daiichi Nuclear Power Station were forced to
leave their homes. Japanese government agencies continue to measure the level of radi‐
oactive contamination in urban areas, agricultural areas, forests, rivers and ocean. A
total of 3,793 observation locations are measured daily throughout the country and used
to chart the current and daily average levels and other measurements of the radioactivity
associated with this disaster (a national radioactivity information list is available online
at http://new.atmc.jp/).
Since the Fukushima Daiichi Nuclear Power Station explosion in 2011, a number of
measurement-based studies have been conducted. Yasunari et al. (2011) applied a
Lagrangian particle dispersion model known as FLEXPART to estimate the total
deposits of 137Cs by integrating daily observations of 137Cs in each prefecture in Japan.
Tonosaki et al. (2012) measured radiation levels in surface soils and found that they
contained greater concentrations of 134Cs and 137Cs relative to deeper soil layers.
However, Hayashi et al. (2012) measured 134Cs and 137Cs contamination in areas
around Mt. Tsukuba, Ibaraki, Japan and found that the levels were higher in forest than
in open areas, such as paddy fields and rural towns. In addition, they found that the litter
layers and surfaces in the forest were significantly contaminated.
Radioactive material released from a nuclear accident follows a process in which it
falls to the ground surface, is partially retained in the soil and vegetation, and to some
extent is then transported into rivers and distributed into bottom sediment (NIES,
2012). In this study, the Universal Soil Loss Equation (USLE) is used to estimate the
transfer of 137Cs through soil erosion into the Abukuma River Basin, which is the closest
river basin to the site of the incident. The USLE is an erosion model designed to predict
the average rate of soil erosion for each feasible alternative combination of crop system
and management practice in association with a specified soil type, rainfall pattern, and
topography (Wischmeier and Smith 1978).
112 C.J. Yi
The Abukuma River runs from southwest to northeast in the Fukushima Prefecture,
which has a population of 1,380,000. The river is 239 km long, and the total river network
is 1,814 km long. The river basin area is 5,405 km2. The river is a first-class fresh water
resource that is surrounded by mountainous areas that have summit heights in excess of
1,000 m and occasionally experience heavy rains (MLIT; Water and Disaster Manage‐
ment Bureau 2003).
2 USLE Application
The USLE is a widely applied erosion model used to estimate soil loss on slopes due to
runoff in specific field areas subjected to particular cropping and management systems
(Warren et al. 2005). Wischmeier and Smith (1978) developed the USLE for conserva‐
tion planners, and the soil equation is defined as follows:
(1)
where
A = soil loss (tons/ha/year),
R = rainfall erosivity factor (tf ∙ m2/ha ∙ h),
K = soil erodibility factor (h/m2),
L = slope length factor,
S = slope steepness factor,
C = cover-management factor, and
P = supporting practice factor (management factor).
The variable A is the calculated soil loss per unit area. It is expressed in the same units
selected for K and includes the same period selected for R. In practice, these parameters
are usually selected such that A can be calculated in tons per hectare per year. However,
other units may be selected if desired. The USLE is an erosion model designed to calcu‐
late long-term soil loss overages from sheet and rill under specified conditions. The
model is also useful for construction sites and other non-agricultural conditions,
although it does not predict deposition or calculate sediment yields from gully, stream
bank, or streambed erosion. Other studies have applied the USLE to Japanese soil (Higa
and Manmoto 2000; Imai and Ishiwatari 2006; Imai and Ishiwatari 2007; Unoki et al.
2010).
R is the rainfall-runoff erosivity factor, and it is calculated from daily precipitation
data. In this study, the Automated Meteorological Data Acquisition System (AMeDAS)
averages of September 2009 were reviewed to determine the R factor, which is defined as
(2)
where I is the rainfall intensity (cm/h) and R is the amount of rainfall (cm). The variable Ei
is defined as follows:
(3)
Prediction of 137Cs-Contaminated Soil Losses by Erosion 113
K is the soil erodibility factor, and it represents the soil loss rate per erosion index
unit for a specified soil type. The soil erodibility factor is measured on a unit plot defined
as a 72.6-foot long, uniform 9 % continuous slope in clean-tilled fallow soil. The K factor
according to soil type is summarized based on work by Imai and Ishiwatari (2007).
L is the slope length factor, and it is the ratio of soil loss from the field slope length
to the soil loss from a 72.6-foot (22.13-meter) slope length under identical conditions.
S is the slope steepness factor, and it is the ratio of the soil loss from the field slope
gradient to that from a 9 % slope under otherwise identical conditions. The S factor may
be evaluated by combining the L factor for each land cell (Wischmeier and Smith 1978):
(4)
where m is 0.5, 0.4, 0.3, and 0.2 for tan β > 0.05, 0.03 < tan β ≤ 0.05, 0.01 < tan β ≤ 0.03,
and tan β ≤ 0.01, respectively (λ is the slope length in meter and ß is the slope angle in
degrees).
C is the cover and management factor, and it is the ratio of the soil loss from an area
with specific cover and management characteristics to that from an identical area under
tilled, continuous fallow conditions. In Fukushima, the cover/management categories
and corresponding C factors are as follows: orchard (0.2), farm (0.4), rice (0.01), live‐
stock (0.01), flowers (0.01), and grass crops (0.05); these values were suggested by the
Okinawa Science and Technology Promotion Center (OSTC) and are based on the
Okinawa Prefecture land-use map (2001).
P is a supporting practice factor (management factor), and it takes on the following
values: 0.3 for orchards, 0.3 for farms, 0.1 for rice, 0.1 for livestock, 0.1 for flowers, and
0.3 for grass crops.
3 Results
Figure 1 depicts the results of the USLE calculations. In the figure, the color changes
progressively among adjacent regions within the range of white to black, with white
indicating no soil loss and black indicating maximum soil loss. The standard USLE
calculations (left panel) predicted an average loss of 184.14 t ∙ yr−1 ∙ ha−1 of 137Cs-
contaminated soil from the Abukuma River Basin. The greatest predicted soil loss,
1762.75 t ∙ yr−1 ∙ ha−1, was concentrated in the steep areas west and north of the center
of the Abukuma River Basin.
After performing the standard USLE calculations, the P factor was replaced by the
137Cs transfer factor, Tp. The Radioactive Waste Management Center (1988) proposed
137Cs transfer implementation coefficients for agricultural products, for example,
0.00070 for tomatoes, 0.012 for beans, 0.0011 for radishes and 0.0010 for apples. The
Tp value is calculated as follows:
(5)
114 C.J. Yi
The right-hand panel in Fig. 1 depicts the distribution of eroded soil contaminated
with 137Cs, which replaced the supporting practice factor as the soil transfer factor. A
maximum soil loss of 449.897 t ∙ yr−1 ∙ ha−1 (an average of 190.65 t ∙ yr−1 ∙ ha−1) was
predicted in the Abukuma River Basin using this calculation. Active soil loss was
predicted upstream of the river, whereas other areas in the hydrology network indicated
minimal or no soil loss.
The Abukuma River Basin is a mountainous region with 77,202 ha of cultivated area
(Abukuma River Basin Risk Reduction Committee 2006), as shown in Fig. 2. In the study
area, an average of 190.65 t ∙ yr−1 ∙ ha−1 (a maximum soil loss of 449.8s97 t ∙ yr−1 ∙ ha−1)
of 137Cs-contaminated soil was predicted by the USLE to be lost through soil erosion.
Fig. 3. 137Cs (2012.12.28) air concentrations overlaid on the results of the 137Cs transfer factor
USLE
Higher soil erosion areas were predicted to occur in association with steep slopes and fallow
soil; these areas were found west and north of the center of the study area. However, soil
erosion occurs widely in the study area near the main streams of the Abukuma River.
Figure 3 shows the 137Cs concentration data from December 28, 2012 overlaid with the
results of the 137Cs transfer factor USLE for the 23 of 59 towns in Fukushima Prefecture
that lie within the Abukuma River Basin. The predicted losses of contaminated soil on the
east side of the river are small compared with those of the west side, as apparent in the
figure from the numerous dark gray to black areas in the west, and the 137Cs concentra‐
tions on the east side of the river are still high. High 137Cs air concentrations gradually
spread out from the east coast to the northwest and the southwest. Relatively lower 137Cs
air concentrations occur along the Abukuma River, which consists largely of flatland areas.
4 Conclusions
137Cs has a 30-year half-life cycle and disappears naturally by radioactive decay. This
study showed that 137Cs transfer can be estimated according to soil erosion rates in
flatland areas that contain 137Cs contaminated soil and where predicted soil loss is less
than in steep areas. Therefore, further attention might be required and the findings
described here can help inform disaster recovery strategies. Intense rainfall may be a
significant factor in the transportation of contaminated soil to downstream areas and
should be monitored in future watershed risk management strategies.
116 C.J. Yi
References
Carter, M.W., Moghissi, A.A.: Three decades of nuclear testing. Health Phys. 33(1), 55–71 (1977)
Cardis, E., Howe, G., Ron, E., Bebeshko, V., Bogdanova, T., Bouville, A., Carr, Z., Chumak, V.,
Davis, S., Demidchik, Y., Drozdovitch, V., Gentner, N., Gudzenko, N., Hatch, M., Ivanov, V.,
Jacob, P., Kapitonova, E., Kenigsberg, Y., Kesminiene, A., Kopecky, K.J., Kryuchkov, V.,
Loos, A., Pinchera, A., Reiners, C., Repachol, M., Shibata, Y., Shore, R.E., Thomas, G.,
Tirmarche, M., Yamashita, S., Zvonova, I.: Cancer consequences of the Chernobyl accident:
20 years on. J. Radiol. Prot. 26, 127–140 (2006)
Gerspar, P.L.: Effect of American Beech trees on the gamma radioactivity in soils. Soil Sci. Soc.
Am. Proc. 34, 318–323 (1970)
Guimarães, M.F., Filho, V.F.N., Ritchie, J.: Application of cesium-137 in a study of soil erosion
ᖺᗘὶᇦ㉥ᅵὶฟ㜵Ṇሗ࿌᭩ Ἀ⦖
Higa, E ( )., Manmoto, H ( ).: USLE . 12
ၨ ▼Ώ ㍤ኵ
( ) (2000). (in Japanese)
Imai, K ( )., Ishiwatari, T ( ).: Regional characteristics of USLE factors based on
ᖹᡂ ᖺᗘ㎰ᴗᅵᮌᏛㅮ₇ㅮ₇せ᪨㞟
literate data. Extended abstract for 2006 the conference of Japanese society of Irrigation,
drainage and rural engineering, 18 (2006).
ၨ ▼Ώ ㍤ኵ ⤫ィ㈨ᩱ➼ࢆ⏝࠸࡚ᩚ⌮ࡋࡓ㒔㐨ᗓ┴ู ᅵ
(in Japanese)
◊✲ᡤ
の . ( ,
.) 249, 1–6 February 2012. (in Japanese)
McHenry, J.R., Ritchie, J.C.: Accumulation of fallout cesium 137 in soils and sediments in
selected watersheds. Water Resour. Res. 9(3), 676–686 (1973)
㈨ᩱ ᖹᡂ ᖺ ᭶
Management Bureau ( ): の の ( )-
4-1, (Report for Abukuma River Basin management) 15 11 , November 2003.
㈈ ள⇕ᖏ⥲ྜ◊✲ᡤ
(in Japanese)
ࡍࡿㄪᰝ◊✲
13 : GIS
(2001). (in Japanese)
Perkins, R.W., Thomas, C.W.: Worldwide fallout. In: Hanson, W.C. (ed.), pp. 53–82. Technical
Information Center, Oak Ridge (1980)
Prediction of 137Cs-Contaminated Soil Losses by Erosion 117
Ritchie, J.C., McHenry, J.R., Gill, A.C.: The distribution of Cs-137 in the litter and upper 10
centimeters of soil under different cover types in northern Mississippi: Health Physics, vol.
22, p. 197 (1972)
Rogowski, A.S., Tamura, T.: Movement of 137Cs by Runoff, Erosion and Infiltration on the
Alluvial Captina Silt Loam. Health Phys. Soc. 11(12), December 1965
Warren, S.D., Mitasova, H., Hohmann, M.G., Landsberger, S., Iskander, F.Y., Ruzycki, T.S.,
Senseman, G.M.: Validation of a 3-D enhancement of the universal soil loss equation for
prediction of soil erosion and sediment deposition. CATENA – Interdisc. J. Soil Sci. Hydrol.
Geomorphol. Focusing Geoecology Landscape Evol. 64, 281–296 (2005)
Okumura, T.: The material flow of radioactive cesium-137 in the U.S. 2000, United States
Environmental Protection Agency (2003)
Yasunari, T.J. Stohl, A., Hayano, R.S., Burkhart, J.F., Eckjardt, S., Yasunari, T.: Cesium-137
deposition and contamination of Japanese soils due to the Fukushima nuclear accident. Proc.
Natl. Aced. Sci. USA (PNAS) 108(49), 6 December 2011
Debo, T.N., Reese, A.J.: Municipal Stormwater Management, 2nd edn. CRC Press, Boca Raton,
ᖹᡂ ᖺᗘ ᾏ㐨㛤Ⓨᢏ⾡◊✲Ⓨ⾲
Hosokawa, H ( ).: USLE の . 53
( 21 ) , February 2010. (in Japanese)
Walton, A.: The distribution in soils of radioactivity from weapons tests. J. Geophys. Res. 68(5),
1485–1496 (1963)
Wischmeier, W.H., Smith, D.D.: Predicting rainfall erosion losses – a guide to conservation
㜿Ṋ㝰ᕝᅪᇦ⥲ྜὶᇦ㜵⅏༠㆟ 㜿Ṋ
planning. U.S. Department of Agriculture, Agriculture handbook No. 537 (1978)
ከ፹య⎔ቃࣔࢹࣜࣥࢢ
National Institute for Environmental Studies of Japan (NIES), G-CIEMS(Grid-Catchment
Integrated Modeling System ( ) (2012). (in Japanese)
Charging and Discharging Characteristics
of a Quantum Well with an Initially Lorentzian
Wave Packet Incident on a DTM Type Potential
1 Introduction
An important piece that governs the operating speed of quantum devices that use
resonant tunneling phenomenon in quantum well structures is the time response
characteristics of the quantum well. Direct tunneling memory (DTM) [1,2] is
an example of such a quantum device. The memory’s 0 and 1 states correspond
respectively to the absence and the presence of electrons in its quantum well
region. The writing time and the retention time of the memory are related,
respectively, to the buildup characteristics and the decay characteristics of P (t),
which is the probability to find an electron in the quantum well region of DTM at
time t. The behavior of P (t) was studied when the electron incident on the DTM
type potential is characterized as a Gaussian wave packet [3]. For related but
different potential structures such as double barrier structures, many references
exist which study P (t) with Gaussian incident packets [4–6].
c Springer Science+Business Media Singapore 2016
S.Y. Ohn and S.D. Chi (Eds.): AsiaSim 2015, CCIS 603, pp. 118–128, 2016.
DOI: 10.1007/978-981-10-2158-9 11
Charging and Discharging Characteristics of a Quantum Well 119
We consider the case where a wave packet is incident on the simplified potential
from the left as shown in Fig. 1. The probability to find an electron in the well
region at time t is given by
L
P (t) = |ψ(x, t)|2 dx, (1)
d
120 Y. Kubo and N. Yamada
where ψ(x, t) is the electron’s wave function, which is the solution to the time
dependent Schrödinger equation
∂ψ(x, t) 2 ∂ 2 ψ(x, t)
i =− + V (x)ψ(x, t) (2)
∂t 2m ∂x2
with a given initial condition ψ(x, 0), where m is the effective mass of the electron
and is the reduced Planck constant (or Dirac’s constant). The wave function
in the well region can be expressed in the following form as the superposition of
plane waves:
∞
1 k2
ψ(x, t) = √ C(k)f (k) sin(k(x − L))e−i 2m t dk, (3)
2π −∞
where f (k), the weighting function for superposition, is the Fourier transform of
the initial wave function, i.e.,
∞
1
f (k) = √ ψ(x, 0)e−ikx dx. (4)
2π −∞
iΔr 1
C(k) C(kr ), Δr = , (5)
k − kr + iΔr θ (kr )
In order for the requirement stated above (5) to be filled, we must choose the
width δx of the incident packet in such a way that the following relation holds.
1
L δx (7)
Δr
Charging and Discharging Characteristics of a Quantum Well 121
With thus determined f (k) and C(k), we use (3) to calculate ψ(x, t), and then use
(1) to obtain P (t), where the k and the x integrations are calculated numerically.
Let us, for now, consider the initial wave packet given by
1 (x − x0 )2
ψG (x, 0) = exp − + ik0 x . (8)
4
2π(δx)2 4(δx)2
This is called a Gaussian wave packet, because |ψG (x, 0)|2 is a normal distribu-
tion. Equations (4) with (8) gives
2
4 2(δx) 2
fG (k) = exp −δx2 (k − k0 ) − i(k − k0 )x0 , (9)
π
where x0 and k0 are, respectively, the peak position and the central wave num-
ber of the initial wave packet, and δx is the standard deviation of the normal
distribution. Figure 2 shows a typical example of PG (t) which we obtained by fol-
lowing the steps described above. The parameters used are: x0 = −10.0 μm, δx =
0 .5 μm, k0 = kr 4 .9482915 nm−1 , d = 1.0 nm, V0 = 3.3 eV, L = 295.984 nm,
and m = 0.33 m0 (m0 is the bare electron mass). The use of these parameter
values ensures that the requirement stated above (5) is met (see Fig. 3). The
value of Δr is 0.029532489 pm−1 .
0.14
0.12
0.1
0.08
0.06
0.04
0.02
0
0 5 10 15 20 25 30
4.93 4.935 4.94 4.945 4.95 4.955 4.96 4.965
Fig. 2. The probability of finding an Fig. 3. Only one sharp resonance peak is
electron in the well region of the DTM covered by the weighting function fG (k).
type potential when the initial wave
packet is given by (8).
where we call ψL (x, 0) given above a Lorentzian wave packet, because |ψL (x, 0)|2
is a Lorentz function δx/π{(x − x0 )2 + δx2 }. Substituting (10) into (4), we have
√
2δx exp{−δx(k − k0 ) − i(k − k0 )x0 } (k ≥ k0 )
fL (k) = (11)
0 (k < k0 ).
Not only for the Gaussian incident packet given by (8), but also for the
Lorentzian incident packet given by (10), the wave number k0 determines the
velocity v0 of the packet peak such that v0 = k0 /m, and δx represents the
width of the packet. It should be mentioned here that, although δx in (8) and
that in (10) are both a measure of the “initial packet width,” the two δx have
slightly different meanings; δx in (8) is the standard deviation of the normal
distribution |ψG (x, 0)|2 , while δx in (10) is the half width at half maximum of
the Lorentz function |ψL (x, 0)|2 .
To make the comparison between PL (t) and PG (t) meaningful, we use the
same parameter values as were used in Sect. 2. The choice of the value of k0 is,
however, exceptional. If we use the same value of k0 as was used in Sect. 2, only
a half of the resonance peak is covered by the weighting function fL (k). To avoid
this, we use a slightly different value of k0 (4.9475501 nm−1 ) to make sure that
the resonance peak is fully covered by fL (k) as shown in Fig. 4. With this careful
choice of the parameter values, we obtain PL (t) shown in Fig. 5 by following the
numerical steps explained in Sect. 2.
0.18
0.16
0.14
0.12
0.1
0.08
0.06
0.04
0.02
0
4.947 4.948 4.949 4.95 4.951 4.952 4.953 4.954 0 5 10 15 20 25 30
Fig. 4. Only one sharp resonance peak is Fig. 5. The probability of finding an
covered by the weighting function fL (k). electron in the well region of the DTM
type potential when the initial wave
packet is given by (10).
Figure 7 shows log PL (t) and log PG (t). It is clear from the figure that PG (t)
shows an exponential decrease in the decay process. We have confirmed that
the exponent of the exponential decay agrees with theory; the slope of log PG (t)
obtained from the numerical data, −0.1025284 ps−1 , is very close to the theo-
retical value −0.1025318 ps−1 (= −2k0 Δr /m). From Figs. 6 and 7, we may say
that PL (t) shows a modulated exponential decay, where an exponential decay
with the same negative exponent as in the case of PG (t) is modulated by an
oscillation of a small amplitude.
Figure 8 is an enlarged view of the modulated exponential decay of PL (t).
Surprisingly, PL (t) increases between t = 11.56 ps and t = 11.85 ps. This implies
that the electron moves to the right in this time interval. The same thing happens
at later times as shown in Fig. 9. The temporary increase of the probability in
the well region means that the electron leaking from the well temporarily moves
backward to the well to re-charge it. This view is supplemented by Fig. 11,
where the flux (probability current density) is shown. The electron is moving
to the right when the flux takes a positive value, while it is moving to the left
when it takes a negative value. Figure 11 shows that, in the time interval where
the temporary increase is observed, the flux is positive, and thus the leaking
electron is moving backward to the well. Although the temporary increase of the
probability observed here is small, it can become much larger depending on the
system parameter values as we will see in the next subsection. Note that a large
temporary increase is of practical significance, because it would cause a memory
error of DTM. The cause of the temporary increase is explored in Sect. 3.4.
0.18 0.14 0
0.16 0.12
0.14 -4
0.1
0.12
0.08 -8
0.1
0.08 0.06 -12
0.06
0.04
0.04 -16
0.02
0.02
0 0 -20
0 5 10 15 20 25 30 0 5 10 15 20 25 30
Fig. 6. The comparison between PL (t) in Fig. 7. Natural log plot of PL (t) and
Fig. 5 and PG (t) in Fig. 2. PG (t).
Let us return to Fig. 7 to see the buildup process where the two curves show
rapid increase. We see there that log PL (t) is a convex function, while log PG (t) is
a concave function. This implies that the functional forms in the buildup process
are quite different between PG (t) and PL (t), although both curves in the buildup
process do not look so different in Fig. 6. We consider that the functional form
of P (t) in the buildup process is sensitive to the functional form of the incident
124 Y. Kubo and N. Yamada
0.094205 0.0007
0.0942
0.0006
0.094195
0.0005
0.09419
0.0004
0.094185
0.0003
0.09418
0.094175 0.0002
0.09417 0.0001
0.094165
0
11.5 11.6 11.7 11.8 11.9 12 80
55 60 65 70 75
160 0.2
0
120
-0.2
80 -0.4
-0.6
40
-0.8
0
-1
-1.2
-40
0 5 10 15 20 25 30 11.5 11.6 11.7 11.8 11.9 12 12.1
Fig. 10. The flux (the probability den- Fig. 11. The flux is positive during the
sity current) as the function of time. time interval when the backflow effect is
observed.
packet. From this viewpoint, it is not surprising that the functional form of PL (t)
seems very different from that of PG (t) in the buildup process, because the shape
of a Lorentzian packet and that of a Gaussian packet are very different.
The probability P (t) of finding an electron in the quantum well region at time
t depends on many system parameters. Here, among them, let us change the
“width” δx of the initial wave packet to see how P (t) is altered by the packet
size, where δx represents the half width at half maximum of the initial probability
density for a Lorentzian packet, while it represents the standard deviation of the
initial probability density for a Gaussian packet. Figures 12, 13, and 14 show
PL (t) for the case of δx = 500 nm, 1500 nm, and 3000 nm, respectively, where
other parameter values are the same as those used in Fig. 5 except for x0 , which
is −100 μm in the present examples.
Charging and Discharging Characteristics of a Quantum Well 125
0.18 0.14
0.16
0.12
0.12
0.08
0.08
0.04 0.04
0
0
40 60 80 100 120
40 60 80 100
Fig. 12. PL (t) with δx = 500 nm. Fig. 13. PL (t) with δx = 1500 nm.
10
9
0.06
8
7
0.04 6
5
4
0.02
3
2
0 1
40 60 80 100 120 500 1000 1500 2000 2500 3000
Fig. 14. PL (t) with δx = 3000 nm. Fig. 15. The dependence of the buildup
time on the initial packet width.
We first find that the temporary increase of PL (t) becomes more distinct as
the packet width becomes large. This means that the experimental detection of
the backflow effect would be easier if energetically monochromatic electrons are
used as incident electrons (due to the position-momentum uncertainty princi-
ple in quantum mechanics, a large packet width means a small uncertainty in
momentum, and thus a small uncertainty in energy of the electrons). As men-
tioned earlier, a large temporary increase of the probability is harmful to DTM.
The numerical examples shown here warn that non-Gaussian incident packets
could cause a fatal memory error; one would not notice this in numerical simu-
lations if only Gaussian initial packets are used in the simulations.
The Figs. 12, 13, and 14 also show that the peak value of PL (t) becomes
small as the packet width becomes large, which means that less electrons enter
the well region as the incident electrons become monochromatic in energy (thus
the experimental detection of the backflow effect becomes difficult if the packet
size is too large). We also notice that PL (t) stands up sharply in the build-up
process when the initial packet width is small. To characterize how sharply it
stands up (i.e., how fast the well is filled with electrons), let us define the buildup
time (charging time) by t2 − t1 , where t2 is the time at which PL (t) reaches its
maximum value and t1 is the time at which it starts to take an appreciable value.
The buildup time has a practical meaning for DTM, because it corresponds to
126 Y. Kubo and N. Yamada
the memory’s writing time. Putting t1 to be the time at which the value of
PL (t) reaches 1 % of its maximum value PL (t2 ), we have calculated the buildup
time for some values of packet width. The result is shown in Fig. 15. The figure
shows that, both in the Gaussian case and in the Lorentzian case, the buildup
time increases as the initial packet size increases. This is reasonable because
a longer object takes a longer time to enter the “room” (the well region). It is
interesting to note that the relationship between the buildup time and the packet
size is almost linear in the Gaussian case, while it is apparently non-linear in the
Lorentzian case.
0.02 0.006
0.016 0.005
0.004
0.012
0.003
0.008
0.002
0.004
0.001
0 0
65 70 75 80 85 90 60 62 64 66 68 70 72 74
Fig. 16. Oscillatory behavior PL (t) in Fig. 17. Oscillatory behavior PL (t) in
the decay process. the decay process.
0.005 0.0009
0.004 0.0008
0.003 0.0007
0.002 0.0006
0.001 0.0005
0 0.0004
60 62 64 66 68 70 72 74 65 65.5 66 66.5 67 67.5 68
Fig. 18. Oscillatory behavior PL (t) in Fig. 19. Oscillatory behavior PL (t) in
the decay process. the decay process.
interference between the waves with wave numbers k0 and kr . The cause of the
counterintuitive backflow effect is, therefore, a purely quantum mechanical one.
The parameter values we used are as follows: As to Figs. 16 and 17:
δx = 3000 nm and kr = 4.94829150000208 nm−1 for both figures; k0 =
4.9475501 nm−1 for Fig. 16 and k0 = 4.94724240001316 nm−1 for Fig. 17. As
to Figs. 18 and 19: δx = 500 nm for both figures; kr = 4.714319700047028 nm−1
and k0 = 4.7130 nm−1 for Fig. 18; kr = 5.160935000005426 nm−1 and k0 =
5.156214700132405 nm−1 for Fig. 19. For Figs. 16, 17, 18, and 19, m = 0.33 m0 ,
x0 = −100000 nm, V0 = 3.3 eV, d = 1 nm, L = 295.984 nm.
Incidentally, although not in the context of DTM, the quantum backflow
effect has also been known for superposition of Gaussian wave functions (see, for
example, Ref. [7]), where the size of the effect is relatively small. The present
study shows that the backflow effect can be made more distinct with a Lorentzian
initial wave function. It also shows that the quantum backflow effect, which is
often studied in a purely academic context, can also be a practical issue in device
applications.
128 Y. Kubo and N. Yamada
4 Summary
We have studied the probability PL (t) to find an electron in the well region of a
DTM type potential when the incident packet is a Lorentzian, and compare the
result with the well-studied case of a Gaussian incident packet.
The obtained PL (t) shows a rapid increase and a slow decay as in the case
of PG (t) for the Gaussian incident packet. In details, however, PL (t) differs from
PG (t) both in the buildup process and in the decay process. In the buildup
process, log PL (t) is a convex function, while log PG (t) is a concave function,
implying that the functional form of PL (t) is very different from that of PG (t).
The buildup time shows a linear dependence on the packet width in the Gaussian
case, but a non-linear dependence in the Lorentzian case. In the decay process,
PL (t) shows a modulated exponential decay. The modulation amplitude increases
as the initial packet width becomes large, and PL (t) eventually begins to oscil-
lates when the packet width exceeds a certain value. The oscillation of PL (t)
periodically causes a temporary increase of the probability. A physical view of
this phenomenon is that the electron leaking from the well periodically changes
its direction of motion and moves backward to the well to re-charge it. The
backflow effect has a practical significance, because it could cause a fatal mem-
ory error of DTM. The backflow effect is counterintuitive and is caused by the
quantum mechanical interference between the waves with wave numbers k0 and
kr , where k0 is the most dominant wave number of the initial wave packet, and
kr is the resonance wave number of the well.
References
1. Kuo, C., King, T.J., Hu, C.: Direct tunneling RAM (DT-RAM) for high density
memory applications. IEEE Electron Dev. Lett. 24, 475–477 (2003)
2. Tsunoda, K., Sato, A., Tashiro, H., Nakanishi, T., Tanaka, H.: Improvement in
retention/program time ratio of direct tunneling memory (DTM) for low power
SoC applications. IEICE Trans. Electron. 88-C(4), 608–613 (2005)
3. Kubo, Y., Yamada, N.: Novel method for simulating quantum buildup process in the
potential well of a DTM potential. In: Proceedings of 33rd JSST Annual Conference
(JSST 2014), pp. 70–71 (2014)
4. Inaba, H., Nakagawa, J., Kurosawa, K., Okuda, M.: Dynamics of resonant tunneling
in double-barrier structures with trapezoidal potential profile. Jpn. J. Appl. Phys.
30, L544–L546 (1991)
5. Hauge, E.H., Støvneng, J.A.: Time-dependent resonant tunneling of wave packets
in the tight-binding model. Phys. Rev. B 44, 13582–13594 (1991)
6. Gong, J., Liang, X.X., Ban, S.L.: Tunneling time of electronic wave packet through
a parabolic quantum well with double barrier. Phys. Stat. Sol. (B) 244, 2064–2071
(2007)
7. Yearsley, J.M., Halliwell, J.J.: An introduction to the quantum backflow effect. J.
Phys.: Conf. Ser. 442, 012055 (2013)
Author Index