ASCE Journal of Computing in Civil Engineering, Volume 14, 2000 PDF
ASCE Journal of Computing in Civil Engineering, Volume 14, 2000 PDF
TECHNICAL PAPERS
River Stage Forecasting in Bangladesh: Neural Network Approach
Shie-Yui Liong, Wee-Han Lim, and Guna N. Paudyal . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
Artificial Neural Network for Measuring Organizational Effectiveness
Sunil K. Sinha and Robert A. McKim . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9
Augmented IFN Learning Model
Shih-Lin Hung and J. C. Jan . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15
Comparison of Neural and Conventional Approaches to Mode Choice Analysis
Tarek Sayed and Abdolmehdi Razavi . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23
Genetic Programming-Based Approach for Structural Optimization
Chee Kiong Soh and Yaowen Yang . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31
Finite State Verification of Intelligent Transportation Systems
John W. Baugh Jr. and GopalaKrishna Reddy Kakivaya . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 38
IntelliGIS: Tool for Representing and Reasoning Spatial Knowledge
Xudong Jia . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 51
Automatic Road Detection in Grayscale Aerial Images
Katherine Treash and Kevin Amaratunga . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 60
Integrating Document Management with Project and Company Data
Dany Hajjar and Simaan M. AbouRizk . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 70
RIVER STAGE FORECASTING IN BANGLADESH:
NEURAL NETWORK APPROACH
By Shie-Yui Liong,1 Wee-Han Lim,2 and Guna N. Paudyal3
ABSTRACT: A relatively new approach, artificial neural network, was demonstrated in this study to be a highly
suitable flow prediction tool yielding a very high degree of water-level prediction accuracy at Dhaka, Bangladesh,
even for up to 7 lead days. The goodness-of-fit R2 value, root-mean-square error, and mean absolute error,
ranging from 0.9164 to 0.9958, 0.0788 to 0.2756 m, and 0.0570 to 0.2050 m, respectively, were obtained from
the training and verification simulation. In addition, the high degree of accuracy is accompanied with very small
computational time. Both results make the artificial neural network a desirable advanced warning forecasting
tool. Sensitivity analysis was also performed to investigate the importance of each of the input neurons. The
sensitivity study suggested a reduction of three from eight initially chosen input neurons. The reduction has
insignificantly affected the prediction accuracy level. The finding enables the policymakers to reduce the un-
necessary data collection at some gauging stations and, thus, results in lower costs.
1
ANN is a computing paradigm designed to mimic the hu-
Assoc. Prof., Dept. of Civ. Engrg., Nat. Univ. of Singapore, 10 Kent man brain and nervous system. ANN has a big role to play in
Ridge Crescent, Singapore 119260.
2
Res. Engr., Dept. of Civ. Engrg., Nat. Univ. of Singapore, 10 Kent
the field of hydrology where complex natural processes dom-
Ridge Crescent, Singapore 119260. inate. The high degree of empiricism and approximation in the
3
Chf. Tech. Adviser, Surface Water Modelling Centre, House No. 476, analysis of hydrologic systems may find the use of ANN
Rd. 32 New DOHS, Dhaka, Bangladesh. highly suitable. In other words, when the possibility of rep-
Note. Editor: Sivand Lakmazaheri. Discussion open until June 1, 2000. resenting the complex relationships between various aspects of
To extend the closing date one month, a written request must be filed the processes in terms of physical or conceptual modeling is
with the ASCE Manager of Journals. The manuscript for this paper was
submitted for review and possible publication on December 16, 1998.
very remote, the ANN may play an important role.
This paper is part of the Journal of Computing in Civil Engineering, ANN problems can be classified into several groups such
Vol. 14, No. 1, January, 2000. 䉷ASCE, ISSN 0887-3801/00/0001-0001– as pattern association, pattern classification, pattern clustering,
0008/$8.00 ⫹ $.50 per page. Paper No. 19885. and pattern mapping. In this study, the problem of capturing
JOURNAL OF COMPUTING IN CIVIL ENGINEERING / JANUARY 2000 / 1
planning and design purposes, but a series of detailed infor-
mation is needed. The physical model also has been applied
quite successfully (Samuels and Reeve 1991; Syme and Pau-
dyal 1994; and SWMC 1996). The model makes use of the
capability of geographic information system technology to re-
trieve useful information such as topography, soil character-
istics, etc. Modeling a large river network such as that in
Bangladesh can pose a potential problem to modelers as large
volumes of data are needed. Model calibration is relatively
difficult because either the required data are not available or
data are expensive and time consuming to collect.
The application of ANN for advanced flood forecasting is a
promising alternative approach because it requires minimal
data. Feedforward NNs are closely related to statistical models
that are a data-driven approach and more suited to forecasting
applications. One of the major strengths of ANNs is their abil-
ity to recognize or correctly classify patterns that have never
been presented to the network before. ANNs appear unique in
their ability to extract the essential features from a training set
FIG. 1. Typical NN Configuration
and use them to identify new inputs or generalization.
The ANNs’ application to simulation and forecasting prob-
lems in water resources is relatively recent, particularly to hy-
the complex relationships between the input and the output drologic problems. French et al. (1992) used the ANN to fore-
patterns falls under the category of pattern mapping. Although cast rainfall patterns in space and time. A series of applications
there are many types of ANNs, by far the most widely used of ANN to catchment runoff forecasting followed. They in-
and well suited to this system is the feedforward NN or the clude works of Karunanithi et al. (1994), Bonafe et al. (1994),
multilayer perceptron, which is organized as layers of com- Hsu et al. (1995), Cheng and Noguchi (1996), and Liong et
puting elements (called neurons) connected by weighted con- al. (1996) to predict flow at the catchment outlet with inputs
nections between layers [Fig. 1(a)]. Typically, there is an input such as rainfall, upstream flow, and/or temperature only. Re-
layer, which receives inputs from the environment, an output cently, Thirumalaiah and Deo (1998) applied ANN to forecast
layer, which outputs the network’s response to its input, and river stage at a downstream station for different lead days us-
one or more intermediate layers called the hidden layers. ing only the most recent river stage at various stations as in-
The action of each neuron is to compute a response from puts. They determine the number of input stations for ANN
the weighted sum net of its inputs Xi and bias from neurons using a trial-and-error method. In the present study, however,
connected to it, using it predetermined activation function a more systematic way to determine the number of input sta-
F(net) [Fig. 1(b)]. The output of a neuron becomes the inputs tions using sensitivity analysis will be presented.
of other neurons in a higher layer. Many types of activation
function are in use; however, the most widely used is the sig-
moid function. Mathematically the above two steps may be
expressed as
冘
N
net = wi Xi ⫹ (1)
i=1
1
F(net) = (2)
1 ⫹ e⫺net
冘
p
[(Xm)i ⫺ (Xs)i]2
冘
R =1⫺
2 i=1
p (4)
[(Xm)i ⫺ (X
¯ m)i]2
i=1
冑冘
p
1
RMSE = [(Xm)i ⫺ (Xs)i]2 (5)
p i=1
冘
p
1
MAE = 兩 (Xm)i ⫺ (Xs)i 兩 (6)
p i=1
SENSITIVITY ANALYSIS
The two sensitivity tests are carried out to determine the
relative significance of each of the input neurons. Both anal-
yses aim to reduce the number of input neurons should they
prove to be insignificant to model performance. Reduction of
input neurons would result in a decrease in unnecessary data
collection that in turn leads to cost reduction.
The first sensitivity test, a step-by-step method, is carried
out on the trained forecaster by varying each of the input neu-
rons, one at a time, at a constant rate. Various constant rates
(⫺5, 5, 10, and 20%) are selected in the study. For every input
neuron, the percentage change in the output, as a result of the
change in the input neuron, is observed. The sensitivity of each
input neuron is calculated by the following:
冘冉 冊
N
1 % change in output
Sensitivity Level of Xi (%) = ⫻ 100
N j=1 % change in input j
(7)
where N (=841) = number of data sets used in the study. FIG. 4. Comparison between Simulated and Measured Water
Results show that the change in output neuron is unaffected Levels with Eight Input Neurons for Verification Set: (a) 1 Lead
by the rate an input neuron (⫺5, 5, 10, or 20%) is varied. Fig. Day; (b) 3 Lead Day; (c) 5 Lead Day; (d) 7 Lead Day
5(a) shows how sensitive the water-level changes at each of
the gauging stations are and how they will affect the changes water level at Dhaka); this unexpected response may imply
in water level at Dhaka. ST 1 (Dhaka itself) and ST 3 are the that water at ST 6 and ST 8 do not flow toward Dhaka at all.
two most sensitive gauging stations. ST 6 and ST 8 have a ST 6 and ST 8 are then eliminated from the input neurons.
reverse impact on Dhaka water level (i.e., increasing water Thus, the total number of input neurons reduces to six. A sim-
levels at ST 6 and/or ST 8 will result in a decrease of the ilar sensitivity test is again carried out on the model with six
4 / JOURNAL OF COMPUTING IN CIVIL ENGINEERING / JANUARY 2000
FIG. 5. Sensitivity Levels of Input Stations for 1 to 7 Lead Days Using Step-by-Step Sensitivity Test: (a) Eight Input Stations; (b) Six
Input Stations; (c) Five Input Stations
input neurons. Results are presented in Fig. 5(b). ST 1 and ST and the output neuron. The input and output neurons are first
3 still are the most sensitive stations, whereas ST 5 displays related through a power function [(8)], that is then transformed
an insignificant reverse impact. The sensitivity test is, once into a logarithmic equation [(9)] as shown below
again, conducted on the model with only five input neurons
(without ST 5). The final sensitivity of each of the five input X1(t) = b0[X1(t ⫺ ␣)]b1[X2(t ⫺ ␣)]b2 ⭈ ⭈ ⭈ [X8(t ⫺ ␣)]b8 (8)
neurons is shown in Fig. 5(c).
The second sensitivity test, a multiple regression based sen- ln[X1(t)] = ln b0 ⫹ b1 ln[X1(t ⫺ ␣)] ⫹ b2 ln[X2(t ⫺ ␣)]
sitivity analysis, is carried out by observing the coefficients of
the log linear regression equation relating the input neurons ⫹ ⭈ ⭈ ⭈ ⫹ b8 ln[X8(t ⫺ ␣)] (9)
where bi = model coefficients estimated using multiple linear both models. This implies that the sensitivity analysis con-
regression. Each of the model coefficients can be interpreted ducted has indeed removed the insignificant gauging stations.
as the sensitivity of the input neuron relative to the output As can be seen in Figs. 7(a–c), the training sets for different
neuron. The larger the coefficient is, the more sensitive is the lead days, as expected, yield higher R2 values, and thus less
corresponding input neuron to the output neuron. RMSE and MAE, than the verification sets. However, a similar
Fig. 6 shows the sensitivity level of each of the eight input conclusion cannot be drawn for maximum error and error’s
neurons. Results from this regression based sensitivity test standard deviation [Figs. 7(d and e)]. As the number of lead
show a similar trend as that in the step-by-step sensitivity test days increases, again as expected, the R2 values decrease in
[i.e. only five gauging stations (ST 1, ST 2, ST 3, ST 4, and both models. Both maximum error and error’s standard devi-
ST 7) are sensitive to Dhaka water level]. ation in the verification sets are smaller than that of the train-
ing sets. This implies that the training data sets have been well
selected to represent the available data sets. Comparison be-
DHAKA RIVER STAGE FORECASTING WITH FIVE tween measured and simulated water levels resulting from the
INPUT NEURONS five-input neuron model yields an almost identical plot as that
of the eight-input neuron model (Fig. 4).
With the five significant input neurons a new NN forecaster A sample of advanced forecasting on the verification set
is then trained. Table 2 lists the R2 values, RMSE, MAE, and using the five-input neuron model is shown in Fig. 8. Fig. 8(a)
other measures of prediction accuracy resulting from the five- shows a comparison between the measured and simulated wa-
input neuron model. The R2 values, RMSE, and MAE for ter levels on the verification set for 7 lead days. Results are
training and verifications range from 0.9164 to 0.9958, 0.0788 also presented in a scatter plot [Fig. 8(b)]. Figs. 8(a and b)
to 0.2756 m, and 0.0570 to 0.2050 m, respectively, meaning show that the simulated river stages at Dhaka mimic their mea-
that very good predictions with relatively small errors have sured counterparts very closely. Most of the prediction errors
been achieved. Figs. 7(a–e) graphically display the differences are within the ⫾0.5 m error range. It should be noted that the
in values (R2 values, RMSE, MAE, maximum error, and er- maximum error occurs at the beginning of the monsoon sea-
ror’s standard deviation, respectively) listed in Tables 1 and 2 son, of the year 1991, at which the flow was very low. The
resulting from the eight- and five-input neuron models. No relatively large error prediction may be due to the small num-
significant difference is observed between the results from ber of training patterns available at this low flow range.
FIG. 8. Simulated and Measured Daily Water Levels for Five-Input Neuron Model at 7 Lead Days—Verification Set: (a) Water-Level
Plot; (b) Scatter Plot
ABSTRACT: Solving engineering problems is a creative, experiential process. An experienced engineer gen-
erally solves a new problem by recalling and reusing some similar instances examined before. According to
such a method, the integrated fuzzy neural network (IFN) learning model was developed and implemented as
a computational model for problem solving. This model has been applied to design problems involving a
complicated steel structure. Computational results indicate that, because of its simplicity, the IFN model can
learn the complicated problems within a reasonable computational time. The learning performance of IFN,
however, relies heavily on the values of some working parameters, selected on a trial-and-error basis. In this
work, we present an augmented IFN learning model by integrating a conventional IFN learning model with two
novel approaches—a correlation analysis in statistics and a self-adjustment in mathematical optimization. This
is done to facilitate the search for appropriate working parameters in the conventional IFN. The augmented IFN
is compared with the conventional IFN using two steel structure design examples. This comparison reveals a
superior learning performance for the augmented IFN learning model. Also, the problem of arbitrary trial-and-
error selection of the working parameters is avoided in the augmented IFN learning model.
冘
M
cally depicted in Fig. 1.
Assume that U is an associated instance base, with N solved dij = diff (Ui,i, Uj,i) = ␣m(wi u mi ⫺ wj u mj )2 (1)
m=1
=
再0
Rmax Rmin ⫺ Rmindij
1
(Rmax ⫺ Rmin)dij
if dij ⱖ Rmax
if Rmin < dij < Rmax
冘
p
assorted from resembling samples in the case of dij less than
or equal to a prescribed value, say t. The elements in At and
k Sk,o Bt are the first and second items, respectively, of these resem-
冘
k=1
Xo = p (COG) (4) bling samples. Next, the accumulative correlation coefficient,
k Ac㛭CORREL(At, Bt, t), is calculated for arrays At and Bt with
k=1 the degree of difference less than or equal to t. Assume that
冘
C for a total P resembling samples with dij less than a prescribed
kmaxS max,o
k t, the arrays At and Bt can be denoted as
冘
k=1
Xo = C (MOM) (5) At = {ak 兩ak = Ui,o 僆 Sij , for dij ⱕ t} = {a1, a2, . . . , ap}
max
k Bt = {bk 兩bk = Uj,o 僆 Sij , for dij ⱕ t} = {b1, b2, . . . , bp}
k=1
冘
p
The reasoning process of the UFN depends on determining 1
the degree of similarity among X and Uj . Consequently, no Cov(At, Bt, t) = (ak ⫺ At)(bk ⫺ Bt) s.t. dij ⱕ t (7)
p k=1
solution can be generated by the UFN reasoning model if the
new instance entirely differs from all instances in the instance where At and Bt are standard errors of arrays At and Bt; At
base, e.g., all dXi are greater than Rmax. In addition, using in- and Bt are the means of At and Bt. The formulas expressed in
appropriate working parameters would allow for the possibility (6) and (7) represent the relationship between the accumulative
that no similar instances can be derived. For the above issues, correlation coefficient to any value of t. An accumulative cor-
the undertrained adaptive L-BFGS supervised neural network relation curve can be plotted as a function of t and Ac㛭COR-
is used as an assistant system to generate an approximate out- REL(At, Bt, t). Fig. 2 demonstrates an accumulative correlation
put for the new instance. curve for two arrays At = sin (x) and Bt = 0.95(x ⫺ x 3/3! ⫹
x 5/5!) such that x = ⫺ ⫹ (/15)i, i = 0 to 30. Shown in Fig.
AUGMENTED IFN LEARNING MODEL 2, the curve falls from one to zero as the value of t increases,
and it resembles the quasi-Z-type fuzzy membership function
In the conventional IFN learning model, working param- defined in (2) for the case of Rmin equal to zero. Note that the
eters (such as wj , ␣m, Rmax, and Rmin) are selected subjectively
by users, and generally, on a trial-and-error basis. Conse-
quently, the learning performance is highly affected by these
parameters, especially Rmax and ␣m. In this work, two novel
approaches are employed for assisting the users to determine
these parameters and weights systematically. One approach,
correlation analysis in statistics, is used to determine the ade-
quate Rmax value in the fuzzy membership function. The other
approach, self-adjustment based on mathematical optimization
theory, is employed to find proper values for weights ␣m.
冘
the two arrays refers to the case in which the instances in the N
two sets are identical and the value of Ac㛭CORREL(At, Bt, t) ⭸E(␣m) 1 ⭸Ei
g= = (11)
equals one. In such a case, no solution to a new instance can ⭸␣m 2N i=1 ⭸␣m
be generated via the UFN reasoning model except for when
冘 冉冘 冊
⫺2
⭸Ei
identical instances exist in the instance base. In order to avoid =2 ( yk ⫺ oik ) f (dip)
this issue here, we set Ac㛭CORREL(At, Bt, t) equal to 0.8 as ⭸␣m k p
冉冘 冘
the lower bound for similarity measurement. The value of t
corresponding to this lower bound is adopted as the appropri- ⫻ opk f ⬘(dip)(u im ⫺ u pm)2 f (dip)
ate value of R max. p p
冘
K
Ei = ( yk ⫺ oki )2 (9)
k=1
where
(g (s))Tg (s)
(s) = , and g (s) = ⫺ⵜE(␣m(s))
(g (s⫺1))Tg (s⫺1)
FIG. 3. Augmented IFN Learning Model
Cw Iy ⫹ EIy GJ
in input, using the self-adjustment approach. where Mp = plastic moment strength; Lp and Lr = limited
• Analysis phase (after learning phase is completed) laterally unbraced length for full plastic bending capacity
Step 6. Present the new (unsolved) instance X to the and for inelastic lateral-torsional buckling, respectively;
UFN reasoning model, and perform a similarity mea- Cb = moment gradient coefficient; Mr = limited buckling
surement between X and instance Ui in the base U using moment; E = elasticity modulus of steel; Cw = warping
a single-layered, lateral-connected competing network. constant; Iy = moment of inertia about the minor axis;
Step 7. If more than one similar instance is found in G = shear modulus of elasticity of steel; and J = torsional
Step 6, generate the solution Xo for the new instance constant.
using fuzzy synthesis approaches in (4) or (5), and go 4. Finally, confirm whether or not the section of selected
to Step 9. Otherwise go to next step. member satisfies the requirements of flexural design
Step 8. Compute the solutions via the undertrained as- strength, 0.9Mn ⱖ Mu, and shear design strength, 0.9Vn
sistant adaptive L-BFGS supervised learning model and ⱖ Vu, where Vn is the nominal shear strength.
go to next step.
Step 9. Feedback the new instance into the base U. The above steps are repeated until a satisfactory lightest mem-
Meanwhile, further learning in the assistant supervised ber is obtained. Eight hundred instances, created according to
learning model is launched offline. the aforementioned design process, are used in the present ex-
ample to train and verify the conventional and augmented IFN
learning models. Of these, 600 are used as training instances
APPLICATIONS and the remaining two hundred (200) are used as verification
instances. Seven decision variables are used as inputs in order
to determine the plastic modulus Z of a lightest W-section steel
Here the novel augmented IFN learning model is applied to beam with compact section. The seven decision variables are
problems of engineering design. Two structural design exam- the yielding strength of steel fy, factored maximum bending
ples are presented to verify the learning performance of the moment Mu, the live load wl, the factored maximum shear
augmented IFN learning model. The first example is a steel force Vu, the span of the beam L, the moment gradient coef-
beam design problem. The second is the preliminary design ficient Cb, and the unbraced length Lb. Notably, the first de-
of steel structure buildings. The two examples are also used cision variable, fy , was not used in the earlier work (Hung and
to train a conventional IFN learning model for comparison. Jan 1999), as the yielding strength of steel was identical for
each instance.
The parameters Rmin and wi are set as constant 10⫺5 and one,
Steel Beam Design Problem respectively. The weights ␣m, however, are initialized as [1, 9,
1, 1, 1, 1, 1] for the seven decision variables. Of these, the
The problem is the design of the lightest W-section simply second weight, ␣2, for decision variable, factored maximum
supported steel beam with compact section, under LRFD Spec- bending moment Mu, is set based on the finding investigated
ification (Manual 1994). The problem was studied in our ear- in the earlier work (Hung and Jan 1999). Using these param-
lier work (Hung and Jan 1999), with those results demonstrat- eters and weights, the augmented IFN learning model is
ing that the conventional IFN learning model is superior to a trained in four different cases with different numbers of train-
stand-alone supervised neural network with L-BFGS learning ing instances in base V. The number of training instances is
algorithm when the number of training instances was large increased from 100 to 400 with an increment of 100. After
enough. This work investigates only the learning performances performing the correlation analysis in four different bases of
of the conventional IFN and the novel augmented IFN learning V, the accumulative correlation coefficients can be computed
models. with respect to any specified value t. With the value of Ac㛭
The LRFD design method, applied to the effect of lateral CORREL(At, Bt, t) equal to 0.8, the values of t(=Rmax) for four
torsional buckling for beam with compact section, is summa- different cases are obtained. Those values are 0.044, 0.034,
rized in the following steps: 0.033, and 0.030, respectively, for 100, 200, 300, and 400
training instances in base V, and are displayed in Fig. 4. In-
terestingly, according to this figure, the greater the number of
1. Determine the factored moment Mu at midspan of a given instances in bases V implies a higher correlation among these
beam. instances. Restated, the value of Rmax decreases with an in-
2. Select one potential member with plastic section modu- crease of the number of training instances. After the value of
lus Z so that the required nominal moment Mn = Zfy sat- Rmax is determined, the self-adjustment approach is launched
isfies 0.9Mn ⱖ Mu, where fy is the specified yield strength to search the appropriate weights ␣m for each decision variable
of the beam. in the input.
3. Estimate whether the lateral supports are close enough After self-adjustment is achieved, a set of weights ␣m is
to design the beam using plastic analysis or whether we obtained, e.g., the weights ␣m are adjusted to values [1.42,
should use the fully plastic moment strength Mp without 6.93, 1.02, 1.01, 1.01, 1.51, 0.73] for the case of V with 400
plastic analysis, for a given unbraced length Lb. The training instances. Note that the first, second, and sixth weights
nominal moment Mn is then obtained according to the (␣1, ␣2, and ␣6) are changed by more than the other weights.
following formulas with different conditions: These adjustments illustrate that the first and sixth decision
JOURNAL OF COMPUTING IN CIVIL ENGINEERING / JANUARY 2000 / 19
for the 200 verification instances. According to this table, the
learning performance is improved as the number of training
instances increases for the augmented and conventional IFN
learning models. This same table reveals that the average er-
rors for verification instances are 3.47% and 4.47% for the
augmented and conventional IFN learning models, respec-
tively. In sum, the learning performance of the augmented IFN
learning model is much better than that of the conventional
IFN learning model in this example.
Based on the newly adjusted weights ␣m, correlation anal-
ysis is next performed for the case of 400 instance bases in V
again. According to those results, the accumulative correlation
coefficients are higher for any specified value t than those ob-
tained from the initial set of weights. For example, the accu-
mulative correlation coefficient is from 0.8 to 0.9 at the value
t equal to 0.03 (Fig. 4).
FIG. 4. Accumulative Correlation Curve for Steel Beam De-
sign Example Preliminary Design of Steel Structural Buildings
In the complete design of a structure, the preliminary design
variables, fy and Cb, are more important than we would be led stage is mainly a creative, experiential process that involves
to believe by a process of trial and error. However, the weight the choice of structure type, selection of the material, and de-
␣2 for the second decision variable, Mu, is adjusted from 9 to termination of the sections of beams and columns in the struc-
6.93. The occurrence reveals that the factored moment, Mu, is ture. An experienced engineer is likely to carry out this stage
the most important decision variable in the input, as compared more quickly than does an inexperienced one. The basic con-
with other decision variables. Then, the 200 verification in- figuration of the structure at this stage should satisfy the spec-
stances are used to verify the learning performance of the con- ified design code, such as LRFD for steel structures. To satisfy
ventional and augmented IFN learning models. The verifica- the prescribed constrains and achieve minimum expenditure
tion in the augmented IFN is based on the newly adjusted for materials and construction, this stage becomes a looped
weights ␣m with the appropriate Rmax. The working parameters optimization decision-making process. Hence, a good initial
and weights used in the conventional IFN, however, are se- development of the basic form, with sections of beams and
lected by trial and error. Fig. 5 compares the computed and columns satisfying the aforementioned constraints, will reduce
desired outputs of the 200 verification instances for the aug- the number of redesign cycles. After a basic structure is de-
mented and conventional IFN learning models. The correlation termined, the structural analysis stage involves analyzing the
coefficients for the computed and desired outputs are 0.997 initial guessed structure and computing the service loads on
and 0.992 for the augmented and conventional IFN learning the members. Also, the maximum lateral drift of the structure
model, respectively. Table 1 summarizes the computing results and the drifts between floors are computed if lateral loads are
considered. The present example involves a complete design
structure that satisfies the conditions that the service loads
should not exceed the strength of the members; the drifts
should be within the prescribed limits, and the structure should
be economical in material (e.g., minimum weight), construc-
tion, and overall cost. In this example, the augmented IFN
learning model is trained to learn how to implement the pre-
liminary design of buildings satisfying the conditions of utility,
safety, and economy in only one design cycle. For simplicity,
only regular buildings with a rectangular layout—such as
most factory buildings—are considered herein. Also, the
beams in every floor have the same sectional dimensions, as
do the columns.
In this example, 416 instances are used. They are randomly
divided into 380 training instances and 36 verification in-
stances. Seven decision variables are used as inputs to deter-
mine the sections of beams and columns of a building that
FIG. 5. Comparison of Computed and Desired Outputs for satisfies the given specifications. The seven decision variables
Verification Instances: Steel Beam Design Example and their limits are described as follows:
TABLE 1. Results of Steel Beam Design Problem 1. Number of stories = [9, 10, 11, 12, 13, 14, 15]
2. Bay length in long-span direction (X direction) = 9 to 12
Average Error (Rmax) for 200
m
Verification Instances
3. Bay length in short-span direction (Y direction) = 6 to 9
Number of training Conventional IFN Augmented IFN m
instances in U(V) learning model learning model 4. Number of bays in both directions = [3, 4, 5]
(1) (2) (3) 5. Seismic zone coefficient = [0.18, 0.23, 0.28, 0.33]
100 (100) 9.33% (0.040) 8.31% (0.044) 6. Live load (kgw/m2) = 200 to 350
200 (200) 8.31% (0.040) 5.50% (0.034) 7. Wall load (kgw/m) = [100, 200]
300 (300) 5.57% (0.040) 4.47% (0.033)
400 (400) 4.92% (0.040) 3.82% (0.030)
600 (400) 4.47% (0.040) 3.47% (0.030)
Other corresponding decision variables used in the stage of
preliminary design are assumed to be constant. In practice, the
20 / JOURNAL OF COMPUTING IN CIVIL ENGINEERING / JANUARY 2000
sections of beams and columns of a building are classified into
certain groups for convenience in construction, instead of sep-
arate consideration being given to each element. Here, a build-
ing with three groups of steel elements in both beams and
columns is considered. The three groups are upper, medium,
and lower. For a building with N stories, these three groups
are defined as upper group: floors from 2N/3 to N; medium
group: floors from N/3 to 2N/3; and lower group: floors from
one to N/3. An instance contains seven decision variable inputs
and six data items as outputs.
The parameters Rmin and wi are set as constant 10⫺5 and one,
respectively. The weights ␣m, however, are initialized as [1, 1,
1, 1, 1, 1, 1] for the seven decision variables. Using these
parameters and weights, correlation analysis in the augmented
IFN learning model is performed first to determine the work-
ing parameter Rmax. Fig. 6 displays the result of correlation
analysis. With the value of Ac㛭CORREL(At, Bt, t) equal to 0.8, FIG. 7. Comparison of Computed and Desired Outputs
the values of t(=Rmax) are obtained. They are 0.12 for beams (Beams) for Verification Instances: Preliminary Design of Steel
and 0.178 for columns, respectively. Structural Buildings
After the fuzzy membership function is defined, the self-
adjustment approach is launched to obtain the adequate
weights ␣m for each decision variable in the input. The weights
␣m for each decision variable are updated from [1, 1, 1, 1, 1,
1, 1] to [1.471, 1.369, 0.416, 0.008, 1.104, 0.825, 0.513] for
beams and to [1.106, 0.997, 0.542, 0.009, 1.232, 0.999, 0.997]
for columns. Interestingly, the weights of the fourth decision
variable in beams and columns are both self-adjusted close to
zero. This observation indicates that this decision variable
(number of bays in both directions) is insignificant in the input.
Consequently, this decision variable can be neglected. The 36
verification instances are used to verify the learning perfor-
mance of the augmented and conventional IFN learning mod-
els, respectively. Notably, the augmented IFN is verified on
the basis of the newly adjusted weights ␣m with the adequate
Rmax. The working parameters and weights used in the con-
ventional IFN, however, are selected on a trial-and-error basis.
Fig. 7 depicts the correlation between the computed and de- FIG. 8. Comparison of Computed and Desired Outputs (Col-
sired outputs of beams for the 36 verification instances for the umns) for Verification Instances: Preliminary Design of Steel
augmented as well as conventional IFN learning models. Sim- Structural Buildings
ilarly, Fig. 8 displays the correlation between the computed
and desired outputs for columns for the thirty-six verification TABLE 2. Results of Preliminary Structural Design Problem
instances using the two IFN learning models. Table 2 sum-
marizes the learning results for thirty-six verification instances. NUMBER OF ITEMS IN OUTPUT FOR
According to this table, the average percentage errors for VERIFICATION INSTANCES
beams and columns are 13.81 and 9.36 for the conventional Conventional IFN Augmented IFN
IFN learning model. However, these errors are reduced to 6.17 Error Learning Model Learning Model
and 6.1 for beams and columns, respectively, for the aug- (%) Beam Column Beam Column
mented IFN learning model. The augmented IFN learning (1) (2) (3) (4) (5)
model significantly improves in terms of learning. This ex-
0⬃5 12 43 52 51
5 ⬃ 10 29 29 39 31
10 ⬃ 15 21 22 13 20
15 ⬃ 20 29 10 3 6
>20 17 4 1 0
Average error for 36 ver-
ification instances 13.81% 9.36% 6.10% 6.17%
Note: Each instance has six items, three beams, and columns, in out-
put.
CONCLUDING REMARKS
This work presents an augmented IFN learning model by
integrating two newly developed approaches into a conven-
FIG. 6. Accumulative Correlation Curves for Example of Pre- tional IFN learning model. These approached are a correlation
liminary Design of Steel Structural Buildings analysis in statistics and self-adjustment in mathematical op-
JOURNAL OF COMPUTING IN CIVIL ENGINEERING / JANUARY 2000 / 21
timization, which collaboratively enhance the learning capa- model for image recognition.’’ Integrated Computer-Aided Engrg.,
bility of the conventional IFN. The first approach, correlation 1(1), 43–55.
Adeli, H., and Hung, S. L. (1995). Machine learning—neural networks,
analysis in statistics, assists users in determining the appro- genetic algorithms, and fuzzy systems. Wiley, New York.
priate working parameter used in the fuzzy membership func- Adeli, H., and Park, H. S. (1995). ‘‘Counterpropagation neural network
tion. The second approach, self-adjustment in mathematical in structural engineering.’’ J. Struct. Engrg., ASCE, 121(8), 1205–
optimization, obtains appropriate weights, systematically, for 1212.
each decision variable in the input of training instance. The Carbonell, J. G. (1981). ‘‘A computational model of analogical problem
augmented IFN learning model proposed herein is applied to solving.’’ Proc., 7th Int. Joint Conf. on Artificial Intelligence, William
Kaufmann, Los Altos, Calif., 1, 147–152.
engineering design problems. Two structural design problems Carpenter, G. A., and Grossberg, S. (1988). ‘‘The ART of adaptive pattern
were addressed to assess the learning performance of the aug- recognition by a self-organizing neural network.’’ IEEE Comp., 21,
mented IFN learning model. Based on the results of this work, 77–88.
we can conclude the following: Elkordy, M. F., Cheng, K. C., and Lee, G. C. (1994). ‘‘A structural dam-
age neural network monitoring system.’’ Microcomp. in Civ. Engrg.,
9(2), 83–96.
1. The problem of arbitrary trial-and-error selection of the Ghaboussi, J., Garrett, J. H., and Wu, X. (1991). ‘‘Knowledge-based mod-
working parameter (Rmax) in fuzzy membership function, eling of material behavior and neural networks.’’ J. Engrg. Mech.,
encountered in the conventional IFN learning model, is ASCE, 117(1), 132–153.
avoided in the newly developed augmented IFN learning Gunaratnam, D. J., and Gero, J. S. (1994). ‘‘Effect of representation on
model. Instead of arbitrary Rmax, the appropriate value of the performance of neural networks in structural engineering applica-
Rmax is determined using correlation analysis in statistics. tions.’’ Microcomp. in Civ. Engrg., 9(2), 97–108.
Hajela, P., and Berke, L. (1991). ‘‘Neurobiological computational modes
Thus, the new learning model provides a more solid sys- in structural analysis and design.’’ Comp. and Struct., 41(4), 657–667.
temic foundation for IFN learning than the conventional Hung, S. L., and Adeli, H. (1991a). ‘‘A model of perceptron learning with
IFN learning model. a hidden layer for engineering design.’’ Neurocomputing, 3(1), 3–14.
2. In the conventional IFN learning model, the weights ␣m, Hung, S. L., and Adeli, H. (1991b). ‘‘A hybrid learning algorithm for
denoting the importance of the mth decision variable in distributed memory multicomputers.’’ Heuristics—J. of Knowledge
Engrg., 4(4), 58–68.
the input, are set on a trial-and-error basis. For compli- Hung, S. L., and Adeli, H. (1992). ‘‘Parallel backpropagation learning
cated problems, the appropriate weights are difficult to algorithms on Cray Y-MP8/864 supercomputer.’’ Neurocomputing,
obtain because of a lack of the relevant heuristic knowl- 5(6), 287–307.
edge. In due course, the value is commonly initialized Hung, S. L., and Adeli, H. (1994a). ‘‘An adaptive conjugate gradient
as one for most of the examples. This problem is avoided learning algorithm for effective training of multilayer neural net-
by the newly developed learning model. Instead of an works.’’ Appl. Math. and Computat., 62(1).
Hung, S. L., and Adeli, H. (1994b). ‘‘A parallel genetic/neural network
assumed constant, the appropriate weights are deter- learning algorithm for MIMD shared memory machines.’’ IEEE Trans.
mined through the self-adjustment approach in mathe- on Neural Networks, 5(6), 900–909.
matical optimization. Therefore, the augmented IFN Hung, S. L., and Lin, Y. L. (1994). ‘‘Application of an L-BFGS neural
learning model provides a more solid mathematical foun- network learning algorithm in engineering analysis and design.’’ Proc.,
dation for neural network learning. 2nd Nat. Conf. on Struct. Engrg., Vol. III, Chinese Society of Structural
Engineering, Taipei, Taiwan, 221–230 (in Chinese).
3. For each training instance, decision variables in the input
Hung, S. L., and Jan, J. C. (1997). ‘‘Machine learning in engineering
are generally selected subjectively by users. As a result, design—an unsupervised fuzzy neural network learning model.’’ Pro-
some trivial decision variables may be adopted in the ceedings of Intelligent Information Systems (IIS’97), IEEE Computer
input for some complicated examples. Based on the self- Society, California, 156–160.
adjustment approach, the importance of a decision vari- Hung, S. L., and Jan, J. C. (1999). ‘‘Machine learning in engineering
able in an input can be derived systematically. Therefore, analysis and design—an integrated fuzzy neural network learning
model.’’ Computer-Aided Civ. and Infrastruct. Engrg., 14(3), 207–219.
insignificant or redundant decision variables in the input Kang, H.-T., and Yoon, C. J. (1994). ‘‘Neural network approaches to aid
can be neglected. simple truss design problems.’’ Microcomp. in Civ. Engrg., 9(3), 211–
4. The results illustrate that the value of the appropriate Rmax 218.
gradually falls with an increase in the number of in- Kishi, N., Chen, W. F., and Goto, Y. (1997). ‘‘Effective length factor of
stances. Consequently, not only is the learning perfor- columns in semirigid and unbraced frames.’’ J. Struct. Engrg., ASCE,
mance for training instances enhanced, but also the per- 123(3), 313–320.
Kitipornchai, S., Wang, C. M., and Trahair, N. S. (1986). ‘‘Buckling of
formance for verification instances is improved. Notably, monosymmetric I-beams under moment gradient.’’ J. Struct. Engrg.,
a small value of Rmax indicates that a strict similarity ASCE, 112(4), 781–799.
measurement is utilized in the UFN reasoning model and Kruse, R., Gebhardt, J., and Klawonn, F. (1994). Foundations of fuzzy
the possibility that no similar instances can be derived systems. Wiley, New York.
for any new instance also increases. Maher, M. L., Balachandran, M. B., and Zhang, D. M. (1995). Case-
based reasoning in design. Lawrence Erlbaum Assoc., Mahwah, N.J.
Manual of steel construction—load and resistance factor design. (1994).
ACKNOWLEDGMENTS American Institute of Steel Construction, Inc., Chicago, Ill.
Rumelhart, D., Hinton, G., and Williams, R. (1986). ‘‘Learning represen-
The writers would like to thank the National Science Council of the tations by back-propagation errors.’’ Parallel distributed processing,
Republic of China for financially supporting this research under Contract Vol. 1, D. Rumelhart et al., eds., MIT Press, Cambridge, Mass., 318–
No. NSC 86-2221-E-009-070. 362.
Stephen, J. E., and Vanluchene, R. D. (1994). ‘‘Integrated assessment of
APPENDIX. REFERENCES seismic damage in structures.’’ Microcomp. in Civ. Engrg., 9(2), 119–
128.
Adeli, H., and Hung, S. L. (1993a). ‘‘A concurrent adaptive conjugate Vanluchene, R. D., and Sun, R. (1990). ‘‘Neural networks in structural
gradient learning on MIMD machines.’’ J. Supercomputer Appl., 7(2), engineering.’’ Microcomp. in Civ. Engrg., 5(3), 207–215.
155–166. Zadeh, L. A. (1978). ‘‘Fuzzy set as a basics for a theory of possibility.’’
Adeli, H., and Hung, S. L. (1993b). ‘‘A fuzzy neural network learning Fuzzy Sets and Sys., 1(1), 3–28.
ABSTRACT: This paper describes a new approach to behavioral mode choice modeling using neurofuzzy mod-
els. The new approach combines the learning ability of artificial neural networks and the transparent nature of
fuzzy logic. The approach is found to be highly adaptive and efficient in investigating nonlinear relationships
among different variables. In addition, the approach only selects the variables that significantly influence the
mode choice and displays the stored knowledge in terms of fuzzy linguistic rules. This allows the modal decision-
making process to be examined and understood in great detail. The neurofuzzy model is tested on the U.S.
freight transport market using information on individual shipper and individual shipments. Shipments are dis-
aggregated at the five-digit Standard Transportation Commodity Code level. Results obtained from this exercise
are compared with similar results obtained from the conventional logit mode choice model and the standard
back-propagation artificial neural network. The advantages of using the neurofuzzy approach are described.
INTRODUCTION to identify and exclude those input variables that do not have
Mode choice modeling is probably the most important ele- a significant contribution. This would lead to a more efficient
ment of transportation planning. It affects the general effi- model. This is particularly important in cases where a large
ciency of travel and the allocation of resources (Ortuzar and number of potential input variables exist but only a subset of
Willumsen 1994). The development of mode choice models them would actually affect the output. Conventional tech-
has recently witnessed significant advances in many fields, in- niques of variable selection such as stepwise modeling (Mas-
cluding passenger and freight transport. A large number of ters 1993) may not produce optimal results.
mathematical models have been used to model the traveler’s It is therefore desired to develop efficient models that are
choice of mode and destination and the shipper’s choice of able to correctly simulate the input/output relationship using a
mode, shipment size, and supply market, among others. Such minimal set of input variables while providing insight into the
models are not only becoming almost intractable but also data decision-making process. The models should also adopt ap-
intensive, difficult to calibrate and update, and intransferable proaches that use a few constraining assumptions as possible
(Abdelwahab and Sayed 1998). These models cover a wide so that they can be successfully used for a wide variety of
range of mathematical complexity and accuracy. As is often applications. It is also beneficial to have the ability to insert
the case, the models that are mathematically manageable, due any available knowledge or expertise into the model if nec-
to the inherent constraints, compromise on their performance. essary.
Linearity and the use of certain distributions are among lim- In this paper, the merits of adopting the neurofuzzy ap-
iting assumptions that handicap some of these conventional proach will be discussed. It will be shown that the proposed
approaches. approach has the potential to provide similar or better accuracy
In a recent paper, Abdelwahab and Sayed (1999) showed compared with ANNs while overcoming their shortcomings.
that artificial neural networks (ANNs) could be successfully In the present study, a bimodal example is used to illustrate
used for mode choice modeling. The results obtained from a the proposed approach, but the extension to multimodal cases
standard back-propagation neural network were compared or similar applications is straightforward.
with similar results from the two most widely used conven-
CONVENTIONAL MODE CHOICE MODELS
tional models, namely, logit and probit functions. ANNs pro-
duced equal or higher predictive accuracy of mode choice As mentioned previously, the two most widely used con-
compared with conventional models. In addition, ANNs were ventional mode choice models are the logit and probit models.
believed to be less data intensive and less time consuming to They both belong to the family of behavioral discrete choice
develop and apply than currently used conventional models. models [see Ben-Akiva and Lerman (1987)]. A full discussion
However, despite the flexibility and the good performance of these models is not within the scope of this study. There-
of ANNs in modeling nonlinear relationships, they have often fore, only relevant aspects of the models will be highlighted.
been criticized for acting as ‘‘black boxes.’’ The knowledge The logit model is perhaps the most widely used mode
contained in the ANN model is kept in the form of a weight choice model, particularly in multimodal contexts (i.e., when
matrix that is hard to interpret and can be misleading at times. choice is made among more than two competing modes). The
The efficiency of ANN models is also another point of con- probability of choosing a mode is usually based on a linear
cern. Because it is not always possible to determine the sig- combination function known as the utility expression ex-
nificance of the input variables in advance, any potential can- pressed as
didate may be included in the model. It is therefore important
Uik = ␣k ⫹ 1Xik1 ⫹ 2Xik2 ⫹ ⭈ ⭈ ⭈ ⫹ j Xik j ⫹ εi = Vik ⫹ εi (1)
1
Asst. Prof., Dept. of Civ. Engrg., Univ. of British Columbia, 2324
Main Mall, Vancouver, BC, Canada V6T 1Z4. E-mail: tsayed@civil.
where Uik = utility of individual i choosing mode k; Vik =
ubc.ca deterministic utility of individual i choosing mode k; ␣k = con-
2
Res. Assoc., Dept. of Civ. Engrg., Univ. of British Columbia, 2324 stant for mode k; X1, X2, . . . , Xj = explanatory variables (can
Main Mall, Vancouver, BC, Canada V6T 1Z4. be generic or mode-specific); 1, 2, . . . , j = corresponding
Note. Associate Editor: Hassan A. Karimi. Discussion open until June calibration coefficients; and εi = random error term. It is as-
1, 2000. To extend the closing date one month, a written request must sumed that each individual chooses the alternative mode that
be filed with the ASCE Manager of Journals. The manuscript for this
paper was submitted for review and possible publication on January 19,
‘‘appears’’ to have the highest utility value. To predict an in-
1999. This paper is part of the Journal of Computing in Civil Engi- dividual’s choice of mode, the utility function should be trans-
neering, Vol. 14, No. 1, January, 2000. 䉷ASCE, ISSN 0887-3801/00/ formed into a probability. The logit model transforms the util-
0001-0023–0030/$8.00 ⫹ $.50 per page. Paper No. 20068. ity function into probability in the following manner:
JOURNAL OF COMPUTING IN CIVIL ENGINEERING / JANUARY 2000 / 23
Pk(i) = (1 ⫹ eVi j⫺Vik)⫺1 (2) is converted to an output according to a transfer function (typ-
ically a sigmoid function).
where Pk(i) = probability that individual i chooses mode k. An There is a wide range of architectures for ANNs among
implicit assumption in the logit transformation function is that which the most widely used architecture, three-layer feedfor-
all error terms are assumed to be independent and identically ward was adopted by Abdelwahab and Sayed (1998) for mode
distributed with a Weibull distribution. This assumption leads choice modeling. In this type of network, there are two dis-
to one of the most fundamental weaknesses of the logit model, tinctive modes, namely, training and query modes. In the train-
the independence of irrelevant alternative property (McFadden ing mode, a training set consisting of input/output patterns is
1978). However, this weakness is more common in multino- presented to the networks, and the weights are found through
mial logit formulations, where similar modes share attributes. an iterative process. The back-propagation learning algorithm
The coefficients of the utility expression are estimated by the (Rumelhart and McClelland 1986) is often used to find the
maximum likelihood method that obtains these coefficients weights such that the difference between the expected outputs
through an iterative search process throughout the data set. and the ones calculated by the network is sufficiently small.
Another technique of transforming the linear utility function After the network is trained, it is presented with the query data
into a bounded probabilistic relationship is described by Fin- (input variables only) to determine how accurately the network
ney (1971) and Daganzo (1979). The probability unit (probit, has simulated the input/output relationship,
in short) is based upon the relationship between the standard
normal distribution and its associated cumulative distribution, NEUROFUZZY APPROACH
which displays a sigmoid curve representing the relationship Fuzzy Logic and Fuzzy Sets
between the utility of choosing a mode and its attributes. In
this case, however, the assumption that the error terms are Fuzzy logic is another area of artificial intelligence that has
independent and identically distributed is relaxed. The probit been successfully applied to an increasing number of appli-
model is considered to be the most versatile discrete choice cations. The concept of fuzzy logic was introduced by Zadeh
model available (Hausman and Wise 1978; Daganzo 1979). Its (1965). Basically, it is a superset of conventional (Boolean)
relatively limited application, compared with the logit model, logic that has been extended to handle imprecise data and the
has been attributed to the computation difficulty involved in concept of partial truth. In fuzzy logic, variables are ‘‘fuzzi-
the estimation of its multivariate normal function. However, fied’’ through use of so-called membership functions that de-
in the binary case, the computation difficulty is significantly fine the membership degree to fuzzy sets, or as they are often
reduced to a level that makes it comparable to the logit for- called ‘‘linguistic variables.’’ A fuzzy subset A of a universe
mulation. A binary probit model can be stated as of discourse U is characterized by a membership function
uA(x), which associates each element x 僆 U a membership
Pk(i) = ⌽((Vik ⫺ Vi j)/) (3) uA(x) in the interval [0, 1] that represents the grade of mem-
bership in A.
where ⌽(⭈) = standardized cumulative normal distribution To each variable, a small number of fuzzy sets (linguistic
function; and = standard deviation of (εk ⫺ εj), which is the variables) are assigned whose membership functions overlap
scale of the utility function that can be set to an arbitrary and cover the necessary range of variation for that variable.
positive value, such as = 1. Consequently, statements that contain linguistic variables (e.g.,
transit fare is inexpensive) may take ‘‘truth values’’ of other
ANNs than ‘‘true’’ or ‘‘false,’’ represented by a number in the range
ANNs are a subfield of the artificial intelligence technology [0,1]. This provides an opportunity to define logical operators
such as ‘‘AND’’ or ‘‘OR’’.
that has gained strong popularity in a rather large array of
One way of representing fuzzy sets is by using B-spline
engineering applications where conventional analytical meth-
functions. B-spline functions (Fig. 1) are piecewise polyno-
ods are difficult to pursue or show inferior performance. Spe-
mials of order k that have been widely used in surface fitting
cifically, ANNs have shown a good potential to successfully
applications. They can be used to represent fuzzy membership
model complex input/output relationships where the presence functions because they have several important properties
of nonlinearity and inconsistent/noisy data adversely affects (Brown and Harris 1994):
other approaches.
An ANN consists of a large number of interconnected com- • The degree of membership can be evaluated using a sim-
putational elements called ‘‘neurons’’ that are organized in a ple, fast, and stable recurrence relationship.
number of layers. The connection between each pair of neu- • The output of the function is nonzero in only a small part
rons is called a link and is associated with a ‘‘weight’’ that is of the input space, which means that knowledge is stored
a numerical estimate of the connection strength. Each neuron locally across only a small number of basis functions.
in a layer receives and processes weighted inputs from neurons • The basis functions form a partition of unity as 兺 pi=1
in the previous layer and transmits its output to neurons in the uAi(x) ⬅ 1, thus producing accurate smooth approxima-
following layer. The weighted summation of inputs to a neuron tion.
to the user but cannot be acquired directly from data. ANNs, y= aiwi = a(x)Tw (4)
on the other hand, have the ability to learn the knowledge from i=1
a set of data, but the knowledge gained is hidden from the where y = output; a(x) = (a1(x), . . . , ap(x)) = vector of basis
user. The concept of neurofuzzy systems has emerged in recent function outputs when presented by input x = (x1, . . . , xn); and
years as researchers have tried to combine the transparent, lin- w = (w1, . . . , wp) = vector of network weights. The power
guistic representation of a fuzzy system with the learning abil- behind this type of network comes from the direct equivalence
ity of an ANN (Brown and Harris 1994). A neurofuzzy system between the union of a set of fuzzy rules in the form
uses an ANN learning algorithm to determine its parameters
(i.e., fuzzy sets and fuzzy rules) by processing data samples. IF (x is Ai) THEN ( y is Bj)cij (5)
冘 冘冘
n n⫺1 n
changes that are considered by the algorithm. Some measure
f (x) = f0 ⫹ fi (xi) ⫹ fi, j (xi, xj) ⫹ ⭈ ⭈ ⭈ ⫹ f1,2,...,n(x) (6) of statistical significance is used as stopping criteria. Among
i=1 i j=i⫹1
many such measures, Bossley et al. (1995) stated that, for
in which f0 represents a constant (the function bias); and other noisy data, the Bayesian statistical significance measure ap-
terms represent univariate, bivariate, and other subfunctions, pears to perform well and therefore is used in this study. The
respectively. In many cases, a majority of the terms, particu- Bayesian statistical significance measure is given by (Brown
larly those of higher dimensions, are zero or negligible. For and Harris 1994)
such cases, the ANOVA representation allows the input/output K = L ln(J ) ⫹ p ln(L) (7)
mapping to be approximated using a limited number of, as
they are often called, ‘‘subnetworks’’ of much lower dimen- where K = performance measure; p = size of current model;
sions. An example of this decomposition is shown in Fig. 3, J = mean square error; and L = number of data pairs used to
where a 5D function is decomposed into a 1D and two 2D train the network.
subnetworks. The ASMOD algorithm has been applied to a wide variety
A fuzzy rule within each subnetwork may have the form: of modeling problems and has been compared with other neu-
rofuzzy methods. The algorithm appears to perform at least as
IF (x2 is large AND x3 is small) THEN (y is small) well as the best of the other methods (Brown and Harris 1994).
Specifically, Brown and Harris (1994) reported the following
with confidence c
desirable properties:
where c = rule confidence. A rule confidence of 0 indicates
that the rule is not contributing to the output, whereas a rule • It has a smaller number of parameters and a reduced im-
confidence of 1 indicates that the rule is completely true. Val- plementation cost.
ues between 0 and 1 allow the rules to partially fire. The num- • A higher ability to generalize exists.
ber of fuzzy rules used in each subnetwork depends on the • The rule base is more transparent because it is smaller.
number of membership functions that are used to fuzzify the
inputs of that subnetwork. In the above example, assuming DATA
that five membership functions are used for each variable, the To study the feasibility and advantages of using the neu-
first, second, and third subnetworks consist of 5, 25, and 25 rofuzzy approach, a bimodal example for shipment of com-
fuzzy rules, respectively. The consequent part of all the rules modities was selected. Data have been extracted from the 1977
will be ‘‘ORed’’ (i.e., summed in this case) together. Commodity Transportation Survey prepared by the U.S. Bu-
ASMOD, or adaptive spline modeling of observation data reau of Census in 1981. The survey has the most disaggregate
(Kavli 1993), is an algorithm that uses the above decomposi- data publicly available on intercity freight movement, in which
each record represents an individual shipment. Data have been
selected randomly in a multistage sampling process. Among
the possible modes of shipment for which data were available,
only rail and Interstate Commerce Commission-regulated mo-
tor carriers for shipments over 500 lb were selected.
The database used for this study consisted of 7,500 records.
Each record holds information on the chosen mode along with
27 attributes. These attributes and their definitions are listed
in Table 1. They represent shipment attributes (e.g., size), mo-
dal attributes (e.g., transit time for each mode), or market at-
tributes (e.g., geographic location). These attributes are used
FIG. 3. Additive ANOVA Decomposition of Neurofuzzy Rule as input or explanatory variables to model the shippers’ de-
Base (Brown and Harris 1995) cision-making process.
26 / JOURNAL OF COMPUTING IN CIVIL ENGINEERING / JANUARY 2000
TABLE 1. Variable Codes and Definitions Conventional Mode Choice Models
Variable The coefficients for the logit model were first examined us-
code Definition ing all 27 variables. Second, a logit model was estimated using
(1) (2) the same training data set but after limiting the number of
mode Mode of transport (1 = rail, 2 = truck) independent variables to as few as possible without compro-
ss Shipment size (lb) mising the predictive ability of the model. The process elim-
numrec Number of records (shipments) on given OD pair inates variables that are not contributing to the explanatory
ton Total tonnage on OD pair (tons)
vl Value (1977 U.S. $/lb) power of the model (as evident from their low t-statistic) and
dn Density (lb/cu ft) variables that were estimated with signs contrary to a priori
sld Shipment is in solid state (1 = solid; 0 otherwise) expectations given the competitive nature of the shipping in-
prt Shipment is in particulate state (1 = particulate; 0 other- dustry. The resulting model, which is called the ‘‘most effi-
wise) cient’’ conventional model, included 12 variables and is re-
lq Shipment is in liquid state (1 = liquid; 0 otherwise) ported in Table 2. It should be noted that the magnitude of the
gas Shipment is in gaseous state (1 = gaseous; 0 otherwise)
frz Shipment requires freezing (1 = yes; 0 = no) coefficients depends on the scale of the respective variable and
shk Shipment requires stock protection (1 = yes; 0 = no) therefore is not useful in analyzing the model.
tmp Shipment requires temperature control (1 = yes; 0 = no)
oth Shipment requires other protection (1 = yes; 0 = no) ANN Mode Choice Models
lf Shelf life (days); 99 means nonperishable
d1 Regional dummy variable (1 = region 1; 0 otherwise) The ANN models were developed using three-layer feed-
d2 Regional dummy variable (1 = region 2; 0 otherwise) forward networks. The number of nodes in the first layer is
d3 Regional dummy variable (1 = region 3; 0 otherwise) equal to the number of input variables (i.e., 27 for the first
d4 Regional dummy variable (1 = region 4; 0 otherwise)
dist Shipping distance in miles
network). The input variables were scaled to be between 0 and
tcost Truck freight charges (cents/cwta) 1. The number of output nodes should be selected equal to the
rrel Reliability of transit time by rail (% on time) number of modes present. However, an exception is the bi-
rcost Rail freight charges (cents/cwt) modal case where the use of one output node is also an option.
rtime Rail transit time (days) The output variable was coded 1 for truck shipments and 0
rld Loss and damage by rail (%) for rail shipments. This reduces the number of weights in-
trel Reliability of transit time by truck (% on time)
tld Loss and damage by truck (%)
volved. After a number of trials, the number of neurons in the
ttime Truck transit time (days) hidden layer was selected as five. A midrange threshold (i.e.,
a 0.5) was used to predict the mode choice for the testing data
cwt = hundredweight.
set. The results of this model are presented in Table 3.
To build the most efficient ANN model, a method proposed
by Abdelwahab and Sayed (1998) was used. They proposed
METHODOLOGY
to examine the importance of each input variable by looking
To compare the three approaches discussed earlier (conven- at the weights between the input nodes and hidden layer. An
tional, ANN, and neurofuzzy), the database was divided into
two data sets. The first data set, called ‘‘training data set,’’ TABLE 2. Results of Most Efficient Logit Mode Choice Model
was used to build a model by each of the three methods. This Logit
data set had 5,000 records that were selected randomly from
Independent Estimated
the original set. The remaining 2,500 records were then used
variable coefficient t-statistic
as ‘‘testing data set’’ to compare the predictive ability of the (1) (2) (3)
developed models. To avoid any empirical bias, the rail-truck
modal shares (i.e., ⬃37 and ⬃63%) of both data sets were const 0.78034 1.85399a
ss ⫺6.91E-05 ⫺21.2102b
kept similar to the modal shares of the original data. Because dn 2.25E-03 4.34128b
the statistical performance of the logit and probit models was vl 7.78E-03 1.6482a
found to be similar (Ghareib 1996; Abdelwahab and Sayed gas 1.14693 2.60439b
1998), only the logit model will be used in this study to rep- frz 1.35583 4.21887b
resent the conventional approach. The following models were Shk ⫺0.98164 ⫺6.52961b
developed and tested: dl 0.3356 6.05226b
lf ⫺3.74E-03 ⫺2.9702b
rcost ⫺7.36E-03 ⫺8.44882b
• The most efficient logit model using a selected set of var- tcost ⫺2.28E-03 ⫺2.89204b
iables rtime 1.07652 8.80011b
ttime ⫺1.20075 ⫺5.48263b
• An ANN model using all 27 variables
• The most efficient ANN model using a selected set of Percent correctly predicted 86.3 —
variables a
Significant at 90% level.
b
• The neurofuzzy model Significant at 95% level.
The Software Neuframe (Neuframe 1997) was used to build TABLE 3. Prediction Results of ANN Model (Using All 27 Input
the neurofuzzy model described in this study. Neuframe is an Variables)
integrated group of intelligence technology tools that include Predicted by ANN
neural networks and neurofuzzy models.
Percent
Observed Rail Truck Total correct
RESULTS (1) (2) (3) (4) (5)
In this section, the results obtained by each of the four mod- Rail 650 255 905 71.8
Truck 23 1,572 1,595 98.6
els are presented. The discussion of the results will then follow Total 673 1,827 2,500 88.9
in the next section.
JOURNAL OF COMPUTING IN CIVIL ENGINEERING / JANUARY 2000 / 27
input variable whose connection weights are all low is prob-
ably not contributing to the network output. A new data set
excluding this variable can then be constructed and tested. If
the network performance is adequate, then this variable can be
excluded. The one by one reduction of the number of outputs
is continued until any further reduction would result in a sig-
nificant drop in the performance.
Several runs were made on the training data set, at each run
one or more variables whose weights are all low were ex-
cluded and the performance compared. After several runs, a
network consisting of 11 input variables was found to give an
overall success rate of 88.6%, which is similar to the network
constructed in the previous run (with all 27 variables). Any
attempts to further reduce the number of variables resulted in
a relatively significant drop in the success rate. The variables
used and the results of the network are given in Tables 4 and
FIG. 4. Neurofuzzy Model
5. A comparison of Tables 3 and 5 reveals that the exclusion
of 16 explanatory variables had almost no effect on the overall
TABLE 6. Fuzzy Rules Used by Neurofuzzy Model
success rate of the model. It is also interesting to note that
only 8 variables are common between Tables 2 and 4. Subnetwork Rule
number number Rule
Neurofuzzy Mode Choice Model (1) (2) (3)
1 1 IF ‘‘Shipment Size’’ is Small AND ‘‘Other Protec-
The training data set was used in a similar way to build and tion’’ is Not Needed
train the neurofuzzy model. A schematic view of the model is THEN ‘‘Selected Mode’’ is Truck (1.00)
2 IF ‘‘Shipment Size’’ is Medium AND ‘‘Other Pro-
shown in Fig. 4. As shown in the figure, only 4 out of the 27 tection’’ is Not Needed
variables were used by the neurofuzzy model. The Bayesian THEN ‘‘Selected Mode’’ is Rail (0.87) OR ‘‘Se-
statistical significance measure was used as the stopping cri- lected Mode’’ is Truck (0.13)
3 IF ‘‘Shipment Size’’ is Large AND ‘‘Other Pro-
teria. Consequently, a very compact model was obtained that tection’’ is Not Needed
only requires 10 fuzzy rules to reach its decision. The first THEN ‘‘Selected Mode’’ is Rail (0.81) OR ‘‘Se-
subnetwork included 6 rules while the other two subnetworks lected Mode’’ is Truck (0.19)
4 IF ‘‘Shipment Size’’ is Small AND ‘‘Other Protec-
included 2 rules each. The 10 rules are listed in Table 6. It is tion’’ is Needed
also interesting to note that 1 of the 4 variables, reliability of THEN ‘‘Selected Mode’’ is Rail (0.63) OR ‘‘Se-
transit time by truck (trel), was not included in any of the lected Mode’’ is Truck (0.37)
previous models. 5 IF ‘‘Shipment Size’’ is Medium AND ‘‘Other Pro-
tection’’ is Needed
The figures in the brackets following each mode in the con- THEN ‘‘Selected Mode’’ is Rail (0.86) OR ‘‘Se-
sequent part represent the corresponding degree of belief. Tri- lected Mode’’ is Truck (0.14)
angular membership functions (k = 2) are used by the model 6 IF ‘‘Shipment Size’’ is Large AND ‘‘Other Pro-
tection’’ is Needed
for all of the linguistic variables. The membership functions THEN ‘‘Selected Mode’’ is Rail (0.79) OR ‘‘Se-
are shown in Fig. 5. lected Mode’’ is Truck (0.21)
7 IF ‘‘Truck Freight Charges’’ is Low
THEN ‘‘Selected Mode’’ is Rail (0.17) OR ‘‘Se-
TABLE 4. Variables Used in Most Efficient ANN Model lected Mode’’ is Truck (0.83)
8 IF ‘‘Truck Freight Charges’’ is High
Variable THEN ‘‘Selected Mode’’ is Rail (1.00)
code Definition 9 IF ‘‘Truck Reliability’’ is Low
(1) (2) THEN ‘‘Selected Mode’’ is Rail (0.74) OR ‘‘Se-
lected Mode’’ is Truck (0.26)
ss Shipment size (lb) 10 IF ‘‘Truck Reliability is High
vl Value (1977 U.S.$/lb) THEN ‘‘Selected Mode’’ is Rail (0.34) OR ‘‘Se-
frz Shipment requires freezing (1 = yes; 0 = no) lected Mode’’ is Truck (0.66)
shk Shipment requires shock protection (1 = yes; 0 = no)
prt Shipment is in particulate state (1 = particulate; 0 other-
wise) The model’s ability to simulate the decision-makers behav-
oth Shipment requires other protection (1 = yes; 0 = no) ior was then assessed using the testing data set. A midrange
dist Shipping distance in miles
rcost Rail freight charges (cents/cwta)
threshold similar to the ANN models was used to define the
tcost Truck freight charges (cents/cwta) output as one of two modes. The result of the assessment is
rtime Rail transit time (days) presented in Table 7.
ttime Truck transit time (days)
a
cwt = hundredweight. Classification Threshold
An important issue in the application of ANN and neuro-
TABLE 5. Prediction Results of Most Efficient ANN Model (Us- fuzzy models to simulate observed behavior, such as the
ing 11 Variables) choice of mode by shippers, is the discretization process, also
Predicted by ANN known as the ‘‘classification threshold,’’ that is, the threshold
to be adopted to transfer the network output (a value between
Percent
0 and 1) into a discrete value (0 or 1). In a discrete model like
Observed Rail Truck Total correct
mode choice, the final results must be classified to belong to
(1) (2) (3) (4) (5)
one and only one category (i.e., mode). For example, a typical
Rail 651 254 905 71.9 neurofuzzy model result may be (0.54, 0.46), and another may
Truck 35 1,560 1,595 97.8
Total 686 1,814 2,500 88.4
be (0.96, 0.04). Although it may be fair to say that the second
result, that is, 96% probability of choosing mode 1 and 4%
28 / JOURNAL OF COMPUTING IN CIVIL ENGINEERING / JANUARY 2000
FIG. 5. Membership Functions
TABLE 7. Predicted Results of Neurofuzzy Model Using Four (percent correctly predicted) for an ANN and the neurofuzzy
Variables model is shown in Table 8.
Predicted by Neurofuzzy Model
Percent DISCUSSION
Observed Rail Truck Total correct
(1) (2) (3) (4) (5) A comparison of the prediction results of the four models
shows that their performances are relatively similar. Although
Rail 608 297 905 67.2
Truck 16 1,579 1,595 99.0 in terms of accuracy no approach shows a clear advantage, the
Total 624 1,876 2,500 87.5 neurofuzzy approach uses the smallest number of input vari-
ables to obtain the same performance. Therefore, it could be
considered as the most efficient among these approaches. In
probability of choosing mode 2, can certainly be classified to addition, the fact that it is capable of building such a model
belong to mode 1, it is not apparent whether the first result automatically is another advantage, while other models require
should end up with the same classification. Furthermore, the a manual, and sometimes subjective, intervention. The trans-
prediction of mode choice in the two examples given above parency of the knowledge gained by the neurofuzzy approach
are obviously not made with equal degrees of certainty. For is also an advantage. The use of linguistic variables makes it
example, if one requires 50% confidence in the prediction re- relatively easy to interpret the rules and if necessary change
sults, then both cases can be classified as mode 1. If one re- them.
quires 90% confidence, then only the (0.96, 0.04) result can Although all models achieved similar performance in the
be classified as mode 1, and the (0.54, 0.46) can only be clas- example used in this study, it is expected that in applications
sified to belong to a fuzzy state, somewhere between the two where the input/output relationship is more complex and in-
modes. The above examples illustrate that the predictive abil- volves higher nonlinearity that the ANN and neurofuzzy ap-
ity of a model decreases with an increasing level of confi- proaches would achieve a higher performance. This is not only
dence. due to the fact that conventional approaches (logit and probit)
In conventional mode choice models, the rate of success, or use linear regression to approximate potentially nonlinear re-
predictive ability of the model is assessed by summing up all lationships but also because in these models each input vari-
probabilities of choosing a particular mode over all observa- able can only affect the output in a univariate manner. This
tions. The sum of probabilities is then compared with the ob- would only be suitable for cases where the output f (x1, x2,
served share of that mode in the testing data set, and a success . . .) can be approximated by an additive (and linear) combi-
rate (or percent correctly predicted) is calculated. With the nation of f (x1), f (x2), etc. In neurofuzzy models, however, each
introduction of neurofuzzy and ANN models, it is now pos- subnetwork may consist of various combinations of as many
sible to fine tune the discretization process to reflect any de- input variables as is necessary up to a maximum number set
sired level of confidence. For comparison purposes, the vari- by the user. The same is implicitly true, of course, for ANN,
ation of the predictive ability, measured by the success rate but it comes at the price of the nontransparency that it causes.
Table 9 provides a comparison of the three approaches.
TABLE 8. Comparison Success Rates by Confidence Level
TABLE 9. Comparison of Three Approaches
Success Rate (% Correctly Predicted)
Comparison criteria Logit ANN Neurofuzzy
Confidence Neurofuzzy
(1) (2) (3) (4)
level Logit model ANN model model
(%) (12 variables) (11 variables) (4 variables) Constraining assumptions pres- IID & error No No
ent distribution
(1) (2) (3) (4) Possibility of usage for more Yes Yes Yes
50 86.3 88.4 87.5 than two modes
60 79.6 85.7 86.6 Ability to model nonlinearity No Yes Yes
Ability to model multivariability No Yes Yes
70 71.8 81.5 80.0
Automatic exclusion of irrele- No No Yes
80 61.2 72.1 69.4 vant inputs
90 43.4 55.6 48.8 Transparency of model To some degree Not transparent Very good
95 32.5 43.3 39.2 Ability to insert expert knowl- No No Yes
99 16.5 32.1 33.8 edge in model
ABSTRACT: The main purpose of this paper is to introduce genetic programming into civil engineering problem
solving. This paper describes a genetic programming-based approach for simultaneous sizing, geometry, and
topology optimization of structures. An encoding strategy is presented to map between the real structures and
the genetic programming parse trees. Numerical results for two examples reveal that the proposed approach is
capable of producing the topology and shape of the desired trusses and the sizing of all the structural components.
Hence, this approach can potentially be a powerful search and optimization technique in civil engineering
problem solving.
FITNESS MEASURE
The fitness measure is what directs the evolutionary process
to a parse tree that constructs the optimal truss. In other words,
the optimal truss arises from fitness.
The penalty approach has been employed to couple with the
constraints of member allowable stresses, joint allowable dis-
placements, and member buckling strength in (2). The reason
for using the penalty approach instead of the rejecting strategy
is that, in constrained optimization problems, the optimum typ-
ically occurs at the boundary between feasible and infeasible
areas, and the penalty approach will force the genetic search
to approach the optimum from both feasible and infeasible
FIG. 2. Improved Encoding Approach: (a) First Step; (b) Sec- regions. The evaluation function of fitness of the following
ond Step; (c) Final Form of Parse Tree form is adopted
JOURNAL OF COMPUTING IN CIVIL ENGINEERING / JANUARY 2000 / 33
冘 冘冋 册
p nj 2
兩Gij兩 The penalty approach is employed to cope with the
⌽(R, A, T) = W(R, A, T) ⫹ cj ⫺1 (5) constraints. The GP run is terminated if the termination
j=1 i=1 兩G aij兩
criterion is satisfied. Here, the maximum number of gen-
G aij = G Lij, when Gij < 0 erations is set as the termination criterion.
4. Encode the structures into the GP parse tree and perform
G aij = G Uij , when Gij ⱖ 0
a genetic operation, that is, reproduction, crossover, and
冋 兩Gij兩
兩G aij兩
⫺1 册 = max 冉 兩Gij兩
兩G aij兩
⫺ 1, 0 冊 mutation.
5. Return to step (2) and repeat the process until the ter-
mination criterion is satisfied.
where cj = the jth penalty parameters ( j = 1, 2, . . . , p) and p
is the total number of constraints. As GP seeks to maximize ILLUSTRATIVE EXAMPLES
the objective function, a rescaling evaluation objective func-
tion can be defined as follows: The two examples presented here are the two well-tested
examples used in structural optimization research. English
fitness units are used in references and in our computing, while the
= 再
Cavg (generation) ⫺ ⌽(R, A, T)
0
when
when
⌽(R, A, T) < Cavg
⌽(R, A, T) ⱖ Cavg
results given in the tables have been converted to SI units.
SOLUTION PROCEDURE
Fig. 3 gives an overview of the GP-based search process
for simultaneous sizing, geometry, and topology optimization.
The process can be divided into the following steps.
1. Initialize the population of the generation.
An individual of the initial population is generated by
randomly selecting functions from the function set and FIG. 4. 10-Member Planar Truss
terminals from the terminal set, constituting a GP parse
tree using functions and terminals selected by the ramped
half-and-half method (Koza 1992).
2. Decode the individuals of the GP parse tree into real
structures and perform FEM analysis of the structures.
If the structure decoded from its GP parse tree is not
a real structure and may not be analyzed by FEM, pe-
nalize the fitness of this individual in the next step and
skip the FEM analysis.
3. Evaluate the fitness value of each individual.
ABSTRACT: Intelligent transportation systems (ITS) represent a major initiative to improve roadway and travel
conditions, not by building roads, but by drawing on new technologies in computing, sensing, and communi-
cations. However, these technologies are extremely complex and, while some demonstration projects have tested
stand-alone systems, the broader architectural and standardization issues are still under development. The use
of complex technologies in a safety-critical system such as ITS and at the scale envisioned requires a very
careful study of the interaction between its various subsystems. This paper describes an approach for modeling
the temporal aspects of ITS systems as well as their logical coordination using finite state verification tools. The
approach is demonstrated by verifying safety properties of an automated traffic signal at a road intersection.
1
Assoc. Prof., Dept. of Civ. Engrg., North Carolina State Univ., Ra- Verification of Safety Properties
leigh, NC 27695.
2
Microsoft Corp., Redmond, WA 98052. A well-defined architecture is precise, unambiguous, and
Note. Editor: Sivand Lakmazaheri. Discussion open until June 1, 2000. complete. It further guarantees that a system following the
To extend the closing date one month, a written request must be filed prescribed protocols does not deadlock, is safe and fair, and
with the ASCE Manager of Journals. The manuscript for this paper was
submitted for review and possible publication on October 22, 1998. This
meets specified timing requirements. The activities required to
paper is part of the Journal of Computing in Civil Engineering, Vol. prove such properties about a given system are collectively
14, No. 1, January, 2000. 䉷ASCE, ISSN 0887-3801/00/0001-0038–0050/ termed ‘‘verification,’’ and cannot be performed manually be-
$8.00 ⫹ $.50 per page. Paper No. 19502. cause of the large number of possible system states to be con-
38 / JOURNAL OF COMPUTING IN CIVIL ENGINEERING / JANUARY 2000
sidered. Even computational approaches to finite state verifi- der information (Bryant 1986; Clarke et al. 1989; Kurshan
cation can be intractable, although recent advances have 1994). Only more recently have tools based on these algo-
enabled the verification of very large systems. Unlike tradi- rithms become available to researchers outside this community
tional discrete-event simulation, finite state verification is com- of theoretical computer scientists. Some of the more prominent
plete in the sense that all possible behaviors, or traces, are tools include the Concurrency Workbench, developed at Ed-
explored and shown correct. In other words, every combina- inburgh and North Carolina State University (Cleaveland and
tion of values for undetermined variables, e.g., timing values, Sims 1996), SMV at Carnegie Mellon University, HyTech at
is implicitly checked, as opposed to only a subset of combi- Cornell University, and SPIN at AT&T Bell Labs.
nations à la Monte Carlo simulation. As might be expected of a relatively new area of study, most
A mathematical statement of the verification problem is as of the case studies in finite state verification have been per-
follows: formed by its researchers. A good overview of these studies,
as well as those in formal methods generally, is given by
Given a system I and its specification S, show that I 兩= S, Clarke and Wing (1996). In civil engineering, a number of
where 兩= is read ‘‘satisfies.’’ verification studies have been performed over the last decade
at North Carolina State University. These studies have con-
Informally, a computing (and communication) system is mod- tributed to an understanding of the role of finite state verifi-
eled as a ‘‘machine’’ that is compared with a behavioral spec- cation in practical applications, as well as approaches such as
ification using a verification tool, as shown in Fig. 1. This semantic minimization and observer processes for verifying
approach is completely automatic in the sense that it is un- complex systems with as many as 1019 states (Elseaidy et al.
necessary to ‘‘give hints’’ to make a proof go through; i.e., 1997). These include the following:
finite state verification is decidable. The verifier simply returns
‘‘true’’ when the specification is satisfied or an error trace
showing actions by the machine that violate the specification. • Active Structural Control—Numerous structural control
Two of the more common approaches to finite state verifi- studies have been performed since the idea was first in-
cation are based on the form of the specification: troduced in the 1970s. The intent is to enhance the ability
of buildings, bridges, transmission towers, and other
• Equivalence-based verification—Both the system and the structures to respond to natural or other disasters, such as
specification are represented as finite state machines. I 兩= strong earthquakes, severe wind, or even terrorist attacks.
S when the system has the same behavior as the specifi- In a series of studies, fault-tolerant multiprocessor sys-
cation; thus, the 兩= relation is a notion of behavioral tems for active structural control have been modeled and
equivalence. Equivalence-based approaches accommodate verified using finite state approaches (Baugh and Elseaidy
a modular approach to verification, where equivalent-but- 1995; Elseaidy et al. 1996, 1997).
simpler machines can be substituted one for another. • Product and Process Data Modeling—Several groups in
However, the specification must capture in its entirety all the Architecture, Engineering, and Construction (AEC) in-
of the properties to be realized by the system. dustry are studying issues in the integration, exchange,
• Model checking—The system is represented as a finite and standardization of information generated during a
state machine and the specification as a formula in a tem- product’s life cycle. Information models are an attempt to
poral logic. I 兩= S when the system satisfies the temporal represent products and processes in a computer-readable
logic formula. In contrast to equivalence-based ap- form that is clear and precise, and that can be subjected
proaches, model checking accommodates the verification to syntactic and semantic checks. Like activity networks,
of ‘‘pieces of a specification’’ at a time, but makes no finite state models can be used for project planning and
provision for modular verification. management, with the added benefit that they readily sup-
port repetitive cycles, synchronization, communication,
In this study we use an equivalence-based approach to veri- and the coordination of events (Chadha et al. 1994; Baugh
fication, overcoming difficulties in giving a complete specifi- and Chadha 1997).
cation through the use of observer processes (Elseaidy et al. • Distributed Systems for Finite Element Analysis—The
1996, 1997), which is subsequently illustrated. prevalence of local area networks has made the combined
processing power of interconnected workstations a com-
Background and Related Work pelling alternative to mainframe computers for compute-
intensive analyses. Managing computation in distributed
In the mid-1980s and early 1990s, finite state modeling systems, however, is considerably more complex than it
emerged as a viable approach to verification with several is in conventional sequential programming. Finite state
breakthroughs that address the so-called ‘‘state explosion’’ models have been used to design and verify the correct-
problem of parallel composition. These include the use of bi- ness of distributed algorithms for structural analysis
nary decision diagrams, localization reduction, and partial or- (Chadha et al. 1999; Baugh and Chadha 1999).
Definition 1 A labeled transition system is a triple 具Q, Act, Hence, two states are bisimulation equivalent when they have
→典 where: transitions to derivative states that are themselves bisimulation
equivalent. As an example, consider system A in Fig. 3: states
• Q is the set of states q and r are bisimilar because they both have c-derivatives to
• Act is the set of actions states s and t, which are bisimilar as they both have d-deriv-
• → 僆 Q ⫻ Act ⫻ Q is the transition relation atives to state u, which is bisimilar to itself. States q and r can
be collapsed, as can states s and t, leading to system B, which
a
A transition a from state q1 to state q2 is written q1 → q2 , is bisimulation equivalent to the original system [Fig. 3(a)].
where q2 is referred to as an a-derivative of q1. An LTS with When states q1 and q2 in a transition system 具Q, Act, →典
a distinguished ‘‘start’’ node is referred to as a rooted LTS are bisimulation equivalent, we write q1 ⬃ q2 . Two rooted
and is defined by the quadruple 具Q, Act, →, q0典, where q0 僆 labeled transition systems 具Q1, Act, →1, q1典 and 具Q2 , Act, →2 ,
Q is the start state. q2典 satisfying Q1 艚 Q2 = 0/ are bisimulation equivalent when
An LTS describes a finite state process by showing the se-
quencing of ‘‘steps’’ in the computation. For example, con-
sider the simple LTSs of Fig. 2. System A begins in state p
and is capable of moving to state q on a and then back to p
on x, yielding the behavior a, x, a, x, etc. Likewise, system B
begins in state r and is capable of moving to state s on y and
then back to r on b, yielding the behavior y, b, y, b, etc.
Although an LTS is required to be in exactly one state at any
given time, systems can be combined with an interleaving se-
mantics as illustrated in system C, which is the composite of
systems A and B. By interleaving, we simply mean that tran-
sitions cannot happen simultaneously; hence, there is no direct
transition between state 具 p, r典 and state 具q, s典.
Most verification techniques for finite state systems rely on
the ability to compare systems for equivalence. When doing
so, the notion of equivalence should be strong enough to allow FIG. 3. Bisimulation Equivalent Systems
Most automatic equivalence-checking algorithms work on denotes an agent P that first performs an action a and then
labeled transition systems, as described above. Direct appli- behaves like the agent Q. For this system, ᏼ = {P, Q} and
cation of this approach would imply the sequence of steps Act = {a}. More interesting behavior can be obtained through
shown in Fig. 4 for verification: Envision a control system, recursion. For example, the agent
devise a finite state machine to model it, enumerate all possible
system states (presumably using a verification tool), and show R = a.b.R
that the traces thereby generated are acceptable. performs the (potentially) infinite sequence of actions a, b, a,
Modeling directly with labeled transition systems, however, b, a, b, . . . , . Given the direct relationship between CCS and
can be both unwieldy and error prone. An alternative is to use labeled transition systems, agent R can be shown to correspond
a more abstract language for representing processes, where a to the LTS in Fig. 5, which is structurally similar to system A
precise semantics is given to each construct so that an equiv- in Fig. 2. By using relabeling to rename actions, a convenient
alent LTS can be generated automatically from the process notational device, agent R can be transformed into an agent
definition. Researchers have devised a number of such lan- that is actually equivalent with system A:
guages, e.g., Calculus of Communicating Systems (CCS) (Mil-
ner 1989), CSP (Hoare 1985), and LOTOS (Bolognesi and R[x/b] ⇔ R = a.x.R
Brinksma 1987), for representing processes. Of these, CCS is
perhaps the best known. where the notation [x/b] substitutes x for b in the definition of
In CCS, processes perform actions and are referred to as R.
Note that in using finite state systems we abstract from some
behaviors and focus on others. In the foregoing example, R
might be an iterative server that accepts and then processes a
connection repeatedly, where a stands for accepting a connec-
tion and b stands for processing it. Alternatively, R might rep-
resent the familiar instructions for using a shampoo: wash,
rinse, repeat. In either case, we abstract from the details of the
actions and focus on their coordination, i.e., a is followed by
b in a repetitive manner.
Of course, languages like CCS allow one to describe and
reason about concurrent systems. Two agents can act in par-
allel, e.g., P兩Q, which is referred to as composite behavior.
Parallel agents can communicate using an action and its com-
plement, represented as a and ā, using the notation (P 兩 Q)\{a}.
For example, the agents Server and Client in Fig. 6 commu-
nicate synchronously over channel x; we adopt the convention
for communication that an action (x) represents reading from
a channel, and that its complement (x̄) represents writing to a
channel. Note the reduction in available transitions from those
of Fig. 2(c) due to agent communication. When two agents
communicate over a channel they synchronize, creating what
is referred to as an internal action, or . This internal action
allows two systems to be compared based only on their ex-
for each of the four signal phases using relabeling. After de- • Fairness: The controller design should be fair to the driv-
tecting a vehicle, it waits for the controller to synchronize with ers approaching the intersection from different directions.
it. For example, a vehicle waiting on the east-west approach It should be noted that efficiency and fairness are com-
trips SensorEW by synchronizing with wew; the controller peting objectives. In other words, it is possible to improve
samples the sensor by synchronizing with sew. If additional throughput through the intersection by giving excessive
vehicles begin lining up, they are not ‘‘detected’’ until the first priority to major street approaches over minor street ap-
gets a green signal and clears the intersection; in other words, proaches. However, such a design may be considered un-
the agent Sensor models a continuous type sensor. fair. One notion of fairness can be characterized by the
statement: A vehicle at an approach should wait at most
Controller x units of time for a green signal, where x is a design
parameter. In contrast to safety properties, fairness can be
The controller is the ‘‘brain’’ behind the operation of a traf- much more difficult to verify.
fic signal and hence implements its design philosophy. Objec-
tives for traffic controller design include: The traffic signal controller is modeled by the TCCS agent
shown in Fig. 12. Control waits for a minimum green time
• Safety: The controller design should guarantee that a traf- of 2 units before sampling the sensors of other phases in a
fic signal implementing its program is free of conflicts. round-robin manner with increasing delay. For example, when
Freedom from conflict can be characterized by the state- the signal becomes green in the east-west direction, the control
ment: A time interval of x units should separate two well- unit starts sampling the sensors for the north-south, east-south,
behaved vehicles entering the intersection from different and south-west directions after 2, 3, and 4 units of time delay,
approaches, where x is a design parameter. respectively. If a north-south vehicle approaches within 2 time
• Efficiency: The controller design should be efficient in the units, the actions below will synchronize with other compo-
sense that the intersection throughput is maximized. nents:
In addition to modeling a single system-wide TMC, as signal. In addition, the approach used for modeling and veri-
above, architectural variations can be proposed and compared fication could serve as a template for verifying other safety-
with respect to system performance, communication band- critical aspects of intelligent transportation systems.
width and latency, and component-level bottlenecks before im-
plementation. These variations might include, for example, dif- ACKNOWLEDGMENTS
ferences in the number and topological arrangement of TMCs. The helpful comments of this paper’s anonymous reviewers signifi-
Finally, we note that the models presented in this paper ab- cantly improved the exposition of our ideas. Support for this research by
stract from the details of data transmission. That is, while syn- the North Carolina Department of Transportation is gratefully acknowl-
chronization and communication cost are captured, the form edged.
of the data with respect to handshaking protocols and the mar-
shalling and unmarshalling of data is ignored; finite state ap- APPENDIX. REFERENCES
proaches are amenable to the verification problems that arise Baugh, J. W., Jr. (1998). ‘‘Design and verification of real-time systems.’’
in such contexts (Clarke and Wing 1996). Artificial intelligence in structural engineering, lecture notes in artifi-
cial intelligence 1454, Ian Smith, ed., Springer, New York, 30–47.
CONCLUSIONS Baugh, J. W., Jr., and Chadha, H. S. (1997). ‘‘Semantic validation of
product and process models.’’ J. Comp. in Civ. Engrg., ASCE, 11(1),
Recent advances in finite state verification have made prac- 26–36.
tical a quantitative approach to the design of ITS components Baugh, J. W., Jr., and Chadha, H. S. (1999). ‘‘Distributed computing
and other real-time control systems (Baugh 1998). Its appeal systems.’’ High performance computing, ASCE, Reston, Va.
lies in its similarity to programming, its high degree of auto- Baugh, J. W., Jr., and Elseaidy, W. M. (1995). ‘‘Real-time software de-
mation, and its ability to scale. With respect to the latter, note velopment with formal models.’’ J. Comp. in Civ. Engrg., ASCE, 9(1),
that a complex system can often be replaced by its specifica- 73–86.
Bolognesi, T., and Brinksma, E. (1987). ‘‘Introduction to the ISO speci-
tion, etc. fication language LOTOS.’’ Comp. Networks and ISDN Sys., 14, 25–
In this study, the technical notion of bisimulation equiva- 59.
lence is used to prove the correctness of a roadway signal Bryant, R. E. (1986). ‘‘Graph-based algorithms for boolean function ma-
controller. The controller design in this traffic signal model nipulation.’’ IEEE Trans. on Comp., C-35(8), 677–691.
avoids starvation by using a round-robin approach to sensor Chadha, H. S., Baugh, J. W., Jr., and Wing, J. M. (1994). ‘‘Formal spec-
sampling with increasing delays. Also, the model incorporates ification of AEC product models.’’ Comput. in Civ. Engrg.: Proc., 1st
Congr., K. Khozeimeh, ed., Vol. 1, ASCE, New York, 571–578.
timing behavior and demonstrates the use of observer pro- Chadha, H. S., Baugh, J. W., Jr., and Wing, J. M. (1999). ‘‘Formal spec-
cesses in verifying properties that abstract from time. Because ification of concurrent systems.’’ Adv. in Engrg. Software, 30(3), 211–
the roadway signal controller satisfies the desired properties, 224.
it could serve as a basis for designing a fully-actuated traffic Clarke, E. M., Long, D. E., and McMillan, K. L. (1989). ‘‘Compositional
ABSTRACT: This paper presents a method for the representation and reasoning of spatial knowledge. Spatial
knowledge is important to decision making in many transportation applications that involve human judgment
and understanding of the spatial nature of the transportation infrastructure. Unfortunately, conventional Knowl-
edge-Based Expert Systems (KBESs) do not have a robust spatial inference engine to construct spatial knowledge
and deduce solutions. The method discussed in the paper integrated a KBES with geographic information system
(GIS) functions using Transmission Control Protocol/Internet Protocol and Remote Procedure Calling technol-
ogies. A software system (called IntelliGIS) implementing the method has been developed. As a case study, a
KBES-based pavement management system was developed using the IntelliGIS. The case study demonstrated
how IntelliGIS represents and reasons about spatial knowledge and how it provides graphical display capabilities
for representing spatial information and derived solutions.
FIG. 2. (a) Spatial Knowledge Representation in IntelliGIS; (b) Spatial Knowledge Reasoning in IntelliGIS
ABSTRACT: Digital aerial photography provides a useful starting point for computerized map generation. Fea-
tures of interest can be extracted using a variety of image-processing techniques, which analyze the image for
characteristics such as edges, texture, shape, and color. In this work, we develop an automatic road detection
system for use on high-resolution grayscale aerial images. Road edges are extracted using a variant of the
Nevatia-Babu edge detector. This is followed by an edge-thinning process and a new edge-linking algorithm
that fills gaps in the extracted edge map. By using a zoned search technique, we are able to design an improved
edge-linking algorithm that is capable of closing both large gaps in long, low-curvature road edges and smaller
gaps that can occur at triple points or intersection points. An edge-pairing algorithm, which is subsequently
applied, takes advantage of the parallel edges of roads to locate the road centers. Results demonstrate that the
current system provides a simple yet effective first stage for a more sophisticated map-generation system.
of tests leads to more noise and broken edges, but at times Fig. 3, which need to be applied to each edge pixel in order
also includes parts of some true edges that the threshold test to thin the edges. One pass of the thinning algorithm compares
does not. However, it is judged that having more continuous, the neighborhoods of each edge pixel with the original, trans-
unbroken edges as output from the edge-detection process pose, and rotations of each of the 17 masks. Also, they must
leads to better final results. Thus, the threshold test is used by be compared starting from the left, right, top, and bottom in
itself to detect edge points. The threshold value is not a fixed order to take full advantage of symmetry. Fig. 4 shows the
value, but is found for a particular image as the value that thinning example from Heipke et al. (1994), illustrating the
retains a certain percentage of the edge energy after thres- process of removing an edge pixel that does not belong in the
holding. The user can adjust this percentage, such that a rea- skeleton. In dealing with the images used for this paper, one
sonable balance between continuous edges and unwanted noise pass through all the checks was enough to satisfactorily thin
is achieved. Edge energy is defined as the sum of the squares the edges. However, if edges are thick enough, the thinning
of the edge values. may have to be repeated several times.
FIG. 7. Results from Test Image #1: (a) After Edge Detection and Thinning; (b) Zoom of 100 ⴛ 100 Block of (a); (c) After Edge Linking;
(d) Zoom of 100 ⴛ 100 Block of (c); (e) Linked Edges and Road Midpoints; (f) Road Midpoints Only
FIG. 8. Results from Test Image #2: (a) After Edge Detection and Thinning; (b) Zoom of 100 ⴛ 100 Block of (a); (c) After Edge Linking;
(d) Zoom of 100 ⴛ 100 Block of (c); (e) Linked Edges and Road Midpoints: (f) Road Midpoints Only
FIG. 9. Results from Test Image #3: (a) After Edge Detection and Thinning; (b) Zoom of 100 ⴛ 100 Block of (a); (c) After Edge Linking;
(d) Zoom of 100 ⴛ 100 Block of (c); (e) Linked Edges and Road Midpoints; (f) Road Midpoints Only
cases of images with difficulties that cause the program to fail either as faded edges or occlusions due to trees or shadows,
or do poorly, there are also some difficult circumstances in the quality of the results declines. This suggests that using a
which this system can still perform quite well. simple three-step process such as that described in this paper
does not always do an adequate job of dealing with the variety
CONCLUSIONS of image characteristics mentioned in the introduction. How-
We have presented a simple and practical scheme for de- ever, this provides a good beginning to a more complex sys-
tecting roads in grayscale aerial images. This scheme combines tem. For example, in Zlotnick and Carnine (1993), the centers
the best features of previous methods (Nevatia-Babu edge de- of roads found by pairing edges are analyzed to find clusters,
tection, edge thinning, and edge pairing) with a new edge- which are then used as seeds to grow the road network. It
linking technique. The edge-linking technique improves on the appears that such a step is necessary if the goal is a complete
scheme presented by Heipke et al. (1994) by using a more road network.
extensive zoned search, which is capable of closing both large Another possible extension of this process is to incorporate
gaps in long low-curvature road edges and smaller gaps that information based on texture or grayscale value. For instance,
can occur at triple points or intersection points. in most cases the grayscale of the roads is significantly dif-
The overall road-detection system was found to perform ferent from that of the background. By looking at the variance
well for images with sharp, continuous edges, such as the first in grayscale value within a small neighborhood of each pixel,
test image. Once discontinuities in the edges are introduced, another tool for extracting road edges is created.
68 / JOURNAL OF COMPUTING IN CIVIL ENGINEERING / JANUARY 2000
A further way to incorporate other types of information may International Society for Optical Engineering, Bellingham, Wash.,
be to make the process semiautomatic instead of fully auto- 2315, 508–518.
Barzohar, M., and Cooper, D. B. (1996). ‘‘Automatic finding of main
matic. The system described here already requires the user to roads in aerial images by using geometric-stochastic models and esti-
input some parameters at the beginning of the program, but mation.’’ IEEE Trans. on Pattern Anal. and Machine Intelligence,
after the initial inputting, the computer proceeds on its own. 18(7), 707–721.
Semiautomatic methods have been devised that allow the com- Farag, A., and Delp, E. J. (1995). ‘‘Edge linking by sequential search.’’
puter and human operator to interact (Heipke et al. 1994). Pattern Recognition, 28(5), 611–633.
Usually a starting point and direction are inputted, as well as Geman, D., Geman, S., Graffigne, C., and Dong, P. (1990). ‘‘Boundary
detection by constrained optimization.’’ IEEE Trans. on Pattern Anal.
widths and lengths of the roads to be found. One can imagine and Machine Intelligence, 12(7), 609–628.
the program may also allow the user to click the mouse on a Gonzalez, R. C. (1987). Digital image processing, 2nd Ed., Addison-
point somewhere inside the road, from which the computer Wesley, Reading, Mass., 344–349.
could determine that road’s texture and/or grayscale. Other Heipke, C., Englisch, A., Speer, T., Stier, S., and Kutka, R. (1994).
road points could be found by comparing texture and grayscale ‘‘Semiautomatic extraction of roads from aerial images.’’ Proc., SPIE
information. This interaction between human and computer ISPRS Commission III Symp., SPIE—International Society for Optical
Engineering, Bellingham, Wash., 2357, 353–360.
may also allow the program to be interrupted if it runs off Law, T., Itoh, H., and Seki, H. (1996). ‘‘Image filtering, edge detection,
course. and edge tracing using fuzzy reasoning,’’ IEEE Trans. on Pattern Anal.
In conclusion, it has been shown that simple steps can form and Machine Intelligence, 18(5), 481–491.
a basis for an effective road detection system. However, ad- Nevatia, R., and Babu, K. R. (1980). ‘‘Linear feature extraction and de-
ditional information must be incorporated in order to develop scription.’’ Comp. Graphics and Image Processing, 13(3), 257–269.
a procedure that satisfactorily deals with the wide variety of Porrill, J. (1990). ‘‘Fitting ellipses and predicting confidence envelopes
using a bias corrected Kalman filter.’’ Image and Vision Computing,
aerial images that need to be processed. 8(1), 37–41.
Scher, A., Shneier, M., and Rosenfeld, A. (1982). ‘‘A method for finding
ACKNOWLEDGMENTS pairs of antiparallel straight lines.’’ IEEE Trans. on Pattern Anal. and
Machine Intelligence, 4(4), 316–323.
This work was partially supported by a grant from the Suksapattana Zhu, M. L., and Yeh, P. S. (1986). ‘‘Automatic road network detection
Foundation, Bangkok, Thailand, to the Intelligent Engineering Systems on aerial photographs.’’ Proc., CVPR’86 IEEE Comp. Soc. Conf. on
Laboratory, Massachusetts Institute of Technology. Comp. Vision and Pattern Recognition, IEEE Computer Society Press,
Washington, D.C., 34–40.
APPENDIX. REFERENCES Zhu, Q., Payne, M., and Riordan, V. (1996). ‘‘Edge linking by a direc-
tional potential function (DPF).’’ Image and Vision Computing, 14(1),
Airault, S., Ruskone, R., and Jamet, O. (1994). ‘‘Road detection from 59–70.
aerial images: A cooperation between local and global methods.’’ Zlotnick, A., and Carnine Jr., P. D. (1993). ‘‘Finding road seeds in aerial
Proc., SPIE Image and Signal Processing for Remote Sensing, SPIE— images.’’ CVGIP: Image Understanding, 57(2), 243–260.
ABSTRACT: Electronic document management (EDM) systems have revolutionized the construction document
storage and retrieval process. An electronically managed document consists of a computer representation of the
main document body, usually in the form of a bitmap, and a reference or indexing structure used to retrieve the
document. In construction, the reference data includes information that relates the document to certain aspects
of the constructed facility and the construction company. It is argued that this relationship needs to be represented
explicitly and in an integrated manner as opposed to the approach used by traditional EDM systems. This paper
presents an integrated approach based on the concept of specialized construction data models. A detailed account
of how this approach was followed to integrate a quality assurance document management system within a steel
fabrication company is also presented.
TABLE 3. Difference in Implementation between Commercial EDMS and Presented Approach for Three Major Features
1. Validation Static validation in the form of user-defined rules or Validation rules are automatically inherited from the business
code. Rules are static in the sense that they don’t rules defined as part of the underlying data model. In other
change as the organization evolves. words, validation is a function of the structure of the doc-
ument itself. As company and project data are altered, so are
the validation rules. More importantly, as the structure of the
data itself changes, the structure of the validation is auto-
matically updated.
2. Complex reference Only advanced systems allow for customization of the With a common data model, all the business objects (employee
data document reference structure. For those systems, the information, material, drawing, etc.) are available for docu-
process usually requires programming skills in a pro- ment designers during the template definition stage.
prietary macrolanguage.
3. Searching based on Object-based searching is almost impossible to imple- With an integrated approach, changes to any portion of the
‘‘object matching’’ ment with EDMS. The structure of the product overall data model—which include the document, project
and company data model would have to be com- and the company information—are automatically reflected
pletely understood by the document designer and in all client systems.
complex data manipulation code would have to be
written.
ABSTRACT: An artificial neural network based methodology is applied for predicting the level of organizational
effectiveness in a construction firm. The methodology uses the competing value approach to identify 14 variables.
These are conceptualized from four general categories of organizational characteristics relevant for examining
effectiveness: structural context; person-oriented processes; strategic means and ends; and organizational flexi-
bility, rules, and regulations. In this study, effectiveness is operationalized as the level of performance in con-
struction projects accomplished by the firm in the past 10 years. Cross-sectional data has been collected from
firms operating in institutional and commercial construction. A multilayer back-propagation neural network based
on the statistical analysis of training data has been developed and trained. Findings show that by applying a
combination of the statistical analysis and artificial neural network to a realistic data set, high prediction accuracy
is possible.
冘
y = f (x) ⫹ e (1)
where f (x) = function of the input variable x; and e = inde- ␦ [s]
pj = f ⬘
[s]
j (I j ) ␦ pk
[s⫹1]
w [s⫹1]
jk (4)
k
pendent noise function with zero mean. A set of input-output
examples xi and yi (i = 1, 2, . . . , N ) is given. The prediction where x [s]
j = actual output of node j in layer s; w ij[s] = weight
problem is: For any given input x find the best estimate of the of the connection between node I at layer (s ⫺ 1) and node j
ouput y. The best estimate can be defined as the estimate that at layer (s); ␦ [s] [s]
pj = measure for the actual error of node j; I j =
minimizes the average error between the real and estimated weighted sum of the inputs of node j in layer s; (t) = time-
outputs. Thus, such a supervised learning problem becomes dependent learning rate; f ( ) = transfer function; m = momen-
equivalent to a mapping or curve-fitting problem in a multi- tum factor (between 0 and 1); and dj, oj = desired and actual
dimensional hyperspace. The traditional statistical approaches activity of node j, respectively (for output nodes only).
to curve fitting, such as the regression analysis, fail to repre- A misclassification measure between the network output and
sent problems with no explicit mathematical model accurately the expert’s is defined as
in a multidimensional space of variables. On the other hand,
冋冘 册
c 1/2
the neural network approach can solve such problems more E= (uik ⫺ u⬘)
ik
2
(5)
effectively. i=1
i=1
{Y ⫺ Y *}2/N (6)
APPENDIX. REFERENCES
Adas, A. A. (1996). ‘‘The level of organizational effectiveness of the
construction firm: A multivariate model-based prediction methodol-
ogy,’’ PhD thesis, University of Waterloo, Waterloo, Ont., Canada.
Adeli, H., and Hung, S. L. (1994). ‘‘An adaptive conjugate gradient learn-
ing algorithm for efficient training of neural networks.’’ Appl. Math.,
62(1), 81–102.
Adeli, H., and Wu, M. (1998). ‘‘Regularization neural network for con-
struction cost estimation.’’ J. Constr. Engrg. and Mgmt., ASCE, 124(1),
18–23.
Flood, I., and Kartam, N. (1994). ‘‘Neural networks in civil engineering.
FIG. 3. Comparison of Training Time for Five Developed Net- II: Systems and application.’’ J. Comp. in Civ. Engrg., ASCE, 8(2),
works 149–165.
TECHNICAL PAPERS
Transient Dynamic Nonlinear Analysis Using MIMD Computer Architectures
J. Sziveri and B. H. V. Topping . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 79
Simplified Fuzzy ARTMAP as Pattern Recognizer
S. Rajasekaran and G. A. Vijayalakshmi Pai . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 92
Java-Based Regulation Broker
Vei-Chung Liang and James H. Garrett Jr.. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 100
Reducing Prediction Error by Transforming Input Data for Neural Networks
Jonathan Jingsheng Shi . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 109
Stochastic Construction Time-Cost Trade-Off Analysis
Chung-Wei Feng, Liang Liu, and Scott A. Burns . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 117
Computer Graphics for Nonlinear Strut-Tie Model Approach
Young Mook Yun . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 127
Computer Imagery and Visualization in Civil Engineering Education
N. Bouchlaghem, W. Sher, and N. Beacham . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 134
Bridge Damage Assessment through Fuzzy Petri Net Based Expert System
Weiling Chiang, Kevin F. R. Liu, and Jonathan Lee. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 141
TRANSIENT DYNAMIC NONLINEAR ANALYSIS USING MIMD
COMPUTER ARCHITECTURES
By J. Sziveri1 and B. H. V. Topping2
ABSTRACT: This paper presents research on the use of homogeneous parallel and heterogeneous distributed
computers for finite-element analysis of transient dynamic problems using a message passing interface. The
appropriate computer architectures are discussed, and this review is the basis for the development of a new
definition of computational efficiency for heterogeneously distributed finite-element analysis. A code for the
transient nonlinear analysis of reinforced concrete plates is profiled. It is demonstrated that although the code
is efficient for homogeneous computing systems a new message passing procedure must be developed for
heterogeneously distributed systems. A new algorithm is developed and tested on a heterogeneous system of
workstations.
FIG. 1. A2 Finite-Element Idealization of Reinforced Concrete Panel Divided into Two, Four, and Eight Subdomains
冘
Ehw = (11)
1
where NE = total number of finite elements in the whole struc- tpar i
t
ture; and NE i = number of elements in the ith subdomain that seq
is processed on the ith processor. On a homogeneous system t iseq is the same for each process,
In the case of a homogeneous system of np processors, t iseq and (11) becomes
is the same for each processor and with the assumption of
冘
np
perfect load balancing when f il = 1/np, (3) becomes equivalent
to (2). But E ih in (3) represents only the efficiency (i.e., the tseq tseq
utilization, of one single processor within the heterogeneous Ehw = = (12)
np tpar np tpar
system). To devise an overall measure of the efficiency for the
whole network, the average of the individual efficiency values which is the same as (2).
for each processor participating in the heterogeneous system
may be calculated as Timing on Unix Platforms
冘 冘
np np
1 1 f il ⭈ t iseq
The timing of program executions on Unix platforms is a
Eh 㛭 ave = E ih = (6) topic of constant debate in the computational literature.
np i =1 np i =1 tpar Whether to measure ‘‘pure’’ CPU time or if wall clock time
is sufficient is an important question. When measuring wall
The only difficulty with this approach is that it does not rep- clock time the main problem is how to deal with ‘‘distur-
resent the overall performance of the whole network well, par- bances’’ created by other processes consuming CPU time
ticularly when the processors have very different capacities. In slices and taking up system resources. This can be caused by
this latter case the most important factor when designing and standard or additional system activities including a wide range
deploying heterogeneous systems of very different compo- of processes (e.g., logging, printing, mail, etc.), daemons, or
nents is that the strongest processors should be utilized the other user application programs. Some researchers prefer to
most; they should run at the highest possible efficiency. The measure execution times when the Unix system excludes these
slow ones just ‘‘assist’’; their full utilization is of little or no disturbances by booting the machine into a different ‘‘init
concern as they do not really affect the overall performance. state’’ where they can restrict general system activities and
This principle may be represented in the efficiency calculation totally exclude network and user processes.
of the whole system by employing a weighting factor to the However in this study these concepts are not followed. The
contributions of the individual components. This weighting Unix workstations are considered ‘‘as is’’ in a state as they
factor would be directly proportional to the power of each are working in a general engineering office. Execution times
processor. Thus instead of (6), the overall efficiency of the are always measured as wall clock times in the concept of
whole heterogeneous system would be written as follows: time to solution values and the machines are in a ‘‘normal
冘冉 冊
state.’’ Thus users can log in and out, other processes can be
冘
f il ⭈ t iseq launched, and system daemons activated at any time. The only
i i pp i
( pp E ) tpar
冘 冘
h provision that was made to achieve some consistency among
Ehw = = (7) the timing result was that the test runs were made at times
pp i pp i when the machines were relatively silent in terms of user ac-
tivity and that there was no great system allocation either. Sev-
where pp i = appropriate measure of the power or capacity of eral measurements were taken in such conditions, and the ob-
the ith processor. viously extreme ones were omitted. The average of the
By concentrating on the finite-element problem to be solved remainder of the values gives the appropriate ‘‘time to solu-
here, the power of each processor can be most easily assessed tion.’’
by looking at the sequential execution times it takes to run the What happens when these conditions are disturbed by some
problem (or a representative problem) on each processor (i.e., arbitrary processes is discussed in the last section on load bal-
the t iseq values). Hence the power or capacity of the processor ancing.
to solve the problem is inversely proportional to these se-
quential time values, and a good implementation for the pp i
Timing Results of PM Parallel Scheme
measures would be
First of all the sequential version of the program was run
i 1 on all the machines that are involved in the present study.
pp = (8)
t iseq These machines are listed in Table 3. They were all running
SUN’s Solaris operating system version 2.5.1, which is a Sys-
The relative power of the workstations need only be assessed tem V Unix. The PCs had the Solaris x86 version installed of
once, when the heterogeneous system is installed. By substi- the same release. The network communication was 10Mb/s
tuting (8) into (7) and reducing the expression with t iseq, the ethernet cable.
overall efficiency formula may be rewritten as follows: The numerical analysis was carried out for a small part of
82 / JOURNAL OF COMPUTING IN CIVIL ENGINEERING / APRIL 2000
TABLE 3. Summary of Unix Workstations Used in This Study half of the sequential time. This would be the ideal case for
Host name Machine type CPU all parallel runs.
(1) (2) (3) These excellent results are due to the characteristics of the
hardware and its good utilization by the software environment.
carme SUN Sparc 10 4 ⫻ TI TMS390Z50 at 50 MHz
(superSparc chip)
The shared memory employed here is a very fast communi-
thebe SUN Sparc 20 1 ⫻ TI TMS390Z50 at 50 MHz cation device; hence it does not impose any serious penalty or
(superSparc chip) overhead on the parallel computation.
iapetus SUN Sparc 4 1 ⫻ Fujitsu MB86904 at 110 MHz
(microSparc II chip)
Running on Distributed Memory Workstation Cluster
elara ALR PC 2 ⫻ Intel Pentium at 90 MHz
(i586 chip)
Using the general-type network connections, a theoretically
rhea ABIT PC 1 ⫻ AMD K5 at 100 MHz
(i586 compatible chip) arbitrary number of processors may be connected to form one
virtual machine. Some of these workstations may themselves
be multiprocessor machines, and thus they can accommodate
TABLE 4. Sequential Execution Times for A2 and A8 Meshes several subdomains. In this case the communication may be
tseq
faster between these subdomains if the software environment
takes advantage of the fast shared memory access. This is not
Host for A2 mesh for A8 mesh the case in this study. When linking the parallel MPI program
name (s) (s) with that version of the MPI library that utilizes Unix sockets
(1) (2) (3) (via the p4 subsystem) as a communication device, every pro-
carme 11.0 49.0 cess instance of the program works according to this version
thebe 9.4 42.0 regardless of whether they are placed on the same machine or
iapetus 12.1 54.0
not.
elara 11.4 51.0
rhea 8.7 39.0 Table 6 shows how the subdomains were mapped onto the
eight processes and which one of these processes was launched
on which machine. The machine names refer back to Table 3
the simulation process using the original sequential version of for the full description of the hardware.
the program, and the execution time tseq was measured. Table The eight subdomain decompositions for the A2 and A8
4 shows these execution times achieved by running the pro- meshes, shown in Figs. 1 and 3, respectively, were used in the
gram for 25 iterations as a benchmark for the A2 mesh and eight processor heterogeneous runs. Fig. 3 shows the gray
the same for mesh A8. scales to identify each subdomain. Again, the same blocks of
These timing results do not include the time required for 25 iterations were measured, and the parallel execution times
the preprocessing of the finite-element mesh (i.e., the domain and the calculated efficiency values are detailed in Table 7.
decomposition ⬇2–3 s), the preparation of subdomain con-
nectivity data for the communication routine (⬇0.1 s), the de- TABLE 5. Parallel Execution Times Measured on Multiproces-
livery of the subdomains to the actual processing nodes sor Workstation Utilizing Shared Memory for Communication
(⬇0.2–0.3 s), nor the setting up time before the iteration com- for A2 and A8 Meshes Using PM Algorithm
mences (⬇0.1 s). However these times are negligible com- Number of tpar E
pared with the long time marching iteration of the transient Mesh subdomains np (s) S (%)
dynamic analysis. (1) (2) (3) (4) (5) (6)
A2 2 2 5.7 1.93 96.5
Running on Shared Memory MP Workstation 4 4 3.3 3.33 83.3
A8 2 2 ⬇24.5 ⬇2 ⬇100
First the MPI program was linked with that version of the 4 4 12.6 3.89 97.2
MPI library that utilizes the shared memory as a communi-
cation device. The program was run on the multiprocessor
workstation ‘‘carme’’ by launching two and four instances of TABLE 6. Placement of Processes (Corresponding to Subdo-
the program to operate on two and four subdomains, respec- mains) onto Different Machines When Using Heterogeneous
tively, of both the A2 and A8 meshes. The running times were Network of Workstations for Eight Subdomains
measured for the same blocks of 25 iterations as quoted in Subdomain number Host name (from Table 3)
Table 4 at the initial stage of the analysis when the material (1) (2)
is mainly in a linear state. The parallel execution times and
1 thebe
the calculated speedup and efficiency values are summarized 2 carme
in Table 5. 3 carme
The terms of np, tpar, S, and E in Table 5 are as defined by 4 carme
(1) and (2). In Table 5 the general trends of parallel processing 5 elara
may be clearly seen as follows: 6 elara
7 rhea
8 iapetus
• When using more processors to work on the same prob-
lem, the efficiency drops due to the increased communi-
cation requirements. TABLE 7. Parallel Execution Times Measured on Network of
• When analyzing larger problems higher computational ef- Workstations Utilizing Unix Sockets for Communication for A2
ficiencies may be achieved as the proportion of commu- and A8 Meshes Using PM Parallelization Scheme
nication shrinks compared to the total computational load. Number of tpar Ehw
Mesh subdomains np (s) S (%)
It is worth noticing that these efficiency values are quite high (1) (2) (3) (4) (5) (6)
even for the smaller problem of the A2 mesh. In the case of
A2 8 8 2.9 3.0–4.2 45.8
the A8 mesh analyzed on two processors the speedup is vir- A8 8 8 8.2 4.8–6.6 72.4
tually 2 as the parallel execution time was almost identical to
JOURNAL OF COMPUTING IN CIVIL ENGINEERING / APRIL 2000 / 83
The terms np, tpar, and S in Table 7 are as defined by (1) ues show that the processors are not utilized properly. This is
and (2). As speedup is not clearly defined for heterogeneous the result of changing the underlying hardware layer of com-
parallel runs, two bounding values are calculated based on the munication from the fast shared memory to the distributed
fastest and the slowest sequential runs measured on the cor- network connection. To study the effect of this in the pro-
responding fastest and slowest processors in the heterogeneous gram’s run-time profile, several time measurement instructions
network. The weighted efficiency of the heterogeneous net- were embedded into the source code. With the aid of these
work Ehw is determined as defined by (11). measurement points, the work of each processor was traced in
It may be noted by comparing Tables 5 and 7 that even detail and the timing results combined into charts.
though the number of processors used in the latter case was Sziveri et al. (1996, 1997) gave full details of this dynamic
increased from four to eight, it did not result in significant explicit analysis algorithm including the PM parallelization
shortening of the parallel execution times. The efficiency val- scheme. This scheme enables interprocessor communication
FIG. 5. Percentage of Time Spent on Each Subtask of PM Parallelization Scheme by Every Processor on A2 Mesh
between pairs (P) of subdomains and multiple (M) subdo- while the communication proceeds. Given the relatively high
mains. In Fig. 4 the major subtasks involved with this algo- communication bandwidth in both the transputer-based hard-
rithm are plotted, and the timing instructions clearly define the ware and the shared memory multiprocessor workstations
borders of these subtasks that constitute an iteration step. compared to their processing capabilities, the presented PM
In Fig. 4 a notation is introduced to reference each subtask communication scheme implemented with this kind of point-
that make up an iteration step. This notation is used in Figs. to-point-type blocking instructions was sufficient to achieve
5 and 6, which are detailed profiling charts of the PM parallel very high computational efficiencies.
algorithm with respect to meshes A2 and A8, respectively. Unfortunately, porting the algorithm to the distributed mem-
These two figures correspond to the parallel runs of Table 7. ory workstation cluster highlighted the weaknesses of this type
Figs. 5 and 6 show the amount of time spent for each sub- of parallel hardware system. The relative communication
task by each processor within one iteration. It is expressed in bandwidth is much less over the standard network facilities.
terms of percentage where 100% is the total time spent on one Moreover, the communication throughput is severely penal-
complete iteration step. ized by very high latency values.
This representation assumes that there is a synchronization These conditions imposed by the hardware environment re-
at the end of each iteration step by the energy communication. duced the efficiency of the parallel application. Clearly a new
This assumption is correct as the length of that communication parallel strategy is required to overcome these limitations.
is very short, and the collective nature of it forms a barrier
within the flow of the program for every processor. New Parallel Algorithm
It is apparent from Figs. 5 and 6 that a very large proportion
of the time is spent on communication (subtasks PM㛭st㛭2, The most obvious strategy when attempting to overcome
PM㛭st㛭3, and PM㛭st㛭4) in each iteration. It is as high as latency is to minimize the actual number of messages to be
40–50% in terms of the smaller A2 mesh and still 20–35% sent by packing as much information into each message buffer
with the bigger A8 mesh. This large communication overhead as possible. The PM communication scheme was developed to
results in the low efficiency values and thus prevents the avoid communication deadlock in the transputer router system
speedup to grow with the number of processors. The next sec- as a first priority. The number of messages to be sent were not
tion details the solution to this problem by introducing a new strictly minimized. When working with shared memory mul-
parallel communication scheme. tiple processor machines or on workstation clusters, message
routing is not a problem. Arbitrary connections are possible
NEW PARALLEL COMMUNICATION SCHEME between any two nodes within the parallel system, and the
MPI software library utilizes these connections. This gives an
The PM communication scheme was first presented by Szi- opportunity to rethink the message layout of the PM scheme.
veri et al. (1996, 1997) and was implemented in Parallel C for The message structure is the simplest between the subdo-
transputer-based hardware. Lämmer et al. (1997) showed how mains if every subdomain sends only one single message to
the Parallel C calls were substituted with standard MPI func- each subdomain to which it is connected. This is best illus-
tion calls. In both implementations these message passing trated in Fig. 7 where the notion of P and M nodes is replaced
functions were point-to-point-type blocking instructions, by groups of intersubdomain nodes that represent the actual
meaning that the current thread of execution could not con- messages to be sent.
tinue until the called communication function had been re- The MPICH library used in this study is a full implemen-
turned. Hence the computational part of the program stops tation of the MPI specification, which contains nonblocking
JOURNAL OF COMPUTING IN CIVIL ENGINEERING / APRIL 2000 / 85
message passing functions as well as blocking ones. This al-
lows the introduction of a special strategy that results in very
significant efficiency improvements of the parallel system on
many parallel hardware platforms. When calling nonblocking
communication functions, the main thread of execution is re-
turned to the caller program before the communication, a send,
or more importantly a receive is completed. It is usually im-
plemented by launching background threads that operate on
reserved buffer spaces in an asynchronous way; that is, these
threads wait for the delays of the communication (caused by
latency and bandwidth characteristics of the hardware)
whereas the main program can continue its execution.
This is where the potential for these kinds of techniques
lies. Because the communication overhead (i.e., the time spent
on communication in a parallel environment) can be ‘‘cov-
ered’’ with useful sequential calculations, it is often referred
to as the ‘‘red-black’’ method in the computational literature
and will be denoted by the abbreviation of RB.
When implementing an RB scheme for the explicit finite-
element problem in question, the first feature to look at is the
profiling figures of the parallel runs. It is apparent from Figs.
5 and 6 that although a significant percentage of time is spent
on communication, the longest subtask within one iteration is
the one denoted by PM㛭st㛭1 and described in Fig. 4. With FIG. 8. Grouping of Elements and Intersubdomain Nodes for
an even more detailed timing it has been determined that ap- New (RB) Parallel Scheme
proximately 99% of the time spent on PM㛭st㛭1 is actually
spent on the calculation of the vector of internal forces N. In sisting of internal subdomain elements, will form the second
the source code of the program this calculation is a loop for group. This concept is illustrated in Fig. 8.
each element contained in the subdomain. This includes the
processing of very complex (nonlinear) material rules as ap- Implementation of RB Parallel Scheme
plied to each integration point of each finite element.
This calculation has to be completed before the exchange The main steps in the implementation of the RB scheme is
of contributions of N can take place between the neighboring described in Fig. 9. In this figure the major subtasks compris-
subdomains for each intersubdomain boundary node. However ing an iteration step of the RB algorithm are plotted and the
it is clear that it is sufficient to calculate the local contributions place of timing instructions in between these subtasks are
to N from those elements that are connected to these intersub- marked. Fig. 9 also introduces a notation to reference these
domain boundary nodes. At this moment the communication subtasks. This notation will be used subsequently in all pro-
may commence, and the integration of the rest of the elements filing figures of the RB parallel communication scheme.
may be continued independently. As may be seen in Fig. 9 after the integration of elements
Thus the overlapping of the communication and some cal- in the first group, which are connected to intersubdomain
culation can be realized in the explicit finite-element problem boundary nodes (within subtask RB㛭st㛭1), all communication
by dividing the elements within each subdomain into two is launched to sum the local contributions of the subdomains
groups. Those elements will be placed into the first group, to the vector N. This means that there are two types of oper-
which are connected to intersubdomain boundary nodes, and ations for each group of intersubdomain nodes (i.e., for each
the rest of the elements, the ‘‘core’’ of the subdomain con- subdomain relation shown in Fig. 8):
To maximize program efficiency all of these communication TABLE 8. Parallel Execution Times Measured on Network of
buffers are dynamically preallocated in advance at program Workstations Utilizing Unix Sockets for Communication for A2
and A8 Meshes Using RB Algorithm
start-up by each instance of the program, and they correspond
to the neighboring subdomains of that instance. Number of tpar Ehw
When the main program is finished with the integration of Mesh subdomains np (s) S (%)
the elements within the core of the subdomain (i.e., subtask (1) (2) (3) (4) (5) (6)
RB㛭st㛭3 is complete), then the execution cannot continue any A2 8 8 2.1 4.1–5.8 63.3
further if the communication has not, as yet, been completed. A8 8 8 6.8 5.7–7.9 87.3
Thus synchronization takes place here, and it is implemented
by an MPI㛭Waitall( ) function call for the sends and a series
of MPI㛭Waitany( ) function calls for the receives. sults presented in Table 8. The notation of the subtasks of the
This function call is different because, for the threads RB communication scheme is as described in Fig. 9.
launched by the send routines, it is simply a check for com- It is apparent in Table 8 and Figs. 10 and 11 that the RB
pletion, and no other work has to be done. However for the communication scheme presented above is much more effi-
threads launched by the receive routines, further action is re- cient than the PM scheme. However, a significant proportion
quired. The MPI㛭Waitany( ) function call returns when any of the execution time is still spent on communication on cer-
of the receives are completed, and the content of the corre- tain processors. If that could be reduced, the program might
sponding receive buffer has to be added to the local system run almost as ideally parallel as for the shared memory plat-
form. An investigation to consider this ideal load balancing is
vector N of the subdomain. In such a way, after all receives
presented in the next section.
are returned for each subdomain relation, the vector of internal
forces N is synchronized between the subdomains. LOAD BALANCING
Timing and Profiling Result of RB Scheme With the introduction of the RB parallel communication
scheme in the previous section, the communication overhead
The program was recompiled and relinked with the MPI (i.e., the time spent on communication in the parallel system)
library as described earlier in the sections entitled Running on has been reduced considerably. However by examining the
Shared Memory MP Workstation and Running on Distributed parallel run-time profiling information in Figs. 10 and 11, it
Memory Workstation Cluster. Running the same examples on is apparent that some processors can still spend as much as
the shared memory MP workstation gives virtually the same 30% of their time on communication, while other nodes in the
results as described in Table 5. But these parallel runs were network spend much less.
almost ‘‘ideal’’; thus no further improvements were expected. As these processors are the fast ones (‘‘thebe’’ and ‘‘rhea,’’
However, running the same examples on the distributed which are P1 and P7 in Figs. 10 and 11) it indicates that this
memory workstation cluster with the same heterogeneous con- is not real work; rather these processors are waiting for the
figuration of workstations as described in Tables 3 and 6, the others to finish. The synchronization of their different speeds
timing results change significantly. In comparison with Table appear as extra communication load on the parallel system.
7 the new timing results for the RB parallel communication Clearly, this problem may be addressed by distributing the
scheme are presented in Table 8. All of the variables in this load onto the processors so that it is proportional to their pro-
table are as described for Table 7. cessing capabilities.
The parallel program runs were examined in detail, and the
profiling information is shown in Figs. 10 and 11. These fig- Energy Communication
ures were generated in a similar way as Figs. 5 and 6. It appears in Figs. 10 and 11 that a significant proportion
The timing ratios in Figs. 10 and 11 correspond to the re- of time is spent on the energy communication on processor
JOURNAL OF COMPUTING IN CIVIL ENGINEERING / APRIL 2000 / 87
FIG. 10. Percentage of Time Spent on Each Subtask of RB Parallelization Scheme by Every Processor on A2 Mesh
FIG. 11. Percentage of Time Spent on Each Subtask of RB Parallelization Scheme by Every Processor on A8 Mesh
P1. This is a collective type of communication as described sults hardly changed from those presented in Table 8. The
by Sziveri et al. (1996, 1997), and for such a course grained characteristics of the profiling shown in Fig. 12 had not
parallel system it should not take that long. To assess the effect changed from the original given in Fig. 11. This was expected
of this communication the source code of the program was as the computational load represented by the energy commu-
modified in such a way that the energy communication and nication, and the energy balance check is negligible compared
the energy balance check was cut out temporarily. The pro- to the load arising from the rest of the iteration. However,
gram was recompiled and run again with the A8 example mesh looking at the profiling data of processor P1 (in Fig. 12) the
on the distributed memory workstation cluster. The timing re- amount of time that was spent on the energy communication
88 / JOURNAL OF COMPUTING IN CIVIL ENGINEERING / APRIL 2000
FIG. 12. Percentage of Time Spent on Each Subtask of RB Parallelization Scheme by Every Processor on A8 Mesh When Omitting
Energy Communication
(Fig. 11) is now spent on waiting for the intersubdomain data As can be seen in Fig. 3, subdomain No. 4 is the one with
exchange to be completed. This clearly indicates that the as- the largest number of elements on the intersubdomain bound-
sumption above is correct (i.e., the difference of the processor ary; consequently the time bar corresponding to subtask RB㛭
speeds appears as an extra communication load). This is noth- st㛭1 is the longest in Fig. 11. Similarly, subdomain No. 3 is
ing else but the faster processor waiting for the slower to fin- the one with the largest number of elements in the core of the
ish. Consequently the load has to be split between the proces- subdomain; consequently the time bar corresponding to sub-
sors in such a way that they can complete their tasks at the task RB㛭st㛭1 is the shortest while the time bar corresponding
same time. to subtask RB㛭st㛭3 is the longest in Fig. 11.
Similar conclusions may be drawn by looking at other pro-
Load Balancing Decomposition cessors and the figure corresponding to the A2 idealization.
When taking the timing measurements shown in Figs. 10
and 11, the material was mainly in a linear state; consequently Adjustment of Decomposition
the computational load on each processor is approximately The computational power of a processor may be estimated
proportional to the number of elements within the subdomain on the basis of sequential run times for a given idealization
allocated to the processor. Thus the decomposition of the fi- (as was suggested earlier in the Speedup and Efficiency Cal-
nite-element mesh may be adjusted on the basis of the com- culation with Heterogeneous Computing section). Table 9
putational power of each processor participating in the het- shows the calculation of the number of elements to be placed
erogeneous network. onto each processor in the heterogeneous network when work-
ing with the A8 mesh. The calculation of the elements required
Intersubdomain Boundary versus Core Elements to load balance the system is based on the sequential execution
Other features shown in Figs. 10 and 11, which confirm times for the same A8 mesh as presented in Table 4.
their correctness and the conclusions above, may be deduced
by looking at the columns corresponding to processors P2, P3, TABLE 9. Redistribution of Elements of A8 Mesh When Using
Heterogeneous Network of Workstations for Eight Subdomains
and P4 in Fig. 11. These three columns correspond to processs
running on the machine carme, which obviously has identical tseq Number of
processors. The load is totally balanced as the number of el- Subdomain Host name (from Table 4) elements for
ements are equal in each subdomain. These conditions result number (from Table 3) (s) processor
in equal execution times for the sequential part of the com- (1) (2) (3) (4)
putation (i.e., RB㛭st㛭1 and RB㛭st㛭3). 1 thebe 42.0 118
This is clearly the case as the sum of the total combined 2 carme 49.0 101
time spent on subtasks RB㛭st㛭1 and RB㛭st㛭3 on each pro- 3 carme 49.0 102
cessor is almost equal. However, it is interesting to notice that 4 carme 49.0 101
the ratio between subtasks RB㛭st㛭1 and RB㛭st㛭3 changes 5 elara 51.0 98
6 elara 51.0 98
according to the ratio of the number of elements in the first 7 rhea 39.0 129
group (which are connected to the intersubdomain boundary 8 iapetus 54.0 93
nodes) to the second group (the core of the subdomain) within Total number of elements in mesh A8 840
each subdomain.
JOURNAL OF COMPUTING IN CIVIL ENGINEERING / APRIL 2000 / 89
run is shown in Fig. 14. It is apparent from Table 10 and Fig.
14 that when some load balancing is introduced into the het-
erogeneous parallel system the execution became very effi-
cient. Working on the given example, the processors spent
<5% of their time on communication, which enables the good
speedup. It also means that the combined computational re-
sources of this parallel system are very well utilized, and this
utilization is expressed by the high efficiency value shown in
Table 10.
FIG. 14. Percentage of Time Spent on Each Subtask of RB Parallelization Scheme by Every Processor on A8 Mesh after Rebalance
of Subdomains
ABSTRACT: Pattern recognition has turned out to be an important aspect of a dominant technology such as
machine intelligence. Domain specific fuzzy-neuro models particularly for the ‘‘black box’’ implementation of
PR applications have been recently investigated. In this paper, Kasuba’s simplified fuzzy adaptive resonance
theory map (ARTMAP) has been discussed as a pattern recognizer/classifier for image processing problems.
The model inherently recognizes only noise free patterns and in the case of patterns with noise or perturbations
(rotation/scaling/translation) misclassifies the images. To tackle this problem, a conventional moment based
rotation/scaling/translation invariant feature extractor has been employed. However, since the conventional fea-
ture extractor is not strictly invariant to most perturbations, certain mathematical modifications have been pro-
posed that have resulted in an excellent performance by the pattern recognizer. The potential of the model has
been demonstrated on two problems, namely, prediction of load from the yield patterns of elastoplastic analysis
of clamped and simply supported plates and prediction of modes from mode shapes.
NN are in general very sensitive to absolute magnitude, and where 0 <  ⱕ 1. Once SFAM has been trained, the equivalent
fluctuations in inputs may tend to swamp the performance of of a ‘‘feedforward’’ pass for an unknown pattern classification
the network while predicting the desired outputs. Hence the consists of passing the input pattern through the complement
need for normalization of inputs so that the inputs correspond coder and into the input layer. The output node activation func-
to the same range of values. tion is evaluated and the winner is the one with the highest
Complement coding is a normalization rule for preserving value. The category of the input pattern is the one with which
amplitude information. Thus it represents the on-response and the winning output node is associated.
the off-response to an input vector. It also represents the pres-
ence of a particular feature in the input pattern and its absence. WORKING OF SFAM
For example, if a is the given input pattern vector of d fea-
tures, the complement coded vector a represents the absence The circle in the square problem requires a system to iden-
of each feature, where a⬘ is defined as tify which pockets of a square lie inside and which lie outside
a circle whose area equals half that of the square. This prob-
a⬘ = 1 ⫺ a (1) lem, specified as a benchmark problem for system perfor-
mance evaluation in the defense advanced research projects
The previous equation is valid because, just as in fuzzy logic, agency (DARPA) artificial neural network technology pro-
all SFAM input values must be within the range of 0–1. gram, has been used to illustrate the workings of SFAM. In
Therefore, the complement coded input vector I internal to this, SFAM is trained with a set of points for a definite number
SFAM is given by the 2D vector of training epochs. The inputs during training are the points
I = (a ⭈ a⬘) = (a1, a 2, . . . , ad , a⬘,
1 a⬘,
2 . . . , a⬘)
d (2) (x, y) and the category to which they belong, namely, inside
the circle (IN) or outside the circle (OUT). The training of
Output Node Activation SFAM has been illustrated in the following examples.
Example 3
Example 4
冕冕
⬁ ⬁
Because there are two weight nodes to decide which node is mpq = x py qf (x, y) dx dy, p, q = 0, 1, 2, . . . , ⬁ (7)
⫺⬁ ⫺⬁
fit enough to learn the new input I, the activation function
values of the two nodes are computed [i.e., T1(I ) = 0.9333 and However, in the case of a spatially discretized M ⫻ N image
T2(I ) = 0.8999]. The node with the highest activation function denoted by f (i, j), (7) is formulated using an approximation
value, W1 in this case, is chosen to learn the new input. Also, of double summations
冘冘
the match function of I with W1 is greater than the vigilance M N
parameter [i.e., (MF(I ) = 0.7) > ( = 0.5)]. However, due to
mpq = i pj qf (i, j) (8)
category mismatch, W1 is not fit to learn the input, and hence i=0 j=0
the next node is to be considered. Before proceeding to the
next node, match tracking is done by updating the vigilance The so-called central moments are given by
冘冘
parameter by a small quantity (i.e., = 0.501). The next node M N
W2 gives MF(I ) = 0.9. Because MF(I ) > and the categories pq = (i ⫺ ˆi ) p( j ⫺ ˆj )qf (i, j) (9)
are also the same, learning occurs in W2 given by the updating i=0 j=0
of W2 as W new
2 = (0.7, 0.9, 0.1, 0.1).
where
Example 5 m10 m01
iˆ = ; jˆ = (10a,b)
m00 m00
I = (0.1, 0.3) and category = IN
Augmented input I = (0.1, 0.3, 0.9, 0.7) The central moments are still sensitive to R and S transfor-
Activation function values of W1 and W2 are T1(I ) = 0.5999 mations. The S invariant may be obtained by further normal-
and T2(I ) = 0.33 izing pq or by forming
from raw measurements. In practical imagery, various geo- ⫺ (30 ⫹ 312)(21 ⫹ 03)[3(30 ⫹ 12)2 ⫺ (21 ⫹ 03)2] (18)
metric distortions or pattern perturbations may be observed in
the image to be classified. Fig. 2 illustrates some example Thus, for a common PR application, which involves rec-
pattern perturbations. It is therefore essential that features that ognition of 2D images, the moment based invariant functions
are invariant to orientations be used for the classification pur- 1 –7 are computed from the raw measurements of every
pose. For 2D images, moments have been used to achieve R, given shape. Those shapes with similar i are classified as
S, and T invariants. Properties of invariance to R, S, and T belonging to the same class. In other words, shapes belonging
transformations may be derived using function of moments. to a class are the perturbed (rotated/scaled/translated) or exact
The moment transformation of an image function f (x, y) is version of the given nominal shape.
given by However, though the set of invariant moments shown are
94 / JOURNAL OF COMPUTING IN CIVIL ENGINEERING / APRIL 2000
invariant to T, despite their being computed discretely, the mo- as shown in Fig. 5. Here X *, Y * are the axes passing through
ments cannot be expected to be strictly invariant under R and the origin O. The dimension of any pixel is taken as a 1 ⫻ 1
S changes due to the ‘‘non 1:1 nature of general discrete square. Considering the unit thickness of the lamina, the image
geometric transformations (including rotation and scale function is analogous to mass density function. Total mass of
changes)’’ (Schalkoff 1989). an image is given as
冕冕
An investigation by the writers revealed that in the defini-
tion of pq, the contribution made by a pixel had been over-
looked. The modified pq definitions will be discussed in the M= f (x*, y*) dx* dy* (19)
succeeding section.
and the mass of pixel
pq REDEFINED
(i, j) = f (x *
i , y*
j ) (20)
Visual image data can be either binary (for monochrome
冘冘
n n
images) or real (for color images). For example, in Fig. 3 the
region R appears in the binary image as region of intensity M= f (x *
i , y*
j ) (21)
j=1 i=1
‘‘1.’’ If we represent an image function as f (x, y), then f (x, y)
is either 0 or 1 depending on whether the pixel is dark or assuming any figure is represented in a square region of n ⫻
bright. On the other hand, in Fig. 4 the intensity is represented n pixels. In general, the moment transform of an image func-
by various shades with 0 ⱕ f (x, y) ⱕ 1, indicating that the tion f (x, y) is given by
density lies anywhere between the ends of a spectrum: very
dark to very bright. This is how color images are represented.
However, the image function f (x, y) is constant over any pixel
region.
mpq = 冕冕 x*py*qf (x*, y*) dx* dy* (22)
Moments are descriptive techniques with an intuitive basis m10 which represents the first moment of the mass of the image
in the study of mechanics of bodies. Consider the real image about the Y *-axis is given by
From Fig. 6
x * = x i* ⫹ x (24)
冘冘
n n
m10 = f (x i*, y *
j )x *
i = Mx̄ *
C (26)
j=1 i=1
And similarly
FIG. 4. Color Image
11: Derivation
Now
冕 y dy = 0
冕冕 xy dx dy = 0
FIG. 7. (a) XY Axes Passing through Center of Mass (C ) of Im- 11 physically means the ‘‘diagonal’’ indication of the quad-
age; (b) Pixel (i, j ) Represented in X, Y-Axes rant with respect to the centroid where the region has more
mass.
冘冘
n n
m01 = f (x *
i , y*
j )y *
j = Mȳ *
C (27) 30, 03: Derivation
冕冕
j=1 i=1
where x̄ *
C and ȳ *
C denote the coordinates of the mass center of 30 = X 3f (X, Y ) dX dY (35)
the image with respect to the origin. Now consider the X, Y-
冕冕
axes passing through the center of the mass of the image as
shown in Fig. 7. Variables xi , yj represent the distance of the 30 for a pixel (i, j) = f (xi , yj) (xi ⫹ x)3 dx dy (36)
center of the pixel (i, j) from the centroidal axes. Now define
the central moments pq as
30 for a pixel (i, j) = f (xi , yj)(x 3i ) (37)
冘冘
n n
where 00 = m 00 = M, the central mass of the image. Also, 10 30 = f (xi , yj)(x 3i )
j=1 i=1
= 01 = 0, since the axes pass through the center of mass of
the image. and represents physically the ‘‘horizontal imbalance’’ of the
The derivations of pq, p, q ⱕ 3 are presented next. location of the CG with respect to half horizontal extent. Sim-
ilarly
冘冘
20, 02: Derivation n n
03 =
冕冕
f (xi , yj)(y3j )
j=1 i=1
20 = X 2f (X, Y ) dX dY (29)
and represents the ‘‘vertical imbalance’’ of the location of the
CG of mass with respect to half vertical extent.
where X = xi ⫹ x.
冕 冕
1/2 1/2 21, 12: Derivation
冕冕
20 of pixel (i, j) = f (xi , yj) dy (xi ⫹ x)2 dx (30)
⫺(1/2) ⫺(1/2)
21 = X 2Yf (X, Y ) dX dY (38)
冘冘 冉 冊
n n
(39a)
1
20 = f (xi , yj) x ⫹
2
i 21 for a pixel (i, j) = f (xi , yj)(x y ) 2
i j (39b)
j=1 i=1 12
In the discrete case
and is physically expressed as the ‘‘horizontal centralness of
冘冘
n n
the mass of the image.’’ Similarly
21 = f (xi , yj)(x 2i yj)
冘冘 冉 冊
n n
1 j=1 i=1
02 = f (xi , yj) y2j ⫹
j=1 i=1 12 and similarly
96 / JOURNAL OF COMPUTING IN CIVIL ENGINEERING / APRIL 2000
冘冘
n n
冘冘 冉 冊
n n
1
20 = f (xi , yj) x 2i ⫹ (43)
j=1 i=1 12
冘冘 冉 冊
n n
1
02 = f (xi , yj) y2j ⫹ (44)
j=1 i=1 12 FIG. 8. Structure of SFAM Based Pattern Recognizer
冘冘
n n
21 = f (xi , yj)(x 2i yj) (45) define the palette. The image processor receives the images
j=1 i=1
that are to be trained (training patterns) or inferred by SFAM
冘冘 (inference patterns).
n n
12 = f (xi , yj)(xi y2j ) (46) The feature extractor obtains the RST invariant features for
j=1 i=1 each image, whether for training or inference. From the images
冘冘
n n surrendered for training or inference, the seven moment based
11 = f (xi , yj)(xi yj) (47) invariant functions, 1 –7, are extracted by the feature ex-
j=1 i=1 tractor for presentation to SFAM. Thus, SFAM works only on
冘冘
n n
the feature vectors of the image and not on whole images. The
SFAM activator functions as two modules: the training module
30 = f (xi , yj)(x 3i ) (48)
j=1 i=1
(SFAM training) and the inference model (SFAM inference).
The feature vectors of the training patterns and the categories
冘冘
n n
to which they belong (image category) are presented to the
03 = f (xi , yj)( y3j ) (49) SFAM’s training module. The only user selectable parameter
for the training session is the vigilance parameter (0 < <
j=1 i=1
The moments listed above are still sensitive to R and S trans- 1). Once the training is complete, the top down weight vectors
formations. Note that the S invariance may be obtained by represent the patterns learned. Next, the feature vectors of the
further normalizing pq as given in (11). images that are to be recognized/classified are presented to the
From (40)–(49) it may be observed that 20 and 02 are inference module. SFAM now begins its classification of im-
different from their conventional definitions of central mo- ages by associating the feature vectors with the top down
ments. In the conventional definition of 20, for example, the weight vectors. The category of the image is output by SFAM.
term The system can handle both symmetric and asymmetric pat-
冘冘 冉冊
n n terns. However, in the case of symmetric patterns it is essential
1 that only distinct portions of the images be trained. This is
f (xi , yj)
j=1 i=1 12 because in the case of doubly symmetric patterns, their RST
invariant feature vectors 2 –7 acquire values that are very
has been neglected. This omission has resulted in a cascading
close to 0, and 1 tends to 1. This consequently results in
effect rendering 20 and 02 and the functions 1, 2, and 6
feature vectors, which are similar, leading to misclassification
incorrectly defined, leading to the misclassification of images
of patterns.
with RST orientations.
Using the normalized central moments and tools of invariant
algebra, a set of seven RST invariant features, similar to those APPLICATION 1: PREDICTION OF LOAD FROM YIELD
shown in (12)–(18), may be derived. Note that 7 is actually PATTERNS OF ELASTOPLASTIC, CLAMPED, SIMPLY
ST invariant and changes its sign for reflection. This is because SUPPORTED PLATES
for reflection about the Y-axis, let us say x becomes ⫺x, for
odd p, pq (i.e., 30, 12) changes sign, and hence 7 changes Benjamin Whang (1969) developed a finite-element dis-
sign for reflection. placement method for elastoplastic analysis of bilinear strain
hardening orthotropic plates and shells, also considering elastic
STRUCTURE OF SFAM BASED unloading. The two basic approaches adopted by Whang in
PATTERN RECOGNIZER the elastoplastic analysis are the initial stiffness approach and
tangent stiffness approach in conjunction with the Huber-
The overall structure of the pattern recognizer is as illus- Mises yield criterion, and the Prandtl-Reuss flow rule in ac-
trated in Fig. 8. The images (patterns) whether monochrome cordance with the strain hardening yield function.
or color are input through the image processor. In this work, Fig. 9 shows the uniformly loaded isotropic plates, one with
the images are engraved on a grid size of 40 ⫻ 40. Also, clamped edges and the other with simply supported edges. The
images may have their colors chosen from a fixed palette. Here following numerical values are used:
the colors that are fixed are white, black, green, yellow, and
magenta. However, any number of colors could be used to t = 1.0, E = 30,000.0, = 0.3, Ep = 300.0, and 0 = 30.0
ABSTRACT: Many engineering activities, such as building design and construction, site redevelopment, and
industrial operations, are governed by environmental regulations that are constantly being revised and published.
Many state environmental protection agencies are now starting to use Web sites to provide access to their
regulations. From the result of a preliminary investigation, many current state regulatory agency Web sites
provide no support, other than browsing, for finding relevant regulations. A Java-based environmental regulation
broker has been designed to help users in locating relevant environmental regulations. Eventually, such broker
software can be provided by regulatory agencies as a standardized support to access their regulations over the
Internet.
INTRODUCTION line regulations, some states used central Web sites, which
were often the state legislation Web sites. Thus, the agencies
The U.S. Environmental Protection Agency (EPA) and state only provided hyperlinks to the corresponding on-line regu-
departments of environmental protection define and execute lations. However, these state agencies lost access control to
environmental policies and regulations for individuals, com- their on-line regulations, which is needed to update, delete,
panies, and corporations to follow. These environmental reg- and classify their regulations. Most state agencies disclaimed
ulations, which are frequently revised, limit the nature and the accuracy of the on-line documents and regulations. The
amounts of air pollution, water pollution, and solid waste that second feature, search mechanism, examined how users found
can be generated from industrial activities. Members of the the relevant regulations on-line. Two major search mechanisms
regulated community request environmental regulations from, were found at the Web sites: classifier-based and full-text
and ask environmental policy questions to, the U.S. EPA and search mechanisms. Sites that provided classifications or cat-
the state agencies. Most often, the reason for asking such ques- egories would generally use air, water, land, and waste as the
tions and requesting regulations is that these people do not only classifiers. Full-text search was used for searching over
want to fail to consider, or violate, the current environmental the on-line regulations or over the entire HTML files of the
regulations and policies that are applicable to their operations. sites. The third feature, permitting service, examined the on-
To comply with the most up-to-date environmental policies line permitting process provided by the agency. The state agen-
and regulations, engineers and lawyers have to locate and de- cies, which were authorized to approve permitting applica-
termine which regulations, and provisions in these regulations, tions, usually put their application forms, frequently asking
apply to their companies’ activities. The regulatory agencies questions (FAQ), and guides on their sites. Several sites pro-
are trying to make their regulations more accessible by adopt- vided wizard-like interactive questions to help users find the
ing new technologies. It is our premise that Web-based support required permits for their business activities. The fourth fea-
for accessing and using regulations can provide an easier way ture, viewing support, examined the formats of the on-line
for users to access the most recent regulations, and for main- regulations. The formats of these regulations included: plain
tainers to maintain and update these regulations (Liang 1998). text, Adobe PDF files, WordPerfect files, Microsoft Word files,
We also suggest that support for searching, in addition to and HTML files. The regulations were sometimes divided into
browsing, is necessary to accelerate the process of locating the chapters and sections. Few state agencies were taking advan-
applicable regulations and provisions for the regulated public. tage of the hyperlink function in the HTML files for cross-
A preliminary investigation has been conducted of the 50 referencing the set of regulations. Several agencies provided
state EPA Web sites. The focus of the investigation was to downloadable PDF files or ZIP files of the regulations. The
discover the usability and the distribution of the on-line reg- last feature, notifying service, examined the ways that the
ulations at these Web sites. Five features were evaluated for agencies notified their community about the changes of the
each state agency’s Web site: regulation on-line service, search regulations. One state agency allowed users to submit their
mechanism, permitting service, viewing support, and notifying e-mail addresses to obtain updated information about their reg-
service. The results are summarized in Table 1. The degree to ulations. Others provided news hyperlinks or labeled the up-
which the area of a circle is filled in Table 1 represents the dated regulations with icons. Several states also provided on-
level that the feature is provided by the state agency’s Web line access to permitting statuses.
site. The first feature, regulation on-line service, represents From this preliminary investigation, we can see that most
whether the Department of Environmental Protection or state of the state agencies have put their regulations on the Web
agency put its regulations on-line. Some state agencies put with different levels of support. Different strategies and meth-
partial sets of regulations on-line because they were relevant ods must be utilized by the users of these systems to search
to permitting processes. Others only put office phone numbers and browse the regulations relevant to their business activities.
on the Web by which requests of information about their reg- However, to help and allow users to locate more easily and
ulations could be made. To provide a unified version of on- effectively the applicable and appropriate regulations and pol-
1 icy documents, a systematic approach for treating these doc-
Res. Asst., Dept. of Civ. and Envir. Engrg., Institute for Complex
Engineered Systems, Carnegie-Mellon Univ., Pittsburgh, PA 15213. uments is necessary. In this paper, we describe a Java-based
2
Prof., Dept. of Civ. and Envir. Engrg., Institute for Complex Engi- regulation broker system that the environmental agencies
neered Systems, Carnegie-Mellon Univ., Pittsburgh, PA. could use to provide on-line regulation search services to their
Note. Editor: Sivand Lakmazaheri. Discussion open until September regulated constituency. In this paper, we will focus on the tech-
1, 2000. To extend the closing date one month, a written request must nical feasibility of the regulation broker. Although important,
be filed with the ASCE Manager of Journals. The manuscript for this
the legal issues of using the broker system will not be ad-
paper was submitted for review and possible publication on December 9,
1998. This paper is part of the Journal of Computing in Civil Engi- dressed here, as the broker is not intended to replace lawyers
neering, Vol. 14, No. 2, April, 2000. 䉷ASCE, ISSN 0887-3801/00/0002- but rather augment the regulated public’s ability to seek out
0100–0108/$8.00 ⫹ $.50 per page. Paper No. 19824. relevant information before seeking legal advice. The history
100 / JOURNAL OF COMPUTING IN CIVIL ENGINEERING / APRIL 2000
TABLE 1. Investigation Results of 50 Environmental State Agency Web Sites
of the regulation broker is first reviewed. The Java-based bro- search engine inside the Inquery search engine was used for
ker system and its architecture are then discussed, and next, a both searching schemes. The regulations in the agency data-
pilot test of this system and the results of this test are pre- base were marked up with standard generalized markup lan-
sented. Finally, the future directions for a standardized regu- guage (SGML) tags for the SGML build program to create the
lation broker system are discussed. internal indexing codes. The SGML tags were used to store
the assigned classifiers with the text, the classification system,
HISTORY OF THE REGULATION BROKER and the logical structural information of the regulations and
provisions.
An environmental regulation broker is a piece of software Using this CGI-based broker, the regulation users were able
that assists engineers and companies to identify the applicable to retrieve applicable regulations from the broker, to traverse
environmental regulations and policies related to their business the hierarchical structure of a selected regulation, and to per-
or engineering activities. The broker takes advantage of the form classifier-based and full-text searches for specific provi-
connectivity of the Internet and delivers the applicable regu- sions. To browse a particular provision by using classifier-
lations to users’ machines. To assist users in identifying the based and full-text searches, the user had to request a list of
applicable regulations, a broker system has to provide search applicable regulations, select one of the applicable regulations,
schemes to locate the relevant regulations and browsing func- request the query form, fill out and send the query form, and
tionality for users to navigate through the contents of docu- then select the desired provision. Several disadvantages were
ments. Two different implementations of such a broker system found to exist with this previous approach: (1) Stateless op-
have been developed. In this section, a common gateway in- erations on client’s machine; (2) heavy workload of server;
terface (CGI) Web-based approach is reviewed. The more re- and (3) lack of editing tools. Each of these disadvantages are
cent system, the Java-based approach, is introduced in the next described in the following sections.
section.
Krofchik (1995) and Stasiak (1996) used CGI scripts and Stateless Operations on Client’s Machine
the Inquery search engine (Callan et al. 1992) to generate the
user interfaces and necessary searching components for a pre- The regulation users had to send HTTP requests to the bro-
liminary broker system. The CGI scripts were used to allow ker server when they wanted to navigate to the next level of
agencies to register their regulations to the broker system, to the classifiers. The users might need to send several HTTP
allow users to retrieve the applicable regulations, and to allow requests to find the document because the Web browsers could
users to do the full-text and classifier-based searches. The not store the navigation status. If users were accessing the
102 / JOURNAL OF COMPUTING IN CIVIL ENGINEERING / APRIL 2000
server via low-speed connections to the Internet, such as by a
modem, they could do nothing with the broker except wait for
results from the broker.
Heavy Workload of Server
Because every user had to send queries for retrieving the
documents and navigating the classification system, the work-
load of the server was made very heavy. The heavy workload
also came about because the broker had to keep track of the
states of the connected client machines. A better approach is
to have users navigate the classification systems by using the
resources of the users’ machines. Then, after users review and
select a set of classifiers, they can send the selected classifier
set to the broker to retrieve the applicable documents. Sup-
porting this usage pattern improves the usability and perfor-
mance of the broker system.
Lack of Editing Tools
In the CGI-based approach, the maintainer of the regulations
had to put SGML tags inside the documents for the search
engine to use when indexing the documents. The tags were
used for recording the associated classifiers and the logical
structure of the documents. The logical structural information
of the provisions was used to record the relationships between
the document files. Having the content of the classification
system mixed with the SGML tags in the document file could
lead to an inconsistency of the classification system. There
were no integrated editing tools available for the maintainers
to use in managing the document files, their structure, and their
classification.
JAVA-BASED VERSION OF BROKER SYSTEM
To address the limitations of the previous CGI-based broker,
a Java-based version of the broker system was developed that FIG. 1. New Scheme for Java-Based Broker System
provides a new user interface, a persistent data storage mech-
anism, and a self-contained search engine. The new scheme
for the broker system is presented in Fig. 1. The user interface to represent these items. The original form of most regulations
has been redesigned and allows users to navigate the entire is a paper hardcopy. These hardcopy documents must be trans-
classification system without sending requests to the Web formed into an electronic form so that the broker system can
server. The user downloads the user subsystem from the broker search through the regulations and present the relevant pro-
Web site. The user subsystem retrieves the whole classification visions to the users. Optical character recognition technology
system from the regulation database and displays it on the can be used to convert the hardcopy text into electronic form.
screen. The user can then perform classifier-based and full-text All the text, images, tables, and formulas must also be stored.
searches. The user subsystem sends the query to the broker Most of the current document formats, such as HTML, RTF,
server and displays the query result, which is the applicable Postscript, Microsoft Word, and PDF can be used to embed
regulation list, on the screen. the tables, images, and formulas within the documents. How-
A database management system (DBMS) is used to separate ever, most of these formats require a specific computer appli-
the classification systems and structural information from the cation to access them. Some applications are expensive and
document files. The DBMS is not only used for storing the some are not Internet-enabled. We chose the HTML format
classification systems and structural information about the doc- since it is supported by the HTTP protocol and the user is able
ument, but also for performing the classifier-based search. A to browse the documents by using the various Web browsers
commercial off-the-shelf search engine is used for performing that are freely available on the Internet. In addition, most word
full-text search. The search engine receives search queries, processors used to compose these documents now provide the
such as strings and keywords, from the user subsystem and functions to convert documents into HTML format. The math-
performs the full-text search over the regulation files. On the ematical equations within the regulations can be converted into
broker server, the bodies of the regulation files are not in the image files embedded in HTML files automatically. The users
database because both the database server and search engine can use their own text editors to modify the regulations with-
have to retrieve and locate these files. out digging into the HTML source files.
Several issues were considered during the design of the The broker maintainers can classify a regulation document
Java-based broker system prototype: (1) The representations with one or many classifiers. This document-classifier rela-
of the regulations and classification systems; (2) the platform tionship information should be stored in the broker system. If
of the prototype broker; and (3) the enabling technologies the information were stored inside the electronic format of the
needed to deliver the requirements of the broker concept. documents, it would increase the complexity of updating the
classifier assignment or the context of the regulations, which
Representations of Regulations and Classification had been discovered in the CGI-based version of this broker
Systems described previously.
Regulations contain text, images, tables, and formulas. We chose to use a relational DBMS to store both the doc-
Therefore, the representation of these documents must be able ument and classifier hierarchies. The hierarchical relationships
JOURNAL OF COMPUTING IN CIVIL ENGINEERING / APRIL 2000 / 103
between the classifiers can be represented in a relational
DMBS (Liang 1998); the relationships between classifiers and
documents can be treated as ‘‘has a’’ or ‘‘has many’’ relation-
ships. Information that needs to be stored in the database con-
sists of locations of document files, classifiers, relations be-
tween classifiers and documents, the classification hierarchy,
the document hierarchy, and the organization hierarchy. We
decided to store the location of the document in the database
but save the document body as one or more files in a directory
of a file system on the broker server. Using this approach, we
need only save the HTML-based layout and format of the reg-
ulations. However, an external full-text search engine is
needed as the DBMS cannot perform string search over the
external HTML files.
PILOT TEST
One objective of this specific research project was to con-
duct a field test with the PADEP in order to evaluate the us-
ability and performance of the broker system. The field test
started after the maintainer subsystem was built. We obtained
27 Air Quality Policy documents in MS Word format from the
PADEP. The documents were converted into HTML files by
using the ‘‘save as HTML’’ command provided by MS Word.
As each of these documents contained chapters and sections,
the documents were dissected based on their sections. Hence,
FIG. 3. Document List Viewer the table of contents of the document was used to represent
the logical structure of the document. For example, the doc-
lations, provisions, and classifications. The maintainer subsys- ument ‘‘Prevention of Significant Deterioration’’ contains 12
tem was built as a Java application. Unlike the Java Applet, a sections. Each section was saved as an HTML file in the di-
Java application can create a direct input and output stream rectory ‘‘prevention㛭of㛭significant㛭deterioration.’’ All the reg-
among the local machine and remote machines. The major ulation directories were saved under the parent directory
differences between a Java Applet and Java application is that ‘‘AirQualityPolicies.’’
the application is considered as a secure program by the Java Working with PADEP engineers, we developed a classifi-
VM. Thus, the Java application can communicate not only cation system for these documents. We used the same classi-
with the local machine and file system, but also with other fication system for classifying the policies and the provisions
remote database servers. In the maintainer subsystem, no index of these policies. The classification system was a four-level-
server connection component is needed since the MS Index deep hierarchy, with a total of 151 classifiers. Each classifier
Server has provided an administrator user interface for updat- was then assigned to the applicable sections of these 27 doc-
ing and indexing the files. The functionality of the DBServer uments used in the pilot study. It took about two complete
of the maintainer subsystem has been extended so that it can days to finish the assignment of classification. The result of
handle updating and deleting commands in SQL. this assignment process was then input into the broker system
The maintainer subsystem also has two user interface win- by using the maintainer subsystem. It took about 4 days to
dows for the maintainers to use in manipulating the regulation input the classifiers into the broker system. One way of in-
information. Additional pop-up menus are provided for main- creasing the document classification speed will be discussed
tainers to access maintenance commands. The maintainer’s in the next section. The following use case [see Jacobson
document viewer provides not only navigation functions but (1992) for the description of a use case] demonstrates how
also maintenance commands to maintain the list of regulations users locate and access the applicable regulations and docu-
known to the broker. The document list viewer of the main- ments in the PADEP Air Quality Policy collection.
ABSTRACT: The primary purpose of data transformation is to modify the distribution of input variables so that
they can better match outputs. The performance of a neural network is often improved through data transfor-
mations. There are three existing data transformation methods: (1) Linear transformation; (2) statistical stan-
dardization; and (3) mathematical functions. This paper presents another data transformation method using
cumulative distribution functions, simply addressed as distribution transformation. This method can transform a
stream of random data distributed in any range to data points uniformly distributed on [0,1]. Therefore, all neural
input variables can be transformed to the same ground-uniform distributions on [0,1]. The transformation can
also serve the specific need of neural computation that requires all input data to be scaled to the range [⫺1,1]
or [0,1]. The paper applies distribution transformation to two examples. Example 1 fits a cowboy hat surface
because it provides a controlled environment for generating accurate input and output data patterns. The results
show that distribution transformation improves the network performance by 50% over linear transformation.
Example 2 is a real tunneling project, in which distribution transformation has reduced the prediction error by
more than 13% compared with linear transformation.
冘
ered and the entire population is transformed to be uniformly N
distributed on [0,1], a continuous and smooth mapping may 兩Actuali ⫺ Predicti兩
MPE = 100 ⭈ (3)
be expected. Of course, the effectiveness of a transformation i=1 Actuali
can be validated by extensive examples.
where N = total number of data patterns; Actuali and Predicti
= actual and predicted values of pattern i, respectively.
CDF and Probability Distribution Function (PDF)
All NN inputs are independent random variables, and out- Error Difference
puts are dependent random variables. In statistics, the char- Error difference is defined as the difference of the absolute
acteristics of a random variable are defined by a PDF or CDF. prediction error between distribution and linear transforma-
It is interesting to notice that CDF is scaled between 0 and tions; that is
1. Inverse transformation method is widely used in Monte
Carlo simulation to generate random distributions from uni- EDi = 兩DErrori兩 ⫺ 兩LErrori兩 (4)
form distribution on [0,1] (Pritsker et al. 1997). Inverse trans- DErrori = Actuali ⫺ DPredicti (5)
formation uses the fundamental information of the CDF F(X).
The random variable R = F(X) is uniformly distributed on the LErrori = Actuali ⫺ LPredicti (6)
unit interval [0,1]. Thus, to generate a random sample of X, where EDi denotes the error difference between the two trans-
we generate a random number r and solve the equation r = formations for the same data pattern; DErrori and LErrori rep-
F(X) for the corresponding value of x = F ⫺1(r). resent the prediction errors of distribution and linear transfor-
The proof for the uniformality of r distribution is straight- mation, respectively; DPredicti and LPredicti = predicted
forward (Bratley et al. 1987) and is based on the following output values from the networks with distribution and linear
reasoning: Let R = F(X) have distribution G. Then, for 0 ⱕ r transformations, respectively.
ⱕ 1, we have The EDi measures the improvement (EDi < 0) or deterio-
G(r) = P[F(X) ⱕ r] = P[X ⱕ F ⫺1(r)] = F(F ⫺1(r)) = r (1) ration (EDi > 0) of distribution transformation over linear
transformation. The values of EDi on the entire data set can
Thus, R is uniformly distributed on [0,1]. be plotted on a scatter graph to enable the reader to visualize
the effect as detailed in the following examples. The averaged
Distribution Transformation EDi value indicates the overall improvement or deterioration
of distribution transformation over linear transformation
冘
Using the relationship between PDF and CDF, any distri- N
bution in any range can be transformed to a uniform distri- 1
MED = EDi (7)
bution on [0,1]. Therefore, this provides a method that can N i=1
transform all NN input random variables to uniform distribu-
where MED is the mean value of the EDi .
tions on [0,1].
Let X be an input variable. Observations corresponding to Example 1: Cowboy Hat Surface
N training patterns are x1, x2, . . . , xN. Using statistical methods
(Decision 1997), a random distribution can be fitted to best Mathematical functions provide a controlled environment to
match the sample with a CDF F(X). A new random variable generate accurate training and test data. They have been
is defined as widely used as a means for validing the performance of an
NN. The first example attempts to fit a cowboy hat surface as
Y = F(X) (2) shown in Fig. 1 and the corresponding function is given as
Then Y is a uniform distribution on [0,1]. The corresponding follows (Choueiki and Mount-Campbell 1999):
transformed data stream can be calculated from yi = F(xi). z = sin(兹x 21 ⫹ x 22) (8)
The term Y is the transformed input variable, and F(X) is the
transformation function, addressed as distribution transforma- Training Network Models
tion.
A BP NN is chosen for this example with three layers: 2
neurons on the input layer corresponding to two inputs (x1 and
PERFORMANCE MEASURES x2), one hidden layer with 22 neurons, and 1 neuron on the
Distribution transformation will be contrasted with linear output layer.
transformation. To examine its effect on the performance of Both x1 and x2 are limited on [⫺3,3]. To guarantee the qual-
an NN, the following traditional and nontraditional measures ity of training data, 500 data points are independently gener-
are used. ated for x1 and x2, respectively, using Monte Carlo simulation
with @Risk (Decision 1997) from the same triangular distri-
Traditional Measures bution (minimal = ⫺3, mode = ⫺3, and maximal = 3) to create
biased input distributions. The values of Z are then calculated
Traditional measures include (1) Correlation coefficient R; from (8).
(2) mean squared error; (3) mean absolute error; (4) minimal The 500 data patterns are treated as the training set.
absolute error; and (5) maximal absolute error. A well trained NeuralShell is employed for all training purpose with net-per-
110 / JOURNAL OF COMPUTING IN CIVIL ENGINEERING / APRIL 2000
FIG. 1. Cowboy Hat Surface
fect function to avoid overfitting (NeuralShell 1998). Two TABLE 1. Performance Measures for Fitting Cowboy Hat Sur-
training scenarios are designed. Scenario 1 scales the 500 input face (Training Set)
data to the range [⫺1,1] with linear transformation. Since Linear Distribution Improvement
NeuralShell has a built-in linear transformation function, the Item transformation transformation (%)
raw data can be directly presented with specifying linear trans- (1) (2) (3) (4)
formation for input data. Scenario 2 is presented with the data Correlation coefficient R 0.993 0.999 —
transformed using the triangular distribution function as fol- Mean squared error 0.003 0.001 66.7
Mean absolute error 0.043 0.018 58.1
lows:
再
Minimal absolute error 0.000 0.000 —
Maximal absolute error 0.233 0.122 —
(X ⫺ min)2 MPE (%) 18.57 7.89 57.5
, min ⱕ X ⱕ mod
(mod ⫺ min)(max ⫺ min)
F(X) =
(max ⫺ X)2
ues, sorted in ascendant order for the 500 patterns, are plotted
1⫺ , mod ⱕ X ⱕ max
(max ⫺ min)(max ⫺ min) versus the predicted output values from distribution trans-
(9) formation Predict(D) and linear transformation Predict(L) in
Fig. 2.
where min = ⫺3; mod = ⫺3; and max = 3. In Fig. 2, it is evident that the predicted values of distri-
Apply (9) to 500 data points of x1 and x2 separately. Trans- bution transformation are much closer to the actual values than
formed input data can be obtained with uniform distribution that of linear transformation. The error differences between
on [0,1]. They constitute the transformed training set for Sce- distribution and linear transformation are computed from (4)
nario 2 together with the original output values z. and are plotted in Fig. 3. The majority of the dots in Fig. 3
Experiments of the two training scenarios are performed lay under 0, which means the absolute predicted errors from
independently. It should be mentioned that the two scenarios distribution transformation are less for the majority of the 500
may need different learning rates, initial weights, and momen- training patterns. This also concludes the error reduction effect
tum values for the best performance even under the same net- of distribution transformation.
work architecture. Therefore, each training scenario must in-
dependently search for the appropriate values of the three Test Network Performance
parameters to achieve the best performance. The two best per- A test data set of 300 patterns is generated again using
forming models, one from each of the two scenarios, are then Monte Carlo simulation from a uniform distribution on [⫺3,3]
contrasted with the performance measures summarized in Ta- for both x1 and x2 with actual z values calculated using (8). It
ble 1. should be mentioned that the same distribution transformation
Table 1 shows that Scenario 2 has a significant improvement function (9) must be used to transform the test data before
(approximately 60%) over Scenario 1. Linear transformation they can be presented to the trained network. Two correspond-
results in an average prediction error of 18.57%, and distri- ing test scenarios are formed with the contrasting results de-
bution transformation results in an average prediction error of tailed in Table 2.
7.89%—an improvement of 57.7%. The mean absolute error Table 2 indicates that distribution transformation can sig-
also shows an improvement of 58.1%. The actual output val- nificantly reduce prediction errors with an overall improve-
JOURNAL OF COMPUTING IN CIVIL ENGINEERING / APRIL 2000 / 111
FIG. 2. Actual Outputs versus Predicted Outputs for Cowboy Function (Training Set)
ment of more than 50% for both the mean absolute error and
MPE. The actual outputs can also be plotted versus the two
sets of predicted outputs corresponding to the two test scenar-
ios.
Fig. 4 also suggests reduced prediction errors resulted from
distribution transformation. The error differences are plotted
in Fig. 5 to show the individual improvement for each pre-
dicting pattern. It is evident that much smaller prediction errors
are generated for the majority of the 300 test patterns.
EXAMPLE 2: A TUNNELING EXAMPLE
Reducing prediction errors has always been the interest of
an NN network application. This example applies the proposed
FIG. 3. Error Differences between Two Transformations for distribution transformation to a real-world construction prob-
Cowboy Hat Surface lem. The Brası́lia Tunnel (6.5-km-long, 9.6 m in diameter) was
excavated in shallow soft ground, as described in detail by
TABLE 2. Performance Measures for Cowboy Hat Surface
Ortigao et al. (1996). The tunneling method had the following
(Training Set) stages: excavation, placement of lattice girders, shotcreting the
primary layer of the lining, closing the invert, and finally plac-
Linear Distribution Improvement ing the final or secondary layer of the lining. In all cases a
Item transformation transformation (%)
clay core was always left to improve the stability of the face.
(1) (2) (3) (4)
Excavation encompassed three construction methods as
Correlation coefficient R 0.909 0.985 —
Mean squared error 0.041 0.008 80.9
listed in Table 3. Method A is the least expensive and is em-
Mean absolute error 0.095 0.046 51.6 ployed in favorable conditions of face stability and negligible
Minimal absolute error 0.000 0.000 — damage to nearby structures. On the other hand, if excavation
Maximal absolute error 1.328 0.522 — conditions deteriorate (such as decreasing thickness of soil
MPE (%) 72.1 34.3 52.4
cover, poor soils are encountered, sensitive nearby structures
exist, or deformation observation indicates that stability may
be decreasing), methods B or C could be selected. An addi-
tional construction method, D, consisting of side drift with
temporary invert, was not included in the analyses, because it
was only used for a very short extension of 20 m in length.
Surface settlement observations took place in 50-m intervals
along the tunnel length at three different excavation stages: at
the face passage, at the closing of the inverted arch, and the
final value after stabilization (Ortigao et al. 1996). In addition,
single settlement marks were deployed at 10-m intervals,
among the fully instrumented stations. The measured stabilized
settlement is plotted in Fig. 6, which appears discontinuous
due to missing or disqualified data points.
Training Network Models
FIG. 4. Actual Outputs versus Predicted Outputs for Cowboy BP neural networks with nine input variables, as listed in
Hat Surface (Test Set) Table 4, were attempted for predicting settlements during tun-
112 / JOURNAL OF COMPUTING IN CIVIL ENGINEERING / APRIL 2000
FIG. 5. Error Differences between Two Transformations for Cowboy Hat Surface (Test Set)
CONCLUSIONS
ABSTRACT: Traditional time-cost trade-off analysis assumes that the time and cost of an option within an
activity are deterministic. However, in reality the time and cost are uncertain. Therefore, in analyzing the time-
cost trade-off problem, uncertainties should be considered when minimizing project duration or cost. Simulation
techniques are useful for analyzing stochastic effects, but a general strategy/algorithm is needed to guide the
analysis to obtain optimal solutions. This paper presents a hybrid approach that combines simulation techniques
and genetic algorithms to solve the time-cost trade-off problem under uncertainty. The results show that genetic
algorithms can be integrated with simulation techniques to provide an efficient and practical means of obtaining
optimal project schedules while assessing the associated risks in terms of time and cost of a construction project.
This new approach provides construction engineers with a new way of analyzing construction time/cost decisions
in a more realistic manner. Historical time/cost data and available options to complete a project can be modeled,
so that construction engineers can identify the best strategies to take to complete the project at minimum time
and cost. Also, what-if scenarios can be explored to decide the desired/optimal time and/or cost in planning and
executing project activities.
trade-off problems by assuming deterministic time and cost, significant overlap. It may be sufficient to use only the mean
at both the activity and the project levels. values of the duration and cost of these options to find the
optimal choice for each activity in such a situation.
Activity Level On the other hand, if there are significant overlaps between
the distributions, using only the mean values is no longer suf-
ficient. For example, in Fig. 4 we cannot say that the duration
Overlap between Distributions
of Option 1 is definitely shorter than the duration of Option
In the traditional deterministic time-cost trade-off problem, 2, even though this is true of the mean values. Therefore, we
only certain values of the duration and cost of the options need a comparison scheme that considers the whole distribu-
within an activity, usually the mean values, are used to eval- tion in order to evaluate the duration and cost of an option.
uate the options. This may be done with the stochastic time-
cost trade-off problem only if there are no significant overlaps Correlation between Duration and Cost
between the distributions of both duration and cost of the op- Because both the duration and cost of an option are deter-
tions. For example, in Fig. 3 both the duration and cost of the mined by the type of resource being used, it is reasonable to
options are distributed within their specific regions without assume that the duration and cost of an option are correlated.
118 / JOURNAL OF COMPUTING IN CIVIL ENGINEERING / APRIL 2000
FIG. 5. Example of Correlation between Duration and Cost of Activity Option
Using only the mean values of the duration and cost does not the solution. As shown in Fig. 6, we can see that d1, 35% >
reflect the correlation between the duration and cost of an op- d2, 35% and c1, 35% < c2, 35%, which explains that neither Solution
tion. The duration and cost of an option can be obtained from 1 nor 2 is dominant. Having information describing the entire
historical data and so can their correlation, which may be ob- distribution allows us to assess risk associated with a particular
tained by using regression analysis. Examples of possible cor- solution.
relation between the cost and duration of an option are shown
in Fig. 5. The first graph in this figure shows that the cost and Correlation between Project Duration and Cost
duration of an option to finish an activity are independent of
each other; the second graph shows that they are negatively Because the project duration and cost are produced by the
correlated. same combination of options, it is reasonable to assume that
the project duration and cost are correlated. However, the real
Correlation between Activities values of the project duration and cost are now known until
the project is finished; the same is true of the correlation be-
Some activities within a project use the same resources, tween the project duration and cost. Nevertheless, we can still
such as heavy equipment; therefore, these activities may be estimate the mean values, variance, and correlation of the
affected by each other. In other words, these activities are cor- project duration and cost according to the results from simu-
related to each other. Such correlated relationships are not re- lation.
flected when using only the mean values of the options to
evaluate the project duration and cost. EXISTING TECHNIQUES
Project Level Existing techniques for construction time-cost trade-off
problems can be categorized into three areas: heuristics, math-
Overlap between Distributions of Project Duration and Cost ematical programming, and simulation. The following sections
briefly describe the strengths and weaknesses of the existing
In solving the traditional deterministic time-cost trade-off techniques.
problem, both the project duration and cost are determined by
the length of the critical path and the sum of the costs of the Heuristic Methods
activities, respectively. As a result, the project duration and
cost are evaluated based on certain values, which are usually Heuristic methods are based on rules of thumb that gener-
the mean values. However, such an approach is not sufficient ally lack mathematical rigor. They provide good solutions but
for stochastic time-cost trade-off problems. For example, in do not guarantee optimality. Examples include Fondahl’s
Fig. 6 both the project duration and cost of Solutions 1 and 2 method (1961), Prager’s structural model (1963), Siemens’s
are normally distributed. If we consider only the mean values effective cost slope model (1971), and Moselhi’s structural
of the project duration and cost, we would conclude that So- stiffness method (1993). In resource-constrained scheduling,
lution 1 dominates Solution 2, because Solution 1 finishes the intensive research has been conducted under the consideration
project more quickly and uses less money than Solution 2 of renewable resource constraints in terms of minimizing
does. However, the above comparison becomes insufficient if project duration. Heuristic methods have enjoyed the majority
the whole distribution of the project duration and cost are of the research effort because they can find good solutions with
taken into consideration. For instance, suppose we compare far less computational effort than other methods. Heuristics
two solutions based on 35 percentile values without consid- can be classified into two types: (1) Serial heuristics, in which
ering the correlation between the project duration and cost of processes are first prioritized and retain their values throughout
JOURNAL OF COMPUTING IN CIVIL ENGINEERING / APRIL 2000 / 119
the scheduling procedure; and (2) parallel heuristics, in which ration or expenditures (Dobin 1985; Weiss 1986). Few studies
process priorities are updated each time a process is scheduled. have been conducted to optimize the project in a stochastic
Examples are Bell and Han (1991), Kahattab and Choobineh network. Examples are Wan’s perturbation analysis (1994) and
(1991), and Boctor (1993). These heuristics provide good so- Kidd (1987). They provide a good strategy to analyze the re-
lutions; however, the performance is problem-dependent, and sults of simulation; however, the above analyses do not at-
good solutions are not guaranteed. tempt to optimize the entire project time-cost trade-off curve.
Heuristic methods, mathematical models, and simulation
Mathematical Programming Models techniques show both strengths and weaknesses. The heuristic
approaches select the processes to be shortened or expanded
Mathematical programming methods convert time-cost based on certain selection criteria, but they do not guarantee
trade-off problems to mathematical models and utilize linear optimal solutions. On the other hand, mathematical models
programming, integer programming, or dynamic programming require great computational effort, and some approaches do
to solve them. Kelly (1961) formulated time-cost trade-off not provide the optimal solution either. In addition, for large-
problems by assuming linear time-cost relationships within scale networks, neither heuristic methods nor mathematical
processes. Other approaches, such as those by Hendrickson programming models can obtain optimal solutions efficiently.
and Au (1989) and Pagnoni (1990), also use linear program- Simulation techniques provide a good estimate for the optimal
ming to solve the time-cost trade-off problem. Linear program- solutions; however, a guide to analyze the result of the sim-
ming approaches are suitable for problems with linear time- ulation must be provided in order to find the solutions effi-
cost relationships but fail to solve those with discrete time-cost ciently. The state of research suggests the need to develop a
relationships. Meyer and Shaffer (1963) and Patterson and more efficient algorithm to conduct optimal time-cost trade-
Huber (1974) solved time-cost problems including both linear off analysis under uncertainties.
and discrete relationships by using mixed integer program-
ming. However, integer programming requires a prohibitive GAs
amount of computational effort once the number of options to
complete an activity becomes too large or the network be- GAs are search algorithms developed by Holland (1975),
comes too complex. Burns et al. (1996) took a hybrid approach which are based on the mechanics of natural selection and
that used (1) linear programming to find a lower bound of the genetics to search through decision space for optimal solutions
trade-off curve; and (2) integer programming to find the exact (Goldberg 1989). The metaphor underlying GAs is natural se-
solution for any desired duration. Robinson (1975), Elmagraby lection. In evolution, the problem each species faces is to
(1993), and De et al. (1995) used dynamic programming to search for beneficial adaptations to the complicated and chang-
solve time-cost trade-off problems for a special class of net- ing environment. In other words, each species has to change
works that can be decomposed to pure series or parallel sub- its chromosome combination to survive in the living world. In
networks. GAs, a string represents a set of decisions (chromosome com-
bination), a potential solution to a problem. Each string is eval-
Simulation uated on its performance with respect to the fitness function
(objective function). The ones with better performance (fitness
Finding the optimal solutions for construction time-cost op- value) are more likely to survive than the ones with worse
timization in a stochastic network has proven to be a difficult
problem to solve. For example, if the durations of the options
were random variables (continuous or discrete), the evaluation
process would involve excessive numerical integration, which
makes it practically impossible. Simulation techniques have
been used to enhance the study of stochastic project networks;
however, many studies focus only on estimating project du- FIG. 7. Example of GA Chromosome Representation
FIG. 18. Mean ⫾1 Standard Deviation of Project Duration versus Mean of Project Cost
FIG. 19. Mean of Project Duration versus Mean ⫾1 Standard Deviation of Project Cost
ABSTRACT: This paper presents an interactive computer graphics program for implementing the nonlinear
strut-tie model approach for the practical design and analysis of disturbed regions in structural concrete. The
graphics program consists of three parts. The first and second parts are considered as the pre- and post-processors
for a nonlinear finite-element analysis of 2D plain concrete and plane truss structures. The results of the first
part provide guidance in the development of strut-tie models and in the evaluation of effective strength levels
for concrete struts. The third part is a design routine for dimensioning and detailing the components of the
developed strut-tie model. This graphics program implementing the nonlinear strut-tie model approach has been
shown to be efficient in the development, analysis, and detailing of strut-tie models for the analysis and design
of structural concrete including disturbed regions.
the nodal zones that connect the elements of the strut-tie model
FIG. 1. Flow Chart for Nonlinear Strut-Tie Model Approach and redirect the internal forces are dimensioned, and the bear-
ing capacities of these zones are then verified by employing
the finite-element nonlinear analysis used for evaluating the
cilitates the accurate consideration of loading and boundary principal stresses and their trajectories.
conditions and permits the updating of the material properties
and geometry of a structure. Although a finite-element nonlin-
ear analysis represents an improvement over linear elastic ap- ANALYSIS AND DESIGN TOOL NL-STM
proaches, it is not intended to model the actual structure for
final conditions. The finite-element nonlinear analysis of 2D NL-STM is a special-purpose computer graphics program
unreinforced concrete at this stage is only intended to help in of about 20,000 lines of Fortran code that is designed for use
the development of a more realistic strut-tie model. with UNIX operating system and HOOPS graphics package
Once a centerline strut-tie model for a given situation is (HOOPS 1997). It was developed as a tool for studying the
developed, a preliminary finite-element analysis of the model behavior of reinforced and prestressed concrete members, as
is conducted to determine the internal forces in its individual well as dimensioning those members. NL-STM consists of
members. After the internal forces are determined, the strut three parts. The first part functions as a pre- and postprocessor
and tie areas are calculated using a linear or nonlinear ap- to the nonlinear finite-element program PLANE for solving
proach that requires a few iterations within the effective stress plane solid problems. The preprocessor is a simple finite-ele-
level limits. The algorithm for calculating strut and tie cross- ment mesh generator that enables the user to draw a finite-
sectional areas using a linear approach is shown in Fig. 2. In element model on the screen. The computer then extracts and
order to evaluate behavior and strength and obtain accurate sends the required input data to PLANE. Thereafter, the post-
strut and tie forces in the analysis and design of the structural processor receives the output data (principal stresses, principal
concrete, a finite-element nonlinear analysis is again conducted angles, and principal stress ratios) from PLANE, which it then
on the strut-tie model itself. In this nonlinear analysis, the displays on the screen. In some cases, the first part can be
actual dimensions of the concrete ties positioned with steel ties bypassed and a strut-tie model developed directly based on the
and steel struts positioned with concrete struts are included in user’s experience. The second part enables the user to develop
the model. In addition, every concrete strut is considered to strut-tie models interactively, based on the results of PLANE.
have its own stress-strain relationship with a peak stress (ef- This part is also considered as a pre- and postprocessor, yet
fective stress) and corresponding peak strain; accordingly, the this time to the nonlinear finite-element program TRUSS for
tangent modulus of elasticity of every strut under an incre- solving plane truss structures. In this part, the computer ex-
mental external load can be used. The stress-strain relationship tracts and sends the required data to TRUSS. The postpro-
of a strut is approximated by the curve-fitting equation (Saenz cessor receives the output data (member forces, member
1964) that describes the stress-strain behavior of concrete un- stresses, and strains) from TRUSS, which it then displays on
der uniaxial compression. The stress-strain relationship of the screen. The third part enables the user to dimension the
every steel tie is assumed to be bilinear, which considers the struts, ties, and nodes. The strut design is based on its internal
strain-hardening of steel. The tangent modulus of elasticity of force and concrete compressive strength, which is a function
a steel tie before and after yielding is obtained using the slopes of the multiaxial state of stress and disturbance of the cracks
of two straight lines. When the stresses of a strut and tie reach and reinforcement. The nodal zone is checked by examining
their peak and ultimate stress, respectively, the stiffness of the its bearing capacity. The node has to be capable of carrying
strut and tie is assumed to be very small, such that these mem- the applied stresses resulting from the struts and ties. The ap-
bers cannot carry any additional forces. plication of the graphics program will be illustrated through
As the last step in the nonlinear strut-tie model approach, the design example outlined in the following section.
128 / JOURNAL OF COMPUTING IN CIVIL ENGINEERING / APRIL 2000
COMPUTER-AIDED DESIGN OF SQUAT SHEAR WALL model was dimensioned and precisely detailed, based on the
internal forces in the struts and ties, material properties of the
A three-story squat shear wall with openings (Paulay and concrete and steel, and current design specifications.
Priestley 1992), as shown in Fig. 3, was designed using the The objective of the preliminary finite-element analysis is
nonlinear strut-tie model approach with the aid of NL-STM. to determine the direction of the compressive principal stresses
The procedure is presented below in three steps. First, the and principal stress ratio of each finite element. The finite-
direction of the compressive principal stresses was determined element types available in NL-STM are three-, four-, and nine-
using a finite-element nonlinear analysis of plain concrete. node isoparametric elements, which give the user more flexi-
Second, a strut-tie model was developed interactively using bility in modeling any type of plane solid problem. Data
the finite-element analysis results. Finally, the developed preparation for a finite-element analysis is a frustrating and
time-consuming task; however, NL-STM provides a finite-
element mesh generator that eliminates these difficulties and
indicates whether the input is right or wrong.
The finite-element model is shown on the screen by a set
of straight lines. The lines are drawn to scale using a mouse.
Thereafter, the element connectivity is automatically calcu-
lated. Since the wall has openings, the finite elements in the
corresponding regions of the openings are deleted, and the
element connectivity is then recalculated accordingly (Fig. 4).
Specifying the load and boundary conditions is the last step
in the preparation of input data for the finite-element analysis
of a plane solid structure. These conditions are defined in the
global coordinate system. The current loading option only in-
cludes concentrated loads at predefined nodes. If external loads
are applied with history, a nonlinear finite-element analysis is
required. Boundary conditions are specified using four types
of supports: hinge, horizontal roller, vertical roller, and roller
FIG. 4. Finite-Element Mesh of Shear Wall after Calculating FIG. 5. Display of Finite-Element Analysis Results of Shear
Element Connectivity Wall: (a) Model I; (b) Model II
ABSTRACT: Higher education institutions in the United Kingdom have invested significantly in the implemen-
tation of communication and information technology in teaching, learning, and assessment of civil and
building engineering—with mixed results. This paper focuses on the use of digital imagery and visualization
materials to improve student understanding. It describes ways in which these materials are being used in the
civil and building engineering curriculum, and, in particular, how distributed performance support systems
(DPSS) can be applied to make more effective use of digital imagery and visualization material. This
paper centers on the extent to which DPSS can be used in a civil and building vocational and continuing
professional development context by tutors in the form of an electronic course delivery tool and by students in
the form of an open-access student information system. This paper then describes how a DPSS approach to
education is being adopted at Loughborough University as part of the CAL-Visual project. After highlighting
the main aims and objectives of the project and describing the system, this paper discusses some of the issues
encountered during the design and implementation of a DPSS and presents some preliminary results from initial
trials.
FIG. 1. Framework for Indexing Imagery and Visualisation Ma- Database Design
terial
One of the project’s objectives is to produce a set of CD-
of a Web-based database), three integrated CBL modules, a ROMs that contains platform-independent visual information.
CAA module, and the development of a suite of electronic The rationale for publishing the database in a Web-based form
lectures containing MTAs. Furthermore, it is proposed to in- on CD-ROM was to produce a product that is maintenance
clude on-the-job support and JIT training for tutors who re- free and able to be transported to a server-based configuration
quire assistance embedding such visual material within their if required. This flexibility will also enable the database to be
electronic lectures. A detailed description of the design and adapted to the needs of end-users and institutions.
implementation of the framework, database, and additional One option considered was to develop the database using a
system components is given in this section. proprietary software tool, such as Microsoft Access. An ad-
vantage of Microsoft Access is that it has a feature that pub-
lishes information in the form of a Web site. This Web site
Detailed Framework Description
can then be accessed on a stand-alone computer and on a Win-
The framework development began by investigating a num- dows 95 Personal Web server or on an NT server.
ber of classifications systems for the indexing of visual ma- Another option was to develop a Windows-based database
terial. It was then decided to use a keyword system based on using a bespoke software tool. Such a tool can be developed
the CI/Sfb system (Construction Index/Samarbetskommittén using a language, such as C⫹⫹. The main advantage is that
för Byggnadsfrâgor; a standard classification system for in- the database could be developed so that it is portable to a
dexing technical and product information in the construction number of platforms. However, the software tool used to ac-
industry used in Europe and the United Kingdom), augmented cess the database would need to be re-engineered and/or re-
with user-defined keywords to form a generic index. Each set compiled to work on each platform. As the system will mainly
of keywords associated with a visual object is to be augmented be used on one type of platform, for example Windows, this
with administrative information to form the metadata. A num- solution would have been suitable. It does, however, require a
ber of indexes have also been created using the visual material large amount of development work to produce a working pro-
and metadata that arrange the visual material into a number of totype.
key subjects. A third option that could meet the platform-independence
This framework is also used as the foundation for the design requirement is using Internet technologies. While this tech-
and implementation of a Web-based database that is to store nology has many advantages over traditional systems, it often
the visual material and its meta-data. Once the metadata and uses a server configuration and common gateway interface
visual material is entered into a Web-based database, the da- scripts to process the information accessible from a database.
tabase can be automatically published along with its indexes. This would therefore be inappropriate if the database is to be
To publish the database as a Web site, on the Internet, intranet, published on a CD-ROM.
or CD-ROM, each image will be stored in an HTML page The actual solution has involved using a combination of the
template. Once the database is published as a Web site, staff options given above. A Microsoft Access database has been
and students can use a number of DPSS components to access created in order to facilitate the entering of metadata and pub-
the visual material through an integrated user interface. lishing the information in the form of a Web site. A hybrid
retrieval system has then been developed using Web-based
System’s Components technologies. This is due to the issue of retrieving this Web-
based information, regardless of whether it exists on a CD-
The system contains a number of integrated components ROM or a server.
(Fig. 2) to provide staff and students with different pathways
in which to access the visual material. Starting with the un- Hybrid Retrieval System
derlying Web-based database, a retrieval system is the main
method of accessing visual material. In doing so, the retrieval Although there are many propriety software tools and lan-
system provides two methods of browsing the database. The guages available for developing Internet and CD-ROM prod-
136 / JOURNAL OF COMPUTING IN CIVIL ENGINEERING / APRIL 2000
FIG. 3. CAL-Visual Retrieval System
ucts, only a few lend themselves to the creation of hybrid text- will be created in the form of a VR environment that uses the
retrieval search engines. Many of the tools (such as WebCD) retrieval system as a central facility for accessing the image
are themselves platform-dependent, difficult to learn and use, materials.
and expensive. Alternatively, the languages are nonstandard
and difficult to learn. For this reason, two prototypes have VR Environment
been developed that use two of the more suitable languages:
Java and JavaScript. This is because Java and JavaScript are Although the textual user interface is considered to be the
based on a common programming language; they are platform- main access point for searching for visual material, the other
independent and are freely available. Although there was a form of access will be through a number of VR models. Each
small number of compatibility problems, when using different model will form a navigational aid and will be developed as
browsers to run Java applets and JavaScript, the latest versions an alternative interface for the retrieval of images from the
of the main two browsers (Internet Explorer and Netscape database. These models will represent a type of building
Communicator) are now compatible. through which the user can navigate and access the informa-
A hybrid retrieval system solution was created for use on a tion in the database (Fig. 4). An initial prototype has been
CD-ROM and a Web server using Java and JavaScript (Fig. developed using QuickTime VR consisting of a model of a
3). Because the components are created and embedded within low-rise building with hot spots allowing the user to zoom
an HTML page, the amount of programming required is re- into particular areas of the building (Fig. 5). This will be used
duced. In the initial phase of the project, a number of retrieval as an interface to access information in the database. For ex-
system prototypes were developed using each of these Web- ample, information on roof construction or defects can be ac-
based approaches to test their feasibility. cessed by clicking on the appropriate hot spot.
Browser compatibility was the main reason for not using
ActiveX. Although ActiveX is widely used to embed objects
into Internet Explorer, it is not commonly used within Net-
scape browsers. This is because using ActiveX within a Net-
scape browser requires the installation of an ActiveX plug-in.
Furthermore, a different plug-in is needed for each version of
the Netscape browser. Unlike ActiveX, both Java and Java-
Script do not require plug-ins.
Overall, the Java-based, retrieval system performed the best
and as a result showed the most potential. The next stage of
the project is to develop a simple textual user interface in the
form of an HTML form to provide staff and students with easy
access to visual material. In addition, a graphical user interface FIG. 4. VR Environment
IMPLEMENTATION STRATEGY
FIG. 6. Screen Display Showing Microsoft Access Form for Marking Up of Images in Database
• Use of the database by tutors to develop electronic lec- vironment. The implementation of additional sets of key-
tures by searching through the database for visual material words for other disciplines could provide an enhancement
and embedding it into PowerPoint presentations. to the system.
• Use of the database to develop CBL modules. As an ex- • The classification system used to mark up visual material
ample a building construction prototype was developed was seen as limiting by some of the users, and as a result
using images from the database. This prototype links vi- an additional field has been added to the database to allow
sual material to a flowchart of construction activities to for the input of user-defined keywords.
visualize specific activities within the construction se- • Some technical aspects that could limit the use of the
quence (Fig. 7). system had to be considered. For example, viewing video
• Visual material for a specific subject was made available clips required a suitable World Wide Web browser con-
to students for access in their own time. This visual ma- taining a video plug-in or player, and the browser used to
terial was linked to textual information and was made access the system needs to support HTML frames and be
available to students through the Internet. Java compliant.
• Finally, there can be limitations on the amount of visual
INITIAL SYSTEM EVALUATION AND RESULTS material to be stored that are imposed by hardware spec-
ifications.
Initial system evaluation and students’ trials have been con-
ducted in order to establish the system’s usability and its ben- Students’ Evaluation
efits to the process of teaching and learning. This was done Initial trials have been conducted with students using the
through workshops involving tutors and evaluation sessions building construction prototype (Fig. 7) as well as some MTAs
with students. developed by tutors using images and videos from the system.
These trials showed improvements in the following areas
System’s Evaluation (based on observations by tutors and students’ feedback):
• Better participation from students during the lecture ses-
• The performance of the system depended on where its sions
different components are located. The system worked fas- • Better students’ motivation during the tutorial sessions
ter on a CD-ROM and an intranet as compared to being • Less time required to complete tutorial exercises
accessed over the Internet. • Availability of visual material for access through the In-
• Some interface improvements were introduced such as the ternet well received by students
use of thumbnails, scalable images, and better feedback • Better retention of information by students
messages during the navigation through the system. • Easier assimilation particularly for those students with
• The thesaurus currently used in the system as it stands is limited practical experience
only suitable for visual material related to the built en-
JOURNAL OF COMPUTING IN CIVIL ENGINEERING / APRIL 2000 / 139
A more detailed evaluation involving a wider audience and Barker, P. G. (1995). ‘‘Interface design for learning.’’ Proc., Comp. Based
more prototypes will be undertaken in the final phase of the Learning in Sci. Int. Conf.
Barker, P. G. (1996a). ‘‘Living books and dynamic electronic libraries.’’
project. Electronic Library, Oxford, U.K., 14(6), 491–502.
Barker, P. G. (1996b). ‘‘Towards real information on demand.’’ Proc.,
CONCLUSIONS ONLINE ’96 Int. Conf., Learning Information (Europe) Limited, U.K.,
261–270.
Communication and information technology play an impor- Barker, P. G., and Beacham, N. A. (1999). ‘‘Making a case for distributed
tant role in the higher education sector encouraged by the in- performance support.’’ Proc., ED-Media ’99, Association for the Ad-
creasing availability and performance of computer technology vancement of Computing in Education, 311–316.
that facilitates the embedding of emergent technologies into Barker, P. G., and Hudson, S. R. G. (1998). ‘‘An evolving model for
multimedia performance support systems.’’ Trends in communication:
teaching and learning. In areas such as the built environment Part 4. Interface technology—Enhancing the quality of life, H. van
where visualization is important to convey concepts and pro- Oostendorp and A. G. Arnold, eds., Boom Publishers, The Netherlands,
cesses to students, the benefits offered by multimedia have 60–73.
been recognized and are being used to improve the process of Barker, P. G., Richards, S., and Banerji, A. (1994). ‘‘Intelligent ap-
teaching and learning. proaches to performance support, ALT-J.’’ J. Assn. of Learning Tech-
This paper discussed the potential of using DPSS as tools nol., Wales, U.K., 2(1), 63–69.
Beacham, N. A. (1998). ‘‘Distributed performance support systems,’’ PhD
for facilitating the development and organization of MTAs thesis, Interactive Systems Research Group, Human-Comp. Interaction
and, in particular, for delivering them using electronic course Lab., School of Computing and Mathematics, University of Teesside,
delivery and electronic open-access student information ser- Middlesbrough, U.K.
vice. Chinowsky, P. S. (1997). ‘‘Introducing multimedia cases into construction
Based upon the principles of DPSS and their implementa- education.’’ Proc., 4th Congr. held in conjunction with A/E/C Systems
tion into mainstream education, this paper discussed the ra- ’97, Computing in Civ. Engrg., Teresa M. Adams, ed., ASCE, New
York, 122–128.
tionale for using such systems within the CAL-Visual project. Collis, B., and Verwijs, C. (1995). ‘‘A human approach to electronic per-
The main aim of this project is to facilitate the use of multi- formance support systems: Hybrid EPSS.’’ British J. Educational Tech-
media in the teaching, learning, and assessment in the field of nol., Oxford, U.K., 35(1), 5–21.
the built environment. A DPSS-based methodology has been Echeverry, D. (1996). ‘‘Multimedia-based instruction of building con-
used within the CAL-Visual framework. This includes a Web- struction.’’ Proc., 3rd Congr. held in conjunction with A/E/C Systems
based generic database of visual aids and a search engine aug- ’96, Computing in Civ. Engrg., Jorge Vanegas and Paul Chinowsky,
eds., ASCE, New York, 972–977.
mented by specialized prototypes dealing with specific topics Finkelstein, J. (1998). ‘‘$D multimedia datascapes: The SSV program.’’
such as building construction and building defects. The system Proc., Int. Conf. on Information Visualisation, IEEE Computer Society
uses different interfaces to access information ranging from a Press, New York, 280–285.
textual-based keyword search to VR models using hot spots Gery, G. J. (1991). Electronic performance support systems—How and
giving access to information in the database. why to remake the workplace through the strategic application of tech-
The project is nearing its completion of the design and im- nology. Weingarten Publications, Boston, Mass.
Hudson, S. R. G. (1998). ‘‘Multimedia performance support systems,’’
plementation phase where a Web-based approach was used to PhD thesis (draft), School of Computing and Mathematics, University
make the system widely available, accessible, and platform- of Teesside, Middlesbrough, U.K.
independent. The main issues related to the visual material Irving, R. H., Higgins, C. A., and Safayeni, F. R. (1986). ‘‘Computerised
included the capturing of textual information to be attached to performance monitoring systems: Use and abuse.’’ Communications of
the images, for this purpose a Microsoft Access form was de- the ACM, 29(8), 794–801.
McGraw, K. L. (1994). ‘‘Performance support systems: Integrating AI,
veloped for the structuring of this information and including
hypermedia and CBT to enhance user performance.’’ J. Artificial In-
it as metadata with the images. telligence in Educ., 5(1), 3–26.
Initial user’s evaluation of the system and trials with stu- Raybould, B. (1995). ‘‘Performance support engineering: An emerging
dents have been conducted to test the usability and perfor- development methodology for enabling organisational learning.’’ Perf.
mance of the system. The final stage of the work will include Improvement Quarterly, Spec. Issue on Electronic Perf. Support Sys.,
a detailed evaluation to measure the effectiveness of embed- 8(1), 7–22.
Riley, D., and Pace, C. (1997). ‘‘Improving teaching efficiency with mul-
ding visual material within a construction-engineering curric- timedia teaching aids.’’ Proc., 4th Congr. held in conjunction with A/
ulum using the CAL-Visual system. E/C Systems ’97, Computing in Civ. Engrg., Teresa M. Adams, ASCE,
New York, 129–135.
APPENDIX. REFERENCES Ring, G. (1994). ‘‘Creating electronic performance support systems for
Australian graduate students in computer education.’’ Proc., Annu.
Aminmansour, A. (1996). ‘‘Utilizing information technologies to better Meeting of the Am. Educational Res. Assn.
educate engineers of tomorrow.’’ Proc., 3rd Congr. held in conjunction Ruggles, C., Underwood, J., and Walker, D. (1995). ‘‘STILE: A hyper-
with A/E/C Systems ’96, Computing in Civ. Engrg., Jorge Vanegas and media approach to the provision of a flexible interdisciplinary resource
Paul Chinowsky, eds., ASCE, New York, 965–971. base for teaching and learning.’’ Innovations in Educ. and Training
Banerji, A. K. (1995). ‘‘Designing electronic performance support sys- Int., London, 32(1), 209–219.
tems,’’ PhD thesis, HCI Lab., University of Teesside, Middlesbrough, Varnadoe, S., and Barron, A. (1994). ‘‘Designing electronic performance
U.K. support systems.’’ J. Interactive Instruction Devel., 6(3), 12–17.
ABSTRACT: The reliability of the techniques adopted for damage assessment is important for bridge manage-
ment systems. It is widely recognized that the use of expert systems for bridge damage assessment is a promising
direction toward bridge management systems. However, several important issues need to be addressed, such as
the management of uncertainty and imprecision, the efficiency of fuzzy rule based reasoning, and the need of
an explanation facility to increase confidence about the assessment results. To address the issues arising from
using expert systems, this paper is aimed at developing an expert system for assessing bridges based on an
expert system shell, which is called the fuzzy Petri net based expert system (FPNES). Major features of FPNES
include the ability to reason using uncertain and imprecise information, knowledge representation through the
use of hierarchical fuzzy Petri nets, a reasoning mechanism based on fuzzy Petri nets, and an explanation of
the reasoning process through the use of hierarchical fuzzy Petri nets. Therefore, this expert system for assessing
bridges does not impose any restriction on the inference mechanism. Furthermore, this approach offers more
informative results than other systems. An application to the damage assessment of the Da-Shi bridge in Taiwan
is used as an illustrative example of FPNES.
where ri and q are of the forms of ‘‘Xi is Fi’’ and ‘‘Y is G,’’
respectively. In Fig. 1, after firing the inference transition t i1,
the tokens will be removed from the input places of t il, a new
token will be deposited into the output place of t i1, and the
fuzzy set and the fuzzy truth value attached to the new token
are derived by three steps: transformation, inference, and com-
position.
Type 2: Aggregation Transition (t a ). An aggregation
transition is used to aggregate the conclusions of several truth-
qualified fuzzy rules that have the same linguistic variable, and
to link the antecedent of a truth-qualified fuzzy rule that also
has the same linguistic variable. For example, there are m
truth-qualified fuzzy rules having the same linguistic variable
in the conclusions, denoted as
(r1 → q11, 1), (r2 → q12, 2), ⭈ ⭈ ⭈ , (rm → q1m, m) (5)
FIG. 1. Modeling Fuzzy Rule Based Reasoning through Fuzzy FIG. 3. Modeling Duplication of Uncertain Fuzzy Token
Petri Nets (a) before Firing Inference Transition t i1 ; (b) after Fir- through Fuzzy Petri Nets (a) before Firing Duplication Transition
ing t i1 t d1 ; (b) after Firing t d1
FIG. 5. (a) Hierarchical Fuzzy Petri Net; (b) Place Status for Pl in Ho
The whole rule base is subdivided into a lot of modules in kens were transformed into these nets in order to fire transi-
order to overcome the complexity arising from too large sizes tions and perform the reasoning mechanism. The results of
of the rule bases and to make the rules well organized. At damage assessment using FPNES for the Da-Shi bridge were
present there are 34 modules in this system, in which 216 expressed in a hierarchical fashion to serve as an explanation
truth-qualified fuzzy rules and 224 recommendations are in- mechanism to facilitate the retrieval of detailed information on
cluded. Seven modules related to the cracking in concrete, five damaged components from the top down to lower levels (Fig.
modules dealing with the delamination in concrete, five mod- 10). The recommendations embedded in rule bases were also
ules concerning the spalling in concrete, two modules about provided on an if-needed basis. As a result, the damage to the
the movement of foundation, and two modules handling to the Da-Shi bridge was evaluated to be severe with strong confi-
scour of support are constructed. Furthermore, two modules dence (i.e., very true) since the overall damage to the super-
are used to evaluate the insufficiency of shear or flexure re- structure was severe with common confidence (i.e., true) and
sistance, and 11 modules are used to assess the damage levels the overall damage to the substructure was also severe with
of components; 216 truth-qualified fuzzy rules and 224 rec-
strong confidence (i.e., very true). The severity of superstruc-
ommendations are accommodated in these modules. As was
ture results from the overall damage to decks was fairly slight
mentioned previously, fuzzy Petri nets’ graphical representa-
tion can help us construct and modify fuzzy rule bases. Fig. with fair confidence (i.e., fairly true), the overall damage to I-
9 shows the fuzzy Petri nets after the transformation of the girders was severe with common confidence (i.e., true), and
rule bases of a crack in an I-girder. the overall damage to diaphragms was very slight with strong
confidence (i.e., very true). Meanwhile, the severity of sub-
EXAMPLE structure as derived from the overall damage of piers was se-
vere with strong confidence (i.e., very true). The FPNES re-
To demonstrate the use of FPNES, the Da-Shi bridge lo- sults not only match the experts’ judgments, but also provide
cated in northern Taiwan was considered. It was rebuilt in more information than experts do, because the explanation
1960 as a simply supported, 12-spanned bridge, 550 m long provided in the system and the confidence level associated
and 7.8 m wide, across the Da-Han river. This bridge consists with the conclusions can be used to justify whether to take the
of 12 decks, 36 prestressed concrete I-girders, 120 diaphragms, recommendations into account.
11 piers, and two abutments. The following recommendations were made with strong
In 1997, this bridge was inspected by the Center of Bridge
confidence (i.e., very true). The observed defects or deficien-
Engineering Research at National Central University. Visual
cies may affect the load-carrying capacity of the bridge. For
inspection revealed that many minor cracks accompanied with
efflorescence spread over eight panels within deck 7. The PCI the safety of the traveling public, traffic on the bridge should
girders S9G1, S9G2, S9G3 in span 9 and S10G1, S10G2, be limited. A second inspection and load capacity evaluation
S10G3 in span 10 had severe flexure, shear cracks, and some should be made. The second inspection is used to identify the
spalls. There were two diaphragms where several spalls were compressive strength of concrete, the carbonation depth in
found. The foundations of piers P1 to P6 were exposed up to concrete, the content of chloride in concrete, the corrosion of
5 m above the ground level, and some of them suffered severe reinforcement, the crack depth in concrete, etc. The load ca-
spalling. Furthermore, pier 2 had rotational movement. pacity evaluation is performed to ascertain the safety load ca-
After executing FPNES for damage assessment of the Da- pacity for the current condition of the bridge. Based on the
Shi bridge, the hierarchical fuzzy Petri nets were constructed results of the load capacity evaluation, a plan can be made for
based on the modularized rule bases, and uncertain fuzzy to- rehabilitating, strengthening, or replacing the bridge.
JOURNAL OF COMPUTING IN CIVIL ENGINEERING / APRIL 2000 / 147
FIG. 10. Results of Damage Assessment for Da-Shi Bridge Using FPNES
TECHNICAL PAPERS
A Representation for Conceptual Design of Buildings
Hugues Rivard and Steven J. Fenves. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 151
Application Framework for Development of Simulation Tools
Dany Hajjar and Simaan M. AbouRizk. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 160
Evaluation of Mobile Mapping Systems for Roadway Data Collection
Hassan A. Karimi, Aemal J. Khattak, and Joseph E. Hummer . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 168
Structural Matrix Computations with Units
Mark A. Austin, Wane-Jang Lin, and Xiaoguang G. Chen. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 174
Forecasting Cyanobacterial Concentrations Using B-Spline Networks
Holger R. Maier, Tarek Sayed, and Barbara J. Lence . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 183
System Dynamics Modeling of Reservoir Operations for Flood Management
Sajjad Ahmad and Slobodan P. Simonovic . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 190
Distributed Genetic Algorithm Model on Network of Personal Computers
M. C. Balla and S. Lingireddy . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 199
Computerized System for Application of Value Engineering Methodology
Sadi Assaf, Osama A. Jannadi, and Ahmed Al-Tamimi . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 206
A REPRESENTATION FOR CONCEPTUAL DESIGN OF BUILDINGS
By Hugues Rivard,1 Associate Member, ASCE, and
Steven J. Fenves,2 Honorary Member, ASCE
ABSTRACT: The building representation for storing design solutions adopted in SEED-Config consists of two
levels of abstraction models defined on top of the object-oriented data model. The first level is an information
model, called the building entity and technology model, which stores design data as they are generated during
conceptual design, supports case-based reasoning, and shares data among all design participants. This model
represents each building entity as a generic container that encompasses its properties, taxonomy, geometry,
composition, relationships, and design knowledge applied in generating it. The second level is a conceptual
model that defines the types of objects, relationships, and data needed to fully represent the information in a
given design domain. The conceptual model specifies the semantics of the design information for the domain
using the syntax defined in the information model. The proposed representation also includes a faceted classi-
fication scheme to define the controlled vocabulary from which indexes are obtained. The representation has the
following advantages: It integrates multiple views, supports design evolution, supports design exploration, and
is extensible.
terior. The AND arc in the figure indicates that all four systems allows each aspect to be designed and analyzed separately.
are necessary. Each system is, in itself, a hierarchical decom- These two subsystems are frequently addressed together by
position of entities organized to address its specific purpose one structural system, but sometimes they may be distinct, and
and provides a unique view of the overall building. The struc- the conceptual model should be able to support both situations.
tural system is the root of all building entities that have struc- For instance, a low-rise industrial building may be designed
tural functions in a building project. It is the repository for with columns for the vertical gravity subsystem and a few
specifications, functional requirements, and constraints that af- braced frames for the vertical lateral subsystem, or it may be
fect the overall structural design. A hierarchical decomposition designed with moment resisting frames that address the func-
for the enclosure system has been proposed in Rivard et al. tion of both subsystems.
(1999). Each of the four structural subsystems consists of a set of
Fig. 3 shows six of the seven levels identified in a structural structural assemblies, which represent the next level of the
system. At the structural massing level, below the systems hierarchy. Structural assemblies are collections of basic ele-
level, the structure of a building can be broken down into ments that represent meaningful concepts in structural engi-
independent structural volumes depending on the building’s neering and that are frequently analyzed independently from
size and complexity. An independent structural volume models the rest of the structure. The decomposition shows seven struc-
a self-contained structural skeleton. Long or irregularly shaped tural assemblies: The roof, which covers the building; the
buildings may have several independent structural volumes floor, which divides a building into stories; the column stack,
that are delimited by expansion joints. An independent struc- which encompasses every aligned column element from the
tural volume may be composed of one or more functional foundation to the roof; the 3D system, which defines a core
spaces. A functional space corresponds to a portion of the vol- or tubelike structure whose primary function is to resist lateral
ume that has a specific use or occupancy that may entail spe- loading by behaving as a 3D cantilever; the frame, which is a
cific structural requirements (e.g., swimming pool or office structural assembly of linear elements connected (simply or
space with minimum column spacing). rigidly) at a finite number of discrete points called nodes; the
The next level of the hierarchy defines the structural sub- truss, which is a structural assembly of linear elements where
systems. An independent structural volume has four essential each element is subject to axial forces only; and the wall,
functional subsystems—the foundation, the horizontal, the which is a vertical plane that may extend beyond several floors
vertical gravity, and the vertical lateral subsystems. The AND and that may resist shear and carry vertical loads. A structural
arc again indicates that all four are necessary. Each subsystem assembly can be made of other structural assemblies; for in-
deals with a specific structural function: The horizontal sub- stance, a roof can be made of trusses, or a core can be made
system is responsible for transferring live, snow, and dead of walls.
loads to the vertical gravity subsystem; the vertical gravity Structural assemblies are decomposed into basic elements,
subsystem is responsible for transferring vertical loads to the which are at the sixth level. The basic elements represent the
foundation subsystem; the vertical lateral subsystem is respon- smallest construction blocks (at the conceptual design stage)
sible for transferring horizontal loads, such as wind and earth- used to assemble the structural system. These elements must
quake loads, to the foundation subsystem; and the foundation be nonintersecting geometrically so that they can be assembled
subsystem is responsible for transferring all loads of the struc- into complex structural systems (Dias 1996). The hierarchical
ture to the ground. The foundation subsystem is not elaborated decomposition has four basic elements; the slab element,
further here. The separation of the two vertical subsystems which is a planar element that makes up a roof or a floor; the
156 / JOURNAL OF COMPUTING IN CIVIL ENGINEERING / JULY 2000
wall element, which is a vertical planar element that makes up identified in the first step. In the BENT model, building enti-
a wall; the linear element, which is a 1D element that has a ties are described with classifiers, DSRs, and attribute-value
length much greater than its cross-sectional dimensions and pairs organized into components. A conceptual model defines
may be used to represent a column, a beam, or a bar of a truss; a specialization of these constructs. Table 1 provides for a
and the node, which is a zero-dimensional element that has number of structural entities a tentative list of components and
position, but no size or form (Meyer 1995). categories of classifiers organized in terms of functional and
A seventh level, not shown in the figure, comprises con- design unit. The behavioral aspect of structures has not been
stituents. A basic element may be broken down into constitu- addressed yet. Items in bold are classifier categories, and plain
ents such as webs and flanges, or concrete and reinforcement. items are components. The table shows that most of the func-
This last level is not discussed further here because it is too tional requirements are defined at the highest level of the hi-
detailed for conceptual structural design. erarchical decomposition, whereas design specifications appear
The hierarchical decomposition of the structural system fol- at the more detailed levels. Note that dimensions, locations,
lows the total-system approach promoted by Lin and Stotes- shapes, as well as topological relations, which are important
bury (1981), where the designer focuses first on the 3D im- design information, are not shown in the table, because they
plications of architectural space-form options, then on the are obtained either from the primary spatial representation or
more-or-less planar subsystems that make up the structure, and from the geometric modeler.
finally on the elaboration and refinement of individual ele- The design unit of the linear element is described here to
ments and connection details. First, 3D structural systems are illustrate the results of the second step. The building entity
decomposed into 2D structural elements such as frames, walls, type ‘‘linear element’’ has one category and three components
and slabs. Then, 2D structural elements are further decom- associated with its design unit. The category ‘‘type of linear
posed into 1D elements or 2D subelements (e.g., beams, col- element’’ classifies the element as a beam, a column, or a
umns, and slab elements). Hence, a complex structural prob- tension member. Each of these classifiers is further specialized
lem is decomposed into simpler subproblems that can be into a hierarchy of classifiers similar to those shown in Fig. 2
considered in a semi-independent fashion. This approach fol- (e.g., a beam is specialized into a joist, a girder, or a girt). The
lows the design philosophy that the whole of a design scheme three design unit components associated with a ‘‘linear ele-
should give rise to the need for details and not vice versa (Lin ment’’ are the end conditions, which defines whether the ends
and Stotesbury 1981). of the linear element are fixed or free to rotate; the material
The hierarchical decomposition can serve as an abstraction properties, which defines the characteristics of the material that
mechanism to present different levels of detail. The designer makes up the linear element; and the section properties, which
can browse through the graph and examine any level of ab-
straction simply by going to a given depth in the graph. Typ- TABLE 1. Structural Entities Described by Components and
ically, the higher levels of the hierarchy are especially useful Categories of Classifiers
in early design stages for defining functional requirements and
for making broad design decisions, whereas the lower levels Building entity FU DU
(1) (2) (3)
are more tangible and often contain more information that is
generated in the detailed design phase (Bjork 1989). The Structural system Seismic performance cate- Main material type
gory
BENT model supports the design process as it unfolds by al- Wind load category
lowing designers to populate the design space in an intuitive Ground snow load category
manner from global system entities through subsystem entities, Soil profile type
Erection time requirement
all the way down to constituent entities. As the design pro- Budget construction cost
ceeds, the description of the design problem becomes increas- Maintenance costs
ingly detailed. Independent structural Most restrictive use/occu- Structural grid
volume pancy
The hierarchical decomposition for a design domain is kept Fire-resistance ratings
in a classification where it is represented by building entity Geometric irregularities
types. An actual building project is modeled with building en- Load combinations
tities linked by containment relationships. Each generic build- Functional spaces Use or occupancy
Minimum column spacing
ing entity is categorized by a classifier instance that corre- Linear element Type of structural action Type of linear element
sponds to one of the building entity types in the hierarchical Distributed load End conditions
decomposition. Fig. 3 shows more building entity types than Material properties
Section properties
may be needed in an actual project. A typical building project Node Point load Type of connection
would be assigned all building entity types defined at the first Restraints
four levels of the hierarchy and possibly only a subset of the
building entity types at the structural assembly and basic ele-
ment levels.
Although the hierarchical decomposition in Fig. 3 represents
an attempt to define all the types of building entities required
in the conceptual structural design stage, some types of entities
may have been overlooked or some designers may want to
include additional types of entities. The BENT model allows
hierarchical decompositions to be extended and modified by
designers. Hence, additional building entity types can be in-
corporated within the hierarchical decomposition as new ones
are identified, or a different hierarchical decomposition can be
modeled altogether.
Components, Classifiers, and Relationships in
Structural Engineering
The second step in elaborating the conceptual model is to FIG. 4. Definition of Three Design Unit Components for Linear
define the type of data required by each building entity type Element
ABSTRACT: Development of specialized construction simulation tools is a time-consuming and complex op-
eration. This has hindered the effective and wide use of simulation in the construction industry. An object-
oriented application framework is a promising software engineering principle, which can help in overcoming
this obstacle. Application frameworks encapsulate reusable code as well as common design patterns for a specific
class of applications so that development of new tools is greatly simplified. This paper describes this new
principle and applies it to the development of a construction simulation object library. A case study performed
to compare the efforts between framework- and nonframework-based simulation tool development revealed that
productivity can be dramatically improved.
longer had to write code directly. Graphical modeling made it CONSTRUCTION SIMULATION
possible to define the simulation model by creating, manipu-
lating, and linking a number of available basic building blocks. Halpin (1977) popularized the use of simulation in construc-
This meant that users no longer had to be proficient in pro- tion research with his development of a system called CY-
gramming. A detailed account of the history of simulation con- CLONE (CYCLic Operation NEtwork). CYCLONE allowed
cepts and systems is detailed in Kreutzer (1986). the user to build models using a set of abstract but simple
There are many ways of modeling a given problem. These constructs. The system became the basis for a wide range of
generally fall into two categories: continuous and discrete- construction simulation research efforts with the objective of
event. With continuous modeling, differential equations are enhancing the basic system functionality. Although CY-
used to describe the progress of a given activity. However, CLONE and its derivatives introduced a wider academic au-
when mathematical modeling is not possible, the discrete- dience to computer simulation, its use in the industry was very
event approach is utilized. limited. This was because CYCLONE proved practical for
Discrete-event simulation views a model as a set of events small-scale applications. Furthermore, the modeling process
and transitions. Transformations are processed as part of event required simulation training—something that the industry was
handlers, and no relevant transformations are assumed to occur not generally ready to invest in. As a result, research turned
during transitions. Entities represent the active elements of the to other concepts that could address these issues and simplify
model. They travel throughout the event network and trigger the modeling process even further.
transformations. An important component of a discrete-event AbouRizk and Hajjar (1998) found that effective transfer of
simulation is the simulation event monitor, which is respon- computer simulation knowledge to the construction industry is
sible for managing the event calendar. The described concept best done through specialization and customization of the
is now illustrated below with a simple example. modeling, analysis, and reporting components of the simula-
Assume that trucks arrive at a concrete batch plant and await tion systems. This approach, known as special purpose simu-
their turn for loading fresh concrete from a mixer. When they lation (SPS), was followed in the successful development and
finish loading, they proceed to their destination. The described implementation of three custom simulation tools. For the mod-
model is illustrated using an event diagram as shown in eling environment, this means that the modeling elements that
Fig. 1. are used to construct simulation models are tailored to the
The ‘‘truck arrival’’ event is first added to the event calendar specific target domain.
during the initialization of the simulation. When the event is The first tool, called AP2-Earth (Hajjar and AbouRizk
processed as part of the event handler, the arrival of the next 1996), allows for the analysis of large earthmoving projects.
truck entity is first scheduled to occur. Then the status of the AP2-Earth was developed with the objective of providing
mixer is checked. If it is busy, then the truck entity is added earthmoving contractors with a system that can automate the
to a queue. Otherwise, the ‘‘truck departure’’ event is sched- traditionally time-consuming and manual process of preparing
uled. During the processing of the Truck Departure event, earthmoving estimates for large and complex projects. The
which occurs when a loading operation has been completed, main interface form shown in Fig. 2 is where the simulation
the status of the waiting queue is checked. If there are any models can be built. Standard earthmoving modeling elements
waiting truck entities, then the first one is removed from the such as sources, trucks, placements, and road segments are
queue so that it can start loading. The transition duration be- used to build general earthmoving scenarios.
tween each truck arrival event is based on the desired inter- The second tool, called CRUISER (Hajjar and AbouRizk
arrival time of trucks. The transition duration between the 1998), can be used for modeling aggregate production plants.
truck arrival and truck departure events depends on the desired It allows users to define the plant layout by creating the crush-
truck loading time. ing components and visually connecting them with conveyers
JOURNAL OF COMPUTING IN CIVIL ENGINEERING / JULY 2000 / 161
FIG. 2. Sample Simulation Model in AP2-Earth
as shown in Fig. 3. The objective of CRUISER is to provide viewed in the form of 2D cross-sectional graphs or 3D surface
managers with a tool to experiment with several possible al- charts.
ternatives for plant design. With this program, ‘‘what-if’’ anal- Although each tool was successfully implemented, it was
ysis can be performed very easily by changing the desired observed that the relatively large initial investment required
parameters and resimulating. This allows plant managers to for their development would hinder the application of the SPS-
test the effect of changing screen sizes, using an alternative based approach to other construction operations. As a result,
pit and countless other design alternatives, by simply changing research efforts were shifted to the identification of software
the CRUISER model and resimulating. The application takes engineering principles that can lead to improvements in the
care of all the tedious analysis that would otherwise be done actual development process of such tools.
by hand.
The third tool was CSD (Hajjar et al. 1998), which allows OBJECT-ORIENTED APPLICATION FRAMEWORKS
for the optimization of construction site dewatering operations.
Overview
The visual modeling environment allows the user to enter var-
ious site data such as the site dimensions, layers, and other Object-oriented application frameworks encapsulate a given
properties. Pumping wells and watch wells are added using design philosophy and a set of components that can be used
simple click and drag mouse operations. Watch wells represent to develop specific applications for a given domain. This
the desired drawdown at various points in the site. The sim- means that overall development of systems becomes a two-
ulation engine of CSD can then be initiated to predict the step approach. Numerous expert developers, familiar with a
resultant water table level throughout the site. Results are specific application domain, first develop the application
162 / JOURNAL OF COMPUTING IN CIVIL ENGINEERING / JULY 2000
framework. Examples of specific application domains are cost is typically an object-oriented library that encapsulates a basic
estimating systems, scheduling systems, and construction sim- tool architecture as well as various specialized components.
ulation systems. The developed application framework cap- These components, also referred to as services, are grouped
tures the broad and recurring business processes, transactions, by their function and represent the basic infrastructure of the
and transformations within the target domain. Once the appli- framework.
cation framework is developed, specific tool implementations Once the framework implementation is complete, it is tested
within the domain can then be developed. For example, a spe- through the implementation of several basic tools. Ideally,
cific implementation based on an estimating framework could some of the tools originally used for the variance analysis are
include an earthworks estimating systems, a bridge estimating reimplemented. This proves the initial practicality and expres-
system, or a multistory commercial building estimating sys- siveness of the developed application framework.
tem. Several issues are related to the deployment of application
Understanding of application frameworks has greatly ad- frameworks after their implementation and testing is com-
vanced in recent years. The following subsection will provide pleted. First, the framework needs to be introduced in a careful
an overview of the basic concepts involved in framework de- manner, so that developers can gradually learn its basic re-
sign, implementation, documentation, and deployment. The in- quirements, features, and components. This process can in-
formation provided is based on application framework design clude rollout sessions, example applications, extensive docu-
and development guidelines as outlined by Froehlich et al. mentation, and tutorials. Second, the distribution means of the
(1998). framework must be addressed. Possible options include source
code or binary code distributions. Third, a process must be
Framework Development established to capture important feedback from the tool de-
The first step in the design of an application framework is velopers. This includes general support information, bug re-
the determination of its intended domain. This helps in defin- ports, and enhancement requests.
ing the scope of the framework and the basic services that it
should provide. Variability analysis is a technique that can be
used to assist in this process. This involves assessing various CONSTRUCTION SIMULATION APPLICATION
components of applications across a number of scenarios and FRAMEWORK
identifying both variable and common features. Scenarios may
include general descriptions, formal process models, or, ide- Overview
ally, actual tools developed for the intended domain. Variabil-
ity analysis results in the identification of the recurring design A framework approach is followed in the development of
patterns and common features. an object-oriented library that can be used for the development
Common features translate into what is referred to as the of individual construction simulation tools. The design of this
‘‘frozen spots’’ of the framework. Frozen spots are framework library is based on a detailed analysis of the three stand-alone
components that tool developers have little or no control over. simulation tools previously described. This variability analysis
Distinct or variable features translate into what is referred to reveals that a large proportion of features were identical in all
as the ‘‘hot spots’’ of the framework. Hot spots are areas in tools; in fact, code had been reused for these features. Fur-
the framework that tool developers can customize for specific thermore, of all the remaining features, the majority do not
tool requirements. vary significantly, and it was possible to generalize their func-
Once the design of the framework is complete, the imple- tionality to standardize their implementation. These common
mentation can begin. Often, the process is iterative rather than features translate into the frozen spots of the framework. The
linear in that implementation may require the redesign of cer- remaining features are genuinely unique to the specific tools.
tain components, as the domain becomes more understood. They are used in the identification of the hot spots of the
During this process, the means by which developers will be framework. The following two subsections will discuss the
customizing the framework are also determined. The end result common and variable features in detail.
CRUISER Redevelopment
In a similar manner, a SOM-based tool, denoted CEM㛭
FIG. 6. Simplified Representation of Simphony Application CRUSH, was developed with the same functionality as
Framework CRUISER. Fourteen modeling elements were developed to al-
ABSTRACT: Mobile mapping systems (MMSs) are based on advanced technologies for automatically collecting
roadway inventory data. A host of advanced technologies, such as the Global Positioning System for collecting
georeferencing data and digital cameras for collecting roadway data, are used in current MMSs. An MMS is
driven on a subject roadway, collecting positional data and digital images of the roadway. Through photogram-
metric software packages, roadway data are extracted from images. Although there have been studies evaluating
the accuracy of positional data collected by MMSs, there have been limited studies of the descriptive data
accuracy they provide. In this paper the results of an evaluation of the accuracy of descriptive inventory data
collected by three different MMSs are discussed. Each system was tested in three different roadway environ-
ments, and five different types of inventory elements were included in each test. The accuracy of collected data
was compared with ground truth data. Analysis of these data indicated that the roadway environment and the
type of inventory element affect the accuracy of descriptive inventory data collected by MMSs.
Experiment
(MMS GPS Digital camera Photogrammetric
number) receiver (pixel resolution) INS DMI software package
(1) (2) (3) (4) (5) (6)
1 Trimble 7400 Pulnix a (768 ⫻ 484) Littona Daytrona Feature Extraction
2 Ashtech Z-12 COHO 4980 RS-170 (640 ⫻ 480) Honeywell Laser Ref III Vehicle ABS VISAT Station
3 Novatel a Sony XC-007 (640 ⫻ 480) INS with 3 gyros and 2 accelerometers a Vehicle ABS RoadView Ruler
a
Indicates exact make or model is proprietary information.
Inventory element type Urban street Two-lane rural Rural interstate Overall
(1) (2) (3) (4) (5)
Lane width ⫺1.32 (7.18) 0.75 (6.37) ⫺0.30 (6.64) ⫺0.31 (6.66)
Traffic sign width 16.87 (12.98) 38.58 (40.92) 8.49 (13.25) 18.98 (26.03)
Traffic sign lateral placement 38.18 (74.80) 10.35 (14.03) 6.97 (8.81) 16.01 (40.78)
Traffic sign height 2.44 (6.20) ⫺1.51 (4.34) ⫺4.95 (10.85) ⫺1.53 (8.06)
Traffic sign support height ⫺2.18 (6.57) ⫺1.95 (7.83) 8.86 (53.96) 2.98 (37.15)
Overall 10.41 (41.12) 6.12 (18.86) 4.65 (29.12) 6.61 (30.87)
Note: Standard deviations are in parentheses.
TABLE 7. Experiment 2 PME Mean and Standard Deviations TABLE 9. Experiment 3 PME Mean and Standard Deviations
Mean Mean
Inventory element type (SD) Inventory element type (SD)
(1) (2) (1) (2)
Lane width ⫺1.62 (8.31) Lane width ⫺0.89 (2.84)
Traffic sign width 4.87 (15.36) Traffic sign width ⫺1.74 (39.75)
Traffic sign lateral placement 13.04 (21.68) Traffic sign lateral placement 3.86 (21.59)
Traffic sign height 9.71 (21.35) Traffic sign height ⫺10.71 (23.61)
Traffic sign support height 2.35 (17.37) Traffic sign support height ⫺5.20 (18.90)
Overall 5.25 (17.91) Overall ⫺1.90 (21.20)
ABSTRACT: Despite the well-known benefits of physical units, matrices, and matrix algebra in engineering
computations, most engineering analysis packages are essentially dimensionless. They simply hold the engineer
responsible for selecting a set of engineering units and making sure their use is consistent. Although this practice
may be satisfactory for the solution of established problem-solving procedures, identifying and correcting un-
intentional errors in computations developed by teams of engineers remains a difficult problem unless units are
an integral part of the computational procedure. This paper begins with a description of the data structures and
algorithms needed to represent and manipulate physical quantity variables and matrices of physical quantities.
The result is a high-level scripting language that offers enhanced type checking for expressions and assignments,
automatic conversion of systems of units, and program control structures for the solution of engineering prob-
lems. These ideas have been implemented in Aladdin, a new computational environment for matrix and finite-
element calculations. The capabilities of Aladdin are demonstrated through the deflection analysis of a cantilever
beam.
PHYSICAL QUANTITIES
shows how physical quantities can be assigned to identifiers.
Aladdin supports three data types: ‘‘character string’’ for Fig. 3 shows the physical quantity content settings for each
variable names, physical quantities, and matrices of physical variable. Variable t represents three seconds. The numerical
quantities for engineering data. A physical quantity is a mea- value for this quantity is 3 and the units name is ‘‘sec.’’ Al-
sure of some quantifiable aspect of the modeled world. In though variables x and y both represent distances of two me-
Aladdin, basic engineering quantities, such as length, mass, ters and have a numerical value of 2, their units names are m
and force, are defined by a numerical value (number itself) and cm, respectively. When Aladdin prints a quantity, the out-
plus physical units. Fig. 2 is a subset of units presented in the put numerical value is the quantity numerical value divided
Unit Conversion Guide (Unit 1990) and shows the primary by the scale factor associated with the units name. Therefore,
base units, supplementary units, and derived units that occur the statements
in structural analysis. The four basic units needed for engi-
neering analysis are length unit L; mass unit M; time unit t;
and temperature unit T. Planar angles are represented by the
supplementary base unit rad.
Derived units are expressed algebraically in terms of base will generate the output
and supplementary units by means of multiplication and di-
vision, namely
units = k ⭈ L␣M t ␥T ␦ ⭈ radε (1) For variable x, the printed value is 2 because 2/1 = 2. For
where ␣, , ␥, ␦, and ε = exponents; and k = scale factor. The variable y, the printed value is 200 because 2/0.01 = 200.
four basic units play the primary role in determining compat- Finally, when Aladdin parses the assignment statement for
ibility of units in physical quantity and matrix operations. Be- gravity, the character sequence ‘‘m/sec2’’ is translated
cause a radian represents the ratio of two distances (i.e., dis- to the set of unit exponents [␣, , ␥, ␦, ε] = [1, 0, ⫺2, 0, 0].
tance around the perimeter of a circle divided by its radius),
most software implementations deal with radians as if they Special Names for Derived Units
were dimensionless entities. Aladdin departs from this trend
by explicitly representing radians and employing a special set A number of derived units have special names that may
of rules for their manipulation during physical quantity and themselves be used to express other derived units in a simpler
matrix operations. The result is enhanced readability of pro- way than in base units. For example, Aladdin uses the symbol
gram output, especially for computations involving displace- N to represent one newton (i.e., [␣, , ␥, ␦, ε] = [1, 1, ⫺2,
ments and rotations of structural components. 0, 0]), and so the force
In addition to a set of units defined by (1), each physical
quantity contains a character string storing the unit’s name and
a flag for representation of various systems of units (e.g., SI, can be simply written 25 N (Fig. 2).
US, and combined SI㛭US). SI㛭US units are those shared by
both US and SI systems, the most important examples being Simplification and Casting of Physical
time, frequency, and planar angle. Quantity Output
The Aladdin units package provides facilities for dynamic
allocation of units, units copying, consistency, checking and Because Aladdin stores all quantities internally with SI
simplification, and units printing. Operations for units conver- units, quantities specified in US units must be converted into
sion are provided, as are facilities to turn units operations on/ their SI counterparts before they can be stored and used in
off. In an effort to keep the scripting language usage and im- calculations. Fortunately, for most sets of units (the important
plementation as simple as possible, all physical quantities are exception being temperature units), only a single scale factor
stored as floating point numbers with double precision accu- is needed to convert a quantity from US into SI and vice versa
racy plus units. Floating point numbers are viewed as physical (e.g., 1 in = 25.4 mm = 25.4E-3 m). Aladdin automates con-
quantities without units (i.e., [␣, , ␥, ␦, ε] = [0, 0, 0, 0, 0]). version between different, but equivalent, units and provides
There are no integer data types in Aladdin. features in its language for physical quantities to be cast and
displayed in a desirable set of units. For example, with the
Physical Quantity and Variable Declarations definition of gravity in place, the statements
Aladdin stores physical quantities as ‘‘unscaled’’ values with
respect to a set of SI units—meter ‘‘m’’ is the reference for
length, kilogram ‘‘kg’’ for mass, second ‘‘sec’’ for time, and
‘‘deg㛭C’’ for temperature. For example, the Aladdin statement generates the output
Both sets of units have the same dimension (i.e., length di- Physical quantity expressions are evaluated in two steps. In
vided by time squared), with the numerical value of the phys- the first step, units within the expression are examined for
ical quantity differing only by a scale factor. The stream of compatibility with respect to the operation to be performed.
characters between (. . .) indicates the desired format for the Incompatible units will trigger the printing of an error message
physical quantity units. and termination of the program execution. Expressions con-
Of course, casting of physical quantity output also works taining compatible units proceed to the second step—evalua-
with special names for derived units. This feature is useful for tion of the numerical expression with appropriate units for the
situations where a particular set of unit exponents has more result.
than one interpretation. For example, the unit exponents [␣, To see how operations of physical quantities works in prac-
, ␥, ␦, ε] = [2, 1, ⫺2, 0, 0] can be interpreted as a moment tice, let q1 be a physical quantity with unit scale factor k1 and
(force times distance) and as work done (force times distance). unit exponents [␣1, 1, ␥1, ␦1, ε1]. Let q2 be a physical quan-
Casting of units allows for these cases to be distinguished, as tity with unit scale factor k2, and unit exponents [␣2, 2, ␥2,
in ␦2, ε2].
1. If the operands all have the same units type (e.g., let us
say all are SI or US units), then the result units will be
PHYSICAL QUANTITY ARITHMETIC set according to the first operand.
2. If the operands have different units type, then the result
Aladdin supports the construction and evaluation of physical units will be set according to the operand, which has the
quantity expressions involving arithmetic, relational, and log- same units type as the environmental units type.
ical operators, thereby enabling a powerful check for the di- 3. When there are two or more operands that have the same
mensional consistency of formulas. Units and physical quan- units type as the environmental units type, the first one
tities are integrated into the Aladdin scripting language via (1). will be the basis for the result units.
JOURNAL OF COMPUTING IN CIVIL ENGINEERING / JULY 2000 / 177
FIG. 3. Data Structure Contents for x, y, t, and gravity
For example, the script of code TABLE 1. Units Arithmetic in Arithmetic Operations
Scale
Description Expression factor Unit exponents
(1) (2) (3) (4)
Addition q1 ⫹ q2 k1 [␣1, 1, ␥1, ␦1, ε1]
Subtraction q1 ⫺ q2 k1 [␣1, 1, ␥1, ␦1, ε1]
Multiplication q1 * q2 k1 ⭈ k2 [␣1 ⫹ ␣2, 1 ⫹ 2, ␥1 ⫹ ␥2, ␦1
generates the output ⫹ ␦2, ε1 ⫹ ε2]
Division q1/q2 k1/k2 [␣1 ⫺ ␣2, 1 ⫺ 2, ␥1 ⫺ ␥2, ␦1
⫺ ␦2, ε1 ⫺ ε2]
Exponential q q
1 2 k 1N a [N␣1, N1, N␥1, N␦1, Nε1]a
a
N is the value of q2.
冘
n
5. Execution of the looping construct begins with the ini-
tializer statement, where we assign 1 cm to variable x.
zij = xik *ykj (2) The condition statement is then evaluated to see if it is
k=1
true or false. In this particular example, x <= 5 cm evaluates
Due to the consistency condition, all terms in the matrix ele- to true and so statements inside the body of the loop are ex-
ment summation must have the same units. That is, the ex- ecuted (i.e., we print out the value of x). When all of the
ponents of the units must satisfy the product constraint aib1c1dj statements inside the for-loop body have finished execut-
JOURNAL OF COMPUTING IN CIVIL ENGINEERING / JULY 2000 / 179
FIG. 6. Single Span Cantilever Beam with Supported Endpoint
ing, the increment statement is executed (i.e., we increase factor corresponding to the beam-end lateral displacements
the value of x by 2 cm) and condition is re-evaluated to and rotations.
see if another loop of computations is required. For this ex-
ample, looping continues for four iterations until the con- Definition of Element Stiffness Matrix
dition evaluates to false. At the end of the for-loop exe- It is well known that if cubic interpolation functions are
cution, the value of x will be 7 cm. You should notice that all used to describe displacements along the beam element, then
冋 册
of the elements in the for-loop statement are dimensionally the beam element stiffness matrix will be
consistent.
⫺3/L
冋 册
2 3/L 1
ALADDIN STACK MACHINE 2EI 3/L 6/L2 3/L ⫺6/L2
k= ⭈
L 1 3/L 2 ⫺3/L
The heart of the Aladdin kernel is a finite-state stack ma- ⫺3/L ⫺6/L2 ⫺3/L 6/L2
chine model that follows in the spirit of work presented by
Kernighan and Pike (1983). Stack machines are suitable for where EI = flexural rigidity of the beam element; and L = its
modeling systems that have a finite number of internal config- length. The abbreviated fragment of Aladdin code
urations or states, and whose behavior evolves as a linear se-
quence of discrete points in time. The stack machine reads
blocks of command statements from either the input file, or
keyboard, and processes them in two phases:
1. High-level input statements are converted into an array defines variables for the beam material and section properties
of equivalent low-level stack machine instructions. and its length. The block of statements
2. The stack machine walks along the array of machine in-
structions and executes the functions pointed to by the
machine instructions.
Destination Array
A straightforward way of assembling these matrices is with Now we can compute the solution to the linear equations
the use of destination arrays, namely stiff*displ = eload, and print the cantilever displace-
ments by simply writing
冋 册冋 册
Element LD[i][j] contains the global degree of freedom [N/rad N/m N/rad N/m N/rad N/m]
for degree of freedom j of beam finite element i. A zero entry
indicates that the beam element degree of freedom is fully displ[1][1] units N* m
fixed and will not be mapped to the global stiffness matrix. displ[2][1] units N
displ[3][1] units N* m
⭈ =
Assembly of Global Stiffness Matrix displ[4][1] units N
displ[5][1] units N* m
The assembly process for the global stiffness matrix may be displ[6][1] units N
冋 册冋册
written k1 ⫹ k2 ⫹ k3, where ki represents the stiffness matrix Equating and rearranging terms in the units equation gives
for beam element i mapped onto the global degrees of free-
dom. The following block of Aladdin code displ[1][1] units rad
displ[2][1] units m
displ[3][1] units rad
=
displ[4][1] units m
displ[5][1] units rad
displ[6][1] units m
In other words, the column units of stiff multiplied by the
row units of displ must match the row units of eload. By
explicitly representing units for the third column as N/rad
(versus N), radian units can be assigned to displ[3][1].
Including radians in the solution vector enhances its readabil-
ity and allows for appropriate conversion among units for pla-
nar angles. The displacements are
ABSTRACT: Artificial neural networks have been used successfully in a number of areas of civil engineering,
including hydrology and water resources engineering. In the vast majority of cases, multilayer perceptrons that
are trained with the back-propagation algorithm are used. One of the major shortcomings of this approach is
that it is difficult to elicit the knowledge about the input/output mapping that is stored in the trained networks.
One way to overcome this problem is to use B-spline associative memory networks (AMNs), because their
connection weights may be interpreted as a set of fuzzy membership functions and hence the relationship between
the model inputs and outputs may be written as a set of fuzzy rules. In this paper, multilayer perceptrons and
AMN models are compared, and their main advantages and disadvantages are discussed. The performance of
both model types is compared in terms of prediction accuracy and model transparency for a particular water
quality case study, the forecasting (4 weeks in advance) of concentrations of the cyanobacterium Anabaena spp.
in the River Murray at Morgan, South Australia. The forecasts obtained using both model types are good. Neither
model clearly outperforms the other, although the forecasts obtained when the B-spline AMN model is used
may be considered slightly better overall. In addition, the B-spline AMN model provides more explicit infor-
mation about the relationship between the model inputs and outputs. The fuzzy rules extracted from the B-spline
AMN model indicate that incidences of Anabaena spp. are likely to occur after the passing of a flood hydrograph
and when water temperatures are high.
INTRODUCTION model developed by Maier et al. (1998) for the same case
study, in terms of prediction accuracy and model transparency.
In recent years, artificial neural networks (Flood and Kartam
1994) have been used successfully in a number of areas of COMPARISON OF MLP WITH B-SPLINE AMNs
civil engineering. In the disciplines of hydrology and water MLPs and AMNs are able to approximate any continuous
resources engineering, they have been applied to the prediction function, given sufficient degrees of freedom (Brown and Har-
and forecasting of rainfall, rainfall-runoff, water levels, and a ris 1994; Hornik et al. 1989). They can be mapped onto three-
number of water quality parameters (Maier and Dandy 2000), layer networks, use feedforward connections, and are trained
as well as the management of water quality systems [e.g., using supervised algorithms. However, there are differences in
Takyi and Lence (1999)]. In the majority of cases, multilayer the structure of the input and hidden layers, resulting in dis-
perceptrons (MLPs) that are trained with the back-propagation tinct modeling capabilities.
algorithm are used (Maier and Dandy 2000). One of the major
shortcomings of this approach is that the knowledge that is Model Structure and Operation
contained in the trained networks is difficult to interpret, be-
The p inputs to MLPs are represented by individual nodes
cause it is distributed across the connection weights in a com-
in the input layer (Fig. 1). The input signals to MLPs are
plex manner. weighted and passed to nodes in one or more hidden layers.
One way to overcome this problem is to use the class of The weighted inputs connected to each hidden layer node are
associative memory networks (AMNs) that have a fuzzy in- summed and passed through a transfer function. The latter is
terpretation and are thus more transparent (Brown and Harris generally a bounded ridge function, such as the logistic or hy-
1995). It has been proven that B-spline AMNs and certain perbolic tangent function. Generally, MLPs are fully connected
types of fuzzy models are learning equivalent, i.e., model co- and their geometry and complexity are defined by the number
efficients are estimated in the same way for both model types of hidden layers and the number of nodes in each of these.
(Brown and Harris 1994). Consequently, B-spline AMNs may In contrast, the input space of AMNs is normalized by a p-
be viewed as a particular type of neurofuzzy model. The ob- dimensional lattice (Fig. 2) (Brown and Harris 1994). Each
jectives of this paper are to: (1) provide a theoretical compar- cell of the lattice represents similar regions of the input space.
ison between AMNs and MLPs; (2) develop a B-spline AMN The geometry of AMNs is limited to one hidden layer. The
model for forecasting (4 weeks in advance) a particular species hidden layer nodes consist of basis functions, which are de-
group of blue-green algae (cyanobacteria) in the River Murray fined on the lattice formed by normalizing the input space.
at Morgan, South Australia; and (3) assess the suitability of The basis functions can take a number of forms, including
the model by comparing its performance with that of the MLP threshold, B-spline, or Gaussian functions. Model structure
1
Postdoctoral Res. Fellow, Dept. of Civ. Engrg., Univ. of British Co-
lumbia, 2324 Main Mall, Vancouver, BC, Canada V6T 1Z4; and Lect.,
Dept. of Civ. and Envir. Engrg., Univ. of Adelaide, Adelaide, S.A. 5005,
Australia. E-mail: [email protected]
2
Asst. Prof., Dept. of Civ. Engrg., Univ. of British Columbia, 2324
Main Mall, Vancouver, BC, Canada V6T 1Z4.
3
Assoc. Prof., Dept. of Civ. Engrg., Univ. of British Columbia, 2324
Main Mall, Vancouver, BC, Canada V6T 1Z4.
Note. Editor: Sivand Lakmazaheri. Discussion open until December 1,
2000. To extend the closing date one month, a written request must be
filed with the ASCE Manager of Journals. The manuscript for this paper
was submitted for review and possible publication on June 18, 1999. This
paper is part of the Journal of Computing in Civil Engineering, Vol.
14, No. 3, July, 2000. 䉷ASCE, ISSN 0887-3801/00/0003-0183–0189/
$8.00 ⫹ $.50 per page. Paper No. 21288. FIG. 1. Structure of Typical MLP Model
冘
p Optimization of the connection weights of AMNs, on the other
y= ai wi (1) hand, is a linear optimization problem. Consequently, conven-
i=1 tional linear learning theory can be used, and convergence re-
sults can be derived (Brown and Harris 1994). In on-line ap-
where y = model output; ai = output from the pth basis func- plications, the weights of AMNs are generally updated
tion; and wi = connection weight associated with ai. iteratively using normalized least mean squares learning rules
As the input space of AMNs is normalized into a p-dimen- (Brown and Harris 1994).
sional lattice, similar network inputs only activate a small
number of basis functions (Brown and Harris 1994). In other Model Development
words, only a small number of weights contribute to the model
output for a given set of model inputs. Consequently, infor- Another difference between MLP and AMN models is the
mation is stored and adapted locally. Only those weights that way in which appropriate model structures are selected. The
contribute to the model output are adjusted during training, type of nonlinear function used in the hidden layer(s) can have
resulting in a fast rate of initial convergence and an ability to a significant impact on the performance of both types of mod-
incorporate new information into the model without overwrit- els. In MLPs, sigmoidal functions are generally used, whereas
ing information that is stored in other regions of the network. the two most commonly used AMN basis functions are B-
In contrast, MLPs have a global response. Adjustment of any splines and Gaussian radial basis functions. A detailed discus-
weight will have an impact on the output obtained for every sion of the relative merits of these two types of functions is
input. In addition, the basis functions may be interpreted as a given by Brown and Harris (1995). The major difference be-
set of fuzzy linguistic rules, making the knowledge that is tween Gaussian and B-spline functions is that the latter have
contained in the model transparent. bounded supports (i.e., the output of the functions is only non-
The major disadvantage of AMNs is that the number of zero if inputs lie in the subset of the input space on which
connection weights is exponentially dependent on the dimen- they are defined), whereas the former do not. Consequently,
sion of the input space (Brown and Harris 1995). Conse- B-spline networks are less flexible than Gaussian networks.
quently, they are not suitable for problems that have a large However, they are more efficient to implement, have greater
number of inputs. However, if it is assumed that the desired transparency, and are more appropriate for generating fuzzy
input/output mapping can be additively decomposed, the di- algorithms.
mensionality of AMN models can be reduced significantly by Parsimony is a major consideration when choosing appro-
modeling the desired relationship as a linear combination of a priate structures for MLP and AMN models. It is desirable to
number of subnetworks. The analysis of variance (ANOVA) use models that have the smallest number of adjustable param-
representation of multivariate functions provides a framework eters (e.g., connection weights), while still being able to ade-
for achieving this, because it represents an n-dimensional func- quately map the desired input/output relationship. Larger net-
tion as follows (Brown and Harris 1995): works have higher computational costs and need more training
冘 冘冘
n n⫺1 n
data to ensure that overfitting does not occur and make it more
difficult to extract knowledge from trained networks (Maier
f (x) = f0 ⫹ fi (xi) ⫹ fi, j (xi, xj ) ⫹ ⭈ ⭈ ⭈ ⫹ f1,2,...,n(x) (2)
i=1 i=1 j=i⫹1
and Dandy 2000). To minimize the number of connection
weights in MLP models, a number of construction and pruning
where f0 represents a constant (the function bias); and the other algorithms have been proposed for optimizing the number of
terms represent univariate, bivariate, and higher-order subfunc- hidden layer nodes (Reed 1993; Kwok and Yeung 1997). In
tions. In many instances, the majority of the higher-order terms addition, several approaches have been suggested for ensuring
are negligible, and the input/output mapping can be approxi- that only relevant model inputs are included (Maier and Dandy
mated by a limited number of subnetworks of reduced dimen- 1997a) as the inclusion of redundant inputs also increases net-
sion. An example of such an additive decomposition is shown work size.
in Fig. 3, where a 5D function is decomposed into one 1D The number of connection weights in B-spline AMNs is a
184 / JOURNAL OF COMPUTING IN CIVIL ENGINEERING / JULY 2000
function of the number of basis functions. The latter is depen-
dent on the dimensionality of the input space, and conse-
quently it is desirable to use the smallest number of model
inputs and the smallest number of basis functions for each of
these. The adaptive spline modeling of observation data
(ASMOD) algorithm (Kavli 1993) is an automatic method for
obtaining the optimal structure of B-spline AMNs. It starts
with a simple model and iteratively refines the complexity of
the model by introducing new input variables, forming mul-
tivariate subnetworks (based on the ANOVA approach) and
adding new basis functions for individual input variables. Fi-
nally, the impact of pruning the network (e.g., removing basis
functions and splitting subnetworks) is investigated in an effort FIG. 4. Schematic Representation of Typical Fuzzy System
[after Bossley et al. (1995)]
to arrive at the simplest model structure. Thus, the ASMOD
algorithm uses the training data to automatically determine the
model inputs, the number and spacing of the basis functions where xk = kth real-valued input; y = output; ri, j = ijth fuzzy
(i.e., the optimum partitioning of the input space), and the rule with corresponding rule confidence ci, j; Aik = univariate
number and dimension of the subnetworks. However, the order membership function for xk, which forms part of the ith mul-
of B-spline functions has to be determined a priori. Higher- tivariate rule premise (or antecedent); and Bj = corresponding
order functions result in smoother model outputs, but increase jth univariate fuzzy membership function for the output. The
computational cost and can lead to overfitting of the data rule confidence is an indication of the strength with which a
(Brown and Harris 1994). Consequently, it is desirable to use particular rule contributes to the output. Examples of rules of
lower-order functions, provided that they are able to model the the above type in the fields of hydrology and water resources
desired relationship adequately. engineering could include
As part of the ASMOD algorithm, models are trained using
IF ‘‘Rainfall Intensity’’ is High AND ‘‘Rainfall Duration’’
a stochastic approximation version of the normalized least
mean squares rule until certain stopping criteria are met at each is Long THEN ‘‘Runoff’’ is Large, c = 0.9
stage of the refinement process (Kavli 1993). The performance
of models of varying complexity is then compared using mea- IF ‘‘Temperature’’ is High AND ‘‘Flow’’ is Low
sures that balance model size, the number of training data, and
THEN ‘‘Dissolved Oxygen’’ is High, c = 0.1
model error. Such measures, including the Bayes information
criterion (Schwarz 1978) and Akaike’s information criterion Fuzzy algorithms are formed by the union (use of the fuzzy
(Akaike 1974), are commonly used in the development of time OR operator) of individual fuzzy rules. The way in which the
series models and have recently been applied to artificial neu- fuzzy operators (IF, THEN, AND, OR) are implemented can have
ral network models [e.g., Doering et al. (1997)]. Alternative a significant impact on model performance. Traditionally, trun-
approaches for automatically determining the optimum struc- cation operators are used for this purpose (Zadeh 1973). However,
ture of B-spline AMNs have also been proposed [e.g., Menken
the use of algebraic operators has recently gained popularity [e.g.,
(1998)]. However, the ASMOD algorithm is used in this re-
Wang and Mendel (1992)] because they generally produce
search because it is well known, has good convergence prop-
smoother outputs (Brown and Harris 1994). If a real-valued out-
erties (Weyer and Kavli 1997), and has been found to perform
well compared with other methods for a wide variety of mod- put is required, a defuzzification process has to be carried out.
eling problems (Brown and Harris 1994). This is generally achieved by either using a mean of maximums
or a center of gravity defuzzification strategy (Brown and Harris
FUZZY INTERPRETATION OF B-SPLINE AMNs 1994).
A schematic of a complete fuzzy system is shown in Fig. 4.
Fuzzy logic was introduced by Zadeh (1965) to enable the
An input/output mapping is obtained by presenting a real-valued
mathematical representation of vague, linguistic concepts such
as ‘‘very high,’’ ‘‘high,’’ ‘‘medium,’’ ‘‘low,’’ etc. This is input to the model, which is fuzzified, operated on using the
achieved by allowing a particular variable to be a partial mem- knowledge stored in a fuzzy rule base, and defuzzified to produce
ber of a particular set. This is in contrast to classical logic, a real-valued output. It should be noted that once all of the com-
where a variable is either a member of a set or it is not. ponents of a particular fuzzy system have been defined, it operates
Generally, a small number of linguistic terms (e.g., high, in a deterministic manner (Procyk and Mamdani 1979). In other
medium, low), referred to as fuzzy sets, are assigned to each words, the input/output relationship is no longer uncertain, but
variable (e.g., river temperature). These fuzzy sets overlap and well defined. Brown and Harris (1994) proved that if fuzzy sys-
cover the necessary range of variation for that variable. The tems use B-spline membership functions, algebraic operators, and
degree of membership (from 0 to 1) of a real-valued input x the center of gravity defuzzification algorithm, they are equivalent
(e.g., temperature of 22⬚C) to a particular fuzzy set A (e.g., to B-spline AMNs. Consequently, the B-spline basis functions
high) is given by a ‘‘membership function’’ uA(x). This trans- may be viewed as fuzzy membership functions and can be used
formation of real-valued inputs into a degree of membership to elicit fuzzy rules. It should be noted that B-spline AMNs are
to a particular fuzzy set is called ‘‘fuzzification.’’ Membership only one type of neurofuzzy system (Nauck et al. 1997). Consid-
functions can take on a variety of forms, including Gaussian eration of other neurofuzzy approaches is beyond the scope of
and B-spline functions (Brown and Harris 1995). this paper, as its objective is to compare the performance of back-
Fuzzification of the variables enables the relationship be- propagation MLPs and B-spline AMNs.
tween the input variable(s) and the output variable to be ex-
pressed in terms of a set of fuzzy rules of the form (Brown
FORECASTING CYANOBACTERIAL
and Harris 1995)
CONCENTRATIONS IN RIVER MURRAY
ri, j: IF (x1 is Ai1 AND x2 is Ai2 AND
The River Murray is South Australia’s major surface water
⭈ ⭈ ⭈ AND xn is Ani ) THEN ( y is Bj ), ci, j (3) resource. Water is pumped several hundred kilometers to major
JOURNAL OF COMPUTING IN CIVIL ENGINEERING / JULY 2000 / 185
iables are dominant (Maier and Dandy 1997b). Consequently,
model transparency is an important consideration.
The available data include weekly values of concentrations
of the cyanobacterium Anabaena spp., turbidity, color, tem-
perature, total phosphorus, soluble phosphorus, oxidized nitro-
gen, and total iron at Morgan as well as weekly flows at Lock
7 (Fig. 5). All data are available from 1985/1986 to 1992/1993.
It should be noted that no data are available for some factors
that can influence the size of cyanobacterial populations in-
cluding the amount of incoming radiation, the concentrations
of algal predators and competing algal species, the thermal
structure of the river, pH, and dissolved oxygen. A time series
plot of Anabaena spp. concentrations at Morgan from 1985/
1986 to 1992/1993 is given in Fig. 6. It can be seen that the
time series exhibits irregular seasonal variations but does not
contain any trends. The same applies for the time series for
the other variables.
ample, it has been shown that water transparency and nutrients dence of Anabaena spp. However, the reverse applies in re-
can have a significant impact on the growth rate of Anabaena lation to the ability of the models to forecast the duration of
spp., although the exact relationship is not well known (Maier this incidence of Anabaena spp.
et al. 1998). However, even when turbidity, soluble phospho- The back-propagation MLP model uses 20 model inputs,
rus, total phosphorus, oxidized nitrogen, and total iron are in- and whereas the same values are used as potential input can-
cluded as model inputs (Table 1, column 3), only flow and didates for the B-spline AMN model, only 4 of these are found
temperature inputs are found to be significant when the to be significant when the ASMOD algorithm is used. Con-
ASMOD algorithm is used. The most likely reason for this is sequently, the B-spline AMN model is more parsimonious than
that there are insufficient training data to cover the full range the back-propagation MLP model, and the transparency of the
of possible environmental conditions. For example, during the former is increased. The model inputs that are found to be
time period covered by the training data, nutrient levels are significant include river flow at times t ⫺ 1, t ⫺ 8, and t ⫺
relatively high and thus unlikely to be limiting (Maier et al. 10, and water temperature at time t ⫺ 1. The fuzzy rules ob-
1998). tained indicate that incidences of Anabaena spp. are most
likely to occur after the passing of a flood hydrograph and
CONCLUSIONS when water temperatures are high. Although this is in agree-
The results of the case study carried out indicate that B- ment with the current level of physical understanding of the
spline AMN and back-propagation MLP models are suitable system, the fuzzy rules do not provide any information about
for obtaining 4-week forecasts of concentrations of the cyano- the relationship between the concentration of Anabaena spp.
bacterium Anabaena spp. in the River Murray at Morgan, and other environmental variables that are considered to be
South Australia. Overall, the forecasts obtained when the B- important (e.g., nutrients, color, and turbidity). The most likely
spline AMN model is used may be considered to be slightly reason for this is that these relationships are not significant in
better than those obtained when the back-propagation MLP is the data that are available for training the model. Conse-
used. The former is able to forecast the first (although smaller) quently, although B-spline AMN models offer a great deal of
of the two incidences of Anabaena spp. that occur during the model transparency by representing the relationships between
validation period accurately, whereas the latter fails to forecast model inputs and outputs in terms of fuzzy rules, the generality
this incidence of Anabaena spp. altogether. The B-spline AMN of the rules obtained is a function of the quality of the avail-
model also outperforms the back-propagation MLP in terms able data. However, if certain relationships are well known a
of forecasting the onset and peak of the second (major) inci- priori (which is not the case in the application considered in
188 / JOURNAL OF COMPUTING IN CIVIL ENGINEERING / JULY 2000
this paper), they can be incorporated into the model with rel- Menken, G. J. (1998). ‘‘Structural adaptation of B-spline networks using
ative ease. genetic algorithms.’’ Engrg. Intelligent Sys. for Electr. Engrg. and Com-
munications, 6(3), 147–152.
Nauck, D., Klawonn, F., and Kruse, R. (1997). Foundations of neuro-
APPENDIX I. REFERENCES fuzzy systems, Wiley, Chichester, England.
Akaike, H. (1974). ‘‘A new look at the statistical model identification.’’ Procyk, T. J., and Mamdani, E. H. (1979). ‘‘A linguistic self-organising
IEEE Trans. on Automatic Control, 19, 716–723. process controller.’’ Automatica, 15, 15–30.
Bossley, K., Brown, M., and Harris, C. (1995). ‘‘Parsimonious neurofuzzy Reed, R. (1993). ‘‘Pruning algorithms—A review.’’ IEEE Trans. on Neu-
modelling.’’ Univ. of Southampton Res. J., Dept. of Electronics and ral Networks, 4(5), 740–747.
Comp. Sci., University of Southampton, Southampton, England. Schwarz, G. (1978). ‘‘Estimating the dimension of a model.’’ Ann. of
Brown, M., and Harris, C. (1994). Neurofuzzy adaptive modelling and Statistics, 6(2), 461–464.
control, Prentice-Hall, Englewood Cliffs, N.J. Takyi, A. K., and Lence, B. J. (1999). ‘‘Surface water quality management
Brown, M., and Harris, C. (1995). ‘‘A perspective and critique of adaptive using multiple-realization chance constraint method.’’ Water Resour.
neurofuzzy systems used for modelling and control applications.’’ Int. Res., 35(5), 1657–1670.
J. Neural Sys., World Scientific, Singapore, 6(2), 197–220. Wang, L. X., and Mendel, J. M. (1992). ‘‘Fuzzy basis functions, universal
Doering, A., Galicki, M., and Witte, H. (1997). ‘‘Structure optimization approximation, and orthogonal least squares learning.’’ IEEE Trans. on
of neural networks with the A*-algorithm.’’ IEEE Trans. on Neural Neural Networks, 3(5), 807–814.
Networks, 8(6), 1434–1445. Weyer, E., and Kavli, T. (1997). ‘‘Theoretical properties of the ASMOD
Flood, I., and Kartam, N. (1994). ‘‘Neural networks in civil engineering. algorithm for empirical modelling.’’ Int. J. Control, Taylor and Francis
I: Principles and understanding.’’ J. Comp. in Civ. Engrg., ASCE, 8(2), Group, London, 67(5), 767–789.
131–148. Zadeh, L. A. (1965). ‘‘Fuzzy sets.’’ Information and Control, 8, 338–353.
French, M., and Recknagel, F. (1994). ‘‘Modeling of algal blooms in Zadeh, L. A. (1973). ‘‘Outline of a new approach to the analysis of com-
freshwater using artificial neural networks.’’ Computer techniques in plex systems and decision processes.’’ IEEE Trans. of Sys. Man and
environmental studies V, P. Zannetti, ed., Computational Mechanics Cybernetics, 3, 28–44.
Publications, Southampton, England, 87–94.
Hornik, K., Stinchcombe, M., and White, H. (1989). ‘‘Multilayer feed- APPENDIX II. NOTATION
forward networks are universal approximators.’’ Neural Networks, 2,
359–366. The following symbols are used in this paper:
Kavli, T. (1993). ‘‘ASMOD—An algorithm for adaptive spline modeling
of observation data.’’ Int. J. Control, Taylor and Francis Group, Lon- A = fuzzy membership function;
don, 58(4), 947–967. Aik = univariate membership function for xk that forms part of
Kwok, T.-Y., and Yeung, D.-Y. (1997). ‘‘Constructive algorithms for the ith multivariate rule premise;
structure learning in feedforward neural networks for regression prob- ai = output from ith basis function;
lems.’’ IEEE Trans. on Neural Networks, 8(3), 630–645. ap = output from pth basis function;
Maier, H. R., and Dandy, G. C. (1997a). ‘‘Determining inputs for neural Bj = jth univariate fuzzy membership function for model out-
network models of multivariate time series.’’ Microcomputers in Civ.
Engrg., 12(5), 353–368.
put;
Maier, H. R., and Dandy, G. C. (1997b). ‘‘Modelling cyanobacteria (blue- cij = rule confidence associated with rij;
green algae) in the River Murray using artificial neural networks.’’ f1,2,...,n = function of order n;
Mathematics and Comp. in Simulation, 43(3-6), 377–386. p = number of nodes in input layer of MLP neural network;
Maier, H. R., and Dandy, G. C. (1999). ‘‘Empirical comparison of various rij = rule relating ith fuzzy input set to jth fuzzy output set;
methods for training feedforward neural networks for salinity fore- t = time;
casting.’’ Water Resour. Res., 35(8), 2591–2596. uA(x) = membership of x in fuzzy set A;
Maier, H. R., and Dandy, G. C. (2000). ‘‘Neural networks for the pre- w = connection weight;
diction and forecasting of water resources variables: A review of mod- wi = connection weight associated with ai;
elling issues and applications.’’ Envir. Modelling and Software, 15(1),
101–123.
wjn = connection weight associated with nth input to node j;
Maier, H. R., Dandy, G. C., and Burch, M. D. (1998). ‘‘Use of artificial xjn = nth input to node j;
neural networks for modelling cyanobacteria Anabaena spp. in the xk = kth model input;
River Murray, South Australia.’’ Ecological Modelling, 105(2/3), 257– y = model output; and
272. yj = output from node j.
ABSTRACT: There exists a strong need to explore simulation techniques that not only represent complex
dynamic systems in a realistic way but also allow the involvement of end users in model development to increase
their confidence in the modeling process. System dynamics, a feedback-based object-oriented simulation ap-
proach, is presented for modeling reservoir operations. The increased speed of model development, the trust
developed in the model due to user participation, the possibility of group model development, and the effective
communication of model results are main strengths of this approach. The ease of model modification in response
to changes in the system and the ability to perform sensitivity analysis make this approach more attractive
compared with systems analysis techniques for modeling reservoir operations. The proposed approach is applied
to the Shellmouth reservoir on the Assiniboine River in Canada. Operating rules are developed for high flow/
flood years to minimize flooding. Alternative operating rules are explored by changing reservoir storage allo-
cation and reservoir outflows. Impacts on the flood management capacity of the reservoir are investigated by
simulating a gated spillway in addition to an existing unregulated spillway. Sensitivity analysis is performed on
the reservoir levels at the start of the flood season and the outflow from the reservoir.
calculation of the flooded area and duration of flooding due to as flow takes five days to reach Winnipeg once released from
the reservoir operation provides information on the effective- dam. With several trials it was found that a time step of one
ness of different operating policies for flood management. day provides the best trade-off between speed of calculation
Simulation techniques are not capable of generating directly and accuracy of results. Similarly, a variation of the delay
an optimal solution to a reservoir operation problem; however, function, used for flood routing, affects the timing and the
by going through several runs of a model with alternative pol- duration of flooding at downstream locations.
icies, an optimal or near optimal operating policy can be iden-
tified. RESULTS AND DISCUSSION
Simulations of the Shellmouth reservoir operation were
Daily variations of the reservoir levels for four major flood
made for the five largest historic floods with natural and gated
events (1974, 1975, 1976, and 1979) are shown in Fig. 6. A
spill scenarios. Model inputs were annual series of daily in-
summary of selected results is provided in Tables 1–3. Re-
flows to the reservoir during the five major flood events.
vised operating rules with natural spill and gated spill are
Model output includes daily variation of the reservoir level,
shown in Fig. 7 along with existing operating rules. Results
daily discharges from the reservoir, total flooded area upstream
show that with revised operating rules it was possible to op-
of the reservoir, discharges and flooded area at seven down-
erate the reservoir with only minor flooding upstream and
stream locations, and diversion to Lake Manitoba at Portage.
downstream for four out of five major flood events. While
Discharges at Headingly were used to estimate the contribution
simulating flood years 1975 and 1976, the spillway was not
of the Assiniboine River towards the flooding of Winnipeg
City. Policy alternatives were explored by changing initial res-
ervoir storage level, both at the start of simulation and at the
start of flooding season. Trials were also made to explore the
effects of changing outflow through the conduit on the varia-
tion of the reservoir level.
Models are developed to answer questions related to rela-
tively uncertain conditions; this is especially true in water re-
sources modeling. SD modeling provides a very convenient
and powerful tool to explore how changes in one system var-
iable impact other variables. For this study, a sensitivity anal-
ysis was performed on Time Step and Delays. Time Step, also
called Delta Time (DT), is the interval of time between model
calculations; thus, DT represents the smallest time interval
over which a change in numerical values of any element in
the model can occur. The Delay function returns a delayed
value of input, using a fixed lag time of delay duration. This FIG. 6. Water Levels in Shellmouth Reservoir for Flood Years
function is important to capture the timing of the flood peak, 1974, 1975, 1976, and 1979
Summary of Results
System dynamics proved to be a very time efficient, user
friendly, and successful approach for modeling the reservoir
operations. The simulation of the reservoir operation verified
that with the revised operating rules, the capability of the
Shellmouth reservoir for flood management can be improved.
For four out of the five largest historic floods, the reservoir
was operated without causing any significant flooding down-
stream or upstream. Due to revision of the operating rules, the
contribution of the Assiniboine River towards the flooding of
Winnipeg City is negligible (discharge at Headingly, which
once reached 170 m3/s, can be reduced to 5.66 m3/s with the
gated spillway option). Number of days when reservoir is full
or at the minimum level is very sensitive to reservoir outflows,
FIG. 9. Reservoir Levels by Varying Initial Reservoir Level at especially over the falling limb of the flood hydrograph. Res-
Start of Simulation: Curve 1 = 422.2 m; Curve 2 = 425.2 m; Curve
ervoir levels during the flood, flooded area upstream as well
3 = 428.3 m; Curve 4 = 429.2 m
as downstream, and duration of flooding are sensitive to res-
ervoir level at the start of the flooding season.
Simulation of the Shellmouth reservoir operation, consid-
ering both gated and unregulated spillway options, suggests
that installation of gates on the spillway will improve the flood
management capacity of the reservoir, especially for large
floods. Installation of gates may be justified, as flooding up-
stream due to the reservoir level rise is insignificant compared
with the flood losses downstream. However a detailed study
incorporating economic analysis and environmental impacts is
recommended before finalizing the decision to put gates on
the spillway.
CONCLUSIONS
The research reported in this paper focused on the simula-
tion of a single multipurpose reservoir for flood management
FIG. 10. Reservoir Levels by Varying Initial Reservoir Level at
purposes using the SD approach. Operating rules were revised
Day 90: Curve 1 = 422.2 m; Curve 2 = 425.2 m; Curve 3 = 428.3 m; for high-flow/flood years to minimize flooding. Impacts on
Curve 4 = 429.2 m flood management capacity of the reservoir were explored by
simulating a gated spillway in addition to the existing unreg-
ulated spillway. Alternative operating rules were explored by
changing the reservoir storage allocation, reservoir levels at
the start of flooding season, and the reservoir outflows.
The proposed SD-based simulation approach is a valuable
alternative to conventional simulation techniques. The in-
creased speed of model development, ease of model structure
modification, ability to perform sensitivity analysis and effec-
tive communication of model results are the main strengths of
the reservoir simulation model based on the SD approach.
However, one limitation is the simplified flood routing as com-
pared with sophisticated hydrodynamic models.
Because of its ease of construction and modification, SD
simulation environments facilitate rapid prototyping and
greatly reduce programming effort. Modeling effort can be di-
FIG. 11. Reservoir Levels by Varying Discharges through
rected to important tasks such as system conceptualization,
Conduit (Falling Limb of Hydrograph at Elevation 429.2–428 m): data collection, gaining input from system operators, and in-
Curve 1 = 0 m3/s; Curve 2 = 11.3 m3/s; Curve 3 = 19.8 m3/s; Curve volving stakeholders.
4 = 34.0 m3/s; Curve 5 = 45.3 m3/s The Shellmouth reservoir simulation model can be fine-
JOURNAL OF COMPUTING IN CIVIL ENGINEERING / JULY 2000 / 197
tuned easily in light of operating experience, or with the help Keys, A. M., and Palmer, R. (1993). ‘‘The role of object oriented simu-
of insight provided by an expert. The SD approach offers a lation models in the drought preparedness studies.’’ Proc., 20th Ann.
Nat. Conf., Water Resour. Plng. and Mgmt. Div., ASCE, New York,
way for operators to participate in the model building process, 479–482.
thus increasing their trust in the model. The operator’s feed- Loucks, D. P., Stedinger, J. R., and Haith, D. A. (1981). Water resources
back provides direction for follow-up simulations and modi- systems planning and management, Princeton Hall, Englewood Cliffs,
fications in the model structure. N.J.
The entire modeling process is open, interactive, and trans- Matthias, R., and Frederick, P. (1994). ‘‘Modeling spatial dynamics of
parent. The model is easy to expand and modify in terms of sea-level rise in a coastal area.’’ Sys. Dyn. Rev., 10(4), 375–389.
Mays, W. L., and Tung, Y.-K. (1992). Hydrosystems engineering and
its internal structure and the amount, type, and resolution of management, McGraw-Hill, New York.
data used. Modifications in both structure and values of pa- Palmer, R. (1994). ‘‘(ACT-ACF) River basin planning study.’’ 具http://
rameters can be made easily as new information that can affect maximus.ce.washington.edu/⬃actacf/典.
the model becomes available. The model can then be rerun, Palmer, R., Hilton, K., and Hahn, M. (1995). ‘‘Willamette river ba-
the results analyzed, and the model improved as necessary. sin reauthorization feasibility study.’’ 具http://www.ce.washington.edu/
Thus, the entire dynamic modeling process facilitates adapta- ⬃willamet/典.
Palmer, R., Keyes, A. M., and Fisher, S. (1993). ‘‘Empowering stake-
tion of a general model to a particular problem. holders through simulation in water resources planning.’’ Water man-
The architecture of the Shellmouth reservoir model is ge- agement for the ’90s, ASCE, New York, 451–454.
neric in nature and can be applied to other reservoirs by re- Raymond, C. S. (1985). ‘‘Dynamic simulation models: how valid are
placing data and operating rules for that particular reservoir. they?’’ Self-report methods of estimating drug use: meeting current
Numerous simulation scenarios, in addition to what has been challenges to validity, A. B. Rouse, J. N. Kozel, and G. L. Richards,
demonstrated in this study, can be tested using the existing eds., National Institute on Drug Abuse, Bethesda, Md.
Russell, S. O., and Campbell, P. F. (1996). ‘‘Reservoir operating rules
framework. As the current model provides information on ex- with fuzzy programming.’’ J. Water Resour. Plng. and Mgmt., ASCE,
tent and duration of flooding, another sector can be added to 122(3), 165–170.
calculate damage to crops or economic losses due to lost op- Simonovic, S. P. (1992). ‘‘Reservoir systems analysis: closing gap be-
portunity of seeding. The model can be extended from a single tween theory and practice.’’ J. Water Resour. Plng. and Mgmt., ASCE,
multipurpose reservoir to a system of reservoirs. 118(3), 262–280.
Simonovic, S. P., and Fahmy, H. (1999). ‘‘A new modeling approach for
water resources policy analysis.’’ Water Resour. Res., 35(1), 295–304.
ACKNOWLEDGMENTS Simonovic, S. P., Fahmy, H., and El-Shorbagy, A. (1997). ‘‘The use of
This work was made possible by an operating grant from the Natural object-oriented modeling for water resources planning in Egypt.’’ Water
Sciences and Engineering Research council of Canada (NSERC). The Resour. Mgmt., 11(4), 243–261.
writers would like to thank the Water Resource Branch, Manitoba De- Tutorial and technical documentation STELLA威 II. (1992). High Perfor-
partment of Natural Resources, for providing the necessary data. mance Systems, Inc., Hanover, N.H.
Water Resources Branch. (1992). ‘‘Shellmouth reservoir study: prelimi-
APPENDIX. REFERENCES nary study of spillway control gates.’’ Internal Rep., Manitoba De-
partment of Natural Resources, Winnipeg, Canada.
Fletcher, E. J. (1998). ‘‘The use of system dynamics as a decision support Water Resources Branch. (1995). ‘‘Assiniboine river flooding & operation
tool for the management of surface water resources.’’ Proc., 1st Int. of Shellmouth dam.’’ Rep. Prepared for the Shellmouth Flood Review
Conf. on New Information Technolo. for Decision-Making in Civ. Committee, Manitoba Department of Natural Resources, Winnipeg,
Engrg., University of Quebec, Montreal, Canada, 909–920. Canada.
Forrester, J. W. (1961). Industrial dynamics, Productivity Press, Portland Wurbs, R. A., and Tibbets, M. N. (1985). ‘‘State-of-the-art review and
Oreg. annotated bibliography of systems analysis techniques applied to res-
Forrester, J. W., and Senge, P. M. (1980). ‘‘Tests for building confidence ervoir operations.’’ Tech. Rep. No. 136, Texas Water Resources Insti-
in system dynamics models.’’ TIMS studies in management science, tute, College Station, Tex.
Vol. 14, A. Legasto, J. Forrester, and J. Lyneis, eds., North Holland Yeh, W. W.-G. (1985). ‘‘Reservoir management and operation models: a
Press, Amsterdam. state-of-the-art-review.’’ Water Resour. Res., 21(12), 1797–1818.
ABSTRACT: Despite recent advances in desktop computing power, certain civil engineering problems continue
to depend on supercomputers and powerful workstations. Although supercomputers and workstations helped in
field testing complicated mathematical models, there appears to be a significant gap in widespread technology
transfer to the industry. The sluggish progress in transfer of technology to the industry may be attributed to the
inability to implement these models on PCs coupled with high costs associated with supercomputers. The paper
reports results from an exploratory research that implemented a complicated optimization model based on a
distributed genetic algorithm on a network of PCs. PCs that formed the network were hardwired using 16bit
10Base-T Ethernet cards and were made accessible using Peer-to-Peer networking capability, which is a built-
in feature of Microsoft Windows 95/NT operating system. The inherent parallelism associated with genetic
algorithms coupled with relatively small data exchange between the computers resulted in a significant reduction
of computational time. The proposed generalized optimization framework, which can be adopted to model several
water resources related problems, is expected to accelerate the transfer of technology to the industry.
INTRODUCTION though GAs are very robust and powerful in obtaining a good
optimal solution, the excessive computational time require-
There was a significant increase in desktop computing ment (a characteristic of a GA) has been a serious hindrance
power beginning in the early 1990s. Affordable PCs with sig- in applying this technology for large-scale real-life problems.
nificant processing speed helped most civil engineering prob- Lack of affordable computational power has significantly
lems become less dependent on mainframe computers. Simul- slowed down the transfer of technology to the industry. This
taneously, there has been a growing interest in developing new prompted the researchers to look into the options of improving
or improved computational tools to solve complicated prob- the computational efficiency of the GA-based optimization
lems. However, because of these modern computational tools, models. Distributed implementation of GAs on an affordable
certain civil engineering problems continue to require large network of PCs is one such option.
computational requirements that cannot be realistically pro- Parallel implementation of GAs is not a new concept. How-
vided by even the most powerful desktop PCs. Unfortunately, ever, most efforts in the past have been focused at imple-
not every user has access to supercomputers or, in some cases, menting the GAs on supercomputers (Hung and Adeli 1994;
even to powerful workstations. The drinking water supply in- Lin et al. 1995; Foo et al. 1997), transputers (Fukuyama et al.
dustry is one such example, wherein lack of affordable com- 1996; Pollmeier et al. 1996; Oliveira et al. 1997; Yang et al.
putational resources prevented optimal utilization of treated 1997), or a network of distributed computing workstations
water and electricity resources. (Davis et al. 1994; York et al. 1994; Areibi and Vannelli 1998;
Past research concerning water supply systems has resulted Vatsa and Wedan 1999). These machines are either beyond the
in abundant technology on optimal utilization of scarce re- reach of most water utilities or too complicated for the per-
sources (Coulbeck et al. 1988a,b; Jowitt and Germanopoulos sonnel in charge of the utilities’ computational resources. In
1992; Ormsbee et al. 1992; Lingireddy 1994; Ormsbee and addition, most of the available literature on parallel computing
Lansey 1994; Ormsbee and Lingireddy 1995a,b). However, addresses noncivil engineering issues.
transfer of technology to the industry has been slow, due pri- This paper reports results from a preliminary study that im-
marily to a lack of affordable computational resources. Al- plemented a GA-based calibration model on a network of PCs.
though recent advances in desktop computational power have The focus of this article is on implementing the calibration
encouraged even small water utilities to utilize simple com- model on a network of PCs for distributed processing, and not
puter models, there appears to be a significant gap in available on details of the calibration model itself. The calibration model
and required desktop computational resources for widespread is described primarily to show the necessity and usefulness of
utilization of advanced computer models. Optimal control of distributed implementation of the model on a network of PCs.
water distribution systems and optimal calibration of distri- The described approach is suitable to other GA-based com-
bution network models are two problems where there is a need puter models as well. Specific examples in water resources
for significant computational power. engineering areas that can benefit from this technology include
A growing interest in the use of genetic algorithms (GAs) the efficient real-time operation of water distribution systems,
for many engineering problems was witnessed in the 1980s the optimal design of transient-pressure suppression devices,
and 1990s. Several GA-based models for solving complex and the optimal operation of multipurpose reservoirs.
civil engineering problems were proposed during this period
(Goldberg and Kuo 1987; Lingireddy 1994, 1996, 1998; CALIBRATION OF WATER DISTRIBUTION NETWORK
Ormsbee and Lingireddy 1995a, 1998; Sen et al. 1992). Al- MODELS
1
Grad. Student, Dept. of Civ. Engrg., Univ. of Kentucky, Lexington, Network models play an important role in the day-to-day
KY 40506-0281. operation of drinking water distribution systems. Recent en-
2
Assoc. Prof., Dept. of Civ. Engrg., Univ. of Kentucky, Lexington, KY. actment of the Safe Drinking Water Act further enhances the
Note. Editor: Sivand Lakmazaheri. Discussion open until December 1, importance of network models as they form the basis for water
2000. To extend the closing date one month, a written request must be quality modeling in pipe networks. Whereas network models
filed with the ASCE Manager of Journals. The manuscript for this paper
was submitted for review and possible publication on March 8, 1999.
are indispensable for safe and efficient operation of water
This paper is part of the Journal of Computing in Civil Engineering, distribution systems, the practical difficulties associated with
Vol. 14, No. 3, July, 2000. 䉷ASCE, ISSN 0887-3801/00/0003-0199– data collection introduce uncertainty in two major parame-
0205/$8.00 ⫹ $.50 per page. Paper No. 20422. ters of a network model—pipe roughness coefficients and wa-
JOURNAL OF COMPUTING IN CIVIL ENGINEERING / JULY 2000 / 199
ter demand at nodes. The most economical way to avoid this mally, the explicit decision variables will include the rough-
problem is to seek indirect estimates for these parameters ness coefficient of each pipe or a group of pipes and the de-
by resorting to a calibration process. Calibration of water dis- mands at each node or a multiplication factor for a group of
tribution systems typically involves adjusting the uncertain pa- nodes. For a given vector of pipe roughness coefficients C
rameters to match measured flow and pressure data collected these constraints can be expressed
during a fire hydrant flow test with model predicted values.
Manual adjustment of uncertain parameters is highly ineffi- Lc ⱕ C ⱕ Uc (5)
cient and often unsuccessful. Use of optimization models is where Lc is lower bound; and Uc is upper bound on decision
considered appropriate for efficient calibration (Lingireddy and variable C. Likewise for a given vector of nodal demands D,
Ormsbee 1998). these constraints can be expressed
An optimization model typically comprises an objective
function that needs to be minimized/maximized subject to a LD ⱕ D ⱕ UD (6)
set of constraints. In the present study, the objective function
is formulated to minimize the square of the difference between where LD is lower bound; and UD is upper bound on decision
observed and predicted values of pressures and flow rates. variable D. Eqs. (1)–(6) constitute the optimization model for
Mathematically, the objective function (fitness function in ge- the optimal calibration of a water distribution system.
netic algorithms terminology) may be expressed The nonlinear nature of the objective function and some of
the constraints warrant the use of a nonlinear solution tech-
冘 冘
I J
nique such as a gradient-based method or a GA-based method.
minimize Z = ␣ (P io ⫺ P ip)2 ⫹  (Q jo ⫺ Q jp)2 (1) A growing interest in adopting underlying principles of natural
i=1 j=1
biological processes into engineering design tools was wit-
where P oi and P pi = observed and predicted pressures at node nessed in the 1980s and 1990s, thereby increasing the robust-
i; Q oj and Q pj = observed and predicted flow rates through pipe ness and efficiency of design tools (Goldberg and Kuo 1987;
j; I and J = number of nodes and number of pipes for which Lingireddy 1998). GAs are one such class of computational
measurements were taken; and ␣ and  = normalizing factors. tools developed to aid mathematical programming problems.
State variables P and Q are implicit functions of decision var- These tools attempt to mimic the mechanics of natural selec-
iables—pipe roughness coefficients and nodal demands. The tion and genetics. In other words, GAs are a class of optimal
objective function in (1) needs to be minimized subject to a search techniques modeled after the process of natural selec-
set of implicit system constraints and a set of explicit bound tion, which performs a number of simultaneous searches in the
constraints as follows. most promising regions of the decision space. Genetic opti-
mization offers a significant advantage over the traditional
Implicit System Constraints nonlinear optimization approaches, in that it attempts to obtain
an optimal solution by continuing to evaluate multiple solution
The implicit system constraints include nodal conservation vectors simultaneously. In addition, GA methods do not re-
of mass and conservation of energy. The nodal conservation quire gradient information. Further, genetic optimization meth-
of mass requires that the sum of flows into or out of any ods employ probabilistic transition rules as opposed to deter-
junction node i minus any external demand Di must be equal ministic rules, which have the advantage of ensuring a robust
to zero. For each junction node i this may be expressed solution methodology. The present study adopted a GA-based
冘
Ni method because of its inherent advantage over a traditional
Qj ⫺ Di = 0 (2) gradient based method (Goldberg 1989).
j=1
Solution Technique
where Ni = number of pipes connected to junction node i.
The conservation of energy constraint requires that the sum The use of a GA-based method assures that the explicit
of the line loss HL and the minor losses HM over any path or bound constraints [(5) and (6)] are satisfied throughout the
loop k, minus any energy added to the liquid by a pump EP, optimization process (Lingireddy and Ormsbee 1998). There-
minus the difference in grade between two points of known fore, the optimization model essentially reduces to (1)–(3).
energy DE is equal to zero. For any loop or path k this may Because (2) and (3) are constraints to the objective function
be expressed [(1)], (1)–(3) constitute a constrained optimization problem.
冘
Nk However, the solution based on a GA may only be applied to
(HLj ⫹ HMj ⫺ EPj) ⫺ DEk = 0 (3) unconstrained optimization problems. One of the popular ways
j=1 of transforming a constrained optimization problem into an
unconstrained one is via the use of penalty functions (Lingi-
where Nk = number of pipes associated with loop or path k. reddy and Elango 1989). In this approach, the objective func-
Herein, it should be emphasized that HLj, HMj, EPj, and DEk tion is augmented (penalized) with the appropriate penalty
are all nonlinear functions of the pipe discharge Q and pipe functions calculated based on constraints that were not satis-
roughness coefficients. For example, when the pipe roughness fied either explicitly or implicitly. However, the penalty func-
is expressed in terms of the popular Hazen-William roughness tion is zero when all constraints are satisfied. The present study
coefficient, the expression for line loss HL is given by employs a bilevel solution approach in which the constraints
10.675LQ1.852
[(2) and (3)] are satisfied outside the optimization framework,
HL = (4) consequently eliminating the need for the explicit transfor-
d1.852C 4.87 mation of the constrained optimization problem into an un-
where L = length of pipe; d = diameter; and C = Hazen-Wil- constrained problem. In the bilevel approach, (2) and (3) are
liam roughness coefficient. removed from the optimization model, solved using a hydrau-
lic simulation model, and the results are linked back to the
Explicit Bound Constraints optimization model. Fig. 1 shows a schematic for the bilevel
approach. A brief description on the methodology that a GA
The explicit bound constraints are used to set limits on the employs in obtaining an optimal solution is provided in the
explicit decision variables of the calibration problem. Nor- following.
200 / JOURNAL OF COMPUTING IN CIVIL ENGINEERING / JULY 2000
FIG. 1. Bilevel Optimization Framework
In general, GAs start with a randomly generated set of de- tion evaluation is where most of the computational time is
cision vectors referred to as population. With regard to net- expended. Although evaluation of objective functions requires
work model calibration, each decision vector comprises a set a great deal of data, most of the data remain unchanged from
of decision variables, which may include group roughness val- one function evaluation to the next. The input variables that
ues and/or demand adjustment factors. Each decision variable change from one evaluation to the next are primarily the de-
in the decision vector is coded as a binary number. With this cision variables, and the number of such variables is typically
binary coding, the entire decision vector is reduced to a string <100 for most optimization models. For example, when cali-
of binary numbers (i.e., 0 and 1). Such strings are analogous brating a water distribution system comprising 10 groups of
to biological chromosomes. From the current set of decision pipes and 10 groups of demands, the number of decision var-
vectors, a subset of decision vectors is selected based on the iables associated with each objective function evaluation is 20.
value of their fitness function (i.e., objective function). Once The objective function evaluations not only require negligible
this subpopulation of decision vectors is selected, a new pop- amounts of variable input data, but they are also independent
ulation of decision vectors (i.e., offspring) is generated by per- of each other. That is, the output generated by one objective
forming the sequential operations of crossover and mutation. function does not influence the output generated by another
The crossover and mutation operations are performed on any objective function. Such independent computer operations
randomly chosen pair of strings from the set of already se- coupled with negligibly small amounts of input requirements
lected best decision vectors. A detailed discussion on GA- and output generated by the objective function make GAs
based optimization models may be found in Goldberg (1989) amenable to distributed computing. Some of the independent
and Lingireddy (1996, 1998). operations may be performed on the other computers, leaving
the primary computer performing important tasks related to
NECESSITY FOR DISTRIBUTED COMPUTING the optimal genetic search. The external computers can pro-
vide essential information to the primary computer with a min-
Although genetic optimization models are more robust than imum of interruption.
the conventional gradient-based nonlinear optimization mod- In the case of the optimal calibration model, after the GA
els, they require a large number of function evaluations. If the generates the decision vectors, some of the decision vectors
optimization model searches through 100 generations and if may be transferred to other computers for distributed process-
each generation has a population of 30 decision vectors, then ing. The results accumulated from all the processors in the
the model has evaluated the objective function 3,000 times. distributed computing network may then be used to generate
Calibration of water distribution systems usually involves the the next set of decision vectors through a genetic search. Such
simultaneous evaluation of several fire hydrant flow tests, thus an approach would speedup the genetic search process of iden-
requiring many hydraulic simulation evaluations to evaluate tifying an optimal solution to a given problem. The process of
the objective function. Even a powerful simulation model such transmitting decision variables to external computers and re-
as KYPIPE (Wood 1995) typically requires about 0.5 CPU ceiving results from external computers may be automated if
second for a distribution system of about 1,000 pipes on an the primary and external computers are networked. The pres-
IBM compatible PC with a Pentium 133-MHz processor. If ent study involved fabrication of a network of PCs, which was
the calibration exercise involves 10 hydrant flow tests, then then used to implement a GA-based optimization model as
the optimization model uses the simulation model 30,000 explained next.
times for 100 generations of search requiring a total CPU time
of about 4 h. Although this may not be a great concern for a METHODOLOGY
calibration exercise, it certainly is a serious drawback when
Network of PCs
genetic optimization models are used for other purposes such
as real-time operation of water distribution systems. The network of PCs used for the present study was designed
Close observation of the proposed calibration model reveals based on Microsoft Windows 95/NT native peer-to-peer net-
that it is highly amenable to distributed computing. Each gen- working protocol. Contrary to a server-based network (such as
eration of the genetic search comprises several decision vec- Novell Network), a Peer-to-Peer network does not demand a
tors (usually between 20 and 50) that need to be evaluated for dedicated server and sets no hierarchy among computers. This
their objective functions independently. In serial computing, is a major advantage of using a peer-to-peer network instead
the objective functions are evaluated considering one decision of a server-based network. Absence of a designated server also
vector at a time. In models based on GAs, the objective func- eliminates the requirement for more expensive components
JOURNAL OF COMPUTING IN CIVIL ENGINEERING / JULY 2000 / 201
FIG. 2. Parallel Implementation of Optimal Calibration Model
such as a large hard disk space, extra memory (RAM), etc. the main-processor or on the sub-processors. It may be noted
The only additional components required to build a peer-to- that a preference was neither intended nor warranted when
peer network are the network cards, connecting cables, and designating a particular PC as a main-processor.
preferably a network hub. Because a peer-to-peer networking When the distributed calibration program is initialized on
capability comes standard with the popular Microsoft Win- the PCNet, all N computers read the basic data pertaining to
dows 95/NT operating system, the cost of buying network the distribution system model being calibrated and perform a
software was also eliminated. preliminary hydraulic analysis that initializes certain variables
The present study employed a ‘‘star’’ type topology in set- for subsequent optimization runs. Next, the main-processor
ting up the computer network. Topology refers to the arrange- randomly generates an initial set of decision vectors, retains
ment or physical layout of computers, cables, and other com- some of the decision vectors, and sends the remaining to the
ponents in the network. If the computers are connected to sub-processors. The main-processor then evaluates the objec-
cable segments that branch out from a single point or hub, the tive function values for the decision vectors that it retained.
topology is known as a star. The advantage of star topology At the same time, the sub-processors evaluate the objective
is that if one of the computers or the cable that connects it to function values for the decision vectors provided by the main-
the hub fails, only the failed computer will not be able to send processor. After completing an objective function evaluation
or receive network data. The rest of the network continues to for all the decision vectors, the sub-processors transmit the
function normally. A 16bit 10Base-T Ethernet network inter- objective function values back to the main-processor for fur-
face card was installed in each computer to be networked. This ther processing. The main-processor, after evaluating its own
was the only hardware modification needed for an ordinary share of decision vectors, gathers the output provided by the
PC to make use of it in a network. The computers with Eth- sub-processors and uses the collective information to produce
ernet cards were then connected in a star type topology using a new generation of decision vectors via crossover and mu-
standard 10Base-T connecting cables and a network hub. The tation operations. This process continues until a prespecified
following section illustrates the implementation of the distrib- convergence criterion has been satisfied. Fig. 2 illustrates the
uted calibration model on a network of N PCs. The network above process in the form of a flowchart.
of PCs hereinafter, will be referred to as the PCNet. One issue that has not been completely explored yet is load
sharing; that is, how to split the decision vectors among the
Implementation Strategy computers in the PCNet? In other words, how many decision
vectors should be retained by the main-processor, and how
Taking advantage of the built-in peer-to-peer networking ca-
many decision vectors should be sent to each sub-processor in
pability of the Windows 95/NT operating system, one of the
the PCNet? The present study, however, divided the decision
PCs in PCNet was designated as a main-processor and the
vectors based on the processor’s speed of each computer while
remaining N-1 as sub-processors. The PCs designated as sub-
allowing extra computational time to the main-processor to
processors receive input (decision vectors) from the main-pro-
perform the crucial crossover and mutation operations. More
cessor, process the data (evaluate the objective function for
information on load balancing issues and strategies may be
each decision vector), and transmit data back to the main-
found in Vatsa and Wedan (1999) and Woo et al. (1997).
processor. On the other hand, the PC designated as a main-
processor not only performs the genetic search to generate
APPLICATIONS
decision vectors but may also evaluate objective function val-
ues for some of the decision vectors. In addition, the main- Applicability and robustness of the proposed parallel opti-
processor sends the decision vectors to the sub-processors and mal calibration model is demonstrated by applying the model
receives the output corresponding to those decision vectors to calibrate three different water distribution systems, A, B,
from the sub-processors. This process was automated by writ- and C. The computer network fabricated for the present study,
ing/reading information (decision vectors, objective function PCNet, was composed of three PCs, X, Y, and Z. PC X had
values, etc.) to and from a few scratch files located either on a 166-MHz Pentium processor, and Y and Z had 133-MHz
202 / JOURNAL OF COMPUTING IN CIVIL ENGINEERING / JULY 2000
Pentium processors. The primary operating system on all three applying the calibration model. Pipelines in the network were
machines was Microsoft Windows 95. Computer Y was des- divided into four groups based on their pipe material and di-
ignated as a main-processor and the remaining two as sub- ameter. A roughness value was designated for each group, and
processors. Based on processing speed alone, computer X all pipes in the group were assigned that roughness value (Ta-
(166-MHz) will get 38% of the decision vectors, while com- ble 1). Fire flows were simulated making use of the KYPIPE
puters Y (133-MHz) and Z (133-MHz) get 31% each. To make hydraulic simulation model, and the corresponding residual
allowance for the extra computational effort required by the pressures were noted. Table 2 shows the simulated data. As-
main-processor, the present study reduced the burden to the suming group roughness values as the unknown decision var-
main-processor (computer Y) by 4% and apportioned that iables, the calibration model was applied to predict them using
equally to the other computers. This resulted in 40, 27, and the simulated flow-pressure data (Table 2). The predictions
33% share of decision vectors for the computers X, Y, and Z, from a good calibration model should be close to the actual
respectively. Although this division reduced waiting time for roughness coefficients with which the flow-pressure data were
the main-processor and sub-processors in sending/receiving simulated. Upper and lower bounds were imposed on group
data, further investigation is warranted in automating this pro- roughness coefficients as listed in Table 1. Group roughness
cess so that there will be a minimum unwanted waiting time values to be predicted by the calibration program that were
for the main-processor and sub-processors. The calibration already known did not have any influence on the optional
model based on a bilevel optimization framework developed search process other than in setting the upper and lower
in this study was installed on all three computers in the PCNet. bounds. The calibration model was designated to have 30 de-
First, the calibration model running stand-alone on individ- cision vectors in each generation.
ual computers was applied to calibrate the distribution sys- First, the calibration model, running on one computer at a
tems. The corresponding computational times were then com- time (i.e., all 30 decision vectors were processed by a single
pared to those required when the model implemented on computer), was applied to obtain the roughness coefficients of
PCNet was applied for calibration. The three water distribution the distribution system. The roughness coefficients predicted
systems employed differed in size and complexity. Distribution by the calibration model are presented in Table 1. As expected,
system A was composed of 52 pipes, system B was composed each run resulted in identical group roughness predictions after
of 435 pipes, and system C was composed of 850 pipes (Figs. going through 35 generations of genetic research. Residual
3, 4, and 5). System A represents a highly skeletonized model pressures computed by the calibration model are presented in
of an existing water distribution system. Systems B and C
represent existing water distribution systems, one serving a
rural community and the other serving a suburban community,
respectively.
The calibration exercise as applied to distribution system A
is described in detail whereas the others follow the procedure.
For distribution system A, being an example network model,
the data required for calibration (such as hydrant flow rates
and corresponding residual pressures) were simulated prior to
FIG. 3. Schematic for Distribution System A TABLE 1. Group Roughness Coefficients for Distribution Sys-
tem A
Roughness Coefficients
ABSTRACT: This paper aims to present an easy, quick, and accurate way of implementing value engineering
(VE) studies by introducing an integrated computerized system for the VE technique and life-cycle cost as a
checkout system. The VE computer program was designed using a database computer program, FoxPro Window.
The computerized version of the VE study consists of several interactive panels, representing the systematic
approach (job plan) of a VE study, through which all the information and data of the VE study will be input.
The program will help the user to come up with the best selected idea, based on certain calculations that will
be done automatically by the VE program. Moreover, the VE case studies that will be investigated by the
program will be stored, and thus the program will also serve as a data bank for these case studies as references
for future studies. The objective of this paper is not to create a smart VE system, but it can be considered a
major step toward this objective.
required to perform the function (i.e., cost related to the basic TABLE 8. Development Phase
function). For example, for a power plant the basic function
Implementation Plan
is to generate power. Expenses related to this will be worth,
and all other expenses will be just cost. The next step is to 1. How should it be implemented?
The proposed case e will be discussed with the manager of engineering
compare the overall system costs to the sum of the worth of and construction for his acquaintance and overcoming any road blocks
the basic functions. The result is defined as the cost-to-worth prior to management presentation
ratio. Based on experience in comparing cost, a cost-to-worth 2. What should be changed and in what sequence?
ratio >2 will usually indicate possibilities for removing cost The proposed case e will be discussed with the manager of engineering
within the project. Having completed the function analysis and and construction for his acquaintance and overcoming any road blocks
cost model for the project, the VE team is now ready to ven- prior to management presentation
3. Who should do it?
ture into the creative phase. In addition, the functional analysis Engineering department will amend required changes prior to package
technique (FAST) is developed in this phase. submittal to procurement department
The diagram shown in Fig. 1 was first used by an engineer 4. How long should it take?
in 1965 (Miles 1972). The FAST diagram is a tool that helps One month
organize the random listing of functions by answering the 5. Any deadline required?
questions How?, Why?, What does it do?, and What must it As per management report forecast
6. Implementation cost, internal hours?
do? This helps the VE team to develop many verb-noun func- 250 man-hours, total cost SR 5,000
tion solutions, which the FAST diagrams can help sort out and
208 / JOURNAL OF COMPUTING IN CIVIL ENGINEERING / JULY 2000
2. Sketch of the original and proposed design ings. This idea led to the development of VE incentive clauses
3. Description of the recommendation and what it involves (VEICs).
4. LCC analysis comparison A VEIC provides a method for the contractor to propose
5. Advantages and disadvantages that each recommenda- changes in contract plans and specifications that will lower
tion entails total costs without degrading performance, maintainability, or
6. Implications and requirements to implement the VE rec- reliability.
ommendation The description and application paragraph of a VEIC must
define the basic aims of the change proposal so that they are
The recommendation phase is the final step, and it is per- clear to the parties.
haps one of the hardest steps in the VE job plan. It is the
phase that involves the presentation of ideas, solutions, or VE in Saudi Arabia
changes in terms of benefits to the user. Recommendations are
essentially a challenge to the original design. The recommen- The technology of VE studies was transferred into the King-
dation phase can be summarized by the word ‘‘salesmanship.’’ dom of Saudi Arabia in the early 1980s. The Military Works
at the Saudi Arabian Ministry of Defense and Aviation was
LCC the first organization in the kingdom to apply the VE concept
to its projects. The general director of the General Directorate
The term LCC refers to the cost implications over the life of Military Works and a member of the Larry Miles Founda-
of the facility; i.e., owning, operating, and maintenance must tion Trust introduced the technique into the kingdom (Al-You-
be provided with the greatest value. It is defined by the Amer- sef 1992).
ican Institute of Architects as any technique that allows as- The year 1981 was the year that VE started to be imple-
sessment of a given solution, or choice among solutions, on mented in Saudi Arabia and began to cease to be an imported
the basis of considering all relevant economic consequences tool in the hands of foreign companies to be used in projects.
over a given period of time (or life cycle). It is further defined Now, internal regulations provide a limited service to other
as the total economic cost of owning and operating a facility, sections and offices of departments, and the signing of a con-
manufactured process, or product. tract for six American VE specialists to instruct and create VE
The LCC analysis reflects the present and future cost of the experts among Saudi engineers is proceeding.
project over its useful life. The two methodologies used to Value analysis should be applied to every government proj-
calculate the LCC are the annualized costs and the present ect, and it must be mandatory—VE may be performed in-
worth cost. house and/or by VE experts of VE private companies under
contract (Al-Otaishan 1985).
VE in Construction Projects The most effective tool for improving the quality of life and
the economy of the country, VE is the discipline of the future.
Designs of construction projects are complex. Regardless of
One cannot afford not to participate in it and to make it part
how capable a designer is, there will always be unnecessary
of our daily life.
cost hidden in his design. The very nature of construction de-
sign demands that countless variables be considered and pulled
together by a certain date. The designer is under pressure to RESEARCH METHODOLOGY
meet that deadline and must ensure that a viable, reliable plan The main objective of this study is to develop a computer
is formulated with that set of circumstances and that the design program to perform the VE analysis. This section discussed
grows and takes form. Once the project is formed after con- the methodology used in designing and building the VE com-
siderable effort, the designer becomes wed to the project as it puter program.
is and is perhaps unable to review it for unnecessary cost. The
VE team has an objective viewpoint, not having participated
in the original design. The VE team may lack the knowledge Selecting Standard VE Approach
of the original designer of the project, but they can still im- To date, various techniques have been used with VE studies.
prove it. Some studies use only five phases, whereas others might use
The designer must understand that unnecessary cost in a as many as nine phases. Analyzing each phase of the job plan
design is not a reflection of his abilities as a professional but in turn shows that some 22 individual techniques are employed
is rather a management problem that needs to be addressed. (Mudge 1988). However, the main VE job plan phases are the
Although there are numerous reasons for poor value, the fol- information phase, function phase, generation phase, and eval-
lowing seem to recur most often (Zimmerman and Hart 1982): uation phase.
To avoid any discrepancy, this study selected a standard VE
• Lack of time systematic approach and job plan. A flowchart of the job plan
• Lack of information used throughout the whole study is shown in Fig. 2. The flow-
• Meaning not clear chart shows a standard job plan that consists of five phases,
• Misconceptions including the information phase, functional phase, creative
• Temporary circumstances that inadvertently become phase, evaluation phase, and development phase.
permanent
• Habits Computerized VE System
• Attitudes
• Politics The computerized version of the VE study consists of sev-
• Lack of fee eral interactive panels through which all the information and
data of the VE study can be input. The computer panels layout
Contractors influence value with the quality of workman- of the VE program is shown in Fig. 2. This program is de-
ship. Their skill in construction and the integrity of their prod- signed for the evaluation of and comparison between one or
uct affects the cost of the facility for years into the future. The more projects to select the most appropriate one in terms of
resources and talents of the suppliers and contractors doing cost and other considerations selected by the user.
work especially for the management is an area for cost sav- In this program, the required information is distributed in
JOURNAL OF COMPUTING IN CIVIL ENGINEERING / JULY 2000 / 209
isting Project’’ to do any revision, modification, editing, etc.
The ‘‘Projects’’ screen is made up of a list that contains a list
of the previously saved projects.
The two following screens will be ‘‘Objectives Or Required
Criteria’’ and ‘‘Goals, Desired Criteria or Features.’’ They are
quite similar in design; each has 12 character fields and one
output area that displays the project title that is found on all
screens. Users are required to enter required and desired cri-
teria for the project. The data entered will be stored in the
corresponding data files.
The ‘‘Cost Data’’ screen (similar to Table 2) is designed to
represent information about the methods suggested for each
project in which the cost analysis of the original design as
well as all alternative designs is performed. The total cost con-
sists of unit price, material cost, labor cost, and installation
cost. Once the user modifies any of the editable data he has
to press the ‘‘Update’’ button, which activates a subprogram
to recalculate the unit price and the total price. Pressing
‘‘Done’’ ensures writing and saving data in their place in the
database files.
The next two screens, namely, ‘‘Present Function’’ and ‘‘Fu-
ture Function,’’ are exactly the same in their design. They ac-
tually are linked to two database files. The first is to add,
display, or edit data in that file, and the other is where some
verbs and nouns are stored.
The ‘‘Generate Ideas’’ screen is designed to provide the
right method or action to be taken. In this screen, two kinds
of displayed fields are shown. These are the ‘‘Study Title’’ and
‘‘Alternatives’’ that were originally entered in the ‘‘Cost Data’’
screen. The user has to add only one term such as ‘‘cover
with’’ or ‘‘build with,’’ and the program will copy it to all
alternatives in their corresponding boxes. The program was
designed to be as flexible as possible by allowing the user to
add more ideas manually on top of the ideas copied from the
‘‘Cost Data’’ screen. The additional ideas will be copied au-
FIG. 2. Job Plan Panels Used in Study
tomatically to the ‘‘Cost Data’’ screen. It is worth mentioning
here that determining the right number of ideas generated ac-
cording to the number of alternatives is done by using the
three groups of database files linked together by a common immediate ‘‘if’’ statement as follows:
field representing the project number.
The first group of database files includes basic information IIF (Empty (Cons-mg 1.particular)
about the project (information phase) such as data regarding
the project team members (study premise), objectives to be Then the ‘‘Feasibility Ranking’’ screen (Fig. 3) is displayed,
accomplished, materials, detailed cost for each material, sum- where each idea that is suggested for every alternative (partic-
mary, and implemented ideas. ular method) is evaluated by giving it a rank out of 10, the
The second group of database files handles the detailed total of these values is calculated (summed up), a ranking ma-
ideas, their advantages and disadvantages, and their rank com- trix is created and sorted (using a sort command), and the rank
pared to each other. The third group of database files is created (1) is given to the maximum value, and so on. This program
for the evaluation of the original and proposed methods using of creating a matrix, sorting, and giving an order or a rank is
weights assigned to each one. performed once the user presses the ‘‘Update’’ command but-
A fourth database file, which is made up of ‘‘memo’’ fields, ton, which is a subprogram included in the program that was
provides on-line help for each screen to guide the user to the developed for the screen.
function and details of each screen. Another file was also cre- The ‘‘Idea Comparison’’ screen (Fig. 4) is designed to write
ated to store information related to function identification, and up to three advantages and three disadvantages for each par-
it includes two character fields. One field is for the verb and ticular case. The study title, the different particulars, and their
the other is for the noun. corresponding rank are displayed as output fields (boxes), and
Upon execution of the VE program, a ‘‘Title’’ screen is dis- only editing is allowed for advantages and disadvantages (in-
played for a few moments. Then the ‘‘Main Menu’’ screen is put fields). One additional key button is added here to allow
displayed. Selection from the ‘‘Main Menu’’ is made through the user to select the number of best choices (ideas) to be
the use of three radio buttons. It has three options, ‘‘New Proj- displayed. Up to seven ideas can be selected. Activating this
ect,’’ ‘‘Revise An Existing Project,’’ or ‘‘Quit.’’ The program key displays a message on the screen, which is created by the
is interactive and all the screens are self explanatory. ‘‘wait window’’ command, and the user enters the number of
It should be pointed out that the screens that follow the main best choices so that the corresponding ideas are displayed.
screen might differ according to the user’s choice. For exam- The ‘‘Weights’’ screen [Fig. 5(a and b)] is designed to per-
ple, if the user selects ‘‘New Project,’’ the ‘‘Screens’’ panel form a comparison among all selected ideas or particulars. The
will be displayed for a few seconds to show the user the var- preferences are to be selected from a pop-up list with three
ious screens he is about to encounter for his VE study. Then fixed entries—major, medium, and minor preference—and
the ‘‘Study Premise’’ screen will be displayed. The ‘‘Projects’’ they are repeated between each of two ideas. This information
screen, however, will appear upon selecting ‘‘Revise An Ex- is stored in database files as an initial step for determining
210 / JOURNAL OF COMPUTING IN CIVIL ENGINEERING / JULY 2000
FIG. 3. Feasibility Ranking
their score on the ‘‘Scoring Matrix’’ screen [Fig. 5(a)]. In this 2, and 1, respectively. Each rank is multiplied by its weight,
screen all of the selected ideas are displayed, along with their and the subtotal is displayed in the total rank boxes. These
scores that were calculated from the ‘‘Assigned Weights’’ subtotals are summed up and listed in the ‘‘Total Ranking’’
screen [Fig. 5(b)], by assigning different scores corresponding column.
to each selected preference displayed as raw scores. Then nor- The next screen is the ‘‘Saving and Life-Cycle Cost’’ screen
malization of these scores is done and displayed in the as- (Fig. 7), and it is designed to compare the cost of the original
signed weight column. So, this is only an output screen, and and proposed ideas and to determine the percentage of savings.
the user does not add or modify any entry, because it is a It also performs the LCC analysis of both original and pro-
result of what he has decided and selected in previous screens. posed methods and demonstrates that savings can be achieved.
The final stage of this phase is in the ‘‘Evaluation Matrix’’ Both the original and proposed methods can be selected from
screen (Fig. 6), where all selected particulars, ideas, and their a ‘‘list from array’’ box, where all methods are copied into an
weights are displayed as output fields, and the user decides array, and the user selects the proper one from this array. The
the rank of each criterion against each of the particulars or ‘‘Calculate’’ button is then pressed to display the correspond-
attributes. Each selection is also made out of a pop-up list ing cost of each method by performing a subprogram that se-
including the choices ‘‘Excellent,’’ ‘‘Very Good,’’ ‘‘Good,’’ lects the value of cost corresponding to the method and stores
‘‘Fair,’’ and ‘‘Poor,’’ which are assigned the values of 5, 4, 3, it in its new field as an original or proposed cost by a series
JOURNAL OF COMPUTING IN CIVIL ENGINEERING / JULY 2000 / 211
FIG. 5. Determining Weights: (a) Scoring Matrix; (b) Assigned Weights
of ‘‘if’’ statements. Also the percentage of the difference be- 4. The expected life time (in years) n is to be input by the
tween costs is displayed, and the percentage of the difference user, as well as the salvage cost (if any).
is calculated and displayed as savings in a percentage. 5. The program then calculates the LCC in two steps:
Upon activating the ‘‘Life Cycle Cost Analysis’’ button, a a. Present work value
subprogram is executed to analyze the LCC for the selected
idea. The LCC analysis is based on the summation of equiv- PW = initial cost ⫹ annual fixed cost(P/A, i%, n)
alent worth of installation, labor, operation, and other costs for ⫹ escalated annual cost(P/A, i1%, n) ⫺ salvage cost
each alternative over its useful life and permits the alternative
having the minimum LCC to be identified. Calculating the b. Annual LCC
LCC in the subprogram is performed as follows (Degarmo et
al. 1984): AW(SR/year) = PW(A/P, i%, n)
The LCC is calculated twice, once for the original method
1. Initial project cost is calculated (installation, material, la- and once for the proposed method. All terms of these equations
bor, etc.). This is obtained from the database file. are expressed as memory variables, and when the LCC per
2. Additional cost is calculated for annual expenditures year is calculated, it is stored in the database file. Two values
(e.g., maintenance cost) with the suitable (determined) are saved, one for the original method and the other for the
interest rate. Both the value and the interest rate are to proposed one.
be input by the user. The program is well designed to give the summary of the
3. Other costs that may have an escalation rate are to be cost analysis with its ‘‘Summary’’ screen and suggestions for
determined. Their interest rate and the escalation rates the implementation of the project with its ‘‘Implementation
are also expected from the user. Plan’’ screen. They are two simple edit screens with scrollable
212 / JOURNAL OF COMPUTING IN CIVIL ENGINEERING / JULY 2000
FIG. 6. Evaluation Matrix
edit boxes to enable the user to write the required information. A VE study was applied to the ground-covering system. The
Finally, using the ‘‘Print All’’ screen, all the information en- information phase included objectives or required criteria,
tered and other details can be printed. goals (i.e., desired criteria or features), and cost data. The
The information included in all database files is shown functional phase consisted of answers to two questions, What
through a series of panels representing the different phases, does it now do? and What must it do? Eleven alternatives were
which includes the information phase, functional phase, crea- proposed in the creative phase. In the judicial phase, the al-
tive phase, judicial phase, and development phase. Each phase ternatives have been tested and ranked against six factors.
is divided into several panels. Those factors were aesthetics, cost to develop, probability of
implementation, time to implement, potential cost, and re-
Case Study: VE of Road Median (Al-Jubail) placement cost. Six of the alternatives, which had high rank-
The Royal Commission in Jubail (Royal Commission 1988), ings, were again evaluated by comparing their advantages and
Saudi Arabia, decided to do modification toward beautification disadvantages. Weights were assigned to each goal (i.e., de-
of some major roads. The work included landscaping, irriga- sired criteria), which were then evaluated against each other
tion, and sidewalk and median paving. The designers recom- using the evaluation matrix. The evaluation matrix showed that
mended the following: Paving the median from the nose of the colored patterned cast-in-place concrete was the best al-
the median to a distance of 150 m with bricks and the rest ternative and that it would save about $600,000, equivalent to
with aggregate; and that sidewalks be designed to be covered 38% of the original cost.
with aggregate. The design has specified paving to be common bond paving
JOURNAL OF COMPUTING IN CIVIL ENGINEERING / JULY 2000 / 213
system, red in color, and meeting the following ASTM stan- This paper provides information to develop an easy and
dard: smart system that will help in VE studies. This study recom-
mends a continuation in this direction. The proposed smart
• Concrete building bricks—C55 system must take into account the various factors that have to
• Sampling and testing bricks and structural tiles—C67 be investigated when performing a VE study. The smart sys-
tem must be so flexible that it can accommodate various types
Trees are designed to be 8.0-m apart in the median and of projects in various fields. In other words, the smart system
sidewalks. The total number of bricks to be used in this project must overcome all the limitations that this program suffers.
is estimated to be 20,000/m2. This report will examine the ACKNOWLEDGMENT
ground-covering system and recommend the most suitable sys-
The writers would like to thank King Fahd University of Petroleum
tem using VE techniques. The VE application to the above
and Minerals for providing various facilities for conducting this study.
project is shown in detail in Tables 1–8. Computer screens
showing ‘‘Feasibility Ranking,’’ ‘‘Idea Generation,’’ ‘‘Scoring APPENDIX. REFERENCES
Matrix,’’ ‘‘Assigned Weights,’’ ‘‘Evaluation Matrix,’’ and
Corcoran, K. R. (1981). Saudi Arabia, keys to business success, McGraw-
‘‘Savings’’ are given in Figs. 3–7, which are part of the above Hill, New York, 16–30.
example. All the screens are self-explaining and interactive. Degarmo, E. P., Sullivan, U. G., and Canada, J. R. (1984). Engineering
economy, 7th Ed.
CONCLUSIONS Miles, L. D. (1972). Techniques of value analysis and engineering, 2nd
Ed., McGraw-Hill, New York, 1–24.
The VE computerized model can help greatly in teaching Mudge, A. E. (1988). Value engineering, a systematic approach, 5th Ed.,
30.
the concept and application of VE. It will guide users of the Al-Otaishan, A. T. (1985). ‘‘Establishing a Saudi Arabian value engi-
technique, and it will speed up the process by handling all neering program.’’ Proc., 2nd Saudi Engrs. Conf., KFUPM, Dhahran,
calculations automatically and in interaction with each other. Saudi Arabia, 373–396.
The VE computerized model will encourage people to use the Parker, D. E. (1985). Value engineering theory, Lawrence D. Miles Value
VE technique, because it will simplify the required procedure. Foundation, Washington, D.C., 62–64.
Another, and probably the most important, feature of this pro- Royal Commission at Jubail. Rep. on Value Engrg. Study, Jubail, Saudi
Arabia.
gram is that it will serve as a database of previous VE case Sadi, A., Al-Musallami, A. I., and Al-Sughaiyer, M. (1996). ‘‘Value en-
studies. The intent was not to develop a smart system, because gineering in public construction projects in Saudi Arabia.’’ Build. Res.
it is beyond the scope of this study; however, the system was and Information, 24(3), 152–156.
designed in such a way that it could be modified and enhanced Al-Tamimi, R. A. (1996). ‘‘A computerized system for the value engi-
in the future into a smart system. The VE computer program neering technique in Saudi Arabia.’’ MS thesis, King Fahd University
was designed using a database computer program, FOXPRO of Petroleum and Minerals, Dhahran, Saudi Arabia.
Al-Yousef, A. (1992). Value engineering, concept and technique, 1st Ed.,
Window. The computerized version of the VE study consists 27.
of several interactive panels, representing the systematic ap- Zimmerman, L. W., and Hart, G. C. (1982). Value engineering, a practical
proach (job plan) of a VE study, through which all information approach for owners, designers and contractors, Van Nostrand Rein-
and data of the VE study will be input. hold, New York, 1–31.
TECHNICAL PAPERS
Interactive Analysis of Spatial Subsurface Data Using GIS-Based Tool
R. L. Parsons and J. D. Frost . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 215
Design Processing of Regular and Nonstandard Structural Components
Mustafa K. Badrah, Iain A. MacLeod, and Bimal Kumar . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 223
Parametric Coordinator for Engineering Design
Chee-Kiong Soh and Zhonghui Wang . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 233
Estimating Labor Productivity Using Probability Inference Neural Network
Ming Lu, S. M. AbouRizk, and Ulrich H. Hermann . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 241
Fuzzy Logic Integrated Genetic Programming for Optimization and Design
Yaowen Yang and Chee Kiong Soh . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 249
Multireservoir Systems Optimization Using Genetic Algorithms: Case Study
Mohd Sharif and Robin Wardlaw . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 255
Computer-Based Application for the Processing of Clients’ Requirements
J. M. Kamara, C. J. Anumba, and N. F. O. Evbuomwan . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 264
TECHNICAL NOTE
Applications of Mathematical Software Packages in Structural Engineering
Education and Practice
Primus V. Mtenga and Lisa K. Spainhour . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 273
END-OF-YEAR INDEX . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 279
INTERACTIVE ANALYSIS OF SPATIAL SUBSURFACE DATA USING
GIS-BASED TOOL
By R. L. Parsons1 and J. D. Frost,2 Members, ASCE
ABSTRACT: Much of the spatial information that is an inherent part of geotechnical and geoenvironmental
investigation programs has been underutilized due to the unavailability of analysis tools designed to take ad-
vantage of the spatial attributes of the information. This paper describes a system developed to provide a
quantitative estimate of the adequacy of coordinate-based information used to make particular decisions, with
particular emphasis on geotechnical and geoenvironmental site investigation information. The system was de-
veloped using a geographic information system (GIS) as a platform that was integrated with an existing com-
mercial geostatistical software package to allow for more comprehensive analysis and optimization capabilities.
冕
z 2
e⫺z /2
Output Generation
The ASCII file generated by Geo-EAS may be converted
into ArcInfo line coverages containing expected value or stan-
dard deviation contours, or into polygon coverages containing
probability or thoroughness polygons. For probability and
FIG. 5. Flowchart Shows Procedure for Evaluating Sensitivity thoroughness coverages, the initial ASCII file is read and the
of Data Set to Changes in Data Values and for Optimizing Addi- probability and thoroughness values are calculated. For each
tional Sampling Resources [after Parsons (1998)] grid cell, after conversion of the relative standard deviation
value to the equivalent absolute standard deviation value (if
been converted for use on Windows NT. The process of anal- required), the normal distribution defined by the expected
ysis is shown in the flowchart in Fig. 5. The system is designed value and standard deviation value is converted to the standard
to operate using menus and virtually eliminates the use of the normal distribution (Dowdy and Weardon 1983). The proba-
ArcInfo command line. GIS-ASSESS is designed to accept bility the actual value is greater than (or less than) the bound-
any type of data that can be represented as a numerical value ary value, b, set by the user is then calculated by solving the
with x, y or x, y, z coordinates. The data may be imported into integral shown in (3) using the regular polygon method (Swo-
GIS-ASSESS from pairs of text files generated in a spread- kowski 1984). If a thoroughness coverage is desired, the val-
sheet or word processor. Specific options have been included ues are calculated at this time. The resulting probability or
to automate the importation of data in selected formats. thoroughness values are then written to an ASCII file com-
The procedure for importing the data into ArcInfo begins patible with ArcInfo and converted to a polygon coverage. The
by generating a point coverage using the first text file. The system is designed such that values are grouped into 10 classes
point coverage is a set of files that contains the information with values 0–10, 10–20 . . . 90–100. Each class is repre-
about the test locations and identification information. The sec- sented by a different color or shading.
ond text file is then imported into a lookup table (LUT). Cre- The polygon coverage that is generated begins as a rectan-
ation of a LUT is automatic for data in one of several standard gular coverage due to restrictions imposed by the geostatistical
formats; LUT tables may also be generated manually for other software. Therefore a module has been included that enables
data sets with relatively little effort. The information in the the user to clip the rectangle using a polygon coverage con-
220 / JOURNAL OF COMPUTING IN CIVIL ENGINEERING / OCTOBER 2000
taining the investigation boundary, such that the coverage of approach to provide meaningful results. For sites where the
investigation adequacy conforms to the boundaries of the in- target parameter may not exist throughout the investigation
vestigation area. area, Parsons (1998) showed that for a site with targets that
were approximately circular, the sample spacing for a grid pat-
System Output tern must be approximately 45% of the target diameter to pro-
duce continuous thoroughness images of moderate quality, and
The coverages generated may be viewed from within the 23% of the target diameter to produce images of good quality.
ARCPLOT module. Menus are provided that enable the user This sampling density is not economical for most geotechnical
to display and overlay polygon, line, and point coverages. projects.
Control of line color, legend display, and text font is provided. For those sites where the sample spacing is too great for the
Grid files containing the actual probability or thoroughness continuous thoroughness approach to be used, GIS-ASSESS
values may also be viewed in 3D from any angle and elevation
at any magnification.
DISCONTINUOUS THOROUGHNESS
Geotechnical investigations are normally performed to de-
termine the value of geotechnical parameters that exist
throughout the subsurface: shear strength and permeability, for
example. The parameter values may vary, but values for them
exist throughout the subsurface. Geotechnical investigations
may also include searches for targets that have discrete bound-
aries such as karst features, perched water, and steel drums.
When attempting to locate this type of target, the sample den- FIG. 7. Calculation of Effective Detection Area Given Tool In-
sity must be sufficiently high for the continuous thoroughness vestigation Radius and Critical Target Size Radius
ACKNOWLEDGMENTS
The research described in this paper has been supported by NSF Grant
No. CMS 9457549. This support is gratefully acknowledged.
APPENDIX. REFERENCES
Adams, T. M. (1994). ‘‘GIS-based subsurface data management.’’ Micro-
computers in Civ. Engrg., 9(4), 305–313.
ASTM. (1998). ‘‘Standard guide to site characterization for engineering
design and construction purposes.’’ D 420-93, West Conshohocken, Pa.,
FIG. 8. Effective Sampling Area for Grid Pattern (Circles) (Tar-
1–7.
get Centroids Are Shown as Dots)
Bowman, D. (1998). ‘‘Civil engineering data meets GIS.’’ J. Comp. in
Civ. Engrg., ASCE, 12(1), 5–7.
has an alternative procedure, referred to as discontinuous thor- Dowdy, S., and Wearden, S. (1983). Statistics for research, Wiley, New
oughness, which is based on site investigation work by Tang York.
(1987) and the original work on thoroughness performed by Englund, E., and Sparks, A. (1991). ‘‘Geo-EAS 1.2.1 user’s guide,’’ EPA
600/8-91/008, U.S. Environmental Protection Agency, Washington,
Tsai (1996) and Tsai and Frost (1996a,b; 1999). Under this D.C.
approach, the adequacy of the investigation is based on the Gilbert, R. B., and McGrath, T. C. (1998). ‘‘Design of site investigation
percentage of the area of interest that is investigated. programs for geotechnical engineering.’’ Uncertainty in modeling and
For some investigations, the depth of the targets has an im- analysis in civil engineering, CRC Press, Boca Raton, Fla.
pact on the adequacy of the investigation because the area Harr, M. E. (1987). Reliability-based design in civil engineering, Mc-
investigated changes for some tools with depth, while for oth- Graw-Hill, New York.
Honjo, Y., and Kuroda, K. (1991). ‘‘A new look at fluctuating geotech-
ers it is constant. For example, tools such as borings or cone nical data for reliability design.’’ Soils and Found., Tokyo, 31(1), 110–
penetrometers sample a cylindrical volume with a constant di- 120.
ameter and a depth that is slightly greater than the maximum Isaaks, E. H., and Srivastava, R. M. (1989). An introduction to applied
depth of the tool. These tools are referred to within GIS- geostatistics, Oxford University Press, New York.
ASSESS as cylindrical tools [Fig. 6(a)]. Other tools, such as Journel, A. G., and Huijbregts, C. J. (1978). Mining geostatistics, Aca-
a magnetometer, sample a volume that can be approximated demic, San Diego.
Luna, R., and Frost, J. D. (1998). ‘‘Spatial liquefaction analysis system.’’
as a sphere (Tsai 1996). For these tools, the area sampled J. Comp. in Civ. Engrg., ASCE, 12(1), 48–56.
changes with depth, as shown in Fig. 6(b). For situations McBrayer, M. C., Wiley, K. E., and Gilbert, R. B. (1998). ‘‘SimSite:
where the target size is known or where targets smaller than Computer model to simulate site investigation for groundwater reme-
a given size may be ignored, an effective sampling radius may diation.’’ Proc., 1st Int. Conf. on Site Characterization, P. K. Robertson
be determined by combining the tool search radius and the and P. W. Mayne, eds., Balkema, Rotterdam, The Netherlands, 189–
minimum target radius, as shown in Fig. 7. 194.
Myers, J. C. (1997). Geostatistical error management, Van Nostrand
Using the test locations, sampling radius, depth of interest, Reinhold, New York.
and minimum target radius, GIS-ASSESS automatically gen- Parsons, R. L. (1998). ‘‘Assessment and optimization of site characteri-
erates an image showing the area investigated and the thor- zation and monitoring activities using geostatistical methods within a
oughness of the investigation. An example of the output gen- geographic information system environment.’’ PhD thesis, Georgia In-
erated using this method is shown in Fig. 8. For this example, stitute of Technology, Atlanta.
the main radius of 133 targets was averaged and combined Parsons, R. L., Frost, J. D., Chameau, J.-L. A., and Tsai, Y. C. (1998).
‘‘GIS-ASSESS: A spatial analysis tool for site investigation planning
with the search radius of the tool to determine the effective and evaluation.’’ Proc., 1st Int. Conf. on Site Characterization, P. K.
detection area for an individual sample. The sample locations Robertson and P. W. Mayne, eds., Balkema, Rotterdam, The Nether-
(circles) are shown with the centroids of the targets (dots). A lands, 251–256.
target was determined to be ‘‘found’’ if the effective detection Star, J., and Estes, J. (1990). Geographic information systems: An intro-
area for a given sample intersected the centroid of a target. duction, Prentice-Hall, Englewood Cliffs, N.J.
Swokowski, E. W. (1984). Calculus with analytic geometry, Prindle, We-
ber & Schmidt, Boston.
CONCLUSIONS Tang, W. H. (1987). ‘‘Updating anomaly statistics—Single anomaly
case.’’ Structural safety, Elsevier Science, Amsterdam, 151–163.
The spatial analysis tool GIS-ASSESS represents a signifi- Tsai, Y. C., and Frost, J. D. (1996a). ‘‘A spatial thoroughness quantifi-
cant contribution to many areas of civil engineering by pro- cation system for environmental site characterization.’’ Proc., 3rd Int.
viding a tool for the analysis of the adequacy of characteri- Symp. on Envir. Geotechnol., Technomic, Lancaster, Pa., 967–976.
zation data sets. The analysis is based on geostatistical Tsai, Y. C., and Frost, J. D. (1996b). ‘‘Reducing uncertainty in environ-
mental site characterization.’’ Proc., 1996 Conf. on Uncertainty in Ge-
methods, and the results are displayed in an easily understood ologic Environment, ASCE, New York, 58(2), 1019–1033.
visual format using the capabilities of GIS. This more sophis- Tsai, Y. C. (1996). ‘‘A quantitative spatial thoroughness methodology for
ticated analysis tool can enable the user to evaluate the sen- environmental site characterization.’’ PhD thesis, School of Civ. and
sitivity of the investigation to changes in data values or the Envir. Engrg., Georgia Institute of Technology, Atlanta.
effect of additional sampling, which may yield unexpected val- Tsai, Y. C., and Frost, J. D. (1999). ‘‘Using GIS and KBS technology for
ues. By performing such analyses, the user may gain a better real-time planning of site characterization activities.’’ Can. Geotech. J.,
Ottawa, 36(2).
understanding of what conclusions may be drawn from the Wild, M. R., and Rouhani, S. (1996). ‘‘Effective use of field screening
data and a measure of how reliable those conclusions are. This techniques in environmental and geotechnical applications.’’ Geostatis-
approach also enables the user to better determine the number, tics for environmental and geotechnical applications, ASTM STP 1283,
type, and location of additional tests in the office prior to the ASTM, West Conshohocken, Pa., 88–101.
ABSTRACT: A design standard can be complemented with a case base of nonstandard designs to form a richer
knowledge source that can be used in the development of a system for processing of both regular and nonstandard
structural designs. This leads to the question of classification of a design context to regular or nonstandard,
which relates to the interpretation of the standards scope. This paper describes a software system, called SADA-
II, which has been developed to address the issues of standards scope characterization and nonstandard com-
ponent design processing. SADA-II also introduces a method for processing of standards provisions that are
written in a generic format (e.g., pertaining to different section shapes) and explores the potential of integrating
the system with a graphics package and other CAD applications. Illustrative examples for the application of
SADA-II to detail structural design are provided.
DEVELOPMENT STRATEGIES
Standards Model (SM)
A design standard can be viewed as a collection of provi-
sions given for specific design issues regarding different types
of design entities. A design issue, here, is a topic in the design
standards involving specific requirements that need to be sat-
isfied during the design synthesis or conformance checking
processes and that may be relevant to a number of design
entities. A design standard is organized in the Standard Re-
quirement Object Hierarchy as a number of classes at different
levels (Fig. 2). Member consideration is an abstract class for
the different member types in a standard, e.g., beam, column,
slab, and base. Each member will have provisions regarding
FIG. 1. SADA-II Architecture
different Design Issue Groups (DIGs), e.g., the ultimate limit
Functionality
SADA-II contains the following modules (Fig. 1): FIG. 2. Standards Model
ILLUSTRATIVE EXAMPLES
Example 1: Standards Applicability Checking of
Inverted-T Beam
The main purpose of this example is to illustrate standards
applicability checking using SADA-II. In this example, stan-
dards applicability is to be checked for an inverted-T precast
concrete beam. The selected standard is BS 8110 (Structural
1985). Beam B8 (Fig. 5) is a part of a three-story parking
garage made of precast concrete. To find out if the design
context is within the scope of the standard, the user should
select appropriate attribute values for data related to the design
FIG. 4. Two-Span Beam Example: (a) Elevation; (b) Cross Sec- context and needed to fire relevant scope rules. These are en-
tions tered through dialogue boxes, such as that shown in Fig. 6,
FIG. 5. B8 Drawing (Inverted-T Beam): (a) Longitudinal Section; (b) Section A-A
which is provided with a commentary to help the user select Matching Process (Design Rules Establishing Module)
the appropriate values. The applicability results would be is-
sued in a report such as that shown in Fig. 7. The applicability The matching process of the design context B8 is carried
of B8 to BS 8110 regarding general scope and member clas- out using the Design Rules Establishing Module. This involves
sification and definition was WSS. The design issues selected selecting attribute values for the design context via dialogue
for this design context are bending moment resistance, shear, boxes, as in the standards applicability checking. The matching
torsion, flexural cracking, and deflection. The applicability of process resulted in exact matching with Case310 (Fig. 8), so
BS 8110 provisions related to bending moment resistance and B8 can be processed using the Design Processing Module.
deflection was PAG. The applicability results of shear and tor-
sion were OSS. The applicability result regarding flexural Design Processing (Component Design Processing Module)
cracking was WSS. Applicable provisions are specified and the
Sizing or conformance checking of a component using the
general classification of the design context B8 was nonstan-
Component Design Processing Module involves two main ac-
dard, since no provisions were found in BS 8110 for some
tivities: data input and processing. The processing results of
design issues.
the Strut and Tie Forces (STF) issue related to the nib part of
B8 are shown in Fig. 9. This shows STF requirements defined
Example 2: Design Synthesis of Nonstandard in the subcase (Scase31212) and their firing results. For ex-
Component (Inverted; T Beam) ample, ‘‘C.6.3.2<<req1-tie-force: ten-boot -links = 292.17’’ re-
fers to the requirement for resisting horizontal tie force.
The same design context described in the previous example C.6.3.2 is the relevant section in the Design Methodology
is used to illustrate the matching and design processing of a Document (see the section entitled ‘‘Case Model’’). And ‘‘ten-
nonstandard component. boot -links = 292.17’’ states the area of the boot links in mm2.
228 / JOURNAL OF COMPUTING IN CIVIL ENGINEERING / OCTOBER 2000
FIG. 8. Matching Results of B8
Table 1 shows the processing results related to the lateral buck- lected column has a circular cross section, the same procedure
ling and deflection of B8, bending moment for section 2, shear explained in this section can be used for any section shape.
for section 1 (Fig. 5), and shear and STF for the nib part. Such The Component Design Processing Module needs to be run
results can be used to synthesize the beam reinforcement, as for the design synthesis of C3. C3 is first instantiated under a
is shown in Fig. 10. relevant PM object by entering preliminary design data such
as dimensions. Since designing the column against axial load
Example 3: Design Synthesis of Circular Column and uniaxial bending moment for this context requires exe-
Using Generic Functions cution of a generic function, the user is referred to use the
external function ‘‘UBend⫹AxialLoad’’ in AutoCAD. The lat-
The main purpose of this example is to illustrate the use of ter is developed for column design against uniaxial bending
a Generic Function for design synthesis. In this example, col- moment and axial load, based on the generic provision 4.3.1.2
umn C3, which is a part of a reinforced concrete frame and of EC2 (EC2 1992), and can be executed from the AutoCAD
subject to an axial load and uniaxial bending moment, needs extended standard toolbar (Fig. 11). The user needs to draw
to be designed according to EC2 (Fig. 11). Although the se- the column section with an initial selection of reinforcement
JOURNAL OF COMPUTING IN CIVIL ENGINEERING / OCTOBER 2000 / 229
TABLE 1. Processing Results for Beam B8
Section Design issue Requirements Requirement processing results
(1) (2) (3) (4)
Section 2 Bending moment Req. 1: BM reinforcement Tension reinforcement area = 2,454 mm2
Req. 2: Lever arm limit Satisfied
Req. 3: Neutral axis limit Satisfied
Section 1 Shear Req. 1: Maximum shear stress Satisfied
Req. 2: Concrete shear stress 0.86 N/mm2
Req. 3: Shear links area-upon-spacing 1.0 mm2/mm
Whole component Lateral buckling Req. 0: Slenderness limit Satisfied
Deflection Req. 0: Deflection Satisfied
Boot Shear Req. 1: Maximum shear stress Satisfied
Req. 2: Shear strength without reinforcement Satisfied
Strut and tie forces Req. 1: Tension Boot links Link reinforcement area = 292.17 mm2
Req. 2: Concrete strength under strut force Satisfied
Req. 3: Load transfer reinforcement Load transfer reinforcement area = 524.75 mm2
ABSTRACT: Coordination of design information is essential for the design team to provide accurate technical
documents for the actual construction. Design changes are inevitable during any ordinary building project. If
the design changes are not properly managed, design conflicts will be created, which will result in more expen-
sive design or even failure in the design-construction process. However, little research has been done to address
the issue and thereby provide tools for the management of design changes. This paper presents a novel approach
that can facilitate the coordination of design information through managing design changes with the help of a
parametric coordinator. Design changes of dimensions between different building components are specially
discussed. The proposed parametric coordinator provides each building component with the linking knowledge
that is described as a dimensional property and a reference property. A group design method is applied to check
the design consistency. Illustrative examples are presented to demonstrate the feasibility of the proposed para-
metric approach for design change management.
software AutoCAD (Autodesk 1995) via DXF file is also example, a component drawing with dimensions is regarded
available in the system. as a geometric model. However, a geometric model contains
only the geometric informations, such as coordinates and di-
Knowledge-Based Constraint Solver mensions, within itself. Such a model does not have any
The proposed knowledge-based constraint solver employs knowledge on how it is related to the other geometric models.
the rule-based graphic reasoning method (Lee and Kim 1996) Thus, additional attributes or properties with the linking
to deduce the geometric configuration, which is associated by knowledge should be attached to facilitate relation control of
the constraints. The built-in knowledge base is made up of the models.
meta-level rules, reasoning rules, and entity-connection pat- In order to clearly describe how the current model’s dimen-
terns. With the constraint solver, constraints can be treated as sions are related to the referred models, the current model
facts to be processed, and a pattern-matching mechanism is should have information on the referred models’ IDs, dimen-
used to search the constraints and fire corresponding rules to sion names, dimension values used, and the way they are ap-
infer the geometric entities. A construction plan is used to re- plied. The source model should provide the unique information
cord the inference procedure. The parametric model is then data, which can be easily accessed by the referring model.
determined by sequentially evaluating the rules in the con- Here the concepts of dimensional property and reference prop-
struction plan together with a set of dimension values. erty are proposed for this purpose. When a geometric model
Take the simple sketch shown in Fig. 2(a) as an example. is attached with a dimensional property and a reference prop-
This sketch can be interpreted by the constraint solver as a erty, such a model is regarded as a constraint model. The terms
conceptual model, as shown in Fig. 2(b), where Pi(i = 1, 6) ‘‘dimension property’’ and ‘‘reference property’’ are defined
represents the point entities, Li (i = 1, 4) the line entities, A1 below.
the arc entity, and di (i = 1, 5) the dimension constraints. The Definition 1: A dimensional property of a constraint model
explicit constraints such as dimensions, and the topological is a document object that records the dimensions’ names of
constraints such as tangency, perpendicular, etc., are recog- the model, the dimensions’ current values and formula rela-
nized by the constraint solver. Examples of representation of tions, and the dimensions’ indicators for the used reference
the constraints as predicate symbols are listed as follows: dimensions. It provides an accessible constraint data that may
be inferred to by the other constraint models.
• On P L: point P is located on line L Definition 2: A reference dimension in a constraint model is
• P㛭fixed P: the location of point P is fixed actually a dimension in another model but is referred to by the
• Tangent L C: line L is tangent to circle C current constraint model.
• Vcoordinate P1P2d: the difference of Y-coordinate be- The reference dimension is used to establish the relationship
tween Points P1 and P2 is d. between different models so as to exert the influence of a
• Vdirection L: the direction of line L is vertical global constraint. With the descriptions of the corresponding
reference dimensions in the referring models, the propagation
An illustrative rule to infer the constraint state of point P2 of design changes along all the related constraint models can
based on point P1, line L1 and dimension d1, is described be- be guaranteed.
low: Definition 3: The descriptions of a set of reference dimen-
sions used by a constraint model are recorded in its reference
P㛭fixed P1
property, which depicts the referred constraint models’ names,
Vcoordinate P1P2d1
dimensions’ names and values, and their representing symbols
On P1L1
in the current constraint model.
On P2L1
Since the dimensional property of a constraint model pro-
Vdirection L1
vides the accessible information data that can be referred to
→ P㛭fixed P2
by the other constraint models, while the reference property
When the dimensions are changed, for example, d1 and d2 are builds the link from the current model to the referred con-
changed to ‘‘72,’’ d3 to ‘‘20,’’ d4 and d5 to ‘‘60,’’ the constraint straint models, the data link can be completely set up for a
solver will start to process the geometric objects using the group of related constraint models. The dimensional property
constraints stored in the conceptual model. The successfully and the reference property are created by a dimension table
inferred rules constitute a construction plan. By evaluating the generator and reference table generator, respectively, which
construction plan with the set of changed dimensions as the will be described in the section on Illustrative Examples.
input data, a new model is generated, as shown in Fig. 2(c).
Group Design Method
Models and Properties for Coordination
In this paper, a model refers to a geometric component or Managing design changes focuses on two aspects: (1) the
part that is made up of geometric entities and constraints; for design change determination, and (2) the design change prop-
JOURNAL OF COMPUTING IN CIVIL ENGINEERING / OCTOBER 2000 / 235
agation. Such considerations are carried out in a model group
pattern.
ABSTRACT: This paper discusses the derivation of a probabilistic neural network classification model and its
application in the construction industry. The probability inference neural network (PINN) model is based on the
same concepts as those of the learning vector quantization method combined with a probabilistic approach. The
classification and prediction networks are combined in an integrated network, which required the development
of a different training and recall algorithm. The topology and algorithm of the developed model was presented
and explained in detail. Portable computer software was developed to implement the training, testing, and recall
for PINN. The PINN was tested on real historical productivity data at a local construction company and compared
to the classic feedforward back-propagation neural network model. This showed marked improvement in per-
formance and accuracy. In addition, the effectiveness of PINN for estimating labor production rates in the context
of the application domain was validated through sensitivity analysis.
puts at Output Layer’’ subsection includes details about Output Zone Setup
the NN outputs.
As discussed in the section ‘‘Introduction of PINN Model,’’
6. Check the NN outputs against the actual outputs of the
the likely range of output values is evenly divided into a num-
training and testing records. If the results are satisfactory,
ber of output zones. The output zone boundary setup is im-
then the NN is declared to have been trained; otherwise,
portant for PINN learning and recall. Wide zones are generally
repeat Stage 1 using different parameters at each layer.
not helpful to the decision maker and hence should be avoided.
Zones that are unacceptably tight may prevent PINN from
Stage 4. Once the model calibration is done, the NN can learning. It requires some trials to obtain reasonable output
be used to recall the output value for any given input vector, zone boundaries, and the following two aspects should be con-
which is similar to Stage 3 using the final results determined sidered:
in Stages 2 and 3. A sample calculation is given in the ‘‘Sam-
ple of Recall Process’’ subsection. • Precision requirement of the user (i.e., the zone width or
subrange that suffices for the user to make decisions)
Data Preprocessing • Distribution of actual output data over the output zones;
a uniform distribution of actual output data over all zones
At the input layer of the PINN model (shown by Fig. 1),
generally yields better results
the number of input nodes corresponds to the dimension of
the input vector. The dimension of the input vector depends PEs at Kohonen Layer
on the number of input factors and the input data types. Three
input data types are used to define NN input factors (i.e., raw, At the Kohonen layer, each output zone contains an equal
rank, and binary). Raw is used simply for quantitative input number of PEs. Each PE is associated with a weight vector
factors, like general expense ratios, winter construction per- [also referred to as a codebook vector (Kohonen 1995)].
centages, or quantities of work. Rank is used to convert sub- Visually, a weight vector is a set of links that emanate from
jective factors, like crew ability ratings, into numeric format. one PE and end at each input node, as illustrated in Fig. 1.
And binary is used to group textual factors into numeric for- Thus the dimension of a weight vector is equal to that of the
mats, like material type and project definition. It should be input vector (the number of input nodes). An output zone at
noted that an input factor of the raw or rank type corresponds the Kohonen layer can be visualized as a chip containing a
to one input node at the input layer and an input factor of the number of pins (PEs). During NN training the orientation of
binary type corresponds to a number of input nodes depending those pins is gradually fine-tuned to capture the underlying
on the number of groups for the factor. For illustration, input statistical patterns within the training data (Kohonen 1995).
factors and data types for the pipe installation NN model are Our experience indicates that the number of PEs assigned
listed in Table 1. A sample record for the pipe installation NN to each class should be close to the average frequency in the
training is listed in Table 2, showing both the raw data and histogram of training data output values (i.e., the average num-
converted NN input data. The NN input data is normalized ber of training samples in one output zone).
and scaled between 0 and 1 at the input layer. These scaled
inputs will be passed forward for NN training. At the Kohonen NN Learning Process
layer all weight vectors are randomly initialized between 0 and Data of all the training records is scaled at the input layer.
1. The scaled input data is fed into the model to calibrate the
JOURNAL OF COMPUTING IN CIVIL ENGINEERING / OCTOBER 2000 / 243
TABLE 1. NN Input Factors and Data Type TABLE 2. NN Data Sample
NN input factor Data Options and remarks NN input factor Raw data NN input data
(1) type (3) (1) (2) (3)
(2) Project location Rural 001
Project location Binary Urban, rural, camp job Administration require- 0.235 0.235
Administration Raw General expense ment
Year of construc- Binary 89 ⬃ 92, 93 ⬃ 94, 95 ⬃ 96, 97 ⬃ 99 Year of construction 94 ⬃ 95 0100
tion Province/state Alberta 10
Province/state Binary Alberta, Saskatchewan Contract type Reimbursable 10
Contract type Binary Reimbursable, lump sum Client Shell 1.037
Client Raw Index derived from historical data Engineering firm Colt 0.682
Engineering firm Raw Index derived from historical data Project manager G. Trunn 1.036
Project manager Raw Index derived from historical data Superintendent J. Yakemchuk 0.982
Superintendent Raw Index derived from historical data Project definition Chemical 10000
Project definition Binary Chemical, cryogenic, gas, refining Work scope Confined to specific area 10
Work scope Binary Confined/scattered Project type Plant upgrade, no shutdown 0100
Project type Binary Upgrade shutdown, grass root, etc. Prefabrication/field 10%, 90%, 0% 10, 90, 0
Prefabrication/field Raw Percentages for prefabrication work
work Average crew size 25 ⬃ 50 0100
Average crew size Binary <25, 25 ⬃ 50, 50 ⬃ 100, >100 Peak crew size 50 ⬃ 100 00100
Peak crew size Binary <25, 25 ⬃ 50, 50 ⬃ 100, 100 ⬃ 150, Unionized Yes 10
>150 Equipment and mate- 9.4 9.4
Unionized Binary Yes, no rial
Equipment and ma- Raw Equipment and material cost per direct Extra work 0.661 0.661
terial manhours Change order 0.019 0.019
Extra work Raw Original project cost/final project cost Drawing and specifica- Excellent 5
Change order Raw Number of change orders per total direct tions quality
manhours Activity location clas- Inside <10-ft high 000100
Drawing and speci- Rank 1 poor, 3 average, 5 excellent sification
fications quality Total quantity (learn- 6,055 6,055
Location classifica- Binary U/G on site, fab shop, A/G on site, etc. ing)
tion Installation quantities 210, 905, 4,940 210, 905, 4,940
Total quantity Raw Total quantity in DialnFt Material type Carbon steel 00100
(learning) Method of installation Hand rigging (%), machine 10, 90
Installation quanti- Raw Quantity for size ranges, <2⬙, 2⬙ ⬃ 16⬙, rigging (%)
ties >16⬙ Pipe supports 0.45 0.45
Material type Binary Alloy, carbon steel, FRP/PVC, etc. Boltups 4.77 4.77
Method of installa- Raw Percentages of hand rigging Valves 1.59 1.59
tion Screwed joints 0 0
Pipe supports Raw Number of pipe supports per foot of Miscellaneous compo- 3.18 3.18
pipe nents
Boltups Raw Number of boltups per foot of pipe Welding impact 1.25 1.25
Valves Raw Number of valves per foot of pipe Season Winter (%), summer (%) 10, 90
Screwed joints Raw Number of screwed joints per foot of Crew ability Average 3
pipe Site working condi- Many problems 2
Miscellaneous com- Raw Installation of miscellaneous components tions
ponents (manhours per foot of pipe) Inspection, safety, and Detailed 2
Welding impact Raw Welding multiplier (miscoding on site) quality
Season Raw Percentages of winter and summer work Overall degree of diffi- Low 2
Crew ability Rank 1 very low, 3 average, 5 very high culty
Site working condi- Rank 1 extreme problems, ⬃ 5 no problem
tions
Inspection, safety, Rank 1 extremely detailed, ⬃ 5 highly tolerant of the training record; Wij = current weight vector of the
and quality global winner PEi; and W⬘ij = updated weight vector of
Overall degree of Rank 1 very low, 3 average, 5 very high the global winner PE i. The repulsion rate is initially set
difficulty between 0 and 1 and is reduced gradually until it ap-
proaches 0 at the end of learning.
weight vectors of the PEs at the Kohonen layer, using the LVQ 2. Following the global competition, an in-zone competi-
algorithm suggested by Kohonen (1995). tion among processing elements occurs only at the output
The learning process involves a number of iterations, each zone into which the actual output value of the training
of which is comprised of the following: record falls. Prior to the in-zone competition, a ‘‘con-
science’’ value is added to each PE’s euclidean distance
and adjusted over the learning iterations, so as to effec-
1. The euclidean distances between the input vector of a tively prevent one PE within a specific output zone from
training record and each PE’s weight vectors are calcu- winning all the time and activate as many PEs as possible
lated. The PE that has the smallest euclidean distance in the learning process. The formulas to calculate the
value is declared to be a global winner. If the global conscience euclidean distance can be found throughout
winner PE does not belong to the same output zone that the pertinent literature (NeuralWare 1993). Readers can
the actual output value of this training record falls into, refer to Appendix I for details. The method adopted here
the weight vector of the global winner PE is penalized is as follows: (1) The ‘‘conscience’’ euclidean distance
according to the following: between each PE’s weight vector and the input vector
W⬘ij = Wij ⫺ RR ⫻ ( X j ⫺ Wij) (1) is calculated; (2) the PE with the shortest ‘‘conscience’’
euclidean distance value is declared to be an in-zone
where RR stands for repulsion rate, which is a learning winner; and (3) only the in-zone winner PE is rewarded
rate to penalize the global winner PE i; X j = input vector using
244 / JOURNAL OF COMPUTING IN CIVIL ENGINEERING / OCTOBER 2000
W⬘ij = Wij ⫹ AR ⫻ ( X j ⫺ Wij) (2) TABLE 4. Updating Weight Vector in First Learning Stage
Input vector, X1 PE1, W11 PE2, W21 PE3, W ⬘31
where AR stands for attraction rate, which is a learning (1) (2) (3) (4)
rate to reward the in-zone winner PE i; X j = input vector
0.3 0.7232 0.0844 ⫺0.0927
of a training record; Wij = current weight vector of the 0.2 0.6772 0.9129 0.3073
in-zone winner PE i; and W⬘ij = updated weight vector of 0.7 0.8780 0.4422 1.1867
the in-zone winner PE i. Like the repulsion rate, the at- 0.8 0.6307 0.8774 0.8353
traction rate is initially set between 0 and 1, and is re- 0.4 0.7523 0.7261 0.3943
duced gradually until it approaches 0 at the end of learn- 0.5 0.8575 0.0439 0.1597
ing iterations. 0.5 0.2088 0.4567 0.2857
0.9 0.6310 0.4542 ⫺0.6881
0.8 0.6065 0.2413 1.0482
0.1 0.8804 0.9188 1.5346
A sample calculation of one learning iteration is presented 0.4 0.8232 0.7465 ⫺0.0004
next to illustrate the learning process. 0.6 0.0485 0.1158 0.8777
As shown in Fig. 1, the dimension of input vector is 12 and
the output range 0–4 is divided into four output zones (i.e.,
0–1, 1–2, 2–3, and 3–4). Each output zone contains three TABLE 5. Updating Weight Vector in Second Learning Stage
PEs. Note that this sample is a simple model and serves for Input vector, X1 PE1, W11 PE2, W21 PE3, W ⬘31
illustration. Problems encountered in a real situation, which (1) (2) (3) (4)
are suitable for the PINN model to solve, are mostly high
0.3 0.7232 0.0844 0.2564
dimensional; the number of input nodes may exceed 100. 0.2 0.6772 0.9129 0.2119
The input vector of a training record is scaled between 0 0.7 0.8780 0.4422 0.7541
and 1, and the weight vector of a PE at the Kohonen layer is 0.8 0.6307 0.8774 0.8039
randomly initialized between 0 and 1. Table 3 shows the input 0.4 0.7523 0.7261 0.3994
vector X 1 and the weight vectors of the three PEs in Zone 1. 0.5 0.8575 0.0439 0.4622
The euclidean distance ED is calculated between the input 0.5 0.2088 0.4567 0.4762
0.9 0.6310 0.4542 0.7235
vector X 1 and each weight vector Wi1 as ED11 = 1.4163, ED21 0.8 0.6065 0.2413 0.8276
= 1.5746, and ED31 = 1.2963. Suppose that PE3 in Zone 1 is 0.1 0.8804 0.9188 0.2594
the global winner PE among all the PEs, which gives a min- 0.4 0.8232 0.7465 0.3555
imum ED of 1.2963. If the actual output of this training record 0.6 0.0485 0.1158 0.6309
does not fall into Zone 1 (i.e., outside the subrange 0–1), then
the weight vector of PE 3 W31 is updated by (1), as shown in
Table 4. Notice that the repulsion rate is set to be 0.8 at the In-Zone Competition Strategy at Kohonen Layer
start of learning in the sample calculation. Kohonen (1995) The in-zone competition at the Kohonen layer occurring in
recommended a smaller initial value such as 0.06 for the re- the recall stage differs from that occurring in the learning
pulsion rate and attraction rate to obtain better results. stage. Once adequate training is complete, PINN is capable of
If the actual output of this training record does fall into Zone mapping the input onto the output. At the Kohonen layer, for
1 (i.e., within the subrange 0–1), then no weight vector is one output zone, the PE that has the shortest global euclidean
updated in the global competition. The learning process steps distance (no conscience value) between its weight vector and
into the second phase. the input vector is declared to be an in-zone winner PE. Only
In the first training iteration, the conscience value for every the in-zone winner PE advances to the Bayesian layer.
PE in Zone 1 is determined to be equal to 0 (Appendix I). So Unlike the GRNN/PNN, which takes the average of a PE’s
the ‘‘conscience’’ euclidean distance value is equal to the orig- euclidean distance within one output zone as the parameter to
inal euclidean distance value. The in-zone competition winner pass forward (Specht 1988), PINN takes the minimum of the
is PE3, which gives the minimum ‘‘conscience’’ ED of 1.2963, PE’s euclidean distance within one output zone as the param-
and its weight vector W31 is updated by (2), as shown in Table eter to pass forward. The reason for the difference is that in
5. GRNN/PNN each PE corresponds to one training record and
The above learning process will iterate through all the train- different numbers of PEs lie in different output zones. In the
ing records for a sufficient number of runs. Notice that during proposed PINN, the PE does not match the training record,
the learning process the repulsion rate, attraction rate, and the and an equal number of PEs dwell in each output zone and
conscience value are dynamically updated to calibrate the work together in the Kohonen layer of PINN to generalize the
weight vectors. underlying patterns within the training records by implement-
ing LVQ.
TABLE 3. Scaled Input Vector and Initial Weight Vectors PDF Approximator at Bayesian Layer
Input vector, X1 PE1, W11 PE2, W21 PE3, W31 As illustrated in Fig. 1, at the Bayesian layer each output
(1) (2) (3) (4) zone only contains the winner PE from the in-zone competi-
0.3 0.7232 0.0844 0.0819 tion at the Kohonen layer in the recall stage. The main com-
0.2 0.6772 0.9129 0.2596 ponents at the Bayesian layer are a kernel function and a soft-
0.7 0.8780 0.4422 0.9704 max activation function that are used to approximate the
0.8 0.6307 0.8774 0.8196
0.4 0.7523 0.7261 0.3969
probability density of one input vector being within each out-
0.5 0.8575 0.0439 0.3109 put zone in the following steps:
0.5 0.2088 0.4567 0.3809
0.9 0.6310 0.4542 0.0177 1. The square of the euclidean distance value of the winner
0.8 0.6065 0.2413 0.9379 PE from each zone is passed into the kernel function,
0.1 0.8804 0.9188 0.8970 which is the Gaussian function of Bayesian methods in
0.4 0.8232 0.7465 0.1775
0.6 0.0485 0.1158 0.7543
statistics as described in Specht (1988) and shown in (3).
If the number of output zones is N, then for each input
JOURNAL OF COMPUTING IN CIVIL ENGINEERING / OCTOBER 2000 / 245
vector X j, the kernel function is evaluated for N times TABLE 7. Recall Calculations at Bayesian Layer
and outputs one q value for each zone. Values for each zone Zone 1 Zone 2 Zone 3 Zone 4
qi = exp[⫺( Wij ⫺ Xj) ( Wij ⫺ Xj)/2 ]
T 2
(3) (1) (2) (3) (4) (5)
Median 0.500 1.500 2.500 3.500
where i = 1, 2, 3, . . . , N ( N = number of output zones); Winner PE’s ED 0.951 2.567 1.843 2.922
Xj = one input vector fed into PINN at the input layer; Gaussian output q 0.636 0.037 0.183 0.014
and ( Wij ⫺ Xj)T( Wij ⫺ Xj) = square of the euclidean Softmax output p 0.731 0.043 0.210 0.016
distance value between the input vector Xj and the win-
ner PE’s weight vector Wij in the output zone i(i = 1, 2,
3, . . . , N ). The value of is a smoothing factor, is the three processing elements. Table 6 lists the scaled input vector
only adjustable parameter of the Gaussian function [(3)], X1 and the weight vectors of the three processing elements in
and controls the shape of the PDF. The greater , the Zone 1. The euclidean distance values ED between the weight
more dispersed the probability density graph. The value vectors and the input vector are calculated as ED11 = 0.9513,
of is critical to PINN’s predicting capability and can ED21 = 1.1670, and ED31 = 1.3249.
be determined through iterative adjustments. In regard to At output Zone 1, PE1 with a minimum ED (0.9513) is the
industrial productivity application, should fall in the in-zone competition winner and proceeds to the Bayesian
range between 0.8 and 1.2. layer. Suppose the winner PEs from the other three zones also
2. The Softmax Activation function (Sarle 1997) as shown are determined in a similar manner and proceed to the Baye-
in (4) makes the sum of the calculation results (q values) sian layer. Table 7 lists their euclidean distance values and
from (3) equal to 1, so that the final output from the outputs from the Gaussian function and the Softmax function.
Bayesian layer can be interpreted as posterior probabil- From Table 7, the probability of output being within Zone
ities ( p values). 1 is 0.731; hence, the mode output value is found to be the
median of Zone 1 (i.e., 0.5). The weighted average output
qi value is obtained by calculating the sum product of the Soft-
冘
pi = N (4) max output ( p values) and median of each output zone (i.e.,
qi 0.5 ⫻ 0.731 ⫹ 1.5 ⫻ 0.043 ⫹ 2.5 ⫻ 0.210⫹ 3.5 ⫻ 0.016 =
i=1 1.011).
where qi = output value from Gaussian function [(3)] for
output zone i (i = 1, 2, 3, . . . , N ); and N = number of IMPLEMENTATION OF PINN MODEL
output zones. Computer software based on the PINN model is developed
for learning and testing in the environment of Microsoft Ac-
Outputs at Output Layer cess 97 and Visual Basic for Applications. Historical piping
At the output layer, the probability distribution predicted by productivity data of 66 projects resulting in 119 records of a
PINN is presented in the form of a PDF graph, which portrays construction company were collected and compiled into NN
the uncertainty of the output value. In addition to the predicted input data for three labor-intensive activities (i.e., pipe instal-
distribution, PINN calculates two point prediction values: lation, pipe welding, and pipe hydro-testing). In the following
sections, pipe installation is used to illustrate the testing and
• Mode value: the median of the output zone or subrange validation of the PINN model.
that has the greatest probability The PINN model for pipe installation has a total of 81 input
• Weighted average value: the sum product of the median nodes. (The input factors and sample data are shown in Tables
and probability of each output zone—the user should 1 and 2). The output range is divided into 20 output zones,
treat this point-value prediction carefully by checking the each with an equal width of 0.72. Ten processing elements are
PDF graph. assigned to each output zone. The attraction rate and repulsion
rate are both equal to 0.06. The smoothing factor of the kernal
function is equal to 0.8.
Sample of Recall Processing
One hundred one records are used for PINN learning, and
A sample PINN recall calculation is given below for illus- 18 records are reserved to test the calibrated model. The learn-
tration. ing process takes 300 iterations.
Suppose that the NN is trained and ready to recall the output
for an input vector. As shown in Fig. 1, the dimension of the Validation of PINN Model on Testing Data
input vector is 12 and the output range 0–4 is divided into
four zones (i.e., 0–1, 1–2, 2–3, and 3–4). Each zone contains The testing of the calibrated network on the 18 unseen
records is summarized in Fig. 2. When measured against the
TABLE 6. Trained PINN Ready to Recall for Given Input Vector actual output values of the test data, for the mode value, the
average absolute error is 0.57 and the maximum absolute error
Input vector, X1 PE1, W11 PE2, W21 PE3, W31 is 2.02 and, for the weighted average value, the average ab-
(1) (2) (3) (4) solute error is 0.75 and the maximum absolute error is 2.23.
0.5 0.8668 0.9681 0.6722 Considering a wide output range of about 15, the error is rea-
0.6 0.3437 0.4966 0.1627 sonable and acceptable.
0.4 0.6202 0.0093 0.5633 To compare the PINN model with a back-propagation NN,
0.9 0.2472 0.2213 0.6362
0.4 0.3469 0.6834 0.4307
the same training records were used to train a three-layer feed-
0.7 0.7326 0.4165 0.5033 forward back-propagation NN, which has 81 input nodes at
0.8 0.6803 0.9864 0.8412 the input layer, 40 hidden nodes at the middle layer, and 1
0.1 0.1366 0.4152 0.7824 output node at the output layer. The training parameters are a
0.3 0.0517 0.5621 0.8916 learning rate equal to 0.8 and a momentum rate equal to 0.4.
0.6 0.9558 0.5190 0.7239 The transfer function is a symmetric logistic function. After
0.9 0.9264 0.5150 0.3458
0.6 0.7467 0.5772 0.0914
training was completed, the testing set of 18 records previ-
ously used to test PINN was fed into the model. The testing
246 / JOURNAL OF COMPUTING IN CIVIL ENGINEERING / OCTOBER 2000
FIG. 2. Comparison of PINN and Feedforward Back-Propagation NN (FF BP NN)
results of the back-propagation NN model compared with that Sensitivity Analysis of PINN Model
of the PINN are shown in Fig. 2. From Fig. 2, it is observed
that the PINN model outperforms the back-propagation NN A recall program based on the trained PINN model for pipe
model in terms of point prediction accuracy, by coming closer installation was developed to validate its effectiveness and ac-
to the actual output values. curacy in the context of the application domain. ‘‘What if’’
JOURNAL OF COMPUTING IN CIVIL ENGINEERING / OCTOBER 2000 / 247
scenarios are tested on the NN model by changing some input APPENDIX I. DESINO’S METHOD TO CALCULATE
factors in order to understand the impact of such changes on CONSCIENCE VALUE
the output values. The response of the NN model is compared
against that of an experienced estimator at the participating Conscience euclidean distance is defined
construction company for the purpose of model validation. D⬘i = Di ⫹ Ci (5)
The base case scenario is taken from one testing record. The
actual difficulty multiplier for this scenario is 1.24. The mode where Di is euclidean distance; and Ci is conscience value
value predicted by the PINN model is 1.181, giving an abso-
lute error of (兩1.181 ⫺ 1.24兩 = 0.059). The weighted average Ci = cf ⫻ D ⫻ (n ⫻ wf ⫺ 1) (6)
value is 1.313, giving an absolute error of (兩1.313 ⫺ 1.24兩 = where D = maximum euclidean distance out of the global com-
0.073). Fig. 3 shows the predicted probability function or dis- petition in the previous supervised learning stage; cf = con-
tribution; the chance of output falling into Zone 2 (0.8 ⬃ 1.5, science factor, which is initially set between 0 and 1 by the
median = 1.181) is 69%. user; n = number of PEs per output zone; and wf = win fre-
In the following validation tests, the actual values remain quency. The initial estimate of the win frequency value wf0 is
unknown, so the responses of the estimator based on personal set to the reciprocal of the PE number per output zone for all
experience and common sense serve as a benchmark to mea- the PEs (i.e., 1/n).
sure the performance of the PINN model. The estimator re- With NN learning ongoing, both conscience factor cf and
sponds with a trend or direction instead of a precise number frequency estimate fe are reduced gradually until they ap-
because there are so many other input factors to take into proach 0 at the end of learning.
account. During the unsupervised learning stage, for the in-zone win-
ner PE, its wf value is updated
Scenario 1
wf ⬘ = (1 ⫺ wf0) ⫻ wf ⫹ wf0 (7)
The location of pipe installation is a major consideration
when the experienced estimator determines the pipe installa- For the in-zone loser PEs, their wf values are updated
tion productivity. The installation location for the base case
scenario is ‘‘piping within a fabrication shop.’’ What if the wf ⬘ = (1 ⫺ wf0) ⫻ wf (8)
location is changed to ‘‘operating plant installation on the
Basically, the above formulas are intended to increase the
site’’? The experienced estimator responds by increasing the
wf values for the winner PE and hence increase its conscience
difficulty multiplier to a certain extent to reflect the unfavor-
value so that the winner PE will have less chances to win again
able job conditions. Response of the PINN model is that the
than the other loser PEs in the learning iterations that follow
mode value increases to 1.903 and the weighted average value
it.
increases to 1.983; the chance of the output value falling into
Zone 3 (1.5 ⬃ 2.3, median = 1.903) increases from 13% in
the base case scenario to 78%. The PINN has taken the same ACKNOWLEDGMENTS
direction as the estimator in the decision process for this sce- The work was funded by a number of construction companies in Al-
nario. berta and the Natural Science and Engineering Research Council of Can-
ada under grant number IRC-195558/96.
Scenario 2
APPENDIX II. REFERENCES
In the base case scenario, the job is done 100% in the winter
in Alberta. What if the job is done 100% in the summer? The AbouRizk, S., and Wales, R. (1997). ‘‘Combined discrete-event/contin-
estimator anticipates a reduction for the difficulty multiplier, uous simulation for project planning.’’ J. Constr. Engrg. and Mgmt.,
which means an increase of productivity level. The response ASCE, 123(1), 11–20.
Chao, L. C., and Skibniewski, M. J. (1994). ‘‘Estimating construction
of the PINN model is that the mode value remains 1.181; productivity: Neural-network-based approach.’’ J. Comp. in Civ.
however, the weighted average value decreases to 1.154. The Engrg., ASCE, 8(2), 234–251.
chance of the output value falling into Zone 2 (0.8 ⬃ 1.5, Karshenas, S., and Feng, X. (1992). ‘‘Application of neural networks in
median = 1.181) decreases significantly from 69 to 44%, earthmoving equipment production estimating.’’ Proc., 8th Conf. Com-
whereas the chance of the output value falling into Zone 1 puting in Civ. Engrg., ASCE, New York, 841–847.
(0.2 ⬃ 0.8, median = 0.5) increases significantly, from 15% Knowles, P. (1997). ‘‘Predicting labor productivity using neural net-
works.’’ MS thesis, University of Alberta, Edmonton, Alta., Canada.
in the base case scenario to 40%. Again the PINN chooses the Kohonen, T. (1995). Self-organizing maps, Springer Ser. in Information
similar course of action as the estimator in the decision process Sci., Springer, London.
for this scenario. Moselhi, O., Hegazy, T., and Fazio, P. (1991). ‘‘Neural networks as tools
in construction.’’ J. Constr. Engrg. and Mgmt., ASCE, 117(4), 606–
CONCLUSIONS 625.
Murtaza, M. B., and Fisher, D. J. (1994). ‘‘Neuromodex—Neural network
The PINN model creates a meaningful representation of a system for modular construction decision making.’’ J. Comp. in Civ.
complex, real-life situation in the problem domain and is ef- Engrg., ASCE, 8(2), 221–233.
fective in dealing with high dimensional input-output mapping NeuralWare. (1993). NeuralWorks advanced reference guide, NeuralWare
Inc., Pittsburgh, Pa.
with multiple influential factors in a probabilistic approach. Portas, J., and AbouRizk, S. M. (1997). ‘‘Neural network model for es-
The application of the PINN model in industrial labor pro- timating construction productivity.’’ J. Constr. Engrg. and Mgmt.,
duction rate estimating helps the estimator choose a course of ASCE, 123(4), 399–410.
action by giving a better understanding of the project infor- Sarle, W. S., ed. (1997). ‘‘Introduction.’’ Neural network FAQ, part 1 of
mation available and the possible outcomes that could occur. 7, periodic posting to Usenet newsgroup 具comp.ai.neural-nets典 URL
Because the probability density of each output zone is pro- 具ftp://ftp.sas.com/pub/neural/faq.html典
Specht, D. F. (1988). ‘‘Probabilistic neural networks for classification,
vided, the predicted distribution and point-prediction values mappings, or associative memory.’’ Proc., IEEE Int. Conf. on Neural
give the estimator much more confidence in the predicted re- Networks, IEEE, Piscatawy, N.J., 1, 525–532.
sult. In combination with personal experience and preferences, Specht, D. F. (1991). ‘‘General regression neural networks.’’ IEEE Trans.
a labor production rate for a new project can be determined. on Neural Networks, 2, 525–576.
ABSTRACT: A fuzzy logic integrated genetic programming (GP) based methodology is proposed to increase
the performance of the GP based approach for structural optimization and design. Fuzzy set theory is employed
to deal with the imprecise and vague information, especially the design constraints, during the structural design
process. A fuzzy logic based decision-making system incorporating expert knowledge and experience is used to
control the iteration process of genetic search. Illustrative examples have been used to demonstrate that, when
comparing the proposed fuzzy logic controlled GP approach with the pure GP method, the proposed new
approach has a higher search efficiency.
FIG. 2. Truss and Its GP Parse Tree: (a) Six-Bar Truss; (b) GP
FIG. 1. General Design Domain Parse Tree of Truss
FIG. 3. Fuzzy Constraint G̃ij with ␣ Membership Degree FIG. 5. Framework of Fuzzy Logic Controller
ABSTRACT: A genetic algorithm approach is presented for the optimization of multireservoir systems. The
approach is demonstrated through application to a reservoir system in Indonesia by considering the existing
development situation in the basin and two future water resource development scenarios. A generic genetic
algorithm model for the optimization of reservoir systems has been developed that is easily transportable to any
reservoir system. This generality is a distinct practical advantage of the genetic algorithm approach. A comparison
of the genetic algorithm results with those produced by discrete differential dynamic programming is also
presented. For each case considered in this study, the genetic algorithm results are very close to the optimum,
and the technique appears to be robust. Contrary to methods based on dynamic programming, discretization of
state variables is not required. Further, there is no requirement for trial state trajectories to initiate the search
using a genetic algorithm. Model sensitivity and generalizations that can be drawn from this and earlier work
by Wardlaw and Sharif are also considered.
Dynamic programming (DP) (Bellman 1957; Bellman and where rt is the release in time step t such that st⫹1 = ft (st , rt);
Dreyfus 1962) has long been recognized as a powerful ap- and t is the time step.
proach in the analysis of water resource systems. The major Another approach that overcomes the dimensionality prob-
attraction of DP is that it can handle nonconvex, nonlinear, lem is the DP successive approximation technique (DPSA).
and discontinuous objective functions without difficulty. Con- Larson (1968) first proposed the technique, and Trott and Yeh
straints on both decision and state variables are easily intro- (1973) used the successive approximation technique in com-
duced. Hall and Buras (1961) was the first to propose the bination with IDP in solution of a six-reservoir problem. Suc-
application of DP to determine optimal returns from reservoir cessive approximation is one-at-a-time optimization, and its
systems. Young (1967) developed optimal operating rules for common drawback is convergence to a local optimum. Exten-
a single reservoir using DP. Extensive review of DP applica- sion to the DPSA technique has been reported by Nopmongcol
tions to reservoir systems can be found in the works of Ya- and Askew (1976), who suggested higher-level combinations,
kowitz (1982) and Yeh (1985). Wurbs (1993) describes various such as two- or three-at-a-time combinations. The technique
computational models that can be used in the analysis of water was demonstrated by application to the four-reservoir problem.
resource systems. The usefulness of DP for multireservoir sys- The requirement that the state variables be discretized is a
tems is, however, limited by the huge demand that it can in- major cause of computational burden and huge memory re-
duce on computational resources. A brief review of optimi- quirements. Murray and Yakowitz (1979) have developed dif-
zation models used in reservoir operation problems is ferential DP for multireservoir control problems. The impor-
presented here to put the work of this paper in context. tant feature of the technique is that discretization of state and
Many variants of DP have been developed to alleviate the decision space is not required. The method requires a quadratic
problem of dimensionality. Larson (1968) obtained the solu- approximation of the objective function. Application of the
tion to the four-reservoir problem using incremental dynamic method to a ten-reservoir problem by Murray and Yakowitz
programming (IDP). Hall et al. (1969) used a different version (1979) has demonstrated its effectiveness. However, the tech-
of IDP and solved a two-reservoir system. Another version, nique requires that the objective function is differentiable and
called discrete differential dynamic programming (DDDP), that the constraints are linear.
was developed by Heidari et al. (1971) and could be seen as A combined LP-DP procedure has been described by Becker
a generalization of IDP. The terms IDP and DDDP have been and Yeh (1974). The procedure has also been used by Takeuchi
used interchangeably in water resources applications. In the and Moreau (1974), Becker et al. (1976), Yeh et al. (1979),
standard IDP of Larson (1968), the number of states to be Yeh and Becker (1982), and Marino and Mohammadi (1983).
tested at each stage is limited to three per state variable. The In a combined LP-DP algorithm, the stage to stage optimiza-
computational burden and memory requirements for such an tion is carried out by LP, and DP is used for determining the
approach are functions of 3n, where n is the number of state optimal policy over a number of periods. The nonlinearities
variables. IDP does overcome the dimensionality problem to are handled by using an iterative technique, such as successive
a large extent, but requires stringent conditions to be satisfied linear programming (SLP). Grygier and Stedinger (1985) de-
for achieving an optimal solution. The IDP method requires scribe the application of SLP, an optimal control algorithm,
that there must exist a function gt (st , st⫹1) such that for every and a combined LP-DP algorithm to a multireservoir system.
pair of states st and st⫹1 The optimal control algorithm executed five times faster than
the SLP, while the LP-DP algorithm took longer to execute
1
Lect., Dept. of Civ. Engrg., Facu. of Engrg. and Technol., Jamia Millia
and produced comparatively inferior solutions. Howson and
Islamia, New Delhi, India. Sancho (1975) developed a progressive optimality algorithm
2
Sr. Lect., Dept. of Civ. and Envir. Engrg., Univ. of Edinburgh, Ed- (POA) for multistate problems. Application of the POA to an
inburgh EH9 3JN, U.K. E-mail: [email protected] example system consisting of four reservoirs in series has been
Note. Editor: Sivand Lakmazaheri. Discussion open until March 1, reported by Turgeon (1981). The most promising aspect of the
2001. To extend the closing date one month, a written request must be approach is that the state variables do not have to be discre-
filed with the ASCE Manager of Journals. The manuscript for this paper
was submitted for review and possible publication on October 5, 1999.
tized. The algorithm, however, required 823 iterations to con-
This paper is part of the Journal of Computing in Civil Engineering, verge to an optimal solution for the example considered. The
Vol. 14, No. 4, October, 2000. 䉷ASCE, ISSN 0887-3801/00/0004-0255– approach has merit in that it overcomes the dimensionality
0263/$8.00 ⫹ $.50 per page. Paper No. 22076. problem.
JOURNAL OF COMPUTING IN CIVIL ENGINEERING / OCTOBER 2000 / 255
Optimal control theory (OCT) also provides an efficient techniques require large computer resources and derivatives of
means of overcoming the dimensionality problem. State and the objective functions. GAs work with numerical values, and
release variables need not be discretized in methods based on there is no requirement for the objective function to be a dif-
OCT. The requirement that the objective function must be dif- ferentiable function. Furthermore, the computing requirements
ferentiable limits the application of the approach to some ex- are minimal and increase only linearly with the dimensionality
tent, however. Storage constraints are not automatically satis- of the system. There are some limitations to the approach,
fied, but can be handled by use of appropriate penalty however. The convergence to an optimum cannot be guaran-
functions. Wasimi and Kitanidis (1983), Papageorgiou (1985), teed. With increasing string length, the likelihood of locating
Georgakakos and Marks (1987), Georgakakos (1989), and the optimum is reduced. Also, with increasing string length
McLaughlin and Velasco (1990) all describe the application of there are difficulties in satisfying the constraints of the prob-
OCT to reservoir optimization problems. lem. This problem of constraint satisfaction is not limited to
In recent years, genetic algorithms (GAs) have developed the GA approach. Other iterative optimization techniques, such
into a powerful optimization approach. They have been ap- as OCT approaches, also face the same problem. The dimen-
plied to a wide spectrum of problems in science and engi- sionality problem associated with DP (Bellman 1957) and the
neering and have proven to be an efficient means by identi- requirements of differentiability of objective functions for gra-
fying optimal solutions. In some cases, GAs have performed dient-based methods make GA an alternative worthy of ex-
better than traditional techniques. For example, Dandy et al. ploring.
(1996) obtained a better solution for the New York City water GAs may be set up in a number of ways, but as yet there
distribution problem using a GA than had been obtained by is no standard procedure. Wardlaw and Sharif (1999) evaluated
any other technique in the past. GAs are being increasingly alternative GA formulation procedures for multireservoir sys-
used in a number of disciplines of civil engineering to solve tems and demonstrated the robustness of the GA approach by
problems that are difficult with traditional techniques. Kou- applying it to a number of problems. The primary objective
mousis and Georgiou (1994) appled a GA for optimization of of this paper is to demonstrate the applicability of the GA
steel truss roofs. Soh and Yang (1996) used GAs in combi- approach for optimization of a real multireservoir system with
nation with fuzzy logic for structural shape optimization prob- a complex objective function. The results produced by a GA
lems. Rajeev and Krishnamoorthy (1992, 1997), Jenkins have been compared to those obtained by the DDDP tech-
(1991, 1992), and Adeli and Cheng (1993, 1994) have also nique.
reported applications of GAs to structural engineering design
problems. GENETIC ALGORITHMS
A methodology based on GAs has been developed by Li In a GA, members of a population of abstract representa-
and Love (1998) for optimizing the layout of construction site tions of candidate solutions to an optimization problem are
level facilities. Navon and McRea (1997) used the GA ap- stochastically selected, recombined, mutated, and then either
proach for selection of construction robots. Feng et al. (1997) eliminated or retained, based on their relative fitness. The GA
appled a GA to the problem of cost-time trade-offs in con- technique has been notably developed by Goldberg (1989) and
struction projects. Applications of GA to transportation engi- Michalewicz (1992), who also give an excellent introduction
neering problems have been described by Ng (1995), Fwa et to the subject.
al. (1996), Liu et al. (1997), and Cheu et al. (1998). The first step in GA formulation is to code the components
GAs have also been used in combination with other tech- of possible solutions in a chromosome. Each chromosome rep-
niques. Soh and Yang (1996) used a fuzzy controlled GA for resents a potential solution consisting of the components of
shape optimization of trusses. Franchini (1996) used a GA in the decision variables (also known as genes) that either form
combination with sequential quadratic programming for the or can be used to evaluate the objective function. The entire
automatic calibration of conceptual rainfall-runoff models. population of such chromosomes represents a generation. For
Many application of hybrid GAs have also been reported by a typical reservoir operation problem, the genes consist of res-
Davis (1991). Dasgupta and Michalewicz (1997) describe var- ervoir releases that are concatenated to form chromosomes.
ious applications of GAs to engineering optimization prob- For the purpose of exposition, suppose there is only a single
lems. reservoir in the system, and that there are 12 months in the
Goldberg and Kuo (1987) were the first to apply GAs to a optimization horizon. The chromosome representing a solution
water resource problem. They solved a pipeline optimization to such a problem consists of 12 genes representing the de-
problem using a GA. Since then, a number of applications of cision variables (releases) of the problem. The fitness of a
GAs to pipe network and other water resource problems have chromosome as a candidate solution is a function of these
been reported in the literature. A brief review of GA applica- genes and is obtained by evaluating the objective function of
tions to water resources problem can be found in the work of the problem. In application to water resource problems, infea-
Wardlaw and Sharif (1999). The literature describing the ap- sible chromosomes may be generated that fail to satisfy the
plication of GAs to water resources problems is not abundant. system constraints, such as continuity and component capaci-
The writers are aware of only a few applications of GAs in ties. The fitness of a chromosome is also a function of problem
this area. Esat and Hall (1994) applied a GA to the four-res- constraints and may be modified through the introduction of
ervoir problem. They concluded that GAs have potential in penalties when constraints are violated.
water resources optimization and that significant savings could The reproduction mechanism in a GA is composed of se-
be achieved in memory and execution times. Fahmy et al. lection, crossover, and mutation. A number of representation,
(1994) applied GAs to the optimization of reservoir systems. crossover, and mutation schemes have been described by
They compared the performance of a GA with DP and con- Goldberg (1989) and Michalewicz (1992). An evaluation of
cluded that GAs have significant potential in reservoir systems alternative GA formulations using different combinations of
optimization. Olivera and Loucks (1997) used GAs to develop these schemes in application to reservoir systems has been
operating policies for multi-reservoir systems, and concluded carried out by Wardlaw and Sharif (1999).
that GAs are a practical and robust method, and could lead to
effective operating policies. SYSTEM DESCRIPTION
The GA approach has the potential of overcoming the di- The Brantas Basin is located in the province of East Java,
mensionality problems associated with DP. Many optimization Indonesia. It has been used as a case study to demonstrate
256 / JOURNAL OF COMPUTING IN CIVIL ENGINEERING / OCTOBER 2000
FIG. 1. Schematic of Brantas Basin
application of the GA approach in the optimization of multi- flow sequences using the GA and the DDDP approaches.
reservoir systems. Fig. 1 is a map of the Brantas Basin. The These control curves were then applied in a simulation model
Kali Brantas is the second largest river in Java, with a total of the Brantas Basin (Wardlaw 1993) to obtain estimates of
catchment area of 12,000 km2. The primary water use in the long-term economic benefits from operation of the reservoir
basin is for irrigated agriculture. The annual diversion require- system.
ments for those irrigation schemes fed by the Brantas main With only Karangkates reservoir in operation, there are large
stream alone is about 1,600 Mm3. The only significant potable deficits in irrigation supply, even with a potable demand of 4
water demand from the K. Brantas is that of the city of Su- m3/s, although the current potable demand is probably of the
rabaya. Demand projections for Surabaya prepared by Mott order of 8 m3/s. The objective of modeling was to assess the
MacDonald (1988) indicated a potable water demand of impact of new development schemes, such as Wonorejo and
around 8 m3/s by the year 2000. Although this demand is a Beng, on the deficits in the irrigation supply while taking into
relatively small part of the total annual irrigation requirements, account the current potable demand for the city of Surabaya.
it is significant during dry periods when it competes with ir-
rigation demands for the available water resources. System Dynamics
Karangkates is the single most important multipurpose res-
ervoir in the Brantas Basin. The effective storage capacity is The dynamics of a multireservoir system can be described
253 Mm3, and although this is less than 4% of mean annual by the following equation:
basin runoff, its regulating effect on dry season flows is sig-
Si (t ⫹ 1) = Si (t) ⫹ Ii (t) ⫹ MRi (t) ⫺ Ei (t) (1)
nificant. The releases from Karangkates reservoir are used for
generating hydropower, for irrigation, and for supplying po- where Si (t) = vector of reservoir storages at time t in reservoirs
table water to the city of Surabaya. Two additional storage i = 1, n; Ii (t) = vector of reservoir inflows in time period t to
reservoirs Wonorejo and Beng (Fig. 1), have been included in reservoirs i = 1, n; Ri (t) = vector of reservoir releases in time
planning for the basin. Wonorejo is currently under construc- period t from reservoirs i = 1, n; Ei (t) = vector of reservoir
tion. evaporation in time period t from reservoirs i = 1, n; and M
= n ⫻ n matrix of indices of reservoir connections. The matrix
Database M depends upon the configuration of the system and consists
of ⫺1, 0, and 1.
The historic inflows and the other data used in this study,
The transformation from stage to stage is governed by (1).
including economic values for power and agricultural produc-
In addition to this, the releases and storages are limited by
tion, come from Mott MacDonald (1988). The DDDP and GA
physical considerations. The system is therefore subject to
models were applied to a single year of data for a critical
constraints expressed as follows:
period with a view to determining control curves for the res-
ervoirs. Critical inflow series with 90% annual exceedence Si,min ⱕ Si (t) ⱕ Si,max (2)
probability were derived. The control curves for Karangkates,
Wonorejo, and Beng reservoirs were developed for these in- Ri,min ⱕ Ri (t) ⱕ Ri,max (3)
冘冘
n N
冘冘
n N
where P1t = penalty for not meeting the irrigation demand for
the delta assumed to be of the following form:
P1t = k ⫻ [d1t ⫺ x1t ]2 (6)
where k is the penalty factor; d1t is irrigation demand at the
delta; and x1t is irrigation supply to the delta in time step t.
Solution Procedure
The solutions to the problems described above were ob- FIG. 2. Progression of Maximum and Average Fitness for Sin-
tained using the GA and the DDDP approach. The GA was gle Reservoir System
EDS1 EDS2
Component DDDP GA DDDP GA
(1) (2) (3) (4) (5)
Irrigation production 50,681 50,905 51,626 51,591
Karangkates energy 32,981 33,058 32,741 32,789
Karangkates capacity 3,398 3,461 3,418 3,428
Total 87,060 87,424 87,785 87,808
ADVANTAGES OF GA APPROACH
Each of the above scenarios was run using the same GA
computer code and required only modification of the input
files. Each reservoir configuration required a different DDDP
code, however, and development of a generic DDDP code
would be difficult. The performances of the GA and DDDP
were similar, and the fact that the GA requires no discretization
of the state and decision variables is a distinct advantage. On
larger systems, the computing resources required by the GA
would be significantly lower than those required by DDDP.
Component DDDP GA
(1) (2) (3)
Net irrigation production 52,045 52,527
Wonorejo irrigation 13,304 13,068
Karangkates energy 32,657 32,821
Karangkates capacity 3,419 3,498
Wonorejo energy 1,510 1,510
Wonorejo capacity 0 0
Beng energy 1,141 1,137
Beng capacity 0 0
Total 104,076 104,561
ABSTRACT: This paper extends previous work on the use of quality function deployment for client require-
ments processing in construction. It describes a computer-based application, ClientPro, which implements a
quality-function-deployment-based model for the definition, analysis, and translation of client requirements into
solution-neutral design specifications. The methodology, development, operation, and evaluation of the prototype
are discussed, and it is concluded that ClientPro is useful in facilitating the systematic definition, analysis, and
mapping of client requirements to design specifications. It also provides for the incorporation and prioritization
of the perspectives represented by clients, the determination of the relative importance of their requirements,
and the early consideration of life-cycle issues affecting a facility. Thus, it can contribute to the ultimate satis-
faction of the requirements of construction clients.
fore the information/data model in Fig. 2 can be described, the decision by a client to commission the refurbishment of
the EXPRESS-G notation needs to be explained. an existing building. The client organization determines the
business need for a project. On the other hand, the business
EXPRESS-G Notation need (e.g., need for more space) influences the type of project
(e.g., refurbishment) as well as the interest groups associated
An EXPRESS-G model is represented by three types of with the process and outcome of that project—the facility. The
graphic symbols, which together form a diagram. These are nature of the client organization and the kind of project will
the definition symbols, relationship symbols, and composition also determine how to procure, operate, and dispose of the
symbols. Definition symbols (i.e., varying forms of rectangular facility. For example, a client organization with a substantial
boxes) are used to denote simple, constructed, defined, and property portfolio can have personnel who are responsible for
entity data types as well as schema declarations. Relationship the acquisition of new property, unlike a one-off client who
symbols (lines) describe relationships that exist among the def- might require considerable assistance from outside consultants.
initions. Relationships are bidirectional, but one of the two The entity, ‘‘solution-neutral specifications,’’ shown in Fig. 2,
directions is emphasized, using an open circle in the empha- is derived from ‘‘other sources of information,’’ ‘‘facility ‘pro-
sized direction. For example, in an entity-attribute relationship, cess’,’’ ‘‘client/project characteristics,’’ ‘‘client business need,’’
the emphasized direction is toward the attribute. For inheri- and ‘‘tertiary client requirements.’’ Solution-neutral specifica-
tance relationships, the emphasized direction is toward the tions, required for the design of the facility that satisfies the
subtype. Composition symbols enable a diagram to be dis- business need of the client, consist of prioritized design attri-
played on more than one page. When there is a relationship butes and target values. The attributes for ‘‘client/project char-
between definitions on separate pages, the relationship line on acteristics,’’ ‘‘client business need,’’ ‘‘facility process,’’ and
the two pages is terminated by an oval (rounded) box that ‘‘other sources of information’’ are listed in Table 2.
contains a page number and a reference number. The page
number is the number of the page where a referenced defini- Selection of Implementation Environment
tion occurs. The reference number is used to distinguish be-
A relational database package, Microsoft (MS) Access, was
tween multiple references on a page. The composition symbol
used in the development of ClientPro. This was due to a num-
on the page where the reference originated contains the name
ber of factors, which included the vital role that database sys-
of the referenced definition (ISO 1994).
tems play in the management of requirements (Fiksel and
Information models described using EXPRESS-G are in-
Hayes-Roth 1993), the need to quickly develop the prototype
dependent of any implementation context and would therefore
at minimal expense, and the fact that MS Access had been
allow flexibility in computer implementation of the model.
used in similar research by Tsang and Jiao (1998).
Fig. 2 shows that ‘‘client requirements,’’ expressed as primary,
secondary, and tertiary requirements, describe the facility that
Development of ClientPro Using MS Access
satisfies the business need of the client. The requirements of
the client consist of information relating to the characteristics The development of ClientPro followed the general proce-
of the client and the project (i.e., ‘‘client/project characteris- dure for developing Access applications. An Access applica-
tics’’); his or her business need (i.e., ‘‘client business need’’); tion is a coordinated set of database objects—tables, forms,
and the acquisition, operation, and disposal of the facility (i.e., macros, queries, reports, and modules—that enable the user
‘‘facility ‘process’ ’’). Client requirements are influenced by to maintain the data contained in a database (McLaren 1996).
‘‘other sources of information’’ in the sense that a change in The development of an Access application involves an eval-
some standards (e.g., energy emission targets) might influence uation of the information needs of the system with particular
266 / JOURNAL OF COMPUTING IN CIVIL ENGINEERING / OCTOBER 2000
TABLE 2. Attributes for Defined Entities in Client Requirements Processing Model
Entity group Entity Attributes
(1) (2) (3)
Client/project characteristics Client details Name, address, business type, contact person name and address, number of employees,
average annual turnover, policy on occupancy/space use
Interest groups Name; type; relationship with client; influence in acquisition and use of facility; effect
of facility acquisition, operation, and use on group
Project details Project name, location, type, facility type, facility objectives
Client business need Use information Activity type, time of day performed, time of year performed, peak use times, required
equipment and furniture
User information User name, user type, user size, relationship with client, activity performed
Facility functions Function verb, function noun, function qualifier, function rationale (i.e., why a specific
function is required)
Facility attributes Attribute name, attribute meaning, attribute rational, function associated with attribute
Facility ‘‘process’’ Acquisition information Available budget, rationale for budget, level of client involvement (risk), rationale for
level of client involvement, client representatives, expected date of completion, ra-
tionale for completion date
Operation information Costs in use, meaning and rationale for cost in use, operation/management strategy,
rationale for operation/management strategy, level of operation/management tech-
nology, rationale for operation/management technology
Disposal information Expected life span, rationale for expected life span, etc.
Other sources of information International standards Standards for expression of user requirements, standards for air capacity for occupants
in specified building types, etc.
Benchmark/other information Operation/maintenance information for existing or similar facilities, etc
TABLE 4. List of Client Requirements for Building Project (Kamara et al. 1999)
Outline requirements Supplementary requirements
(1) (2)
Prestigious building, particularly with respect to appearance Single point of access to the building and to each lab space is required to
Building that provides an open/welcoming environment promote security through staff monitoring. This will be reinforced by
Building that establishes security as a priority: limited points of entry/exit; controlled access systems (e.g., SWIPE cards or key button pads).
switch card operation (if affordable); main entrance reception for visitors Preference was expressed for entry to the building from the south where car
Building to be networked throughout and to include a mix of heavy engi- parking is to be provided. This could be integrated into a courtyard space
neering laboratories; computer labs; support for small-to-medium enter- with soft landscaping.
prises; offices; teaching space as available North elevation should address the streetscape of a major road as it is a
Engineering laboratories to be located in zones with emphasis on open plan- major frontage for the university.
ning, flexible space, and movable equipment wherever possible; light- The new facility will become the front door to the School of Science and
weight partitioning as appropriate Technology and should present an appropriate image.
Small-to-medium provision to be various and as available Existing school is spread over a number of buildings and lacks any focal
Office accommodation based normally on two persons per room; senior staff point or communal space. Opportunities exist in this project to address
in individual rooms this.
reference to the flow of data and information and the design information model (Fig. 2), which also provided the basis for
of the database objects that will solve the information prob- specifying the relationships between tables. Each table repre-
lems. sented an entity, corresponding to the defined entities in Table
The development of tables, queries, forms, reports, and 2. The entity attributes constituted the ‘‘fields’’ (or columns)
macros for ClientPro was carried out using the facilities pro- of respective tables. Queries facilitated the development of
vided within Access. The design of tables was based on the features that facilitated the analysis of requirements, the prior-
JOURNAL OF COMPUTING IN CIVIL ENGINEERING / OCTOBER 2000 / 267
itization of client interest groups, tertiary client requirements,
and design attributes. Command buttons and customized menu
items were used in the development of the user interface. Table
3 provides a summary of how the desired features of ClientPro
were implemented in the development process.
This involves the input of information on the project; the Analyzing Client Requirements
client; interest groups represented by the client; proposed use
and users; the functions and attributes of the facility; and in- The analysis of client requirements involves the structuring
formation on the acquisition, operation, and disposal of the of requirements into primary, secondary, and tertiary require-
facility. Fig. 3 shows the forms for entering, viewing, and ed- ments; the prioritization of client interest groups; and the pri-
iting data about the functions and attributes of the facility. oritizing of tertiary requirements, using the appropriate forms
Facility functions are defined using a ‘‘verb,’’ ‘‘noun,’’ and a provided in ClientPro.
‘‘function qualifier.’’ Fig. 3 shows a function of the facility—
‘‘provide security to its users’’—that was derived from the list Structuring of Client Requirements
of requirements in Table 4. The rationale for this function is
The structuring of requirements is an iterative process,
that ‘‘the building will be used by a wide range of people;
which involves grouping requirements into a hierarchy from
adequate security is therefore required.’’ The attribute ‘‘wel-
general to specific, using the form shown on the left side of
coming to visitors’’ in the lower half of Fig. 4 is also derived
Fig. 4. On the top part of this form, previously defined func-
from Table 4. This process allows the ‘‘raw’’ requirements of
tions and attributes of the facility can be viewed. Other pre-
the client to be restated in a format that allows understanding
viously defined information on the use of the facility, interests
of what his or her needs are and the rationale behind these
groups, etc., can also be accessed by using the menu to open
needs. The entry of other information related to the ‘‘define
the relevant forms. The most general descriptions are primary
client requirements’’ function follows the same procedure as
requirements, and these are decomposed into greater levels of
for functions and attributes. Reports on the information entered
detail. To facilitate the structuring process, new statements can
on each form can be viewed by clicking on the appropriate
be introduced, or existing ones rephrased. The decomposition
button.
from general (primary) to specific (secondary or tertiary) is
based on the ability of the lower level requirement to enhance
the implementation of the higher level requirement. On the
other hand, the higher level requirement (e.g., primary require-
ments) provides justification for the lower level requirement
(secondary requirement). This means that, when a statement is
being considered to be a primary requirement (a group cate-
gory), the question ‘‘how?’’ is asked to determine how it can
be implemented, or which requirements can fall into that
group. Consistency between a secondary requirement and a
primary requirement is maintained by asking ‘‘why?’’ That is,
the primary requirement must provide a justification for the
associated secondary requirements. This process is continued
until the processing team is satisfied that they have a hierarchy
of primary, secondary, and tertiary requirements that ade-
quately represent the wishes and expectations of the client. The
right side of Fig. 4 shows the report of the combined list of
primary secondary and tertiary requirements.
ABSTRACT: The use of mathematical software packages provides a number of benefits to an engineering user.
In general, the packages provide a platform that supports iterative design and parametric analysis through a
flexible, transparent interface combined with extensive computing power. This enables an engineering user to
develop design equations that are based on fundamental mechanics theories, rather than relying on the ‘‘black-
box’’ approach of most commercial design packages. As an example, a closed-form solution for obtaining
effective length factors for the design of stepped columns is presented. In the example a series of formula is
used to demonstrate the transparency of Mathcad, including the ability of using real engineering units in the
calculations, formulas as they may appear in textbooks or in codes, and ability to hide and password protect
some areas. This facilitates easier automation of the design and design checking processes. Most commercial
structural design packages can be classified as black-box packages. The analyst inputs data at one end only to
receive results at the other without fully appreciating the process the input data have gone through. This phe-
nomenon has the tendency of reducing the engineer to a technician, blindly implementing the ideas of the
software designer. The Mathcad package discussed in this paper and similar mathematical packages returns the
engineer to being in control of the design process.
⫺ 1,260) (7)
2S
Pcr = (9)
K 12L2
冉冑 冊
FIG. 2. Model of Stepwise Uniform Column
S
K1 = (10)
• The ratio  of the length of the bottom segment to the L2Pcr
total length of the column
• The ratio of the load at the stepped level to the load at A substitution of (2) into (10) will yield
the top of the column
兹[S 2(Q4 ⫹ 612) ⫹ DS(Q5 ⫺ 156) ⫹ D 2(Q6 ⫹ 12)]
K1 = (11)
Thus, the modified section will be L in length and have a 兹252[S 2(Q1 ⫹ 24) ⫹ DS(Q2 ⫺ 6) ⫹ D 2(Q3 ⫹ 1)]
moment of inertia equal to ␣I. The reference section will be
(1 ⫺ )L long and have a moment of inertia I, as shown in The axial load that will be acting in the modified section will
Fig. 2(b). be (1 ⫹ )P. The modification factor for this section will be
At one end, a rotational restraint with spring constant kR is such that
provided. The value of kR is set equal to zero for a pinned end
and infinity for a fixed end condition. To facilitate computation 2␣S
(1 ⫹ )Pcr = (12)
in Mathcad, a very large value, on the order of 1020 is assumed K 22L2
to be equivalent to infinity. A semirigid end condition is rep-
resented by an intermediate spring constant value. Solving for K2 yields
At the pinned end of the column (assumed to be the refer-
ence section) a compressive force P is applied. At the stepped
level of the column, a force P is applied. The application of
K2 = 冉冑 冑 冊 冑
S
L2Pcr
␣
1⫹
= K1
␣
1⫹
(13)
冕 冕
L L
factors for the stepped column that were programmed in the
Mathcad example.
S( y⬙)2 dx ⫹ ␣S( y⬙)2 dx
0 L
Mathcad is a Windows-based product that displays both text
冕 冕
Pcr = L L
(1) and equations in a ‘‘what you see is what you get’’ (WYSI-
WYG) format. New users find the program intuitive and easy
( y⬘) dx ⫹
2
( y⬘) dx
2
0 0
to use. As changes are made to the files, the results are updated
automatically, making parametric studies easy to formulate.
The external work done by load P will be as a result of the Graphs are easy to create and can be frequently used to sum-
deformation of the entire length of the column. On the other marize and clarify results. Users can easily add notes, modify
hand, the external work done by the load at the stepped level equations, and add supplemental checks to verify the pro-
P will be due to deformation in the segment of length L gram’s results.
[Fig. 2(b)]. This explains the integration limits in the denom- Mathcad and comparable mathematical software packages
inator of (1). can be used to keep engineers and students in charge of their
Using a fourth-order polynomial for y(x) in (1) will yield designs, as opposed to the black-box design software output.
Mathcad is an excellent tool that can be used to bridge the
252S [S 2(Q1 ⫹ 24) ⫹ DS(Q2 ⫺ 6) ⫹ D 2(Q3 ⫹ 1)] gap between the traditional educational approach and the au-
Pcr =
L2 [S 2(Q4 ⫹ 612) ⫹ DS(Q5 ⫺ 156) ⫹ D 2(Q6 ⫹ 12)] tomated black-box industry approach. Unlike programming
(2) languages, Mathcad does not require extensive training in
commands, syntax, or structured programming. For engineer-
where S = E ⭈ I, where E = modulus of elasticity of the material ing applications, all that is needed is knowledge of how to
and I = moment of inertia about the buckling axis; and D = solve the problem in a step-by-step flowchart manner, a clas-
kR ⭈ L, where kR = rotational stiffness of the partial restraint at sical classroom approach.
the end of the member In the classroom and in practice, students or engineers using
Q1 = (␣ ⫺ 1)3(1442 ⫺ 360 ⫹ 240) (3) Mathcad can solve problems, step by step, just as they would
in a pencil and paper solution. Equations, notations, and sym-
Q2 = (␣ ⫺ 1) (⫺96 ⫹ 270 ⫺ 240 ⫹ 60)
2 3 2
(4) bols are used just as in the text and lectures. As problems
become more complex, previous models and problems can be
Q3 = (␣ ⫺ 1)(164 ⫺ 503 ⫹ 552 ⫺ 25 ⫹ 5) (5) reused and combined to build more comprehensive problem
JOURNAL OF COMPUTING IN CIVIL ENGINEERING / OCTOBER 2000 / 275
solutions. Equations and assumptions are visible in engineer- APPENDIX I. STEPPED COLUMN: MATHCAD
ing notation, unlike ordinary spreadsheets. Text is frequently EXAMPLE OF MTENGA AND SPAINHOUR
used to outline the design process as well as reference appli-
cable theory, standards, specifications, and literature. User Specified Units
The benefits of solving engineering problems in Mathcad N ⬅ 1 newton; KN ⬅ 1,000 N; kip = 1,000 lbf;
are numerous. Users can easily build upon their earlier work
by incorporating previous models into new projects. At the end psf ⬅ lbf/ft ; MPa ⬅ N/mm ;
2 2
ksi ⬅ 1,000 psi;
of the project, the users will have a working model of the plf ⬅ lbf/ft
entire analysis and/or design. Students not only learn how to
solve each piece of the design, they can understand the prob- Note 1: One of the most powerful characteristics of Math-
lem from an overall perspective. Showing the results both nu- cad is the ability to use units in the calculations (both built-in
merically and graphically in Mathcad reinforces the engineer- units and user-specified units). When the formulas’ noncom-
ing concepts and theories. Once the model is complete, patible units are inputted, an error will appear, thus forcing
students have a working program. Written in engineering no- the engineer or student to review the input information. Black-
tation, they can reuse the model in other classes and as a ver- box packages and even other mathematical packages permit
ification tool to check other design programs. ‘‘garbage in garbage out. Moreover, one can mix both metric
and customary units (at appropriate magnitudes, e.g., Fy = 50
ksi entered as Fy = 345 MPa) in the same worksheet and end
MATHCAD IMPLEMENTATION OF up with the same design results.
STEPPED-COLUMN PROBLEM
Input Data
Given below is a Mathcad implementation of the stepped-
column problem discussed above. In the interest of saving Note 2: With Mathcad it is possible to hide and password
space, some areas are hidden. However, unlike black-box pro- protect some regions of the worksheet. Normally one will have
grams, the contents of these hidden areas can be viewed by the ‘‘Input Data’’ area visible, which is the case in this ex-
double clicking on the name icon of the area. For security ample.
reasons some areas may be password protected; thus, to view
the contents of these areas, one may need to have the correct Material Data
password. As seen in the example (Appendix I, notes 1–6),
text can be inserted within the calculation, as is the case in a Modulus of elasticity
paper and pencil worksheet, thus creating a well-documented E := 200 KN/mm2
report. All this can be done without much impairment in the
ability and sophistication of the developed application. Data Yield stress of material (A572 grade 50 steel)
can be imported from or exported to external files and input
Fy := 345 N/mm2
and output arrays can be created, thus creating an application
capable of examining several design alternatives. Residual stress
Fr := 69 N/mm2
CONCLUSIONS
Bending strength reduction factor (per LRFD)
The use of mathematical software packages provides a num- b := 0.90
ber of benefits to an engineering user. In general, the packages
provide a platform that supports iterative design and paramet- Compression strength reduction factor (per LRFD)
ric analysis through a flexible, transparent interface combined
c := 0.85
with extensive computing power. The following benefits need
to be emphasized
Geometry Data (Fig. 1)
Length of top section (reference section)
• Ability to incorporate the actual engineering units in the
calculation Ltop := 4 m
• Ability to see the formulas on the worksheet in the exact
form as they may appear in a textbook or a code Length of bottom section
• Ability to import and export data to external files Lbot := 8 m
Effective length of top section for buckling about weak axis
These characteristics of Mathcad allow students, teachers,
and engineers to develop sophisticated computer models that L1y := 4 m
solve engineering problems in scientific notation. Complete
Effective length of bottom section for buckling about weak
solutions to engineering problems can be developed for a axis
homework problem or to simply check the results of other
programs. Mathcad allows students and teachers to focus on L 2y := 4 m
the engineering theory and procedures while letting the soft-
ware handle the calculations along with the units of measure. Maximum length of laterally unbraced segment in top portion
As the engineering profession continues to move toward au- Lbtop := 4 m
tomated design software, engineers must find better tools that
they can use to verify engineering calculations. Engineering Maximum length of laterally unbraced segment in bottom por-
graduates who possess portfolios of working engineering mod- tion
els, produced in Mathcad as part of their college course work, Lbbot := 8 m
will be much better prepared to enter the automated paradigm
now present in the engineering profession. Rotational restraint spring constant
276 / JOURNAL OF COMPUTING IN CIVIL ENGINEERING / OCTOBER 2000
k1 := 0.0 ⭈ 1020 KN ⭈ m Q4i := 12 ⭈ ⭈  ⭈ (240 ⭈ 6 ⫺ 840 ⭈ 5 ⫹ 756 ⭈ 4 ⫹ 210 ⭈ 3
Total length of column ⫺ 420 ⭈ 2 ⫹ 105)
L := Ltop ⫹ Lbot Q5i := ⭈ 2 ⭈ ((⫺1,920 ⭈ 5 ⫹ 7,560 ⭈ 4 ⫺ 9,576 ⭈ 3 ⫹ 2,940 ⭈ 2
Fraction of column that is modified ⫹ 2,100 ⭈  ⫺ 1,260))
Lbot Q6i := ⭈ 3 ⭈ (320 ⭈ 4 ⫺ 1,400 ⭈ 3 ⫹ 2,247 ⭈ 2 ⫺ 1,575 ⭈  ⫹ 420)
 :=
L
The equivalent effective length factor for the top portion of
Derived variable the column [(11)]
D := k1 ⭈ L
⭈ 兹(Si)2 ⭈ (Q4i ⫹ 612) ⫹ D ⭈ Si ⭈ (Q5i ⫺ 156) ⫹ D2 ⭈ (Q6i ⫹ 12)
Kmtopi :=
Loading Data 兹252 ⭈ [(Si)2 ⭈ (Q1i ⫹ 24) ⫹ D ⭈ Si ⭈ (Q2i ⫺ 6) ⫹ D2 ⭈ (Q3i ⫹ 1)]
Factored compression force at top of column The equivalent effective length factor for the bottom portion
of the column [(13)]
冑
Putop := 80 KN
␣i
Factored compression force at the stepped level of column Kmboti := Kmtopi ⭈
(1 ⫹ )
Pustp := 480 KN
Ratio of the compression forces
Note 5: The design of the column sections are in accor-
Pustp dance with the LRFD approach. Again, in the interest of short-
:=
Putop ening this paper, the area is hidden.
Maximum factored bending moment in the top section
Mutop := 160 KN ⭈ m
Note 6: The combined force index is shown for the various
Maximum factored bending moment in the top section trial sections. The section is adequate only when both indices
are equal or less than 1.0.
Mubot := 220 KN ⭈ m
Moment modification factor for nonuniform moment (LRFD, Top portion of column
Eq. F1-3: top section
Cb 㛭 top := 1