Optimization Theory, Methods, and Applications in Engineering (PDFDrive)
Optimization Theory, Methods, and Applications in Engineering (PDFDrive)
Problems in
Engineering
Theory, Methods, and Applications
Special Issue
Optimization Theory, Methods, and Applications in Engineering
This is a special issue published in “Mathematical Problems in Engineering.” All articles are open access articles distributed
under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any
medium, provided the original work is properly cited.
Editorial Board
Mohamed Abd El Aziz, Egypt Horst Ecker, Austria Tomasz Kapitaniak, Poland
E. M. Abdel-Rahman, Canada M. Onder Efe, Turkey Hamid Reza Karimi, Norway
Rashid K. Abu Al-Rub, USA Elmetwally Elabbasy, Egypt Metin O. Kaya, Turkey
Salvatore Alfonzetti, Italy Alex Elias-Zuniga, Mexico Nikolaos Kazantzis, USA
Igor Andrianov, Germany Anders Eriksson, Sweden Farzad Khani, Iran
Sebastian Anita, Romania Vedat S. Erturk, Turkey Kristian Krabbenhoft, Australia
W. Assawinchaichote, Thailand Qi Fan, USA Jurgen Kurths, Germany
Erwei Bai, USA Moez Feki, Tunisia Claude Lamarque, France
Ezzat G. Bakhoum, USA Ricardo Femat, Mexico F. Lamnabhi-Lagarrigue, France
José Manoel Balthazar, Brazil Rolf Findeisen, Germany Marek Lefik, Poland
Rasajit K. Bera, India R. A. Fontes Valente, Portugal Stefano Lenci, Italy
Jonathan N. Blakely, USA C. R. Fuerte-Esquivel, Mexico Roman Lewandowski, Poland
Stefano Boccaletti, Spain Zoran Gajic, USA Ming Li, China
Daniela Boso, Italy Ugo Galvanetto, Italy Shanling Li, Canada
M. Boutayeb, France Xin-Lin Gao, USA Tao Li, China
Michael J. Brennan, UK Furong Gao, Hong Kong Jian Li, China
John Burns, USA Behrouz Gatmiri, Iran Shihua Li, China
Salvatore Caddemi, Italy Oleg V. Gendelman, Israel Teh-Lu Liao, Taiwan
Piermarco Cannarsa, Italy Didier Georges, France P. Liatsis, UK
Jose E. Capilla, Spain P. Batista Gonçalves, Brazil Jui-Sheng Lin, Taiwan
Carlo Cattani, Italy Oded Gottlieb, Israel Shueei M. Lin, Taiwan
Marcelo M. Cavalcanti, Brazil Fabrizio Greco, Italy Yuji Liu, China
Diego J. Celentano, Chile Quang Phuc Ha, Australia Wanquan Liu, Australia
Mohammed Chadli, France M. R. Hajj, USA Bin Liu, Australia
Arindam Chakraborty, USA Thomas Hanne, Switzerland Paolo Lonetti, Italy
Yong-Kui Chang, China Tasawar Hayat, Pakistan V. C. Loukopoulos, Greece
Michael J. Chappell, UK Katica R. Hedrih, Serbia Junguo Lu, China
Kui Fu Chen, China M.I. Herreros, Spain Chien-Yu Lu, Taiwan
Xinkai Chen, Japan Wei-Chiang Hong, Taiwan Alexei Mailybaev, Brazil
Kue-Hong Chen, Taiwan J. Horacek, Czech Republic Manoranjan Maiti, India
Jyh Horng Chou, Taiwan Chuangxia Huang, China O. Daniel Makinde, South Africa
Slim Choura, Tunisia Gordon Huang, Canada R. Martinez-Guerra, Mexico
Cesar Cruz-Hernandez, Mexico Yi Feng Hung, Taiwan Driss Mehdi, France
Swagatam Das, India Hai-Feng Huo, China Roderick Melnik, Canada
Filippo de Monte, Italy Asier Ibeas, Spain Xinzhu Meng, China
Maria de Pinho, Portugal Anuar Ishak, Malaysia Y. Vladimirovich Mikhlin, Ukraine
Antonio Desimone, Italy Reza Jazar, Australia G. Milovanovic, Serbia
Yannis Dimakopoulos, Greece Zhijian Ji, China Ebrahim Momoniat, South Africa
Baocang Ding, China J. Jiang, China Trung Nguyen Thoi, Vietnam
Joao B. R. Do Val, Brazil J. J. Judice, Portugal Hung Nguyen-Xuan, Vietnam
Daoyi Dong, Australia Tadeusz Kaczorek, Poland Ben T. Nohara, Japan
Balram Dubey, India Tamas Kalmar-Nagy, USA Anthony Nouy, France
Sotiris K. Ntouyas, Greece Massimo Scalia, Italy Youqing Wang, China
Gerard Olivar, Colombia Mohamed A. Seddeek, Egypt Yongqi Wang, Germany
Claudio Padra, Argentina Alexander P. Seyranian, Russia Cheng C. Wang, Taiwan
Francesco Pellicano, Italy Leonid Shaikhet, Ukraine Moran Wang, USA
Matjaz Perc, Slovenia Cheng Shao, China Yijing Wang, China
Vu Ngoc Phat, Vietnam Daichao Sheng, Australia Gerhard-Wilhelm Weber, Turkey
A. Pogromsky, The Netherlands Tony Sheu, Taiwan Jeroen A. S. Witteveen, USA
Seppo Pohjolainen, Finland Jian-Jun Shu, Singapore Kwok-Wo Wong, Hong Kong
Stanislav Potapenko, Canada Zhan Shu, UK Ligang Wu, China
Sergio Preidikman, USA Dan Simon, USA Zheng-Guang Wu, China
Carsten Proppe, Germany Luciano Simoni, Italy Wang Xing-yuan, China
Hector Puebla, Mexico Grigori M. Sisoev, UK X. Frank Xu, USA
Justo Puerto, Spain Christos H. Skiadas, Greece Xuping Xu, USA
Dane Quinn, USA Davide Spinello, Canada Jun-Juh Yan, Taiwan
K. Ramamani Rajagopal, USA Sri Sridharan, USA Xing-Gang Yan, UK
Gianluca Ranzi, Australia Rolf Stenberg, Finland Suh-Yuh Yang, Taiwan
Sivaguru Ravindran, USA Changyin Sun, China Mahmoud T. Yassen, Egypt
G. Rega, Italy Jitao Sun, China Mohammad I. Younis, USA
Pedro Ribeiro, Portugal Xi-Ming Sun, China Huang Yuan, Germany
J. Rodellar, Spain Andrzej Swierniak, Poland S. P. Yung, Hong Kong
R. Rodriguez-Lopez, Spain Allen Tannenbaum, USA Ion Zaballa, Spain
A. J. Rodriguez-Luis, Spain Cristian Toma, Romania Arturo Zavala-Rio, Mexico
Ignacio Romero, Spain Irina N. Trendafilova, UK Ashraf M. Zenkour, Saudi Arabia
Hamid Ronagh, Australia Alberto Trevisani, Italy Yingwei Zhang, USA
Carla Roque, Portugal Jung-Fa Tsai, Taiwan Xu Zhang, China
Rubén Ruiz Garcı́a, Spain John Tsinias, Greece Lu Zhen, China
Manouchehr Salehi, Iran Kuppalapalle Vajravelu, USA Liancun Zheng, China
Miguel A. F. Sanjuán, Spain Victoria Vampa, Argentina Jian Guo Zhou, UK
Ilmar Ferreira Santos, Denmark Josep Vehi, Spain Zexuan Zhu, China
Nickolas S. Sapidis, Greece Stefano Vidoli, Italy Mustapha Zidi, France
Bozidar Sarler, Slovenia Xiaojun Wang, China
Andrey V. Savkin, Australia Dan Wang, China
Contents
Optimization Theory, Methods, and Applications in Engineering, Jung-Fa Tsai,
John Gunnar Carlsson, Dongdong Ge, Yi-Chung Hu, and Jianming Shi
Volume 2012, Article ID 759548, 7 pages
A Stone Resource Assignment Model under the Fuzzy Environment, Liming Yao, Jiuping Xu,
and Feng Guo
Volume 2012, Article ID 265837, 26 pages
Dynamic Programming and Heuristic for Stochastic Uncapacitated Lot-Sizing Problems with
Incremental Quantity Discount, Yuli Zhang, Shiji Song, Cheng Wu, and Wenjun Yin
Volume 2012, Article ID 582323, 21 pages
Sparse Signal Recovery via ECME Thresholding Pursuits, Heping Song and Guoli Wang
Volume 2012, Article ID 478931, 22 pages
Variable Neighborhood Search for Parallel Machines Scheduling Problem with Step
Deteriorating Jobs, Wenming Cheng, Peng Guo, Zeqiang Zhang, Ming Zeng, and Jian Liang
Volume 2012, Article ID 928312, 20 pages
A Hybrid Algorithm Based on ACO and PSO for Capacitated Vehicle Routing Problems,
Yucheng Kao, Ming-Hsien Chen, and Yi-Ting Huang
Volume 2012, Article ID 726564, 17 pages
New Bounds for Ternary Covering Arrays Using a Parallel Simulated Annealing,
Himer Avila-George, Jose Torres-Jimenez, and Vicente Hernández
Volume 2012, Article ID 897027, 19 pages
A VNS Metaheuristic with Stochastic Steps for Max 3-Cut and Max 3-Section, Ai-fan Ling
Volume 2012, Article ID 475018, 16 pages
Mixed Mortar Element Method for P1NC /P0 Element and Its Multigrid Method for
the Incompressible Stokes Problem, Yaqin Jiang and Jinru Chen
Volume 2012, Article ID 979307, 18 pages
Applying Neural Networks to Prices Prediction of Crude Oil Futures, John Wei-Shan Hu,
Yi-Chung Hu, and Ricky Ray-Wen Lin
Volume 2012, Article ID 959040, 12 pages
A Novel Method for Technology Forecasting and Developing R&D Strategy of Building
Integrated Photovoltaic Technology Industry, Yu-Jing Chiu and Tao-Ming Ying
Volume 2012, Article ID 273530, 24 pages
Solving the Tractor and Semi-Trailer Routing Problem Based on a Heuristic Approach,
Hongqi Li, Yue Lu, Jun Zhang, and Tianyi Wang
Volume 2012, Article ID 182584, 12 pages
Adaptive Method for Solving Optimal Control Problem with State and Control Variables,
Louadj Kahina and Aidene Mohamed
Volume 2012, Article ID 209329, 15 pages
A Hybrid Network Model to Extract Key Criteria and Its Application for Brand Equity
Evaluation, Chin-Yi Chen and Chung-Wei Li
Volume 2012, Article ID 971303, 14 pages
Hybrid Optimization Approach for the Design of Mechanisms Using a New Error Estimator,
A. Sedano, R. Sancibrian, A. de Juan, F. Viadero, and F. Egaña
Volume 2012, Article ID 151590, 20 pages
A Label Correcting Algorithm for Partial Disassembly Sequences in the Production Planning
for End-of-Life Products, Pei-Fang Jennifer Tsai
Volume 2012, Article ID 569429, 13 pages
Improved Degree Search Algorithms in Unstructured P2P Networks, Guole Liu, Haipeng Peng,
Lixiang Li, Yixian Yang, and Qun Luo
Volume 2012, Article ID 923023, 18 pages
The Number of Students Needed for Undecided Programs at a College from the Supply-Chain
Viewpoint, Jin-Ling Lin, Jy-Hsin Lin, and Kao-Shing Hwang
Volume 2012, Article ID 276519, 12 pages
A Hybrid Genetic Algorithm for the Multiple Crossdocks Problem, Zhaowei Miao, Ke Fu,
and Feng Yang
Volume 2012, Article ID 316908, 18 pages
A Nonlinear Multiobjective Bilevel Model for Minimum Cost Network Flow Problem in
a Large-Scale Construction Project, Jiuping Xu, Yan Tu, and Ziqiang Zeng
Volume 2012, Article ID 463976, 40 pages
Applying Hierarchical Bayesian Neural Network in Failure Time Prediction, Ling-Jing Kao and
Hsin-Fen Chen
Volume 2012, Article ID 953848, 11 pages
A Fuzzy Dropper for Proportional Loss Rate Differentiation under Wireless Network with
a Multi-State Channel, Yu-Chin Szu
Volume 2012, Article ID 827137, 16 pages
A Modified PSO Algorithm for Minimizing the Total Costs of Resources in MRCPSP,
Mohammad Khalilzadeh, Fereydoon Kianfar, Ali Shirzadeh Chaleshtari, Shahram Shadrokh,
and Mohammad Ranjbar
Volume 2012, Article ID 365697, 18 pages
Comment on “Highly Efficient Sigma Point Filter for Spacecraft Attitude and Rate
Estimation”, Baiqing Hu, Lubin Chang, An Li, and Fangjun Qin
Volume 2012, Article ID 170391, 5 pages
A Filter Algorithm with Inexact Line Search, Meiling Liu, Xueqian Li, and Qinmin Wu
Volume 2012, Article ID 349178, 20 pages
Editorial
Optimization Theory, Methods,
and Applications in Engineering
Copyright q 2012 Jung-Fa Tsai et al. This is an open access article distributed under the Creative
Commons Attribution License, which permits unrestricted use, distribution, and reproduction in
any medium, provided the original work is properly cited.
Over years of development, optimization theory and methods have grown in their ability to
handle various practical problems. In light of advances in computing systems, optimization
approaches have become one of the most promising techniques for engineering applications.
To close the gap between optimization theory and the practice of engineering, this special
issue intends to provide the details of recent advances of optimization sciences and promote
the applications of optimization methods in engineering. This special issue also provides a
forum for researchers and practitioners to review and disseminate quality research work on
optimization approaches and their applications in the context of engineering and to identify
critical issues for further developments.
The papers accepted in the special issue include original research articles as well as
review articles on all aspects of optimization including deterministic approaches, continuous,
mixed-integer and discrete optimization, stochastic optimization, particle swarm optimiza-
tion, neural network, simulated annealing, genetic algorithm, and hybrid methods. Some of
the papers are dedicated to the development of advanced optimization methods for direct
or indirect use in engineering problems such as network, scheduling, production planning,
industrial engineering, and manufacturing systems. Contributions containing computational
2 Mathematical Problems in Engineering
issues, search strategies, and modeling and solution techniques to practical problems such
as multicriteria decision making and management information system are also involved.
According to the characteristics of the accepted papers, the special issue is organized as the
following four parts and each part is composed of several important papers to the part’s
scope.
Deterministic Optimization
The paper “Global sufficient optimality conditions for a special cubic minimization problem”
by X. Zhang et al. presents some sufficient global optimality conditions for a special
cubic minimization problem with box constraints or binary constraints by extending the
global subdifferential approach proposed by V. Jeyakumar in 2006. Numerical examples
demonstrate that the optimality conditions can effectively be used for identifying global
minimizers of the certain nonconvex cubic minimization problem.
Baiqing Hu, Lubin Chang, An Li, and Fangjun Qin propose a comment on “Highly
efficient sigma point filter for spacecraft attitude and rate estimation” by Fan and Zeng 2009. In
the comment, the geometric simplex sigma points GSSPs can be derived from a numerical
integration formula of degree 2 and can be generalized for different degrees, that is, the GSSP
can be derived through the orthogonal transformation from the basic points set of degree 2.
Moreover, their method can be used to construct more accurate sigma points set for certain
dynamic problems.
“Adaptive method for solving optimal control problem with state and control variables” by
L. Kahina and A. Mohamed solves the problem of optimal control with state and control
variables by adaptive method and technology of linear programming. The obtained results
show that it is possible to construct very fast solving algorithms based on the adequate
consideration of the dynamic structure of the problem in question.
M. Liu and X. Li propose a filter algorithm with inexact line search for nonlinear
programming problems that ensures superlinear local convergence without second order
correction steps. In their paper “A filter algorithm with inexact line search” a filter is constructed
by employing the norm of the gradient of the Lagrangian function to the infeasibility
measure. Numerical experiences show the efficiency of their filter algorithm.
The paper “Predictor-corrector primal-dual interior point method for solving economic
dispatch problems: a postoptimization analysis” by A. R. Balbo et al. proposes a predictor-
corrector primal-dual interior point method involving line search procedures to solve the
economic dispatch problem. Their method applies the Fibonacci search technique in the
predictor step and an Armijo line search in the corrector step. The comparative results with
other methods described in the literature demonstrate the efficiency of their method.
The paper “An optimal classification method for biological and medical data” by Y.-H. Huang
et al. proposes a union of hyper spheres by the mixed-integer nonlinear program to classify
biological and medical datasets and a piecewise linearization technique is used to reformulate
the nonlinear program to obtain a global optimum. Numerical examples illustrate that the
proposed method is computationally more efficient than current methods.
“A review of deterministic optimization methods in engineering and management” by M.-
H. Lin et al. introduces recent advances in deterministic methods for solving signomial
programming problems and mixed-integer nonlinear programming problems. A number
of important applications in engineering and management are also reviewed to reveal the
usefulness of the optimization methods.
Mathematical Problems in Engineering 3
Heuristic Algorithms
The paper “Opposition-based barebones particle swarm for constrained nonlinear optimization
problems” by H. Wang presents a modified barebones particle swarm optimization to solve
constrained nonlinear optimization problems and simulation results show that the presented
approach achieves a promising performance.
In “Multithreshold segmentation based on artificial Immune systems,” E. Cuevas et al.
present an algorithm for multi-threshold segmentation which is based on the artificial
immune systems technique. The clonal selection algorithm-based method shows a fast con-
vergence and a low sensitivity to initial conditions and improves complex time-consuming
computations commonly required by gradient-based methods.
In “New bounds for ternary covering arrays using a parallel simulated annealing” by H.
Avila-George et al. three parallel approaches for simulated annealing: the independent,
semiindependent, and cooperative searches are applied to the covering array construction
problem. The empirical evidences indicate that cooperative approach offers the best execution
times and the same bounds compared to the independent and semiindependent approaches.
In “Applying hierarchical bayesian neural network in failure time prediction,” L.-J. Kao and
H.-F. Chen apply the Hierarchical Bayesian neural network HBNN approach to predict the
failure time and utilize the Gibbs sampler of Markov Chain Monte Carlo to estimate statistical
model parameters. The results of sensitivity analysis show that HBNN can provide not only
the predictive distribution but also the heterogeneous parameter estimates for each path.
In “Applying neural networks to prices prediction of crude oil futures,” J. Wei-Shan Hu
et al. attempt to accurately forecast prices of crude oil futures by adopting three popular
neural networks methods including the multilayer perceptron, the Elman recurrent neural
network, and recurrent fuzzy neural network RFNN. Experimental results indicate that the
use of neural networks to forecast the crude oil futures prices is appropriate and consistent
learning is achieved by employing different training times. Moreover, the RFNN outperforms
the other two neural networks in forecasting crude oil futures prices.
The paper “Solving the tractor and semi-trailer routing problem based on a heuristic
approach” by H. Li et al. develops a heuristic algorithm to solve the tractor and semi-trailer
routing problem TSRP. The computational study shows that their method takes relatively
little time to obtain satisfactory solutions.
L. Yao and J. Xu tackle a stone resource assignment problem with the aim of decreasing
dust and waste water emissions. In their paper “A stone resource assignment model under
the fuzzy environment” a bilevel multiobjective optimization model with fuzzy coefficients is
constructed. They design a fuzzy simulation-based improved simulated annealing algorithm
FS-ISA to search for Pareto optimal solutions. They also present a case study to demonstrate
the practicality and efficiency of their model.
“A modified PSO algorithm for minimizing the total costs of resources in MRCPSP” by M.
Khalilzadeh et al. introduces a multimode resource-constrained project scheduling problem
with finish-to-start precedence relations among project activities, considering renewable
and nonrenewable resource costs. The authors formulate this problem as a mixed-integer
programming model and present a metaheuristic algorithm based on a modified particle
swarm optimization approach. Experimental results reveal the effectiveness and efficiency
of their algorithm.
The paper “A nonlinear multiobjective bilevel model for minimum cost network flow problem
in a large-scale construction project” by J. Xu et al. deals with a minimum cost network flow
problem in a large-scale construction project using a nonlinear multiobjective bilevel model
4 Mathematical Problems in Engineering
with birandom variables. Results and analysis are presented to highlight the performances of
the proposed method which is more effective and efficient compared to a genetic algorithm
and a simulated annealing algorithm.
“Sparse signal recovery via ECME thresholding pursuits” by H. Song and G. Wang
develops ECME thresholding pursuits EMTPs for sparse signal recovery. Two effective
support detection strategies hard thresholding and dynamic thresholding are devised
for the sparse signals with components having a fast decaying distribution of nonzero
components. The experimental studies are presented to demonstrate that EMTP offers an
appealing alternative to state-of-the-art algorithms for sparse signal recovery.
Hybrid Methods
The paper “Solving constrained global optimization problems by using hybrid evolutionary
computing and artificial life approaches” by J.-Y. Wu presents a hybrid real-coded genetic
algorithm with a particle swarm optimization PSO algorithm and a hybrid artificial
immune algorithm with a PSO algorithm for solving 13 constrained global optimization
problems, including 6 nonlinear programming and 6 generalized polynomial programming
optimization problems. Experimental results indicate that the proposed algorithms converge
to a global optimum solution.
In “Combining diffusion and grey models based on evolutionary optimization algorithms
to forecast motherboard shipments,” F.-K. Wang et al. develop a combined model based on
the rolling Grey model and the Bass diffusion model to forecast motherboard shipments.
The results indicate that the proposed model using a hybrid algorithm outperforms other
methods for the fitting and forecasting processes in terms of mean absolute percentage error.
The paper “Hybrid optimization approach for the design of mechanisms using a new error
estimator” by A. Sedano et al. proposes a hybrid optimization algorithm that combines the
advantages of both stochastic and deterministic approaches to the design of mechanisms.
Two engineering applications, a four-bar linkage and an injection machine, are presented to
demonstrate the accuracy, robustness and efficiency of their proposed approach.
“A Hybrid genetic algorithm for the multiple crossdocks problem” by Z. Miao et al. considers
multiple crossdocks problem through fixed transportation schedules with time windows,
capacity, and penalty. The authors prove that the problem is NP-hard in the strong sense and
develop a hybrid genetic algorithm integrating greedy technique and variable neighborhood
search method to solve the problem effectively and efficiently.
“A new hybrid Nelder-Mead Particle swarm optimization for Coordination optimization of
directional overcurrent relays” by A. Liu and M.-T. Yang proposes a new hybrid Nelder-Mead
simplex search method and particle swarm optimization proposed NM-PSO algorithm to
solve the directional overcurrent relays coordination optimization problem. The findings
demonstrate that the performance of the proposed NM-PSO is better than that of PSO and
original NM-PSO in terms of computation speed, rate of convergence, and feasibility.
“A VNS metaheuristic with stochastic steps for Max 3-cut and Max 3-section” by A.-f. Ling
proposes a local search algorithm and a variable neighborhood global search algorithm with
two stochastic search steps to obtain the global solution by establishing a neighborhood
structure of the Max 3-cut problem. Numerical results show that the proposed heuristic
algorithm can obtain efficiently the high-quality solutions and has the better numerical
performance than 0.836-approximate algorithm for the NP-Hard Max 3-cut and Max 3-section
problems.
Mathematical Problems in Engineering 5
The paper “Dynamic programming and heuristic for stochastic uncapacitated lot-sizing
problems with incremental quantity discount” by Y. Zhanga et al. considers the stochastic
uncapacitated lot-sizing problems with incremental quantity discount where the uncertain
parameters are supposed to evolve as discrete time stochastic processes. The obtained results
reveal that the presented algorithm outperforms the commercial solver CPLEX and other
heuristics in both quality of solution and run time.
“A hybrid algorithm based on ACO and PSO for capacitated vehicle routing problems”
by Y. Kao et al. proposes a new hybrid algorithm based on two main swarm intelligence
approaches, ant colony optimization ACO and particle swarm optimization PSO, for solv-
ing capacitated vehicle routing problems. Computational results show that the performance
of the proposed method is competitive in terms of solution quality when compared with
existing ACO-and PSO-based approaches.
Challenging Applications
The paper “Goal-programming-driven genetic algorithm model for wireless access point deployment
optimization” by C.-S. Wang and C.-T. Chang proposes a method that integrates a goal-
programming-driven model and a genetic algorithm to resolve the multiple objectives appro-
priate wireless access point deployment problem. Three experiment results demonstrate the
utility and stability of the proposed method.
S.-Y. Lin et al. paper “A cost-effective planning graph approach for large-scale web service
composition” proposes a novel cost-effective Web service composition mechanism, utilizes
planning graph based on backward search algorithm to find multiple feasible solutions, and
recommends a best composition solution according to the lowest service cost.
In “A fuzzy dropper for proportional loss rate differentiation under wireless network with
a multi-state channel,” Y.-C. Szu proposes a novel packet dropper for fuzzy controlling
the proportional loss rate differentiation in a wireless network with multi-state channel.
Simulation results reveal that the fuzzy proportional loss rate dropper does achieve accurate
loss rate proportion, lower queuing delay and loss rate, and higher throughput, compared
with other methods in the wireless environment.
The paper “Evaluating the performance of taiwan homestay using analytic network process”
by Y.-C. Hu and J.-H. Wang develops and constructs a set of evaluation indicators tailor-
made for homestay sector through discussion of literatures and interviewing experts so that
the evaluation framework would be more comprehensive and more practical.
The paper “Improved degree search algorithms in unstructured P2P networks” by G. Liu et
al. proposes two memory function degree search algorithms: memory function maximum
degree algorithm and memory function preference degree algorithm and studies their
performance including the search success rate and the search message quantity in scale-free
networks, random graph networks, and small-world networks.
The paper “Variable neighborhood search for parallel machines scheduling problem with
step-deteriorating jobs” by W. Cheng et al. studies a scheduling problem of minimizing the
total completion time on identical parallel machines where the processing time of a job is a
step function of its starting time and a deteriorating date that is individual to all jobs. The
computational results show that the proposed approaches obtain near-optimal solutions in a
reasonable computational time even for large-sized problems.
The paper “Solving packing problems by a distributed global optimization algorithm” by
N.-Z. Hu et al. develops a novel method to convert the nonlinear objective function in a
packing program into an increasing function with a single variable and two fixed parameters.
6 Mathematical Problems in Engineering
The transformed linear program is then decomposed into several subproblems by specifying
various parameter values, which is solvable simultaneously by a distributed computation
algorithm to obtain a global optimum.
The paper “A hybrid network model to extract key criteria and its application for brand equity
evaluation” by C.-Y. Chen and C.-W. Li develops a hybrid model based on social network
analysis and maximum mean deentropy algorithms for extracting evaluation criteria which
considers substitutions between the criteria. The effectiveness and feasibility of the hybrid
model are demonstrated by examples of evaluating brand equity.
The paper “Mixed mortar element method for P1NC /P0 element and its multigrid method
for the incompressible stokes problem” by Y. Jiang and J. Chen discusses a mortar-type P1NC /P0
element method for the incompressible Stokes problem. The inf-sup condition and the
optimal error estimate are proved. The study also proposes a W-cycle multigrid for solving
the discrete problem and demonstrates the optimal convergence of the multigrid method.
In “Quality improvement and robust design methods to a pharmaceutical research and
development” by B. R. Cho and S. Shin, new robust design experimental and optimization
models for time-oriented data are applied to the pharmaceutical production research and
development. Compared to the traditional Taguchi optimization model, the proposed experi-
mental methodology is particularly useful for experiments with time-oriented pharmaceuti-
cal characteristics.
The paper “A novel method for technology forecasting and developing R&D strategy of
building integrated photovoltaic technology industry” by Y.-J. Chiu and T.-M. Ying proposes
a hybrid approach to explore the life cycle of building-integrated photovoltaic BIPV
technology and develop the R&D strategy of related industries. The proposed approach
comprises patent analysis, logistic growth model analysis, and patent matrix map analysis.
The authors also provide three-dimensional matrix of degree of protection, R&D capability,
and benefit creation to select R&D strategies for BIPV industry.
J.-H. Lin et al. propose a systematic method to analyze student recruitment numbers
for future needs, based on the concept of material resource planning MRP. In their paper
“The number of students needed for undeclared programs at a college from the supply-chain view-
point,” the relationship between a curricular structure tree and the associated commonalities
is studied and a quantified model of commonality and recruitment planning for appropriate
curriculum design is proposed. The authors use two simple examples to illustrate the
implementation of MRP in analysis of the replenishment levels in an education system.
The paper “A selection approach for optimized problem-solving process by grey relational
utility model and multi-criteria decision analysis” by C.-K. Ke and M.-Y. Wu proposes an
approach to assist workers in determining the optimal selection order of candidate actions
based on a Grey relational utility model and a multi-criteria decision analysis. Experimental
results from analyzing a high-tech company’s knowledge base log demonstrate that their
selection approach is effective.
M.-H. Lin and H.-J. Hsu addresses an incentive pricing problem for relaying services
in multi-hop cellular networks. In their paper “Optimal incentive pricing on relaying services for
maximizing connection availability in multi-hop cellular networks” a mathematical programming
model is constructed and then solved to determine an optimal incentive price for each
intermediate node providing relaying services. The computational results demonstrate that
their proposed approach maximizes connection availability of the networks compared to
fixed-rate or location-based methods.
“A two stage dea to analyze the effect of entrance deregulation on iranian insurers: a robust
approach” by S. G. J. Naini et al. analyzes technical efficiency for Iranian insurance companies
Mathematical Problems in Engineering 7
between 2003 and 2010, a period that insurers experienced intense volatility due to the
entrance deregulation of the market by two-stage data envelopment analysis. The major
results show that ownership type and failure to meet the risk management rules are the main
drivers of efficiency.
“A label correcting algorithm for partial disassembly sequences in the production planning for
end-of-life products” by P.-F. Tsai investigates a single period partial disassembly optimization
problem to generate an optimal disassembly sequence in product recovery of the end-of-life
products. The study presents a heuristic procedure that utilizes a label correcting algorithm
which is a polynomial-time algorithm to find a solution. Numerical examples are also used
to demonstrate the effectiveness of the proposed solution procedure.
Finally, the guest editors would like to thank the authors for their contributions, the
referees for their time and energy in reviewing the paper, and all the staff of Hindawi involved
in the preparation of this special issue for their support to publish the articles related to this
subject. Without their cooperation, it would have not been possible to edit this special issue.
Jung-Fa Tsai
John Gunnar Carlsson
Dongdong Ge
Yi-Chung Hu
Jianming Shi
Hindawi Publishing Corporation
Mathematical Problems in Engineering
Volume 2012, Article ID 265837, 26 pages
doi:10.1155/2012/265837
Research Article
A Stone Resource Assignment Model under
the Fuzzy Environment
Copyright q 2012 Liming Yao et al. This is an open access article distributed under the Creative
Commons Attribution License, which permits unrestricted use, distribution, and reproduction in
any medium, provided the original work is properly cited.
This paper proposes a bilevel multiobjective optimization model with fuzzy coefficients to tackle
a stone resource assignment problem with the aim of decreasing dust and waste water emissions.
On the upper level, the local government wants to assign a reasonable exploitation amount to each
stone plant so as to minimize total emissions and maximize employment and economic profit.
On the lower level, stone plants must reasonably assign stone resources to produce different stone
products under the exploitation constraint. To deal with inherent uncertainties, the object functions
and constraints are defuzzified using a possibility measure. A fuzzy simulation-based improved
simulated annealing algorithm FS-ISA is designed to search for the Pareto optimal solutions.
Finally, a case study is presented to demonstrate the practicality and efficiency of the model.
Results and a comparison analysis are presented to highlight the performance of the optimization
method, which proves to be very efficient compared with other algorithms.
1. Introduction
The dust and the waste water from the stone industry can cause serious damage to the
regional ecological environment. The overexploitation and the stone processing have resulted
in the vegetation decrement, and the pollution of air and water in those areas with rich
stone resources. The annual amount of waste generated include 700,000 tons of slurry
waste as well as 1 million tons of solid waste. The consequent dumping of this waste
in open areas has created several environmental problems and has negatively impacted
agriculture, local inhabitants, and groundwater 1. Therefore, it is urgent to normalize
the quarrying and processing of the stone resource. Some technologies are introduced to
save energy and reduce the emission in the stone industry by many scholars 2, 3. Some
other scholars 4–6 considered the use of the marble powder to reduce the waste, but few
literatures discussed the quantitative relationship between the emission and the exploiting
2 Mathematical Problems in Engineering
and processing amount. In fact, a reasonable assignment of stone resources could significantly
reduce the emissions. This paper considers the government as the upper level and the stone
plant as the lower level to develop a bi-level model. For the stone industry, the objectives
of the government authority are to minimize environmental pollution and maximize social
employment and economic revenue. This can be achieved by optimizing the amount of stone
extracted and exploited between the participating plants, which are assumed to cooperate
and act as a lower-level decision maker. Noting that industrial symbiosis implicitly requires
the cooperative behavior of the participants 7, 8, the government can influence the plants
by imposing disincentives by assigning different amounts to stone plants according to their
production scale and clean technology level. The plants operate independently of each other.
Each plant has its own goals, which are to maximize the profit from the sale of nano
calcium carbonate, marble products, granite slabs, and man-made slabs and to minimize the
emissions of stone dust and waste water.
To develop the bi-level optimization model for assigning the stone resources, some
emission coefficients have to be effectively estimated. It is usually difficult to collect the
exact data of emissions of stone dust and waste water when exploiting the stone mine and
processing stone products. The fuzzy number is an efficient tool to describe the variables
without crisp information. The membership function of fuzzy sets can be used to describe
the possibility that emission coefficients take the value according to the experience of those
people in the stone industry. Actually, there has been some studies describing the uncertainty
by fuzzy sets. For example, Petrovic et al. 9 used fuzzy sets to describe the customer
demand, supply deliveries along the SC and the external or market supply, and develop a
supply chain model with fuzzy coefficients. Lee and Yao 10 fuzzify the demand quantity
and the production quantity per day to solve the economic production quantity. These studies
inspire us to use the fuzzy sets to interpret the vague and imprecise about the emissions of
stone dust and waste water. For the fuzzy bi-level optimization problem, a satisfactory near-
optimal or “satisficing” solution can be reached by providing tolerances in the objective
functions and constraints and by defining corresponding degrees of satisfaction through
membership functions to indicate the preference of the decision makers which is typical of
decision making in a fuzzy environment 11. The followers then communicate their results
to the leader, who modifies his goals and control variables if the original tolerances are not
met. The process continues iteratively until a solution which satisfies the goals of both leader
and follower is reached.
A bi-level multiobjective model with fuzzy coefficients is always an NP hard problem,
and it is especially difficult for nonlinear bi-level programming under a fuzzy environment
to find a numerical solution. Some existing methods mainly focus on metaheuristics which
include the genetic algorithm 12, the simulated annealing 13, and the hybrid tabu-ascent
algorithm 14. However, as these need to be designed for single-objective problems with
crisp coefficients, it is difficult to find a usual or normal pattern for a bi-level model with
fuzzy coefficients. This paper proposes an improved simulated annealing based on a fuzzy
simulation to search for a Pareto optimal solution after a possibilistic check. The following
sections of this paper are organized as follows. In Section 2, the reason a bi-level multi-
objective model is used to optimize the stone industry is explained. The process of data
fuzzification is introduced in detail. A possibilistic bi-level multi-objective programming
model is developed. In Section 3, a fuzzy simulation-based improved simulated algorithm
is proposed to solve the bi-level multi-objective programming model with fuzzy coefficients.
In Section 4, a practical case is presented to show the significance of the proposed models and
algorithms. Finally, conclusions are given in Section 5.
Mathematical Problems in Engineering 3
Minimum pollution
Maximum revenue
Followers
··· ···
Stone-material Stone-material Stone-material
plant 1 plant i plant N
2. Mathematical Modelling
In order to develop the mathematical model, some basic background and descriptions are
introduced.
Primary Secondary
Transportation to Transportation
cutting/product Finishing cutting/product Packing
processing facility to storage
shaping shaping
Transportation to
Exploring and crushing at
stone mine Stone dust
Fuzzy emission coe fficient crushing facility
economic profit as their first goal. Due to a limitation on the amount that can be extracted and
the environmental protection requirements, they also have to consider minimizing emissions.
At the same time under a government policy and according to capacity, they also need to
think about employment. In addition, it is necessary to increase the investment to improve
emission reduction capacity to satisfy sustainable development requirements. Considering
the above, the problem should be regarded as a bi-level optimization model in which the
government authority is the upper-level decision maker and the stone plants are the lower-
level decision makers. It is assumed that there is a perfect exchange of information between
all the participants such that the objectives and constraints are known.
As shown in Figure 2, the granite is first exploited from the stone mine and cut into the
primary products such as granite slabs and man-made slabs; then, these are processed into
the floor or other products. Scrap materials are usually processed into fine powder calcium
carbonate and nano calcium carbonate to meet market demand. During the complete process,
a great deal of stone dust and waste water are produced. Since it is technically difficult to
collect the exact data of emissions, we usually make a rough estimation by the difference
of weight before and after exploiting and processing and then look for the possibility for
every weight by the professional advices in the stone industry. Therefore, the fuzzy number
is an efficient tool to describe this situation by its membership function. Actually, the fuzzy
environment has been successfully studied and applied in many areas, such as flow shop
scheduling problem 17, supply chain problem 18, and assignment problem 19. These
studies show the necessity of considering fuzzy environment in practical problems. It is
also the motivation for considering fuzzy environment in the stone resources assignment
problems.
1 Emission of stone dust and waste water is proportional to the amount of stone
processed into products.
2 Employment level is also proportional to the amount of stone processed into
products.
Mathematical Problems in Engineering 5
3 The constant cost of product j only exists when the stone-material plant produces
product j.
4 Since the government endows different subsidies to plants, it is assumed that each
plant has its own tax rate Si and the tax is proportional to the turnover of all stone
products.
The notations are used to describe the subsidy model in the investigation are referred
to in the Abbreviations Section.
Step 2. Find the smallest, middle, and largest data in S; denote them Eds , Edm , and Edl ,
respectively.
Step 3. Compute the left and right slops for the data in Str by the following equations;
respectively,
⎧
⎪
⎪ Eds Edm − 2x Eds Edm
⎪
⎨ , if Eds ≤ x ≤ ,
Eds Edm 2
Edα x 2.1
⎪ 2x − Eds Edm
⎪ Eds Edm
⎪
⎩ , if ≤ x ≤ Edm ,
2Edm 2
⎧
⎪
⎪ Edl Edm − 2x Edl Edm
⎪
⎨ , if Eds ≤ x ≤ ,
Edl Edm 2
Edβ x 2.2
⎪
⎪ 2x − Edl Edm Edl Edm
⎪
⎩ , if ≤ x ≤ Edm .
2Edm 2
Then we get the set of left slops EdL {α | Edα x, x ∈ Str , x ≤ Edm } and the set of left slops
EdR {β | Edβ x, x ∈ Str , x ≥ Edm }.
6 Mathematical Problems in Engineering
⎧
⎪
⎪ 0, if x < Eds ,
⎪
⎪ α
⎪
⎪ x − Eds Eds Edm
⎪
⎪ 2α−1
, if Eds ≤ x ≤ ,
⎪
⎪ Edm − Eds 2
⎪
⎪
⎪
⎪ α
⎪
⎪1 − 2α−1 Edm − x
⎪
⎪ , if
Eds Edm
< x < Edm ,
⎪
⎪ Edm − Eds 2
⎪
⎪
⎨
x
μEd 2.3
⎪
⎪ x − Edm β Edl Edm
⎪
⎪ 1 − 2β−1
, if Edm ≤ x < ,
⎪
⎪ Edl − Edm
⎪
⎪ 2
⎪
⎪
⎪
⎪
⎪
⎪ β
⎪
⎪ Edl − x Edl Edm
⎪2β−1
⎪ ≤ x ≤ Edl ,
⎪
⎪ , if
⎪
⎪ Edl − Edm 2
⎩
0, if x > Edl ,
x; α, β,
Step 5. Take all the data in Sv in the above equation and compute the membership μEd
where x ∈ Sv , α ∈ EdL and β ∈ EdR .
Step 6. Carry out the crossover validation test proposed by Kohavi 26. Compute the
memberships of xi ∈ Sv for any combination α, β ∈ EdL , EdR . Then compute the
percentage of correct results by the following equation:
1
PCCv 100 × δ xi , μEd
xi ; α, β , 2.4
NV
xi ,μEd
xi ;α,β∈Sv
where PCCv denotes the percentage of correct results over the validation set Sv , NV
is the number of data points in validation set Sv , and δμEd xi ; α1 , β1 , μEd
xi ; α2 , β2
1 if μEd xi ; α1 , β1 μEd xi ; α2 , β2 , while δμEd
xi ; α1 , β1 , μEd
xi ; α2 , β2 0 if
xi ; α1 , β1 /
μEd μEd
xi ; α2 , β2 .
Step 7. Find the combination α, β by which the largest percentage of correct results can be
obtained when carrying out the crossover validation test with each other. Then we get the
membership function.
min F 1 2.5
subject to s.t.
⎧ ⎫
⎨ m m n ⎬
Pos i Yi
Ed ij Xij ewij Xij ≤ F 1
ed ≥ δ1U , 2.6
⎩ i1 i1 j1
⎭
where Pos is the possibility measure proposed by Dubois and Prade 28 and δU is the
possibilistic level representing the possibility that decision makers achieve the minimum
objective. All fuzzy arithmetic in 2.6 and the following equations come from the operation
proposed by Kaufmann and Gupta 29.
To achieve maximum employment F2 which consisted of constant workers Pi and
variable workers pij Xij , the following objective function is obtained:
⎛ ⎞
m n
max F2 ⎝ pij Xij Pi ⎠ . 2.7
i1 j1
To achieve the maximum economic output which can be obtained by multiplying unit
amount cj , conversion rate θij , and amount of stone Xij , the following objective function
is obtained:
⎛ ⎞
m n
max F3 Si ⎝ cj θij Xij ⎠. 2.8
i1 j1
Generally, some mandatory conditions must be satisfied when the government makes
a decision. These are listed as follows.
8 Mathematical Problems in Engineering
m
The total exploration quantity i1 Yi cannot exceed the upper limitation RU of the
total stone resources in the region:
m
Yi ≤ RU . 2.9
i1
⎧ ⎫
⎨m n ⎬
Pos ewij Xij ≤ EW U ≥ δ3U . 2.11
⎩ i1 j1 ⎭
The output of some products mi1 θij Xij should meet the market demand Dj . For
L
example, the nano calcium carbonate is very popular in many areas, so the stone plants
should provide enough output to meet the demand:
m
θij Xij ≥ DjL ∀j. 2.12
i1
Every plant also wishes to achieve minimum emissions. However, since the emissions
ij and ewij are fuzzy numbers, it is usually difficult to determine the precise minimum
ed
Mathematical Problems in Engineering 9
2
emissions, and decision makers only require a minimum objective H i under some
possibilistic level σiL . Hence, the possibilistic constraint is as follows:
2
min H i 2.15
subject to
⎧ ⎫
⎨ m n 2
⎬
Pos ij Xij ewij Xij ≤
ed Hi ≥ σiL , 2.16
⎩ i1 j1 ⎭
where σiL is the possibilistic level under which decision makers require the minimum
objective.
Since production in all the plants is influenced by government policy and market
demand, there are some conditions that need to be satisfied.
The amount used for production nj1 Xij should not exceed the total limitation Yi :
n
Xij ≤ Yi . 2.17
j1
n
The inventory amount Yi − j1 Xij should not exceed the maximum limitation
IViU :
n
Yi − Xij ≤ IViU . 2.18
j1
The production cost which consisted of two parts including product cost nj1 fXij
and total inventory cost nj1 hi Yi − nj1 Xij should not exceed the predetermined level
P CiU : ⎛ ⎞
n n
f Xij hi ⎝Yi − Xij ⎠ ≤ P CiU . 2.19
j1 j1
Some products θij Xij should not be less than the lowest production level PijL in
plant i:
⎧
⎪
⎪ min F 1
⎛ ⎞
⎪
⎪
⎪
⎪ m n
⎪
⎪ ⎝ pij Xij Pi ⎠
⎪
⎪ max F2
⎪
⎪
⎪
⎪ i1
⎛j1
⎞
⎪
⎪
⎪
⎪
⎪
⎪
m n
⎪
⎪
⎪ max F3 Si ⎝ cj θij Xij ⎠
⎪
⎪
⎪
⎪ ⎧ ⎧
i1 j1
⎫
⎪
⎪ ⎪ ⎨ m n ⎬
⎪
⎪ ⎪ m
⎪
⎪ ⎪
⎪ i Yi ij Xij ewij Xij ≤ F 1 ≥ δU
⎪
⎪ ⎪
⎪ Pos Ed ed
⎪
⎪ ⎪
⎪ ⎩ i1 ⎭ 1
⎪
⎪ ⎪
⎪ i1 j1
⎪
⎪ ⎪
⎪
⎪
⎪ ⎪
⎪ ⎧ ⎫
⎪
⎪ ⎪
⎪ ⎨m ⎬
⎪
⎪ ⎪
⎪ m n
⎪
⎪ ⎪
⎪ i Yi ij Xij ≤ EDU ≥ δU
⎪
⎪ ⎪
⎪ Pos
⎩ i1
Ed ed
⎭ 2
⎪
⎪ ⎪
⎪
⎪
⎪ ⎪
⎪
i1 j1
⎪
⎪ ⎪
⎪
⎪
⎪ ⎪
⎪ ⎧ ⎫
⎪
⎪ ⎪
⎪ ⎨m n ⎬
⎪
⎪ ⎪
⎪
⎪
⎪ ⎪
⎪ Pos ew X ≤ EW U
≥ δ3U
⎪
⎪ ⎪
⎪ ⎩ i1 j1 ij ij
⎭
⎪
⎪ ⎪
⎪
⎪
⎨ ⎪
⎪
⎪
⎪
⎪
⎪ 2.21
⎪ ⎪
⎪
m
⎪
⎪ ⎪
⎪ θij Xij ≥ DjL ∀j
⎪
⎪ ⎪
⎪
⎪
⎪ ⎪
⎪⎧ ⎛ ⎞
⎪
⎪ ⎪
⎪
i1
⎪
⎪ ⎨⎪⎪ n n n
⎪ s.t. ⎪⎪
⎪
⎪
⎪ ⎪
⎪ max Hi1 cj θij Xij − f Xij − hi ⎝Yi − Xij ⎠
⎪
⎪ ⎪
⎪ ⎪
⎪
⎪
⎪ ⎪
⎪ ⎪
⎪ j1 j1 j1
⎪
⎪ ⎪
⎪ ⎪
⎪
⎪
⎪ ⎪
⎪ ⎪
⎪
2
⎪
⎪ ⎪
⎪ ⎪
⎪ min ⎧ Hi ⎧ ⎫
⎪
⎪ ⎪
⎪ ⎪
⎪
⎪
⎪ ⎪
⎪ ⎪
⎪ ⎪
⎪ ⎨ m n ⎬
⎪
⎪ ⎪
⎪ ⎪
⎪ ⎪
⎪ Pos ij Xij ewij Xij ≤ H 2 ≥ σ L
⎪
⎪ ⎪
⎪ ⎪
⎪ ⎪
⎪ ed
⎪
⎪ ⎪
⎪⎪⎪ ⎪
⎪ ⎩ i1 j1 i
⎭ i
⎪
⎪ ⎪
⎪ ⎪
⎪ ⎪
⎪
⎪
⎪ ⎪
⎪ ⎨ ⎪
⎪
⎪
⎪ ⎪
⎪ ⎪
⎪
n
⎪
⎪ ⎪
⎪ ⎪
⎪ Xij ≤ Yi
⎪
⎪ ⎪
⎪⎪⎪ ⎪
⎪ j1
⎪
⎪ ⎪
⎪ ⎪
⎪ ⎪
⎨
⎪
⎪ ⎪
⎪ ⎪
⎪
⎪
⎪ ⎪
⎪ ⎪
⎪ s.t.
n
⎪
⎪ ⎪
⎪ ⎪
⎪ ⎪ Y − Xij ≤ IViU
⎪
⎪ ⎪
⎪ ⎪
⎪ ⎪
⎪
i
⎪
⎪ ⎪⎪
⎪ ⎪ ⎪
⎪ j1
⎛ ⎞
⎪
⎪ ⎪⎪
⎪ ⎪ ⎪
⎪
⎪
⎪ ⎪
⎪ ⎪
⎪ ⎪
⎪ n n
⎪
⎪ ⎪
⎪ ⎪
⎪ ⎪
⎪
⎪
⎪ ⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪ f Xij hi ⎝Yi − Xij ⎠ ≤ P CiU
⎪
⎪ ⎪
⎪ ⎪ ⎪
⎪
⎪ ⎪ ⎪ ⎪
⎪ ⎪⎪
j1 j1
⎪ ⎪ ⎪ ⎪
⎪
⎩ ⎩⎪⎩ ⎩θij Xij ≥ P L .
ij
3. Solution Approach
Generally, bi-level programming is an NP-hard problem, and it is difficult to determine
an optimal solution 30–32. In the proposed model, decision makers on the upper and
lower levels have to face more than two conflicting objectives and then make a decision
under a fuzzy environment. This significantly increases the difficulty of finding an optimal
strategy for both the upper and lower levels. Therefore, the fuzzy simulation-based improved
Mathematical Problems in Engineering 11
simulated annealing FS-ISA is designed to solve the bi-level optimization model with fuzzy
coefficients.
If mi1
n
j1 κij Xij ≤ EW
U
and L < μ, then we set L μ. Repeat this process N times. The
value L is regarded as an estimation of the possibility. Then the process for constrain check
can be summarized in Procedure 2.
simulated annealing algorithms including SMOSA, UMOSA, PSA, and WMOSA to solve
multiobjective optimization of constrained problems with varying degree of complexity and
then proposed a new algorithm PDMOSA. Sanghamitra et al. 41 proposed a simulated
annealing-based multiobjective optimization algorithm AMOSA that incorporates the
concept of archive in order to provide a set of trade-off solutions for the problem under
consideration.
In the following part, we will incorporate the fuzzy simulation into the SMOSA
algorithm proposed by Suppapitnarm and Parks 16 and use the interactive method
to search the Pareto optimal solution for the bi-level multiobjective optimization with
fuzzy possibilistic constraints. Take the problem A.10 as an example and denote Xi
Xi1 , Xi2 , . . . , Xin and Y Y1 , Y2 , . . . , Ym , the process of FS-ISA can be summarized in
Procedure 3.
Above all, the whole procedure of FS-ISA for bi-level multiobjective optimization
problems with fuzzy coefficients is described in Figure 3.
4. A Case Study
In the following, a practical example in China is introduced to demonstrate the complete
modelling and algorithm process.
Mathematical Problems in Engineering 13
approach proposed by Xu and Li 15. If G1∗ is more optimal than G1 , let G G1∗ . If not G G1
Step 9. Periodically, restart with a randomly selected solution from the Pareto set.
While periodically restarting with the archived solutions, Suppapitnarm et al. 16
have recommended biasing towards the extreme ends of the trade-off surface;
Step 10. Periodically reduce the temperature by using a problem-dependent annealing schedule
Step 11. Repeat steps 2–10, until a predefined number of iterations is carried out.
Initialization: t0 and Y
(check the feasibility by fuzzy simulation)
Satisfy Y Output
terminal conditions? results
Feasibility N
tk+1 = λ∗ tk check ?
Y
Check
min(1, exp(−∆f/tk )) N
≥ RND(0, 1)
Y
1
Accept Y as a new input for the lower
level, and solve the programming to get
∗ ∗
the new solution G1 = (X 1 , Y 1 )
to satisfy the constrained index in a short time due to uncertainty, the possibility of holding
the two constraints should not be less than 0.9 which indicates that the possibilistic levels δ2U
and δ3U for the government should also be 0.9. For total emissions, the environmental sector
requires the minimum objective to be under the possibilistic level δ1U 0.85. As the demand
and the price of the four stone products sharply increase, the government requires that their
output from all the plants should at least satisfy the basic market demand DjL j 1, . . . , 4 as
Mathematical Problems in Engineering 15
Granite Man-made
Stone
slab material slab
Nano
Granite calcium
sand carbonate
Stone processing
in Table 2. Each stone products’ unit price is in Table 2. For the 10 stone plants, the inventory
and the production upper limitations for each plant are listed in Table 1. The possibilistic
level δiL that plant i needs to obtain for minimum emissions is in Table 1. Since every plant
has a different capacity for controlling emissions, the fixed and unit variable cost, emission
coefficients, and constant costs are different as outlined in Table 3. The transformation rate θij
and the lower limitation of product j in plant i are also listed in Table 3.
16 Mathematical Problems in Engineering
Parameters
Stone plants
Edi kg/m3 Pi Person Si hi Yuan/m3 i M m
IV U 3
i M Yuan
P CU
Kai Quan 24.5, 25.7, 27.8 65 0.3655 0.32 4.5 1350
Feng Huang 21.2, 26.8, 29.3 60 0.3705 0.35 6.2 1670
Li Du 19.2, 23.8, 25.7 65 0.3735 0.33 4.3 1410
Hong Yuan 22.6, 27.9, 28.3 62 0.3615 0.28 7.8 1630
Xiang Zong 21.0, 23.5, 27.1 60 0.3675 0.38 2.3 1000
Ji Cheng 20.6, 23.3, 27.9 55 0.3485 0.42 2.6 1100
Hui Huang 18.3, 21.2, 25.2 60 0.3525 0.35 5.4 1650
Hong Yun 19.2, 24.8, 29.1 60 0.3725 0.36 4.6 1320
De Sheng 25.4, 28.1, 30.3 70 0.3615 0.33 5.6 1230
Guo Jian 27.2, 29.3, 32.8 80 0.3435 0.41 6.2 1670
Stone products
Parameters
Nano calcium carbonates Granite slabs Granite sand Man-made composite slabs
cj 1325 Yuan/ton 65 Yuan/m2 25 Yuan/ton 30 Yuan/m2
Dj 8.5 × 107 ton 3.06 × 108 m2 5.13 × 106 ton 1.56 × 108 m2
Taking all the numerical values into 2.21 and setting the initial temperature T0 500,
the last temperature is 0 and the cooling method is 1 decrement once. The neighbourhood
can be developed as Yi1 Yi0 rh and Xij1 Xij0 rh, where r is a random number in
−1,1 and h is the step length here h 2.0. After a simulation of many cycles, the Pareto
optimal solution and theobjective value are determined as shown in Tables 4 and 5. The
results illustrate that although some plants have the highest productive efficiency, their high
emission coefficient will result in the low exploiting quotas such as Kai Quan, Guo Jian, and
De Sheng. On the other hand, stone plants will tend to produce the high value-added but low
emission products due to the environmental pressure and the limitation of exploiting quotas,
such as nano calcium carbonates and man-made composite slabs. However, stone plants will
abundantly produce the traditional products such as granite slabs because of the huge cost of
those new products.
Parameters
Stone plants Stone products
Pij tij Cij θij PijL edij ewij
NPCC 0.03 650 286 1.03 7.0 2.21, 3.42, 5.23 3.15, 3.42, 4.21
GSl 0.01 38 367 3.62 262.2 21.2, 22.3, 25.6 23.1, 25.2, 27.9
Kai Quan
GSa 0.01 10 80 0.95 0 26.4, 28.5, 32.7 0.47, 1.32, 1.44
MmCS 0.02 22 440 8.34 35.5 1.57, 2.68, 4.39 2.32, 3.18, 3.69
NPCC 0.03 580 254 1.12 7.2 2.68, 3.67, 4.59 2.86, 3.39, 6.14
GSl 0.01 44 380 3.14 232.3 24.7, 26.3, 27.6 24.5, 25.9, 27.0
Feng Huang
GSa 0.01 16 68 0.90 2.4 27.6, 29.0, 33.9 0.58, 1.96, 2.40
MmCS 0.02 24 448 8.11 43.1 2.03, 3.56, 4.57 1.84, 2.68, 3.57
NPCC 0.03 620 272 0.85 9.9 1.86, 2.31, 4.03 2.85, 3.57, 4.37
GSl 0.01 40 392 2.67 279.0 17.4, 19.3, 20.2 25.8, 26.3, 27.1
Li Du
GSa 0.01 13 88 0.93 1.9 23.5, 26.5, 28.7 0.78, 1.56, 2.23
MmCS 0.02 26 380 8.35 67.4 1.45, 2.23, 4.05 3.10, 3.50, 3.87
NPCC 0.03 685 310 0.78 0 3.05, 4.21, 5.67 2.86, 3.33, 4.45
GSl 0.01 42 370 3.78 298.8 20.8, 23.3, 23.9 22.8, 24.8, 25.1
Hong Yuan
GSa 0.01 13 86 0.92 2.1 26.5, 28.3, 30.4 0.68, 1.45, 1.67
MmCS 0.02 23 380 8.26 22.0 1.57, 2.68, 4.39 2.30, 3.32, 4.22
NPCC 0.03 632 267 1.26 10.2 2.17, 3.33, 4.78 3.04, 3.57, 4.32
GSl 0.01 44 380 3.82 307.2 20.3, 23.5, 26.7 24.5, 26.3, 27.4
Xiang Zong
GSa 0.01 12 82 0.91 1.9 26.4, 27.3, 29.3 0.47, 1.23, 1.78
MmCS 0.02 20 350 8.27 39.4 1.46, 2.79, 3.45 2.11, 3.26, 4.45
NPCC 0.03 630 264 1.26 11.1 2.14, 3.39, 5.46 2.80, 3.24, 3.70
GSl 0.01 42 354 3.46 344.5 22.5, 23.8, 24.7 22.9, 25.1, 26.3
Ji Cheng
GSa 0.01 12 85 0.92 2.2 25.8, 27.9, 28.9 0.54, 1.26, 1.87
MmCS 0.02 23 350 8.26 0 1.76, 2.77, 4.25 2.78, 3.32, 4.45
NPCC 0.03 635 260 1.12 12.7 2.15, 3.56, 5.00 3.04, 3.37, 3.89
GSl 0.01 35 340 3.87 451.7 20.7, 23.2, 24.7 24.3, 26.3, 26.9
Hui Huang
GSa 0.01 15 85 0.90 0 28.6, 29.9, 33.4 0.58, 1.40, 1.72
MmCS 0.02 20 350 8.26 43.6 1.68, 2.70, 4.25 2.68, 3.27, 3.54
NPCC 0.03 660 290 1.00 10.0 2.54, 3.68, 5.42 3.35, 3.56, 4.04
GSl 0.01 40 380 3.54 331.3 22.3, 23.5, 24.2 21.5, 24.6, 26.2
Hong Yun
GSa 0.01 12 85 0.90 2.4 25.3, 26.7, 30.3 0.85, 1.56, 1.78
MmCS 0.02 24 385 8.42 0 1.33, 2.55, 4.72 2.24, 3.76, 3.87
NPCC 0.03 630 276 1.12 8.0 2.35, 3.67, 4.68 3.27, 4.18, 4.92
GSl 0.01 42 383 3.11 203.5 20.5, 21.4, 23.8 22.7, 24.7, 26.8
De Sheng
GSa 0.01 13 85 0.92 0 23.6, 25.2, 28.6 0.87, 1.63, 1.72
MmCS 0.02 25 378 8.02 48.0 1.67, 2.78, 4.23 2.45, 3.95, 4.51
NPCC 0.03 780 320 1.88 12.5 3.14, 4.37, 7.86 4.22, 4.78, 5.39
GSl 0.01 40 380 3.62 210.4 21.2, 22.3, 25.6 23.1, 25.2, 27.9
Guo Jian
GSa 0.01 13 86 0.91 0 27.3, 29.4, 33.8 0.68, 1.46, 1.57
MmCS 0.02 22 367 8.13 51.1 1.32, 2.59, 4.21 2.36, 3.67, 4.82
NPCC: Nano calcium carbonates; GSl: granite slabs; GSa: granite sand; MmCS: man-made composite slabs.
18 Mathematical Problems in Engineering
Stone products
Stone plants
Total Nano calcium carbonates Graniteslabs Granitesand Man-made composite slabs
Kai Quan 85.2 6.82 72.40 1.70 4.30
Feng Huang 88.6 6.47 74.16 2.66 5.32
Li Du 126.2 11.61 104.49 2.02 8.08
Hong Yuan 91.7 7.70 79.05 2.29 2.66
Xiang Zong 95.4 8.11 80.42 2.10 4.77
Ji Cheng 112.9 8.81 99.58 2.37 2.15
Hui Huang 135.4 11.37 116.71 2.03 5.28
Hong Yun 112.5 10.01 93.60 2.70 6.19
De Sheng 79.8 7.18 65.44 1.20 5.99
Guo Jian 72.3 6.65 58.13 1.23 6.29
Notation F1∗ F2 F3 H11 H21 H31 H41 H51 H61 H71 H81 H91
δiU 0.95 66289 12841 61240 5443 5281 9722 5648 6304 7403 10387 7930 5419
δiU 0.90 68362 13216 62530 5587 5362 9910 5753 6421 7489 11253 8016 5578
δiU 0.85 69137 13781 63110 5612 5374 9983 5842 6511 7570 11891 8117 5632
1
Notation H10 H12∗ H22∗ H32∗ H42∗ H52∗ H62∗ H72∗ H82∗ H92∗ 2∗
H10 —
δiU 0.95 3990 3524 3994 4894 3887 4085 4953 5867 4624 3112 2866 —
δiU 0.90 4114 3678 4953 4930 3922 4137 4953 5952 4731 3220 2917 —
δiU 0.85 3990 3524 3994 4894 3887 4236 5078 6013 4827 3315 3013 —
increases, the government requirements are more strict and hence the total emissions decrease
and the government tax revenue increases.
Similarly, for the following level, if the possibilistic levels δiL i 1, 2, . . . , 10 decrease,
the plants pay less attention to the stone dust and waste water emissions resulting in an
increase in profit and consequently more emission.
Stone products
Stone plants
Nano calcium Granite Granite Man-made
Total
carbonates slabs sand composite slabs
Kai Quan 1.32% 0.25% −0.56% 0.32% 0.08%
Feng Huang 0.83% −0.27% −0.12% 1.12% 0.43%
Li Du −1.49% 1.04% −1.65% 0.57% −0.13%
Hong Yuan 0.17% 0.12% −0.31% 0.09% 0.12%
Xiang Zong −0.28% 0.15% 0.36% 0.35% 1.25%
Ji Cheng 0.36% 0.22% 0.18% −0.24% 0.35%
Hui Huang −0.21% 0.13% 0.27% 0.28% −0.12%
Hong Yun 0.22% 0.33% −0.08% 0.12% 0.24%
De Sheng −0.34% −0.11% 0.16% 0.27% 0.34%
Guo Jian 0.62% 1.04% 0.20% 0.03% 0.38%
model for bi-level multiobjective optimization problems with fuzzy coefficients. At the
same time, it is found from Table 7 that the average computational time by ISA is less
than the time by FS-ISA. It is also reasonable because the process of fuzzy simulation for
possibilistic constraint will spend much time to get the approximate value. However, not
all possibilistic constraints can be directly converted into crisp ones. Lemma A.1 is efficient
only for the special membership functions such as the triangular and trapezoidal fuzzy
numbers.
To illustrate that FS-ISA is suitable for this kind of fuzzy bi-level model, the results are
compared with a genetic algorithm GA. GA is one of the most popular algorithms. Many
scholars also made the comparison between SA and GA in solving bi-level optimization
problems 12, 42, 43. They regard that different data scales will result in huge differences
on the computational efficiency. To ensure the fairness, we also design the GA based on the
fuzzy simulation for the bi-level multiobjective optimization with fuzzy coefficients. We set
the chromosome number 20, the crossover rate 0.6, the mutation rate 0.8, and the iterative
number 500. The average computing time and memory are listed in Table 7. Experiments
show that the similar optimal results can be obtained by both FS-ISA and FS-GA, but the
computational efficiency is different when the number of stone resources and stone plants
20 Mathematical Problems in Engineering
changes. It is found that when the number of stone resources and stone plants is small, FS-
ISA is more efficient than GA in solving the bi-level multiobjective optimization and much
more computational effort is needed for FS-GA to achieve the same optimal solution as
FS-ISA. However, when the data scale is large, FS-GA can reach a more optimal solution
at the expense of more computation time. The result is in accordance with the findings by Xu
et al. 43. Of course, if the fuzzy bi-level multi-objective optimization model can be easily
converted into the crisp model, we can obtain a more accurate solution and spend less time
by ISA than that by FS-ISA.
5. Conclusions
In this paper, we have developed a bi-level multi-objective optimization model with
possibilistic constraints under the fuzzy environment. In the model, the government is
considered as the leader level for minimizing the emissions of the stone dust and the waste
water and maximizing the employment and economic growth, and then stone plants are
considered as the follower level for maximizing the profit and minimizing the emissions.
Then we propose an algorithm FS-ISA to solve the model. Finally, a practical case proves that
the proposed model and algorithm are efficient.
Although the model proposed in this paper should be helpful for solving some real-
world problems, it only dealt with by the possibilistic constraints. If DM has different
purposes such as maximizing the possibility that the predetermined goals are achieved, we
can apply dependent-chance constraint to deal with it. In further research to be undertaken,
a detailed analysis will be given.
Appendix
Lemma A.1. Assume that Ed ij , and ewij (i 1, 2, . . . , m; j 1, 2, . . . , n) are L-R fuzzy
i , ed
numbers with the following membership functions:
⎧
⎪
⎪ Edi − t
⎪
⎪ L , t < Edi , αEd
⎪
⎪ i > 0,
⎨ αEd
i
i t
μEd A.1
⎪
⎪
⎪
⎪ t − Edi
⎪
⎪ , t ≥ Edi , βiEd > 0,
⎩R
βiEd
⎧ ⎛ ⎞
⎪
⎪ ed − t
⎪
⎪ L⎝
ij
⎠, t < edij , αed > 0,
⎪
⎪ ij
⎪
⎨ α ed
ij
ij t
μed ⎛ ⎞ A.2
⎪
⎪
⎪
⎪ t − ed
⎪
⎪ ⎝ ij
⎠, t ≥ edij , βed > 0,
⎪
⎩R ij
βijed
Mathematical Problems in Engineering 21
⎧
⎪
⎪ ewij − t
⎪
⎪ L , t < ewij , αew
⎪
⎪ ij > 0,
⎨ αew
ij
μewij t A.3
⎪
⎪
⎪
⎪ t − ewij
⎪
⎪ t ≥ ewij , βijew > 0,
⎩R βijew
,
where αEd Ed ed ed
i , βi are positive numbers expressing the left and right spreads of Ed, αij , βij are positive
and αew , βew are positive numbers expressing the
numbers expressing the left and right spreads of ed, ij ij
left and right spreads of ew, i 1, 2, . . . , m, j 1, 2, . . . , n. Reference functions L, R : 0, 1 → 0, 1
with L1 R1 0 and L0 R0 1 are nonincreasing, continuous functions. Then one has
Pos{ m i1 Edi Yi j1 edij Xij ew ij Xij ≤ F 1 } ≥ δ1 if and only if
m n U
i1
⎛ ⎞
m m n m m n
F1 ≥ Edi Yi edij ewij Xij − L −1
δ1U ⎝ αEd αed αew Xij ⎠.
i Yi ij ij
i1 i1 j1 i1 i1 j1
A.4
Proof. Let ω ∈ 0, 1 be any positive real number and LEdi − x/αEd
i Ledij − y/αij
ed
−1 −1 −1
x Edi − αEd
i L ω, y edij − αed
ij L ω, z ewij − αew
ij L ω.
A.5
m m n
t xYi yXij zXij
i1 i1 j1
⎡ ⎤ ⎛ ⎞
m m n m m n
⎣ Edi Yi edij ewij Xij ⎦ − ⎝ αEd
i Yi αed
ij αew
ij Xij ⎠L−1 ω.
i1 i1 j1 i1 i1 j1
A.6
22 Mathematical Problems in Engineering
Therefore, we have
⎛ ⎞
m m n
⎜ i1 Edi Yi i1 j1 edij ewij Xij − t ⎟
L⎝ ⎠ ω. A.7
m m n
i1 αEd
i Yi i1 j1 αed
ij αew
ij Xij
⎛ ⎞
m m n
t− Edi Yi edij ewij Xij
⎜ i1 i1 j1 ⎟
R⎝ ⎠ ω. A.8
m m n
i1 βiEd Yi i1 j1 βijed ew
βij Xij
m
i1 Edi Yi j1 edij Xij ew ij Xij is also a L-R fuzzy
m n
Hence, it is easily found that i1
m m n
j1 αij ijce Xij and the right spread
ed
number with the left spread Ed
i1 αi Yi i1 αew
m m n
j1 βij βij Xij . According to the definition of possibility measure
ed ew
βiEd Yi
i1 i1
proposed by Dubois and Prade 28, it can be obtained as follows:
⎧ ⎫
⎨ m m n ⎬
Pos i Yi
Ed ij Xij ewij Xij ≤ F 1
ed ≥ δ1U
⎩ i1 i1 j1
⎭
⎛ ⎞
m m n
⎜ i1 Edi Yi i1 j1 edij ewij Xij − F 1 ⎟
⇐⇒ L⎝ ⎠ ≥ δ1U
m m n ed ew
i1 αEd
i Yi i1 j1 αij αij Xij
m
Edi Yi m n
edij ewij Xij − F 1
⇐⇒
i1 i1 j1
≤ L−1 δ1U A.9
m m n ed
i1 αEd
i Yi i1 j1 αij αew
ij Xij
m m n
⇐⇒ Edi Yi edij ewij Xij
i1 i1 j1
⎛ ⎞
m m n
− L−1 δ1U ⎝ αEd
i Yi αed
ij αew
ij Xij ⎠ ≤ F 1 .
i1 i1 j1
From Lemma A.1, the model 2.21 is equivalent to the following bi-level multi-
objective programming problem:
⎧ ⎛ ⎞
⎪
⎪ m m n m m n
⎪
⎪ min F1∗ edij ewij Xij −L−1 δ1U ⎝ αEd αed αew Xij ⎠
⎪
⎪ Edi Yi i Yi
⎪
⎪
ij ij
⎪
⎪ i1
⎛ i1 j1
⎞ i1 i1 j1
⎪
⎪
⎪
⎪ m n
⎪
⎪ ⎝ pij Xij Pi ⎠
⎪
⎪ max F2
⎪
⎪
⎪
⎪ i1
⎛j1
⎞
⎪
⎪
⎪
⎪ m n
⎪
⎪
⎪
⎪ max F3 Si ⎝ cj θij Xij ⎠
⎪
⎪
⎪
⎪ ⎧ i1 j1
⎛ ⎞
⎪
⎪
⎪
⎪ ⎪
⎪ m m n m m n
⎪
⎪ ⎪
⎪
⎪
⎪
⎪
⎪ Edi Yi edij Xij − L δ2 ⎝ αi Yi
−1 U Ed
αij Xij ⎠ ≤ EDU
ed
⎪
⎪ ⎪
⎪ i1
⎪
⎪ ⎪
⎪ i1 j1 i1 i1 j1
⎪
⎪ ⎪
⎪ m n
⎪
⎪ ⎪
⎪
m n
⎪
⎪ ⎪
⎪ ewij Xij − L−1 δ3U αewij Xij ≤ EW
U
⎪
⎪ ⎪
⎪ i1 j1
⎪
⎪ ⎪
⎪ i1 j1
⎪
⎪ ⎪
⎪m
⎨ ⎪
⎪
⎪
⎪ θij Xij ≥ Dj ∀j L
⎪
⎪
⎪
⎪ ⎪
⎪ ⎧
i1 ⎛ ⎞
⎪
⎪ ⎪
⎪⎪
⎪
⎪ ⎪
⎪ ⎪ n n n
⎪
⎪ ⎪
⎪ ⎪
⎪ cj θij Xij − f Xij − hi ⎝Yi − Xij ⎠
⎪
⎪ ⎪
⎪ ⎪
⎪ max Hi1
⎪
⎪ ⎪
⎪ ⎪
⎪
⎪
⎪ ⎨⎪⎪ j1 j1 j1
⎪
⎪ ⎪
⎪ n n
⎪
⎪ s.t. ⎪
⎪ −1
⎪
⎪ ⎪
⎪ ⎪
⎪ min H 2∗
ed ew X − L δ U
α ed
α ew
Xij
⎪
⎪ ⎪
⎪ ⎪
⎪ i ij ij ij 3 ij ij
⎪
⎪ ⎪
⎪ ⎪
⎪ ⎧ j1 j1
⎪
⎪ ⎪⎪
⎪ ⎪
⎪
⎪ ⎪
⎪ ⎪
⎪ ⎪
⎪
n
⎪
⎪ ⎪
⎪ ⎨ ⎪
⎪ Xij ≤ Yi
⎪
⎪ ⎪
⎪ ⎪
⎪
⎪
⎪ ⎪
⎪ ⎪
⎪ j1
⎪
⎪ ⎪⎪
⎪ ⎪ ⎪
⎪
⎪
⎪ ⎪ ⎪ ⎪ n
⎪
⎪ ⎪⎪
⎪
⎪ ⎪
⎪
⎪
⎪Yi − Xij ≤ IViU
⎨
⎪
⎪ ⎪
⎪ ⎪
⎪
⎪
⎪ ⎪
⎪ ⎪
⎪
⎪
⎪ ⎪
⎪ ⎪ s.t. ⎪
⎪
j1
⎛ ⎞
⎪
⎪ ⎪
⎪ ⎪
⎪ ⎪
⎪
⎪
⎪ ⎪⎪
⎪ ⎪ ⎪
⎪
n n
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪ f Xij hi ⎝Yi − Xij ⎠ ≤ P CiU
⎪
⎪ ⎪
⎪ ⎪
⎪ ⎪
⎪
⎪
⎪ ⎪
⎪ ⎪
⎪ ⎪
⎪
j1 j1
⎪
⎩ ⎩⎪
⎪ ⎩ ⎪
⎩θij Xij ≥ P L .
ij
A.10
Abbreviations
Indices
Parameters
ewij : Waste water emissions coefficient when that plant i produces product j
pij : Employment coefficient that plant i produces product j
Pi : Basic employment that plant i needs
Si : Unit tax rate that plant i pays to the government
cj : Unit price of product j
Decision variables
Acknowledgments
The work is supported by the Key Program of National Natural Science Foundation of
China Grant no. 70831005 and also supported by the “985” Program of Sichuan University
Innovative Research Base for Economic Development and Management.
References
1 M. Al-Jabari and H. Sawalha, “Treating stone cutting waste by flocculation-sedimentation,” in
Proceedings of the Sustainable Environmental Sanitation and Water Services Conference, 28th WEDC
Conference, Calcutta, India, 2002.
2 K. Nasserdine, Z. Mimi, B. Bevan, and B. Elian, “Environmental management of the stone cutting
industry,” Journal of Environmental Management, vol. 90, no. 1, pp. 466–470, 2009.
3 N. Almeida, F. Branco, and J. R. Santos, “Recycling of stone slurry in industrial activities: application
to concrete mixtures,” Building and Environment, vol. 42, no. 2, pp. 810–819, 2007.
Mathematical Problems in Engineering 25
4 V. Corinaldesi, G. Moriconi, and T. R. Naik, “Characterization of marble powder for its use in mortar
and concrete,” Construction and Building Materials, vol. 24, no. 1, pp. 113–117, 2010.
5 F. Saboya Jr., G. C. Xavier, and J. Alexandre, “The use of the powder marble by-product to enhance
the properties of brick ceramic,” Construction and Building Materials, vol. 21, no. 10, pp. 1950–1960,
2007.
6 N. Bilgin, H. A. Yepremb, S. Arslan et al., “Use of waste marble powder in brick industry,”
Construction and Building Materials, vol. 29, pp. 449–457, 2012.
7 A. Singh and H. H. Lou, “Hierarchical pareto optimization for the sustainable development of
industrial ecosystems,” Industrial and Engineering Chemistry Research, vol. 45, no. 9, pp. 3265–3279,
2006.
8 S. R. Lim and M. P. Jong, “Cooperative water network system to reduce carbon footprint,”
Environmental Science and Technology, vol. 42, no. 16, pp. 6230–6236, 2008.
9 D. Petrovic, R. Roy, and R. Petrovic, “Supply chain modelling using fuzzy sets,” International Journal
of Production Economics, vol. 59, no. 1, pp. 443–453, 1999.
10 H. M. Lee and J. S. Yao, “Economic production quantity for fuzzy demand quantity and fuzzy
production quantity,” European Journal of Operational Research, vol. 109, no. 1, pp. 203–211, 1998.
11 R. E. Bellman and L. A. Zadeh, “Decision-making in a fuzzy environment,” Management Science, vol.
17, pp. B141–B164, 1970.
12 S. R. Hejazi, A. Memariani, G. Jahanshahloo, and M. M. Sepehri, “Linear bilevel programming
solution by genetic algorithm,” Computers & Operations Research, vol. 29, no. 13, pp. 1913–1925, 2002.
13 M. Al-Salamah, “Temporary workforce planning with firm contracts: a model and a simulated
annealing heuristic,” Mathematical Problems in Engineering, vol. 2011, Article ID 209693, 18 pages, 2011.
14 M. Gendreau, P. Marcotte, and G. Savard, “A hybrid tabu-ascent algorithm for the linear bilevel
programming problem,” Journal of Global Optimization, vol. 8, no. 3, pp. 217–233, 1996.
15 J. Li, J. Xu, and M. Gen, “A class of multiobjective linear programming model with fuzzy random
coefficients,” Mathematical and Computer Modelling, vol. 44, no. 11-12, pp. 1097–1113, 2006.
16 A. Suppapitnarm and G. T. Parks, Simulated Annealing: An Alternative Approach to True Multiobjective
Optimization, Morgan Kaufmann, 1999.
17 J. S. Yao and F. T. Lin, “Constructing a fuzzy flow-shop sequencing model based on statistical data,”
International Journal of Approximate Reasoning, vol. 29, no. 3, pp. 215–234, 2002.
18 J. S. Yao and K. Wu, “Consumer surplus and producer surplus for fuzzy demand and fuzzy supply,”
Fuzzy Sets and Systems, vol. 103, no. 3, pp. 421–426, 1999.
19 C. J. Lin and U. P. Wen, “A labeling algorithm for the fuzzy assignment problem,” Fuzzy Sets and
Systems, vol. 142, no. 3, pp. 373–391, 2004.
20 J. Xu, L. Yao, and X. Zhao, “A multi-objective chance-constrained network optimal model with
random fuzzy coefficients and its application to logistics distribution center location problem,” Fuzzy
Optimization and Decision Making, vol. 10, no. 3, pp. 255–285, 2011.
21 J. Xu and Z. Zeng, “A discrete time optimal control model with uncertainty for dynamic machine
allocation problem and its application to manufacturing and construction industries,” Applied
Mathematical Modelling,, vol. 36, no. 8, pp. 3513–3544, 2012.
22 J. C. Bezdek, “Fuzzy models—what are they, and why?” IEEE Transactions on Fuzzy Systems, vol. 1,
no. 1, pp. 1–6, 1993.
23 D. Dubois and H. Prade, Fuzzy Sets and Systems, Academic Press, New York, NY, USA, 1980.
24 U. Thole, H. J. Zimmermann, and P. Zysno, “On the suitability of minimum and product operators
for the intersection of fuzzy sets,” Fuzzy Sets and Systems, vol. 2, no. 2, pp. 167–180, 1979.
25 B. Kosko, Neural Networks and Fuzzy Systems, Prentice Hall, Englewood Cliffs, NJ, USA, 1992.
26 R. Kohavi, “A study of cross-validation and bootstrap for accuracy estimation and model selection,”
International Joint Conference on Artificial Intelligence, vol. 14, pp. 1137–1145, 1995.
27 J. Xu and X. Zhou, Fuzzy-Like Multiple Objective Decision Making, Springer, 2011.
28 D. Dubois and H. Prade, “Operations on fuzzy numbers,” International Journal of Systems Science, vol.
9, no. 6, pp. 613–626, 1978.
29 A. Kaufmann and M. M. Gupta, Introduction to Fuzzy Arithmetic, Van Nostrand Reinhold, New York,
NY, USA, 1985.
30 J. F. Bard, “Some properties of the bilevel programming problem,” Journal of Optimization Theory and
Applications, vol. 68, no. 2, pp. 371–378, 1991.
31 P. Hansen, B. Jaumard, and G. Savard, “New branch-and-bound rules for linear bilevel program-
ming,” Society for Industrial and Applied Mathematics, vol. 13, no. 5, pp. 1194–1217, 1992.
26 Mathematical Problems in Engineering
32 O. Ben-Ayed and C. E. Blair, “Computational difficulties of bilevel linear programming,” Operations
Research, vol. 38, no. 3, pp. 556–560, 1990.
33 S. Kirkpatrick Jr., C. D. Gelatt,, and M. P. Vecchi, “Optimization by simulated annealing,” American
Association for the Advancement of Science, vol. 220, no. 4598, pp. 671–680, 1983.
34 S. Kirkpatrick, “Optimization by simulated annealing: quantitative studies,” Journal of Statistical
Physics, vol. 34, no. 5-6, pp. 975–986, 1984.
35 N. Metropolis, A. W. Rosenbluth, M. N. Rosenbluth, A. H. Teller, and E. Teller, “Equation of state
calculations by fast computing machines,” The Journal of Chemical Physics, vol. 21, no. 6, pp. 1087–
1092, 1953.
36 P. J. M. van Laarhoven, Theoretical and Computational Aspects of Simulated Annealing, vol. 51, Stichting
Mathematisch Centrum Centrum voor Wiskunde en Informatica, Amsterdam, The Netherlands, 1988.
37 J. M. Steel, “A review of Laarhoven,” Journal of the American Statistical Association, vol. 85, pp. 596–596,
1990.
38 A. Suppapitnarm, K. A. Seffen, G. T. Parks, and P. J. Clarkson, “Simulated annealing algorithm for
multiobjective optimization,” Engineering Optimization, vol. 33, no. 1, pp. 59–85, 2000.
39 B. Suman, “Study of simulated annealing based algorithms for multiobjective optimization of a
constrained problem,” Computers and Chemical Engineering, vol. 28, no. 9, pp. 1849–1871, 2004.
40 B. Suman and P. Kumar, “A survey of simulated annealing as a tool for single and multiobjective
optimization,” Journal of the Operational Research Society, vol. 57, no. 10, pp. 1143–1160, 2006.
41 B. Sanghamitra, S. Sriparna, M. Ujjwal, and D. Kalyanmoy, “A simulated annealing-based
multiobjective optimization algorithm: AMOSA,” IEEE Transactions on Evolutionary Computation, vol.
12, no. 3, pp. 269–283, 2008.
42 A. Sadegheih, “Scheduling problem using genetic algorithm, simulated annealing and the effects of
parameter values on GA performance,” Applied Mathematical Modelling, vol. 30, no. 2, pp. 147–154,
2006.
43 T. Xu, H. Wei, and G. Hu, “Study on continuous network design problem using simulated annealing
and genetic algorithm,” Expert Systems with Applications, vol. 36, no. 2, pp. 1322–1328, 2009.
Hindawi Publishing Corporation
Mathematical Problems in Engineering
Volume 2012, Article ID 582323, 21 pages
doi:10.1155/2012/582323
Research Article
Dynamic Programming and Heuristic for
Stochastic Uncapacitated Lot-Sizing Problems
with Incremental Quantity Discount
Copyright q 2012 Yuli Zhang et al. This is an open access article distributed under the Creative
Commons Attribution License, which permits unrestricted use, distribution, and reproduction in
any medium, provided the original work is properly cited.
The stochastic uncapacitated lot-sizing problems with incremental quantity discount have been
studied in this paper. First, a multistage stochastic mixed integer model is established by the
scenario analysis approach and an equivalent reformulation is obtained through proper relaxation
under the decreasing unit order price assumption. The proposed reformulation allows us to extend
the production-path property to this framework, and furthermore we provide a more accurate
characterization of the optimal solution. Then, a backward dynamic programming algorithm is
developed to obtain the optimal solution and considering its exponential computation complexity
in term of time stages, we design a new rolling horizon heuristic based on the proposed property.
Comparisons with the commercial solver CPLEX and other heuristics indicate better performance
of our proposed algorithms in both quality of solution and run time.
1. Introduction
The lot-sizing problems have been the subject of intensive research in the last few decades.
The basic definition of single-item lot-sizing problems can be stated as follows: the order
production, inventory and backlog quantities in each period should be determined to
meet the deterministic but dynamic demand over a finite time horizon. The objective is
to minimize the total costs, which consist of the fixed setup cost, order cost and inventory
cost. Different quantity discount policies, such as all-units quantity discount and incremental
quantity discount, have been widely executed in practice and thus have also been introduced
into the lot-sizing problems.
Although the deterministic planning and scheduling models have been intensively
studied, in practice there are many different sources of uncertainties, such as customer
2 Mathematical Problems in Engineering
demand, production lead-time, and product price, which make information that will be
needed in subsequent decision stages unavailable to the decision maker. In such cases, the
solution provided by a deterministic model may be of little value in terms of applicability
of the model’s recommendations, see Beraldi et al. 1. Thus the stochastic version of lot-
sizing problems has been studied recently and with the advent of stochastic programming,
the classical deterministic lot-sizing models have been extended to scenario-based multistage
stochastic mixed integer programming.
Wagner and Whitin 2 first introduced the single-item dynamic lot-sizing problems
without backlogging. They proposed a dynamic programming approach based on the
Wagner-Whitin W-W property, that is, no production is undertaken if inventory is
available. Because W-W property holds for the nondiscount problem, many heuristics
have been developed based on this property; however, in the all-units quantity discount
case, the property does not necessarily hold due to its discontinuous piecewise linear cost
structure by Hu and Munson 3. For the modified all-units discount problem, Chan et
al. 4 demonstrated that a zero-inventory-ordering policy based on the W-W property
exists, whose cost is no more than 4/3 times the optimal cost. Federgruen and Lee 5
characterized structural properties of optimal solutions for both all-units and incremental
quantity discount, and proposed dynamic programming algorithms with complexity OT 3
and OT 2 , respectively, and with T being the number of periods in the planning horizon.
For deterministic capacitated lot-sizing problem with general piecewise linear cost function,
Shaw and Wagelmans 6 presented a dynamic programming procedure with complexity
OT 2 qd, where T is the number of periods, d is the average demand and q is the
average number of pieces required to represent the production cost function. Note that this
pseudopolynomial time algorithm is based on the assumption that the demand is an integral
value.
For the stochastic version problem, although Ahmed et al. 7 showed that the W-
W property does not hold for the stochastic lot-sizing problems several modified W-W
properties have been presented for different versions of the stochastic lot-sizing problems:
Guan and Miller 8 studied the stochastic uncapacitated lot-sizing problems with uncertain
parameters and introduced the production path property for the optimal solution. Further,
the production path property was extended to the stochastic capacitated lot-sizing problems
with backlogging in Guan and Miller 8 and the dynamic programming algorithms based
on this property have been presented. For the stochastic lot-sizing problems with random
lead times, Huang and Küçükyavuz 9 presented the Semi-Wagner-Whitin property under
assumption that no order crosses in time.
Besides the algorithms based on the extended W-W properties, Lulli and Sen 10
and Ahmed et al. 7 have presented branch-and-price and branch-and-bound algorithms,
respectively to solve the proposed multistage integer programming. Although such branch-
and-bound-based B&B-based methods have broad application prospects for general integer
programming, special structure properties of the stochastic lot-sizing problems have not been
explored in order to design customized algorithms, and only computational results for small-
size stochastic batch-sizing problems have been reported in Lulli and Sen 10. Other heuristic
methods, such as the fix and relax, have also been redesigned to solve particular stochastic
lot-sizing problems, see 1, 11. We also refer the reader to the recent literature review of
12–14.
To the best of our knowledge, little research has been reported on the stochastic lot-
sizing problems with incremental quantity discount SLP-IQD. However as it is reported in
the survey by Munson and Rosenblatt 15, 83% of the buyers received quantity discounts
Mathematical Problems in Engineering 3
for most of the items they purchased and 37% of orders involved either the offer or receipt
of some incremental quantity discounts; thus, the study on quantity discount is of great
importance in practice. In this paper, a multistage stochastic mixed-integer programming
model is established and under decreasing unit order price assumption, an equivalent
relaxed formulation is obtained. The reformulation provides the possibility of extending the
production path property for optimal solution of SLP-IQD. The extended production path
property is not only a direct extension to the case with incremental quantity discount, but
also provides a more accurate characterization for the optimal solution. Then, a backward
dynamic programming algorithm has been developed. Although it can obtain optimal
solutions in polynomial time in terms of the number of nodes, it has an exponential
computational complexity in terms of the number of stages. Thus, a new rolling horizon
heuristic which makes use of the extended production path property and has flexible
parameters settings is presented in order to balance the desired solution quality and run time.
Numerical experiments have been implemented to explore the proper parameters settings
and validate the effectiveness of the proposed algorithms by comparison with the CPLEX
11.1 solver and other heuristics.
The remainder of the paper is organized as follows. In Section 2, we first introduce
the deterministic formulation and then formulate the general multistage stochastic mixed
integer model for the stochastic uncapacitated lot-sizing problems with incremental quantity
discount SULP-IQD. An equivalent reformulation is proposed under the decreasing unit
order price assumption. In Section 3, the extended production path property is proven and
a backward dynamic programming algorithm and a rolling horizon heuristic with flexible
parameters settings are developed. Computational results are reported and discussed in
Section 4. Section 5 presents conclusions.
2. Mathematical Model
2.1. Deterministic Lot-Sizing Problems with Incremental Quantity Discount
First, we will establish an mathematical model for the deterministic uncapacitated lot-sizing
problems with incremental quantity discount DULS-IQD. Considering a planning horizon
of T time periods stages, at each period t, the nonnegative demand dt , variable inventory
holding cost ht , the variable setup cost ct , and piecewise order cost function ft · at period
t ∈ {1, . . . , T } are given. Variable st denotes the inventory quantity at the end of period t, and
variable xt and Boolean variable yt denote the order quantity and fixed charge variable at
period t, respectively.
The incremental quantity discount cost structure is given as follows:
⎧
⎪
⎪ pt,1 for the first Qt,1 units,
⎪
⎪
⎪
⎨pt,2 for the next Qt,2 − Qt,1 units,
the unit order cost . .. 2.1
⎪
⎪..
⎪
⎪ .
⎪
⎩p
t,Kt for each unit in excess of Qt,Kt −1 units,
where Kt denotes the number of price intervals. Suppose that Qt,0 0 and Qt,Kt ∞.
The decreasing unit order price assumption, that is, pt,k > pt,k1 , k 1, . . . , Kt − 1 with
4 Mathematical Problems in Engineering
0 Qt,0 < Qt,1 < · · · < Qt,k < Qt,k1 < · · · < Qt,Kt ∞ always naturally holds in practice. The
piecewise order cost function is given as
pt,1 xt , if Qt,0 ≤ xt < Qt,1 ,
ft xt k−1 2.2
j1 pt,j − pt,j1 Qt,j pt,k xt , if Qt,k−1 ≤ xt < Qt,k , k 2, . . . , Kt .
T
min : ft xt ht st ct yt 2.3
t1
subject to
st−1 xt − dt st , t 1, . . . , T, 2.4
xt ≤ Myt , t 1, . . . , T, 2.5
xt ≥ 0, st ≥ 0, s0 0 t 1, . . . , T, 2.6
The objective function 2.3 minimizes the sum of inventory cost, setup cost and
piecewise order cost. Constraints 2.4 guarantee that the dynamic demands in every period
are met. Constraints 2.5 are the capacity constraints, and here we assume that M is a
sufficiently large upper bound on xt . Constraints 2.6 ensure that there is no backlogging
and order quantity variables are nonnegative. Without loss of generality, we suppose that
there is no initial inventory. Setup variable yt is defined as a binary variable in constraints
2.7.
Here using the above notation, the deterministic model in the above subsection can
be extended to the stochastic environments by replacing the stage index t ∈ {1, . . . , T } with
node index n ∈ N. The multistage stochastic problem formulation, denoted by SP, is given as
follows:
SP min : θn fn xn hn sn cn yn 2.8
n∈N
subject to
san xn − dn sn , n ∈ N,
xn ≤ Myn , n ∈ N,
2.9
xn ≥ 0, sn ≥ 0, sa0 0, n ∈ N,
yn ∈ {0, 1}, n ∈ N.
xn,k ≥ 0, n ∈ N, k 1, . . . , Kn ,
0, xn,k 0, 2.10
yn,k
1, Qn,k−1 < xn,k ≤ Qn,k .
Δ Δ
For model brevity, we introduce the following notations: pn,k θn pn,k , hn θn hn , and
Δ
cn,k θn {cn k−1
j1 {pn,j −pn,j1 Qn,j }}. Note that the setup cost also increases correspondingly
in index k, that is, cn,k < cn,k1 for each node n ∈ N. An equivalent MIP formulation, ESP1, is
given as follows:
K
n
ESP1 min : pn,k xn,k cn,k yn,k hn sn 2.11
n∈N k1
6 Mathematical Problems in Engineering
subject to
Kn
san xn,k − dn sn , n ∈ N, 2.12
k1
0 by constraints 2.13, 2.14, and the optimality. Note that if k 1 xn,k xn,kn 0, then
Kn ∗ ∗
y
k 1 n,k
yn,k n
0. Thus, by checking constraints in SP, we construct a corresponding
feasible solution of SP as {x∗n xn,k ∗
n
, y ∗n yn,k
∗
n
, s∗n s∗n | n ∈ N}. By the definition of pn,k , cn,k
∗ ∗
and hn , the constructed feasible solution of SP has the same objective cost to {xn,k , yn,k , s∗n |
n ∈ N, k 1, . . . , Kn }.
For each optimal solution of SP, {xn∗ , yn∗ , s∗n | n ∈ N}, we define the corresponding
solution for ESP1 as follows: for given n ∈ N, if Qn,k−1 < xn∗ ≤ Qn,k , then {x∗n,k xn∗ , y ∗n,k
1, s∗n s∗n , x∗n,k y∗n,k 0, k ∗ ∗ ∗
/ k} and if xn∗ 0, then {xn,k 0, y n,k 0, sn s∗n , k ∈
{1, . . . , Kn }}. By checking its feasibility and the objective cost in ESP1, we conclude that the
constructed solution is also feasible in ESP1 and has same objective cost to {xn∗ , yn∗ , s∗n | n ∈
N}.
Further, by relaxing the constraints on Boolean variables and order quantity variables
in 2.13-2.14, the following formulation, denoted by ESP2, and proposition can be obtained.
K
n
ESP2 min : pn,k xn,k cn,k yn,k hn sn 2.17
n∈N k1
subject to
Kn
san xn,k − dn sn , n ∈ N,
k1
Proposition 2.2. Under the decreasing unit order price assumption, formulation ESP2 is equivalent
to ESP1 and SP.
Proof. Because the feasible set of the relaxed formulation ESP2 contains the feasible set of
∗ ∗
ESP1, the optimal solutions of ESP2 must be proven feasible for ESP1. Let {xn,k , yn,k , s∗n | n ∈
∗
N, k 1, . . . , Kn } be an optimal solution of ESP2. First note that if yn,k 1, we must have
∗
xn,k > 0 otherwise it contradicts the optimality since cn,k > 0.
n
Second, it is asserted that the relaxed constraints K y ≤ 1, n ∈ N always hold for
k1 n,k
∗ ∗
every optimal solution of ESP2. Suppose that yn,l yn,m 1, n ∈ N, 1 ≤ l /
m ≤ Kn ; thus,
∗ ∗ ∗ ∗
xn,l > 0 and xn,m > 0. Without loss of generality, suppose that m > l and Qn,r−1 ≤ xn,l xn,m <
Qn,r , 1 ≤ r ≤ Kn , then
∗ ∗
cn,r pn,r xn,l xn,m
⎧ ⎫
⎨ r−1
⎬
∗ ∗
θn cn pn,j − pn,j1 Qn,j pn,r xn,l xn,m
⎩ j1
⎭
⎧ ⎫
⎨ r−1 ⎬
∗ ∗ ∗ ∗
θn pn,j − pn,j1 Qn,j pn,r xn,l xn,m − pn,m xn,l xn,m
⎩jm ⎭
∗ ∗
cn,m pn,m xn,l xn,m
⎧ ⎫
⎨ r−1 ⎬
∗ ∗ ∗ ∗ ∗ ∗
≤ θn pn,j − pn,j1 xn,l xn,m pn,r xn,l xn,m − pn,m xn,l xn,m
⎩jm ⎭
∗ ∗
cn,m pn,m xn,l xn,m
∗ ∗
cn,m pn,m xn,l xn,m
∗ ∗
< cn,m pn,m xn,m pn,l xn,l
∗ ∗
< cn,m pn,m xn,m cn,l pn,l xn,l ,
2.19
r−1
where jm {pn,j − pn,j1 pn,m − pn,r in the third inequality.
∗
From the above analysis, a better solution can always be obtained by setting xn,r
Kn ∗ ∗
Kn ∗ ∗ ∗
x
j1 n,j ,
yn,r 1, where Qn,r−1 ≤
x
j1 n,j ≤ Q n,r , and
xn,k
0, yn,k
0, k
/ r, which
contradicts the optimality.
Third, we prove that the relaxed constraints Qn,k−1 yn,k < xn,k ≤ Qn,k yn,k , n ∈ N, k
∗ ∗
1, . . . , Kn also hold for the optimal solution. Assume that xn,l > 0 and xn,m 0 for all m / l, but
8 Mathematical Problems in Engineering
∗
the inequality constraint Qn,l−1 ≤ xn,l ≤ Qn,l does not hold. Without loss of generality, suppose
∗
that Qn,r−1 < xn,l ≤ Qn,r ≤ Qn,l−1 where r ≤ l − 1, then
⎧ ⎫
⎨ l−1
⎬
∗ ∗
cn,l pn,l xn,l θn cn pn,j − pn,j1 Qn,j pn,l xn,l
⎩ j1
⎭
⎧ ⎫
⎨ l−1 ⎬ 2.20
∗ ∗
cn,r pn,r xn,l θn pn,j − pn,j1 Qn,j pn,l − pn,r xn,l
⎩ jr ⎭
∗
> cn,r pn,r xn,l .
∗
Therefore, we reach a contradiction because we obtain a better solution by setting xn,r
∗ ∗ ∗ ∗
xn,l , yn,r 1, and xn,k 0, yn,k 0 for k /
r. Since the optimal solution of ESP2 satisfies all the
constraints of ESP1, and ESP2 is obtained from ESP1 by relaxing its constraints, this implies
that ESP2 is equivalent to ESP1 and SP.
Proposition 3.1 extended production-path property. For any instance of SULS-IDQ, there
∗
exists an optimal solution x∗ , y∗ , s∗ , such that for each node n ∈ N, if xn,k > 0, then there exist
∗
m ∈ Dn and Qn,k−1 < dn,m − san ≤ Qn,k such that
Mathematical Problems in Engineering 9
∗ ∗
1 xn,i 0, yn,i 0, for all i 1, . . . , Kn , i /
k;
∗
2 xn,k dn,m − s∗an , yn,k
∗
1;
∗ ∗
3 xl,i 0, yl,i 0, for all l ∈ P n, am \ {n}, i 1, . . . , Kl .
Note that under assumption that all lead time is equal to 1 and by similar arguments,
the second optimal condition can be regarded as an extension of the Semi-Wagner-Whitin
Property in Huang and Küçükyavuz 9; however, the third optimal condition provides new
restrictions which narrow the scope of the expected optimal solutions. In the next section, an
improved backward dynamic programming algorithm and a rolling horizon heuristic based
on the extended production-path property will be presented.
For any optimal solution of ESP2, x, y, s, if xn,k > 0, we introduce the following
definition:
Kl Km
NP n m | m ∈ Dn, xl,k 0, ∀l ∈ P n, am \ {n}, xm,k > 0 ,
k1 k1
3.1
Kl
NNP n m | m ∈ Dn ∩ L, xl,k 0, ∀l ∈ P n, m \ {n} ,
k1
where set NP n contains the first node with positive order quantity after node n, and set
NNP n contains these leaf nodes from which to node n no positive order has been placed
except node n. Note that if a certain positive order quantity is properly transferred between
node n and all the nodes in set NP n, the feasibility can be kept and only the cost for nodes
in the following set is changed:
Proof. First, it is asserted that there exists at least an optimal solution by Weierstrass’ theorem
n
since the feasible set is compact note that K x can be bounded by maxm∈Dn∪L {dn,m }
k1 n,k
and the objective function is continuous. Then for any given optimal solution of ESP2,
∗
x∗ , y∗ , s∗ , if xn,k > 0, the first part holds from Proposition 2.2. Next, we show that an optimal
solution holding the above property can always be constructed from any given optimal
solution of ESP2 by adjusting the variable’s value.
We scan the optimal solution x∗ , y∗ , s∗ from stage 1 to stage T . Assume that there
∗
exists a node n with xn,k > 0 not holding the property, but for nodes at stage t < tn the
conclusion holds, the following approach adjusts variables’ value assigned to nodes in Dn
to satisfy the property without changing variables’ value before stage tn. For any feasible
solution x, y, s of ESP2, we define the object cost function as
K
n
F x, y, s pn,k xn,k cn,k yn,k hn sn . 3.3
n∈N k1
10 Mathematical Problems in Engineering
F x∗ , y∗ , s∗
⎧ ⎫
⎨ Km ⎬
∗ ∗
pm,j xm,j cm,j ym,j hm s∗m ∗
pn,k xn,k ∗
cn,k yn,k
m∈N\φn
⎩ j1 ⎭ 3.4
hn s∗n hm s∗m ∗
pm,jm xm,j m
∗
cm,jm ym,j m
hm s∗m ,
m∈φn{{n}∪NP n} m∈NP n
where jm is the unique positive order quantity at node m, m ∈ NP n, and the node set N
is divided into four subsets: {n}, NP n, φn \ {{n} ∪ NP n}, and N \ φn. For any node
m ∈ φn \ {{n} ∪ NP n}, no order is placed.
∗
Since xm,j m
> 0 for m ∈ NP n by definition and s∗m > 0 for m ∈ NNP n otherwise it
contradicts the assumption, we choose small positive scalar ε satisfying
∗
0 < ε ≤ xn,k , 3.5
x∗n,k xn,k
∗
− ε, s∗m s∗m − ε for m ∈ φn \ NP n,
x∗m,jm xm,j
∗
m
ε for n ∈ NP n, 3.8
∗ ∗
xn,k xn,k ε, s∗m s∗m ε for m ∈ φn \ NP n,
∗ ∗
xm,j m
xm,j m
−ε for n ∈ NP n 3.9
and keep the value of other variables unchanged. For given ε satisfying constraints 3.5–
3.7, the objective costs of x∗ , y ∗ , s∗ and x∗ , y∗ , s∗ are
F x ∗ , y ∗ , s∗
⎧ ⎫
⎨ Km ⎬
∗ ∗
pm,j xm,j cm,j ym,j hm s∗m ∗
pn,k xn,k ∗
− ε cn,k yn,k hn s∗n − ε
m∈N\φn
⎩ j1 ⎭
hm s∗m − ε ∗
pm,jm xm,j m
ε c m,j y ∗
m m,jm
hm s∗
m
m∈φn\{{n}∪NP n} m∈NP n
F x∗ , y∗ , s∗ − pn,k ε − hn ε − hm ε pm,jm ε
m∈φn\{{n}∪NP n} m∈NP n
F x∗ , y∗ , s∗ − pn,k ε − hm ε pm,jm ε,
m∈φn\NP n m∈NP n
Mathematical Problems in Engineering 11
F x∗ , y∗ , s∗ pn,k ε hn ε hm ε − pm,jm ε.
m∈φn\{{n}∪NP n} m∈NP n
F x∗ , y∗ , s∗ pn,k ε hm ε − pm,jm ε.
m∈φn\NP n m∈NP n
3.10
Note that the first equality comes from the definition of x∗ , y ∗ , s∗ , x∗ , y∗ , s∗ , and F function.
The second one is obtained by comparing with the value of Fx∗ , y∗ , s∗ and the last one
is obtained by rearranging the terms in the former, where hn ε m∈φn\{{n}∪NP n} hm ε
m∈φn\NP n hm ε.
Since x∗ , y∗ , s∗ is an optimal solution of ESP2, thus we must have pn,k
∗ ∗ ∗
m∈φn\NP n hm − ∗ , y∗ , s∗ are also optimal solution
m∈NP n pm,jm 0 and x , y , s , x
for ESP2. Now we increase ε so that at least one of the following cases occur.
Case 1. If equality in 3.5 holds, then we have eliminated the undesired positive order
at node n and x∗ , y ∗ , s∗ will be used to construct an eventual solution by a similar method.
Case 2. If equality in 3.6 for a certain m holds, then there exists m ∈ Dn such that
x∗n,k dn,m − s∗an and x∗l,k 0, y ∗l,k 0 for all l ∈ P n, am \ {n}, k 1, . . . , Kl . Next,
x∗ , y ∗ , s∗ will be used to construct an eventual solution by a similar method.
Case 3. If equality in 3.7 for a certain m holds, then we apply the above analysis to
x∗ , y∗ , s∗ . Since there are finite nodes in Dm, eventually case 1 or case 2 will occur in the
worst case NP n ∅ after finite steps.
Thus, the optimal solution holding the proposed property can always be constructed
after finite steps.
Hn, s: value function at node n when its initial inventory is san s, that is,
m
Hn, s min{ m∈Dn { K k1
{pm,k xm,k cm,k ym,k } hm sm }} subject to constraints
2.18 and san s:
HP n, s: production value function at n when its initial inventory is san s and
Kn
x > 0:
k1 n,k
HNP n, s: nonproduction value function at n when its initial inventory is san s
n
and K x 0.
k1 n,k
12 Mathematical Problems in Engineering
From Proposition 3.1, if a positive order quantity is made at node n when the initial
inventory is s, then there exists a node j ∈ Dn, such that xn dn,j − s. Therefore, the
following equations hold:
HP n, s
⎧ ⎫
⎨ ⎬
min cn,k pn,k dn,j − s hn dn,j − dn H m, dn,j − dn ,
j∈Dn: Qn,k−1 <dn,j −s≤Qn,k ⎩ ⎭
m∈Cn
3.11
To obtain the exact optimal solution of SP, it is not necessary to completely characterize
the value function H0, s. We only need to calculate its values at possible positive
discontinuous points. That is, with zero initial inventory assumption for node n ∈ N \ {0},
we only need to evaluate Hn, s for s d0,m − d0,an ≥ 0, m ∈ N since dn,m − dn d0,m − d0,n
in 3.11. Thus, we give the following dynamic programming for SULS-IQD:
Without loss of generality, we assume | Cn | C for all n ∈ N in the following
analysis. Dynamic programming demonstrates a straight method to obtain the optimal
solution. Although Guan and Miller 8 showed that general SULS without incremental
quantity discount can be solved in O| V 0|2 max{log V 0, C}, the above algorithm runs in
exponential time in terms of T since V 0 CT − 1/C − 1. The exponential computational
time encourages us to develop a more effective algorithm for the problems with large
numbers of stages see Algorithm 1.
proposed RHH based on the extended production-path property stems from the proposed
dynamic programming algorithm. In the heuristic, nonproduction strategy is developed
to take advantage of the accurate characterization of the optimal solution, and then look-
forward and look-backward heuristic strategies are developed to reduce the computation
complexity of the complete enumeration in evaluation of the value function. The computation
complexity is also analyzed to demonstrate its advantage for the problems with a large
planning horizon.
if m ∈ Dn \ {n}, then skip step1 and set Hn HNP n, sm at step3. 3.14
HP n, s
⎧ ⎫
⎨ ⎬
min cn,k pn,k dn,j − s hn dn,j − dn H m, d0,j − d0,n ,
j∈DFT n:Qn,k−1 ≤dn,j −s≤Qn,k ⎩ ⎭
m∈Cn
3.15
where DFT n {j | j ∈ Dn and tj − tn ≤ FT} denotes the set of node n’s
descendants within FT generations. The problem is how to select proper FT. It is obvious that
increasing FT will not only improve the quality of solution, but will also increase computation
time. Thus, the quality of solution and run time can be balanced by properly selecting FT.
The performance of the proposed RHH with different FT settings will be tested by numerical
experiments in the next section.
and this iteration for sm can be skipped too. Look-backward strategy with parameter BF
backward time stage can be stated as follows:
Lemma 3.2. For each node n ∈ N \ {0}, the rolling horizon heuristic with FT and BT only needs to
evaluate the following values:
H n, d0,m − d0,an : d0,m − d0,an ≥ 0, m ∈ RHFT,BT n , 3.17
Proof. This lemma is proven by induction from nodes at stage t 1 to leaf nodes at stage T .
Since positive order must be made at node 0 due to the zero initial inventory assumption, thus
for node n at stage tn 2 by 3.15 possible initial inventory san at node n could only be a
certain value in set {d0,m −d0,an : d0,m −d0,an ≥ 0, m ∈ RHFT,BT n} {d0,m −d0,0 : d0,m −d0,0 ≥
0, m ∈ DFT 0}. Suppose the lemma holds for node an at stage t, where 2 ≤ t ≤ T , that is,
the set of initial inventory for node an is given by {d0,m − d0,aan : d0,m − d0,aan ≥ 0, m ∈
RHFT,BT an}.
Now consider the possible initial inventory at node n.
Case 1. If a positive order has been made at node an, by 3.15 in the look-forward
strategy, the set of initial inventory at nodes n is given by {d0,m − d0,an : d0,m − d0,an ≥ 0, m ∈
DFT an}.
Case 2. if no order has been made at node an, the set of possible initial inventory
at node n is given by {d0,m − d0,an : d0,m − d0,an ≥ 0, m ∈ RHFT,BT an}. However, note
that by look-backward strategy 3.16 we skip the calculation of HNP an, d0,m − d0,an for
those m such that tan − tm BF; thus, in this case we only need to consider initial
inventory value in set {d0,m − d0,an ≥ 0, m ∈ ∪{DFT m : m ∈ P 0, aan, tan − tm ≤
FT BT − 1}} {d0,m − d0,an ≥ 0, m ∈ ∪{DFT m : m ∈ P 0, aan, tn − tm ≤ FT BT}}
at node n.
Combine the above two cases, the conclusion holds for node n.
Figure 1 gives an intuitive example of a balanced scenario tree with C 2 at each node.
For the rolling horizon heuristic with FT 2 and BT 1, given RHH2,1 an for node an,
consider the initial inventory set for node n. If a positive order has been made at node an,
{d0,j − d0,an : d0,j − d0,an ≥ 0, j ∈ DFT an and tj tn FT − 1} will be introduced
by 3.15. By the look-backward strategy 3.16, we do not need to calculate Hn, sm , where
tm tn − BT − 1. Other values will be inherited if no order has been made at node an.
Mathematical Problems in Engineering 15
.
.
.
.
.
.
.
.
.
.
.
.
m
Stage 1
a(m)
n
.
.. .
..
a(n)
.
a(a(n))
..
a(j) j
RH (a(n))
Stage t− BT −1
.
RH (n) .
.
Stage t− BT Stage t Stage t+ FT −1
Note that | RHFT,BT n | is no more than FT BTCFT where we assume that | Cn | C for
all n ∈ N.
By summarizing the above analysis, the rolling horizon heuristic is given as in
Algorithm 2.
Next the computation complexity of RHH is analyzed. For given node n at stage t,
there exist at most | RHFT,BT n | possible initial inventories sm . For each given sm , it takes
OKn | Cn | | DFT n | time to complete calculation of HP n, sm at step 1. Step 2 and step
3 will be completed in OCn time. Thus the total run time can be estimated as assume that
Kn K for all n ∈ N
CFT1 − 1
KCFT BTCFT |D0|
C−1
16 Mathematical Problems in Engineering
CFT1
KFT BT CFT1 − 1 |D0|
C−1
≤ KFT BTC2FT1 |D0|.
3.18
Proposition 3.3. For any instance of SULS-IDQ, the rolling horizon heuristic with parameter FT
and BT runs in OK FT BT C2FT1 | D0 | time.
The above analysis can be applied to the dynamic programming algorithm, thus the
total run time for DP is given by
T
KC|D0| Ct−1 1 C1 · · · CT −t 3.19
t1
C CT − 1
K |D0| TC −
T
.
C−1 C−1
By comparison of computational complexity in 3.18 and 3.19, we observe that RHH works
more efficiently for large T . Effective parameters settings of RHH will be further explored by
numerical experiments in the next section.
4. Computational Experiment
In this section, the computational results on both DP and RHH are reported. In the
computational analysis, we first concentrate on identifying proper settings of parameters FT
and BT for RHH by comparison of its relative error gap and run time with DP’s. Then, DP and
RHH are compared with the CPLEX solver and other heuristics for the lot-sizing problems
with fixed charge.
Mathematical Problems in Engineering 17
Parameter Distribution
hn,k Uniform 1, 2
pn,1 Uniform 1, 2
cn,1 Uniform 3, 5
Qn,k Uniform 5, 15
dn,k Truncated normal 8, σ 2
RHH1 RHH2 DP
Instance cv
RE % CPU sec RE % CPU sec Value CPU sec
G1 − 1 0.05 0.289 0.016 0.0 0.094 12968.28 0.100
G1 − 2 0.15 0.168 0.015 0.0 0.110 12991.93 0.985
G1 − 3 0.25 0.135 0.015 0.0 0.110 13082.23 0.969
G2 − 1 0.05 0.039 0.079 0.0 0.219 25694.23 4.672
G2 − 2 0.15 0.015 0.079 0.0 0.219 26111.22 4.672
G2 − 3 0.25 0.005 0.078 0.0 0.218 26447.11 4.703
G3 − 1 0.05 0.009 0.172 0.0 1.109 51470.76 22.047
G3 − 2 0.15 0.008 0.172 0.0 1.109 51722.50 22.109
G3 − 3 0.25 0.006 0.156 0.0 1.110 52772.90 22.172
G4 − 1 0.05 2.431 0.047 0.509 0.156 8322.42 0.656
G4 − 2 0.15 1.529 0.031 0.195 0.125 8998.02 0.641
G4 − 3 0.25 1.391 0.047 0.158 0.125 9126.63 0.640
G5 − 1 0.05 2.877 0.140 0.635 0.516 24898.40 7.344
G5 − 2 0.15 1.869 0.125 0.409 0.516 26456.16 7.359
G5 − 3 0.25 1.212 0.141 0.190 0.500 27716.36 7.375
G6 − 1 0.05 2.950 0.438 0.214 5.937 74120.22 81.484
G6 − 2 0.15 1.879 0.453 0.064 5.969 79096.31 81.563
G6 − 3 0.25 1.434 0.453 0.054 6.047 82641.24 81.844
G7 − 1 0.05 7.894 0.031 1.270 0.093 7814.61 0.828
G7 − 2 0.15 5.590 0.016 0.783 0.093 8551.74 0.813
G7 − 3 0.25 4.350 0.015 0.435 0.094 9024.62 0.812
G8 − 1 0.05 1.697 0.437 0.247 2.219 30614.77 16.969
G8 − 2 0.15 1.001 0.453 0.075 2.234 33982.89 16.938
G8 − 3 0.25 0.597 0.437 0.053 2.234 36099.06 17.032
time in seconds for DP, and the percentage relative error RE% and CPU time for RHH1
and RHH2 , where
For RHH1 , the computational results show that RHH1 is able to solve all the instances in the
shortest time with no more than 3% relative error except, for instance, G7. The solution quality
and run time are further improved when the number of stages T becomes larger and the worst
results of G7 can be interpreted as the look-forward time horizon being too small to obtain
near optimal solution. For RHH2 , the relative error has been further decreased compared
with RHH1 . Although the run time has been increased, RHH2 still has advantage in run time
compared with DP especially for large problem instances. We also note that as the coefficient
of variation increases, the solutions quality and run time have been constantly improved for
all problem instances.
Then in the second experiment, we concentrate on the comparison of solution quality
and computation time with standard MIP solver CPLEX version 11.1 and another heuristic
dynamic slope scaling procedure DSSP. DSSP proposed by Kim and Pardalos 17 is a
Mathematical Problems in Engineering 19
newly developed effective heuristic algorithm that provides good-quality solution to the
concave piecewise linear optimization problem and among the heuristics it works well in
practice 18 The initial solutions and updating schemes are implemented in accordance with
the recommendations in 18 and the stopping criterion is given as follows: if xk − xk−1 ≤ ε
or the iteration reaches MaxIteration, then it stops, where MaxIteration 20. To compare
the solution quality for given time limit, we set the time limit of CPLEX solver to 5 time of
the total run time of DP. We report in Table 4 the percentage relative error and CPU time of
RHH its parameters are set according to RHH2 , DSSP and CPLEX and the optimal objective
function value and CPU time of DP. Table 4 shows that the proposed RHH outperforms the
DSSP in both the quality of solution and run time for almost all the test instances, and the
CPLEX solver fails to find optimal solution within the given time.
In summary, the proposed DP can solve the SULP-IQD efficiently compared with the
standard CPLEX solver and by properly setting the parameters, we obtain effective RHH
which outperforms the DSSP heuristic for the tested instances. The computational results
also show that RHH performs better for problem instances with a larger number of stages
and high coefficient of variation.
20 Mathematical Problems in Engineering
5. Conclusion
In this paper, we study the stochastic uncapacitated lot-sizing problems with incremental
quantity discount where the uncertain parameters are supposed to evolve as discrete time
stochastic processes. First, we establish the original stochastic formulation by scenario
analysis approach. Another two formulations are built by introducing auxiliary variables
and relaxing certain constraints. Then, it is proven that under the decreasing unit order
price assumption, the relaxed formulation is equivalent to the original one. Based on this
reformulation, the extended production-path property is presented for the SULP-IQD and
it enhances the ability to further refine the desired optimal solution by providing a more
accurate characterization.
To obtain the exact optimal solution, a dynamic programming algorithm is developed.
Although the dynamic programming algorithm has the polynomial-time computational
complexity in terms of the number of nodes, it runs exponentially in terms of the number
of stages. Thus, a new rolling horizon heuristic is further designed which contains three
types of strategies to reduce the computational time. The nonproduction strategy is designed
based on the accurate characterization of the optimal solution, and the look-forward and
look-backward strategy is developed to overcome the complete enumeration calculations in
the production and nonproduction value function. Numerical experiments are carried out to
identify proper parameters settings of the proposed RHH and to evaluate the performance
of the proposed algorithms by comparison with the CPLEX solver and DSSP heuristic. The
computational results of a large group of problem instances with different parameters setting
suggest that DP outperforms the CPLEX solver in run time required for obtaining optimal
solution and the proposed RHH demonstrates satisfactory run time and solution quality
compared with DSSP heuristic; moreover, as the computational complexity analysis suggests,
the performance of RHH is better for problems with a greater number of stages.
Since the main difficulties for the stochastic lot-sizing problems stem from the setup
cost and uncertain parameters, it will be an area of future research to analyze the properties
of the problem and present effective algorithms for the stochastic lot-sizing problems with
complex setup requirements, such as setup carryovers by Buschkühl et al. 14 and sequence-
dependent setup costs by Beraldi et al. 1.
Acknowledgments
The paper is supported by the National Nature Science Foundation of China NSFC no.
60874071 and 60834004 and Research Fund for the Doctoral Program of Higher Education of
China RFDP no. 20090002110035.
References
1 P. Beraldi, G. Ghiani, A. Grieco, and E. Guerriero, “Fix and relax heuristic for a stochastic lot-sizing
problem,” Computational Optimization and Applications, vol. 33, no. 2-3, pp. 303–318, 2006.
2 H. M. Wagner and T. M. Whitin, “Dynamic version of the economic lot size model,” Management
Science, vol. 5, pp. 89–96, 1958.
3 J. Hu and C. L. Munson, “Dynamic demand lot-sizing rules for incremental quantity discounts,”
Journal of the Operational Research Society, vol. 53, no. 8, pp. 855–863, 2002.
4 L. M. A. Chan, A. Muriel, Z. J. Shen, and D. S. Levi, “On the effectiveness of zero-inventory-ordering
policies for the economic lot-sizing model with a class of piecewise linear cost structures,” Operations
Research, vol. 50, no. 6, pp. 1058–1067, 2002.
Mathematical Problems in Engineering 21
5 A. Federgruen and C. Y. Lee, “The dynamic lot size model with quantity discount,” Naval Research
Logistics, vol. 37, no. 5, pp. 707–713, 1990.
6 D. X. Shaw and A. P. M. Wagelmans, “An algorithm for single-item capacitated economic lot sizing
with piecewise linear production costs and general holding costs,” Management Science, vol. 44, no. 6,
pp. 831–838, 1998.
7 S. Ahmed, A. J. King, and G. Parija, “A multi-stage stochastic integer programming approach for
capacity expansion under uncertainty,” Journal of Global Optimization, vol. 26, no. 1, pp. 3–24, 2003.
8 Y. P. Guan and A. J. Miller, “Polynomial-time algorithms for stochastic uncapacitated lot-sizing
problems,” Operations Research, vol. 56, no. 5, pp. 1172–1183, 2008.
9 K. Huang and S. Küçükyavuz, “On stochastic lot-sizing problems with random lead times,”
Operations Research Letters, vol. 36, no. 3, pp. 303–308, 2008.
10 G. Lulli and S. Sen, “A branch-and-price algorithm for multistage stochastic integer programming
with application to stochastic batch-sizing problems,” Management Science, vol. 50, no. 6, pp. 786–796,
2004.
11 P. Beraldi, G. Ghiani, E. Guerriero, and A. Grieco, “Scenario-based planning for lot-sizing and
scheduling with uncertain processing times,” International Journal of Production Economics, vol. 101,
no. 1, pp. 140–149, 2006.
12 D. Quadt and H. Kuhn, “Capacitated lot-sizing with extensions: a review,” A Quarterly Journal of
Operations Research, vol. 6, no. 1, pp. 61–83, 2008.
13 J. Raf and D. Zeger, “Modeling industrial lot sizing problems: a review,” International Journal of
Production Research, vol. 46, no. 6, pp. 1619–1643, 2008.
14 L. Buschkühl, F. Sahling, S. Helber, and H. Tempelmeier, “Dynamic capacitated lot-sizing problems:
a classification and review of solution approaches,” OR Spectrum, vol. 32, no. 2, pp. 231–261, 2010.
15 C. L. Munson and M. J. Rosenblatt, “Theories and realities of quantity discounts: an exploratory
study,” Production and Operations Management, vol. 7, no. 4, pp. 352–369, 1998.
16 P. Beraldi, G. Ghiani, A. Grieco, and E. Guerriero, “Rolling-horizon and fix-and-relax heuristics for
the parallel machine lot-sizing and scheduling problem with sequence-dependent set-up costs,”
Computers and Operations Research, vol. 35, no. 11, pp. 3644–3656, 2008.
17 D. Kim and P. M. Pardalos, “A solution approach to the fixed charge network flow problem using a
dynamic slope scaling procedure,” Operations Research Letters, vol. 24, no. 4, pp. 195–203, 1999.
18 S. Lawphongpanich, “Dynamic slope scaling procedure and lagrangian relaxation with subproblem
approximation,” Journal of Global Optimization, vol. 35, no. 1, pp. 121–130, 2006.
Hindawi Publishing Corporation
Mathematical Problems in Engineering
Volume 2012, Article ID 478931, 22 pages
doi:10.1155/2012/478931
Research Article
Sparse Signal Recovery via ECME
Thresholding Pursuits
Copyright q 2012 H. Song and G. Wang. This is an open access article distributed under the
Creative Commons Attribution License, which permits unrestricted use, distribution, and
reproduction in any medium, provided the original work is properly cited.
The emerging theory of compressive sensing CS provides a new sparse signal processing
paradigm for reconstructing sparse signals from the undersampled linear measurements. Recently,
numerous algorithms have been developed to solve convex optimization problems for CS sparse
signal recovery. However, in some certain circumstances, greedy algorithms exhibit superior
performance than convex methods. This paper is a followup to the recent paper of Wang and Yin
2010, who refine BP reconstructions via iterative support detection ISD. The heuristic idea of
ISD was applied to greedy algorithms. We developed two approaches for accelerating the ECME
iteration. The described algorithms, named ECME thresholding pursuits EMTP, introduced two
greedy strategies that each iteration detects a support set I by thresholding the result of the ECME
iteration and estimates the reconstructed signal by solving a truncated least-squares problem on
the support set I. Two effective support detection strategies are devised for the sparse signals
with components having a fast decaying distribution of nonzero components. The experimental
studies are presented to demonstrate that EMTP offers an appealing alternative to state-of-the-art
algorithms for sparse signal recovery.
1. Introduction
Sparsity exploiting has recently received a great amount of attention in the applied
mathematics and signal processing community. Sparse signal processing algorithms have
been developed for various applications. A recent Proceedings of the IEEE special issue
on applications of sparse representation and compressive sensing devoted to this topic.
Some of the exciting developments addressed in 1–7. Compressed sensing, also known as
compressive sensing, compressive sampling CS 8, 9, has been the subject of significant
activity in sparse signal processing where one seeks to recover efficiently a sparse unknown
signal vector of dimension n via a much smaller number m of undersampled linear
2 Mathematical Problems in Engineering
measurements. For k-sparse unknown signal x0 ∈ Rn , the sparse signal recovery is intimately
related to solving an underdetermined system of linear equations y Ax0 with sparseness
constraint
where x is true signal to be recovered, x0 is used to denote the number of nonzero
components of x, A ∈ Rm×n is the measurement matrix, and y ∈ Rm is the measurement
vector. The key insight in the pioneering work on CS 8, 9 is to replace 0 by 1 for the P0
problem due to nonconvexity and combinatorial effect. In 10, it is the basis pursuit problem
Sections 2.1 and 2.2. Then, we review the initial work of Iterative Support Detection ISD
in Section 2.3. In Section 3, we describe the proposed approaches. After that, Section 4 details
the experimental results. Finally, we conclude this paper in Section 5.
1.1. Notation
We introduce the notation used in this paper.
i xt : the algorithms described in this paper are iterative and the reconstructed signal
x in current iteration t is denoted as xt . The same convention is used for other
vectors and matrices.
ii I, AI : index set I, the matrix AI denotes the submatrix of A containing only those
columns of A with indexes in I. The same convention is used for vectors.
iii 1, n \ I: the complement of set I in set {1, 2, . . . , n}.
iv suppx: the support set of a vector x, that is, the index set corresponding to the
nonzeros of x, suppx {i : xi
/ 0}.
v Hk x: the hard thresholding that sets all but the largest in magnitude k elements
of a vector x to zero.
vi |x|, xp , xT : the absolute value, p norm, and transpose of a vector x, respectively.
vii A† : the Moore-Penrose pseudoinverse of matrix A ∈ Rm×n . A† AT AAT −1 for
m ≤ n; A† AT A−1 AT for m ≥ n.
2. Related Works
2.1. One-Stage Thresholding
Qiu and Dogandzic derived an expectation conditional maximization either ECME iteration
based on a probabilistic framework 29
−1
t
x Hk xt−1 A T
AA T
y − Ax t−1
. 2.1
Note that ECME iteration reduces to IHT step when the measurement matrix A has
orthonormal rows i.e., AAT is the identity matrix. These one-stage thresholding
algorithms e.g., IHT 27 and ECME 29 are guaranteed to recover sparse signals and
converge with limited iterations. However, OST is not the empirical choice for practical
applications due to slow convergence. To this end, Qiu and Dogandzic proposed an
acceleration method, termed double overrelaxation DORE thresholding scheme 28, 29,
to improve the convergence speed of the ECME algorithm. DORE utilizes two overrelaxation
steps:
t
x1 xt α1 xt − xt−1 ,
2.2
t t t
x2 x1 α2 x1 − xt−2 ,
4 Mathematical Problems in Engineering
where α1 , α2 are the line search parameters. Finally, an additional hard thresholding step
t
xt Hk x2 2.3
ensures that the resulting signal is guaranteed to be k-sparse. In addition, Qiu and Dogandzic
further presented an automatic double overrelaxation ADORE thresholding method for
conditions that the underlying sparsity level is not available.
y AI xI . 2.4
Step 1 support detection. Detect the support set I of the signal x, that is, select atoms of
measurement matrix A which have been used to generate y; in other words, determine active
atoms in sparse representation of a signal x. In some literature, this step also is called basis
selection or atom selection.
Step 2 signal estimation. Update the signal x using the least-squares solution on the
detected support set I. xI arg minz {y − AI z22 , suppz ⊆ I}, x1,n\I 0.
Many algorithms e.g., Orthogonal Matching Pursuit OMP 23, Subapace Pursuit
SP 24, Compressed Sensing Matching Pursuit CoSaMP 25, and Gradient Pursuits GP
32 developed various approaches for Step 1 or Step 2.
The reliability of ISD relies on the support detection. Wang and Yin devised serval detection
strategies for the sparse signals with components having a fast decaying distribution of
nonzero components called fast decaying signals 30. One of the detection strategies is
based on thresholding we use ISD defined by 2.6 to refer the implementation algorithm in
the following context:
t
I t i : xi > βt maxxt , β ∈ 0, 1. 2.6
t †
xI t AI t y,
3.2
t
x1,n\I t 0.
Step 4 halting. Check the stopping condition and, if it is not yet time to stop, increment
t t 1 and return to Step 2. If it is time to stop, the recovered signal x has nonzero entries in
t
support set I t and corresponding support vector lies in xI t .
Remarks. 1 EMTP updates the leading elements of xt using a least-squares solution on the
support set I. However, DORE updates all the elements of xt using double overrelaxation
Mathematical Problems in Engineering 7
Sparsity = 20, detected (total = 7, good = 6, Sparsity = 20, detected (total = 12, good = 11,
bad = 1, miss = 14), RelErr = 5.86e−001 bad = 1, miss = 9), RelErr = 3.74e−001
2 2
1.5 1.5
1 1
0.5 0.5
0 0
−0.5 −0.5
−1 −1
−1.5 −1.5
−2 −2
20 60 100 140 180 20 60 100 140 180
Sparsity = 20, detected (total = 18, good = 14, Sparsity = 20, detected (total = 21, good = 18,
bad = 4, miss = 6), RelErr = 1.93e−001 bad = 3, miss = 2), RelErr = 3.72e−002
2 2
1.5 1.5
1 1
0.5 0.5
0 0
−0.5 −0.5
−1 −1
−1.5 −1.5
−2 −2
20 60 100 140 180 20 60 100 140 180
Sparsity = 20, detected (total = 19, good = 19, Sparsity = 20, detected (total = 20, good = 20,
bad = 0, miss = 1), RelErr = 1.9e−002 bad = 0, miss = 0), RelErr = 1.4e−015
2 2
1.5 1.5
1 1
0.5 0.5
0 0
−0.5 −0.5
−1 −1
−1.5 −1.5
−2 −2
20 60 100 140 180 20 60 100 140 180
Figure 1: An EMTP-β demo that recovers a Gaussian-distributed sparse vector with n 200, m 80, k
20, β 0.5.
8 Mathematical Problems in Engineering
SNR (dB)
150
100
50
0
−50
0.1 0.15 0.2 0.25 0.3 0.35 0.4 0.45 0.5
m/n
a
n = 400, k = 20, m = 40 : 10: 200
25
Number of iterations
20
15
10
5
0
0.1 0.15 0.2 0.25 0.3 0.35 0.4 0.45 0.5
m/n
Figure 2: Influence of thresholding parameter β for sparse Gaussian signals with n 400, k 20:
comparisons in terms of SNR dB and number of iterations.
steps. Finally, DORE needs a hard thresholding to ensure the new approximation is k-sparse.
EMTP-β demands no prior knowledge about the underlying sparsity level k.
2 EMTP is a special case of TST and utilizes OST in the support detection stage. It is
different from OMP, SP, and CoSaMP. When the ECME iteration reduces to IHT step, that is,
the reference for support detection xt xt−1 AT r t−1 is the gradient descent step for least-
squares, OMP, SP and CoSaMP use the correlation between the residual signal and the atoms
of the measurement matrix A i.e., AT r t−1 to detect support set. AT r t−1 is the negative
gradient for least squares. It is clear that ECME iteration provides the estimation for the
underlying sparse signal. We can employ the heuristic idea of ISD to devise various support
detection strategies depending on the underlying sparse signal distribution. The dynamic
thresholding method can be performed without the sparsity level k. EMTP directly detects
the support of the underlying sparse signal by referencing the ECME iteration while OMP, SP,
and CoSaMP augment the support by picking out the leading values of the negative gradient.
OMP each step spots one index into the support, so it requires more iterations than EMTP.
SP and CoSaMP spot several indexes into the support, so they need an additional step i.e.,
orthogonal projection to make sure the recovered signal is k-sparse. Like ISD, EMTP has an
appealing self-correction capacity. An EMTP demo is presented in Figure 1.
3 Like other greedy pursuits such as Subapace Pursuit 24 and CoSaMP 25,
EMTP-k fixes the cardinality of support set I and removes previous false detections.
However, EMTP-β refines the support set I which is not necessarily increasing and nested
over the iterations.
Mathematical Problems in Engineering 9
n = 400, m = 80, k = 4 : 2 : 40
350
300
250
SNR (dB)
200
150
100
50
0
0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4 0.45 0.5
k/m
n = 400, m = 80, k = 4 : 2 : 40
30
Number of iterations
25
20
15
10
5
0
0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4 0.45 0.5
k/m
Figure 3: Influence of thresholding parameter β for sparse Gaussian signals with n 400, m 80:
comparisons in terms of SNR dB and number of iterations.
4 The dynamic thresholding strategy used in EMTP-β is inspired by ISD 30. It finds
significant nonzeros by comparing a threshold rather than maintaining fixed number k of
items. It is appealing for the conditions that the underlying sparsity level k is not available.
5 EMTP resembles ISD 30. EMTP and ISD have the same idea in support detection
step with iterative behavior. However, the support detection step is based on different sparse
recovery methods, ECME and BP, respectively. EMTP updates the reconstruction signal using
a least-squares solution on the detected support set I. However, ISD iteratively refines the BP
solution on the complement of the detected support set I.
6 EMTP-β with large β obtains high-quality reconstruction from a small number of
measurements. However, because support set I grows slowly, EMTP-β takes a larger number
of iterations to be discussed in Section 4. EMTP-k wrongly detected elements can often be
pruned out in later iterations.
7 Like ISD as discussed in 30, EMTP-β only performs well for fast decaying
signals. It does not work on sparse signals that decay slowly or have no decay at all e.g.,
trinary and binary sparse signals. EMTP-k performs worse for the sparse signals that
nonzero elements have similar magnitude 31, 33. However, we can apply EMTP to non-
fast-decaying signals via linear or nonlinear premapping 34, 35.
10 Mathematical Problems in Engineering
SNR (dB)
150
100
50
0
−50
0.1 0.15 0.2 0.25 0.3 0.35 0.4
m/n
30
25
20
15
10
5
0
0.1 0.15 0.2 0.25 0.3 0.35 0.4
m/n
Figure 4: Influence of thresholding parameter β for sparse Gaussian signals with n 2000, k 100:
comparisons in terms of SNR dB and number of iterations.
4. Experimental Results
To assess the performance of the proposed approaches, we conduct numerical
experiments on computer simulations. All algorithms were implemented and tested
in MATLAB v7.6 running on Windows XP with 2.53 GHz Intel Celeron CPU and
2 GB of memory. We compared the proposed approaches with the accelerated
ADORE/DORE approaches. The code of ADORE/DORE is available on the authors
homepage http://home.eng.iastate.edu/∼ald/DORE.htm. We initialized x with zero
sparse signal. The search length of ADORE was set to 1. All results were averaged
Mathematical Problems in Engineering 11
SNR (dB)
150
100
50
0
0 0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4 0.45 0.5
k/m
35
30
25
20
15
10
5
0 0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4 0.45 0.5
k/m
Figure 5: Influence of thresholding parameter β for sparse Gaussian signals with n 2000, m 400:
comparisons in terms of SNR dB and number of iterations.
30
Number of iterations
25
20
15
10
5
200 400 600 800 1000 1200 1400 1600 1800 2000 2200
n
DORE
EMTP-k
EMTP-β = 0.8
Figure 6: Number of iterations as a function of problem size with fixed ratios of m/n 0.4, k/m 0.2, n
200 : 400 : 2200.
12 Mathematical Problems in Engineering
70
60
Number of iterations
50
40
30
20
10
200 400 600 800 1000 1200 1400 1600 1800 2000 2200
n
DORE
EMTP-k
EMTP-β = 0.8
Figure 7: Number of iterations as a function of problem size with fixed ratios of m/n 0.3, k/m 0.3, n
200 : 400 : 2200.
over 100 times Monte Carlo problem instances. We used our unoptimized code
http://cs-notes.googlecode.com/files/EMTP.zip. The least-squares solution x
argminy−Ax22 was implemented using MATLAB pseudoinverse function by x pinvA∗
y. The MATLAB code was partially adapted from 30, 33, 36. In all the experiments, the mea-
surement matrix A was generated by uniform spherical ensemble USE, that is, we generate
the measurement matrix by sampling each entry independently from the standard norm dis-
tribution and then normalize each column to have unit norm. The MATLAB code is given by
A randnm, n; A A./repmatsqrtsumA.∧ 2, m 1.
The underlying k-sparse vectors were generated by randomly selecting a support set
of size k and each entry in the support set is sampled uniformly from a specific distribution.
The sparse signals were generated in MATLAB by
x zerosn, 1; support randpermn; xsupport1 : k v;
and the nonzeros v generated by following code.
The sparse Gaussian signals were generated in MATLAB by
v randnk, 1;
The sparse Laplacian signals were generated by
z randk, 1;
v zerosk, 1;
in z < 0.5;
ip z > 0.5;
1
vin ∗ log2 ∗ zin;
lambda
1
vip − ∗ log2 ∗ 1 − zip.
lambda
The power-law decaying signals were generated by
v signrandnk, 1. ∗ (1 : k). \∧ −1/lambda .
Mathematical Problems in Engineering 13
SNR (dB)
100
0
−100
0.1 0.15 0.2 0.25 0.3 0.35 0.4 0.45 0.5
m/n
800
600
400
200
0.1 0.15 0.2 0.25 0.3 0.35 0.4 0.45 0.5
ADORE
b
Number of iterations
100
50
0
0.1 0.15 0.2 0.25 0.3 0.35 0.4 0.45 0.5
m/n
DORE
EMTP-β = 0.5
EMTP-k
c
Figure 8: Sparse Gaussian signals with n 400, k 20: comparisons in terms of SNR dB and number of
iterations.
The variable lambda controls the rate of decay. We set lambda 10 and lambda 0.5
for sparse Laplacian signals and power-law decaying signals, respectively.
For fair comparison, we stopped iterations once the relative error fell below a certain
convergence tolerance or the number of iterations is greater than 100. The convergence
tolerance is given by
y − r t 2
≤ 10−6 . 4.1
y 2
n = 400, m = 80, k = 4 : 2 : 40
350
300
250
SNR (dB)
200
150
100
50
0
0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4 0.45 0.5
k/m
n = 400, m = 80, k = 4 : 2 : 40
800
Number of iterations
600
400
200
0
0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4 0.45 0.5
k/m
Figure 9: Sparse Gaussian signals with n 400, m 80: comparisons in terms of SNR dB and number of
iterations.
SNR (dB)
150
100
50
0
0.1 0.15 0.2 0.25 0.3 0.35 0.4
m/n
1000
800
600
400
200
0
0.1 0.15 0.2 0.25 0.3 0.35 0.4
m/n
Figure 10: Sparse Gaussian signals with n 2000, k 100: comparisons in terms of SNR dB and number
of iterations.
with m/n 0.3, k/m 0.3, n 200 : 400 : 2200, DORE requires more iterations, EMTP-k
needs significantly fewer iterations with increasing the problem size, and EMTP-β requires
relatively stable iterations.
SNR (dB)
150
100
50
0
0 0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4 0.45 0.5
k/m
1000
800
600
400
200
0
0 0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4 0.45 0.5
k/m
Figure 11: Sparse Gaussian signals with n 2000, m 400: comparisons in terms of SNR dB and number
of iterations.
Figures 15 and 16. The latter tests come to similar conclusions as the first test. These figures
do not fully capture the performance of EMTP-β since we only set β 0.5; however EMTP-β
achieved almost superior performance than ADORE/DORE.
SNR (dB)
150
100
50
0
0.2 0.25 0.3 0.35 0.4 0.45 0.5 0.55 0.6
m/n
800
600
400
200
0
0.2 0.25 0.3 0.35 0.4 0.45 0.5 0.55 0.6
m/n
Figure 12: Sparse Laplacian signals with n 400, k 20: comparisons in terms of SNR dB and number
of iterations.
SNR (dB)
150
100
50
0
0.1 0.15 0.2 0.25 0.3 0.35 0.4
m/n
1000
800
600
400
200
0
0.1 0.15 0.2 0.25 0.3 0.35 0.4
m/n
Figure 13: Sparse Laplacian signals with n 2000, k 100: comparisons in terms of SNR dB and number
of iterations.
4.5. Summary
To sum up, we have compared EMTP with ADORE/DORE. EMTP has significant advantages
over the accelerated OST in terms of SNR and number of iterations. EMTP can significantly
reduce the number of iterations required and achieves significantly higher SNR. For low
indeterminacy level m/n, EMTP requires fewer measurements. EMTP-β can work with no
prior knowledge of the underlying sparsity level k yet achieves recovery performance better
than ADORE. Furthermore, we compared EMTP with the state-of-the-art greedy algorithm
SP in terms of phase transitions. Among all methods, EMTP-β appeared to be the best.
SNR (dB)
150
100
50
0
0.2 0.25 0.3 0.35 0.4 0.45 0.5 0.55 0.6
m/n
800
600
400
200
0
0.2 0.25 0.3 0.35 0.4 0.45 0.5 0.55 0.6
m/n
Figure 14: Power-law decaying signals with n 400, k 20: comparisons in terms of SNR dB and
number of iterations.
then turn to the problem of reducing the computational cost of each iteration, especially for
large scale applications. Future research includes three directions.
1 EMTP requires the precomputation and storage of the pseudoinverse matrix A† .
So it is the advisable choice that replacing ECME iteration by IHT step in the
support detection stage. IHT does not require the computation and storage of
the pseudoinverse matrix A† , which leads to significant advantage for large scale
applications.
2 In signal estimation stage, that is, update reconstructed signal by solving least-
squares problem, EMTP uses the orthogonal projection, that is, by calculating
† †
xI AI y where AI is the pseudoinverse of AI . We will approximate the orthogonal
projection efficiently by gradient pursuits 32.
3 This paper devotes efforts to devise computational algorithms which are exper-
imentally reproducible and provides empirical studies as useful guidelines for
practical applications. Further, future investigations will test other OST methods
as the reference and sophisticated thresholding methods for support detection.
20 Mathematical Problems in Engineering
SNR (dB)
150
100
50
0
0.1 0.15 0.2 0.25 0.3 0.35 0.4
m/n
1000
800
600
400
200
0
0.1 0.15 0.2 0.25 0.3 0.35 0.4
m/n
Figure 15: Power-law decaying signals with n 2000, k 100: comparisons in terms of SNR dB and
number of iterations.
0.7
0.6
0.5
0.35
0.4
0.3
0.2
0.1
0
0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4 0.45 0.5
Sparsity k/m
DORE
SP
EMTP-β = 0.8
Figure 16: Probability of exact recovery for DORE, SP, and EMTP as a function of problem sparsity and
four problem indeterminacies from the thickest to thinnest lines: δ m/n 0.05, 0.2, 0.35, 0.5.
Mathematical Problems in Engineering 21
0.55
0.5
0.45
0.4
Sparsity k/m
0.35
0.3
0.25
0.2
0.15
0.1
0.05
0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4 0.45 0.5
Indeterminacy m/n
DORE
SP
EMTP-β = 0.8
Figure 17: Comparison of phase transitions of DORE, SP, and EMTP for Gaussian-distributed sparse
vectors.
Acknowledgment
This work was supported by the National Science Foundation of China under Grant no.
61074167, 61170126, and the Scientific Research Foundation for Advanced Talents by the
Jiangsu University, no. 12JDG050.
References
1 M. Elad, M. A. T. Figueiredo, and Y. Ma, “On the role of sparse and redundant representations in
image processing,” Proceedings of the IEEE, vol. 98, no. 6, pp. 972–982, 2010.
2 M. J. Fadili, J. L. Starck, J. Bobin, and Y. Moudden, “Image decomposition and separation using sparse
representations: an overview,” Proceedings of the IEEE, vol. 98, no. 6, pp. 983–994, 2010.
3 M. D. Plumbley, T. Blumensath, L. Daudet, R. Gribonval, and M. E. Davies, “Sparse representations
in audio and music: from coding to source separation,” Proceedings of the IEEE, vol. 98, no. 6, pp.
995–1005, 2010.
4 L. C. Potter, E. Ertin, J. T. Parker, and M. Çetin, “Sparsity and compressed sensing in radar imaging,”
Proceedings of the IEEE, vol. 98, no. 6, pp. 1006–1020, 2010.
5 J. Wright, Y. Ma, J. Mairal, G. Sapiro, T. S. Huang, and S. Yan, “Sparse representation for computer
vision and pattern recognition,” Proceedings of the IEEE, vol. 98, no. 6, pp. 1031–1044, 2010.
6 A. Y. Yang, M. Gastpar, R. Bajcsy, and S. S. Sastry, “Distributed sensor perception via sparse
representation,” Proceedings of the IEEE, vol. 98, no. 6, pp. 1077–1088, 2010.
7 R. Robucci, J. D. Gray, L. K. Chiu, J. Romberg, and P. Hasler, “Compressive sensing on a CMOS
separable-transform image sensor,” Proceedings of the IEEE, vol. 98, no. 6, pp. 1089–1101, 2010.
8 E. J. Candès, J. Romberg, and T. Tao, “Robust uncertainty principles: exact signal reconstruction from
highly incomplete frequency information,” IEEE Transactions on Information Theory, vol. 52, no. 2, pp.
489–509, 2006.
9 D. L. Donoho, “Compressed sensing,” IEEE Transactions on Information Theory, vol. 52, no. 4, pp. 1289–
1306, 2006.
10 S. S. Chen, D. L. Donoho, and M. A. Saunders, “Atomic decomposition by basis pursuit,” SIAM
Review, vol. 43, no. 1, pp. 129–159, 2001.
22 Mathematical Problems in Engineering
11 A. M. Bruckstein, D. L. Donoho, and M. Elad, “From sparse solutions of systems of equations to sparse
modeling of signals and images,” SIAM Review, vol. 51, no. 1, pp. 34–81, 2009.
12 J. A. Tropp and S. J. Wright, “Computational methods for sparse solution of linear inverse problems,”
Proceedings of the IEEE, vol. 98, no. 6, pp. 948–958, 2010.
13 B. K. Natarajan, “Sparse approximate solutions to linear systems,” SIAM Journal on Computing, vol.
24, no. 2, pp. 227–234, 1995.
14 R. Tibshirani, “Regression shrinkage and selection via the lasso,” Journal of the Royal Statistical Society,
Series B, vol. 58, no. 1, pp. 267–288, 1996.
15 B. Efron, T. Hastie, I. Johnstone, and R. Tibshirani, “Least angle regression,” The Annals of Statistics,
vol. 32, no. 2, pp. 407–499, 2004.
16 M. E. Tipping, “Sparse Bayesian learning and the relevance vector machine,” Journal of Machine
Learning Research, vol. 1, no. 3, pp. 211–244, 2001.
17 D. P. Wipf and B. D. Rao, “Sparse Bayesian learning for basis selection,” IEEE Transactions on Signal
Processing, vol. 52, no. 8, pp. 2153–2164, 2004.
18 S. Ji, Y. Xue, and L. Carin, “Bayesian compressive sensing,” IEEE Transactions on Signal Processing, vol.
56, no. 6, pp. 2346–2356, 2008.
19 S. D. Babacan, R. Molina, and A. K. Katsaggelos, “Bayesian compressive sensing using Laplace
priors,” IEEE Transactions on Image Processing, vol. 19, no. 1, pp. 53–63, 2010.
20 S. G. Mallat and Z. Zhang, “Matching pursuits with time-frequency dictionaries,” IEEE Transactions
on Signal Processing, vol. 41, no. 12, pp. 3397–3415, 1993.
21 Y. C. Pati, R. Rezaiifar, and P. S. Krishnaprasad, “Orthogonal matching pursuit: recursive function
approximation with applications to wavelet decomposition,” in Proceedings of the 27th Asilomar
Conference on Signals, Systems & Computers, vol. 1, pp. 40–44, Pacific Grove, Calif , USA, November
1993.
22 J. A. Tropp, “Greed is good: algorithmic results for sparse approximation,” IEEE Transactions on
Information Theory, vol. 50, no. 10, pp. 2231–2242, 2004.
23 J. A. Tropp and A. C. Gilbert, “Signal recovery from random measurements via orthogonal matching
pursuit,” IEEE Transactions on Information Theory, vol. 53, no. 12, pp. 4655–4666, 2007.
24 W. Dai and O. Milenkovic, “Subspace pursuit for compressive sensing signal reconstruction,” IEEE
Transactions on Information Theory, vol. 55, no. 5, pp. 2230–2249, 2009.
25 D. Needell and J. A. Tropp, “CoSaMP: iterative signal recovery from incomplete and inaccurate
samples,” Applied and Computational Harmonic Analysis, vol. 26, no. 3, pp. 301–321, 2009.
26 I. Daubechies, M. Defrise, and C. De Mol, “An iterative thresholding algorithm for linear inverse
problems with a sparsity constraint,” Communications on Pure and Applied Mathematics, vol. 57, no. 11,
pp. 1413–1457, 2004.
27 T. Blumensath and M. E. Davies, “Iterative hard thresholding for compressed sensing,” Applied and
Computational Harmonic Analysis, vol. 27, no. 3, pp. 265–274, 2009.
28 K. Qiu and A. Dogandzic, “Double overrelaxation thresholding methods for sparse signal
reconstruction,” in Proceedings of the 44th Annual Conference on Information Sciences and Systems (CISS
’10), Princeton, NJ, USA, March 2010.
29 K. Qiu and A. Dogandzic, “Sparse signal reconstruction via ECME hard thresholding,” IEEE
Transactions on Signal Processing, vol. 60, no. 9, pp. 4551–4569, 2012.
30 Y. Wang and W. Yin, “Sparse signal reconstruction via iterative support detection,” SIAM Journal on
Imaging Sciences, vol. 3, no. 3, pp. 462–491, 2010.
31 A. Maleki and D. L. Donoho, “Optimally tuned iterative reconstruction algorithms for compressed
sensing,” IEEE Journal on Selected Topics in Signal Processing, vol. 4, no. 2, pp. 330–341, 2010.
32 T. Blumensath and M. E. Davies, “Gradient pursuits,” IEEE Transactions on Signal Processing, vol. 56,
no. 6, pp. 2370–2382, 2008.
33 B. Sturm, “A study on sparse vector distributions and recovery from compressed sensing,”
http://arxiv.org/abs/1103.6246.
34 X. Zhang, J. Wen, Y. Han, and J. Villasenor, “An improved compressive sensing reconstruction
algorithm using linear/non-linear mapping,” in Proceedings of the Information Theory and Applications
Workshop (ITA ’11), pp. 146–152, San Diego, Calif, USA, February 2011.
35 X. Zhang, Z. Chen, J. Wen, J. Ma, Y. Han, and J. Villasenor, “A compressive sensing reconstruction
algorithm for trinary and binary sparse signals using pre-mapping,” in Proceedings of the Data
Compression Conference (DCC ’11), pp. 203–212, Snowbird, Utah, USA, March 2011.
36 V. Stodden, L. Carlin, D. Donoho et al., “SparseLab: seeking sparse solutions to linear systems of
equations,” Tech. Rep., Stanford University, 2011.
Hindawi Publishing Corporation
Mathematical Problems in Engineering
Volume 2012, Article ID 928312, 20 pages
doi:10.1155/2012/928312
Research Article
Variable Neighborhood Search for
Parallel Machines Scheduling Problem with
Step Deteriorating Jobs
Copyright q 2012 Wenming Cheng et al. This is an open access article distributed under the
Creative Commons Attribution License, which permits unrestricted use, distribution, and
reproduction in any medium, provided the original work is properly cited.
In many real scheduling environments, a job processed later needs longer time than the same
job when it starts earlier. This phenomenon is known as scheduling with deteriorating jobs to
many industrial applications. In this paper, we study a scheduling problem of minimizing the
total completion time on identical parallel machines where the processing time of a job is a
step function of its starting time and a deteriorating date that is individual to all jobs. Firstly, a
mixed integer programming model is presented for the problem. And then, a modified weight-
combination search algorithm and a variable neighborhood search are employed to yield optimal
or near-optimal schedule. To evaluate the performance of the proposed algorithms, computational
experiments are performed on randomly generated test instances. Finally, computational results
show that the proposed approaches obtain near-optimal solutions in a reasonable computational
time even for large-sized problems.
1. Introduction
Scheduling is a form of decision making that plays a crucial role in manufacturing and
service systems. It began to be taken seriously in manufacturing at the beginning of 20th
century, and since then has been received the attention of many researchers for years. In
the traditional scheduling problems, most research in the literature was usually conducted
under the assumption that the processing time of a job is known in advance and remains
constant throughout the whole decision process. However, there are many practical situations
where the processing times of jobs are not constant but increasing overtime such that the
later a job starts, the longer it takes to process. This phenomenon is known as scheduling
2 Mathematical Problems in Engineering
2. Literature Review
We discuss related work with respect to the scheduling problem with piecewise-deteriorating
jobs and with regard to VNS methods in scheduling, especially for parallel machines.
Mathematical Problems in Engineering 3
Since many different problems are encountered in production and service envi-
ronments where the processing time of jobs or tasks can be modeled as the piecewise-
deteriorating function, many active researchers in the field of scheduling have perceived
the promising advantages of considering the piecewise-deteriorating jobs in scheduling
problems. Kunnathur and Gupta 27 firstly assumed that job processing times are piecewise-
linear function of their starting times with two pieces and addressed five heuristics and two
optimizing algorithms based on dynamic programming and branch and bound techniques.
However, Sundararaghavan and Kunnathur 22 considered that the processing time can
be modeled by step function and gave some solvable cases for minimizing the sum of the
weighted completion times in single machine scheduling. In addition, Mosheiov 28 studied
the problem of stepwise deteriorating jobs for minimizing makespan, proved that the case of
single machine step-deterioration is NP-complete, and suggested heuristic methods in single
and multimachine scheduling environments.
Cheng and Ding 24 studied a piecewise model where the job processing time
deteriorates as a step function if its starting time exceeds a given threshold, and presented
NP-complete proofs for minimizing makespan, flow time, and weighted flow time in
single machine scheduling. Jeng and Lin 29 presented a pseudopolynomial time dynamic
programming algorithm for a single-machine scheduling problem. Then, they described an
efficient branch and bound algorithm based on two dominance rules and a lower bound for
deriving optimal solution. Subsequently, a single machine scheduling problem of minimizing
the total completion time was solved by branch and bound algorithms 30. Owing to the
complexity of the problem, He et al. 31 proposed a weight combination search algorithm
to yield a near-optimal solution. They also developed a branch and bound algorithm
incorporated with several properties and a lower bound to obtain the optimal solution.
Afterwards, Layegh et al. 32 applied a memetic algorithm based on three dominance
properties to minimize the total weighted completion time on a single machine under step-
deterioration.
With regards to the piecewise-linear-deterioration model, Kovalyov and Kubiak 33
investigated that the job processing time can be expressed by a piecewise-linear-increasing
function with three pieces and presented a fully polynomial approximation scheme for
minimizing makespan of single machine scheduling problem. Then Kubiak and van de Velde
23 proved that the problem is NP-hard, and proposed a branch and bound algorithm
that solves instances with up to 100 jobs in a reasonable amount of time. Alternatively, the
processing job times was expressed in a piecewise-linear-decreasing function of their start
times. The single machine scheduling problem was researched by Cheng et al. 34. Moslehi
and Jafari 35 dealt with the same problem proposed by 33 with the minimization of the
number of tardy jobs. Owing to the complexity of the studied problem, they proposed a
heuristic algorithm with On2 and a branch and bound algorithm.
As far as reviewed, there are some literature focused on single machine scheduling
problem with piecewise-deteriorating jobs. However, to the best of our knowledge, parallel
machines scheduling problem with piecewise deteriorating jobs has not been considered in
the existing literature. And it still lacks some effective approaches to solve the intractable
problem, especially metaheuristics. Therefore, we intend to propose a VNS approach for
parallel machines scheduling problem with step-deteriorating jobs.
VNS approaches have successfully applied to solve several scheduling problems.
Gupta and Smith 36 described a new hybrid of VNS for single machine total tardiness
scheduling with sequence-dependence setups. Paula et al. 37 applied a VNS algorithm to
solve large instances of parallel machines scheduling problems with sequence-dependent
4 Mathematical Problems in Engineering
times. Anghinolfi and Paolucci 38 contributed to the design of a hybrid meta-heuristic
approach which integrates several features from tabu search TS, simulated annealing SA,
and VNS for a generalized parallel machine total tardiness scheduling problem. Moreover,
Driessel and Mönch 39 proposed several variants of VNS schemes for the same problem
considered precedence constraints to minimize the total weighted tardiness of all jobs.
Behnamian et al. 40 proposed a hybrid meta-heuristic method which combines some
advantages of ant colony optimization, SA and VNS for the minimization of makespan in
parallel machine scheduling problem with sequence-dependent times. C.-L. Chen and C.-L.
Chen 41 proposed several hybrid metaheuristics for unrelated parallel machine scheduling
with sequence dependent setup times by integrating VNS and TS principles. Furthermore,
some job shop scheduling problems were solved by VNS 42–45. A latest systemic survey of
state-of-the-art development of VNS can be found in 46.
3. Problem Formulation
The problem under consideration is to schedule n independent jobs with step-deterioration
effects, noncommon deadlines and varying processing times on m identical parallel
machines. The jobs are available for processing at time zero and the machines are available
in the whole process. No job preemption is permitted. For each job Ji , there is a normal
processing time ai and associated with a deteriorating date di . If the starting point of its
process is less than or equal to its deteriorating date, then it only requires a normal processing
time ai . Otherwise, it requires an extra processing time bi , which is called the deteriorating
penalty. Thus, the actual processing time pi of job Ji depends on its starting time si and
deteriorating date di , and can be defined as a step-function: pi ai if si ≤ di ; pi ai
bi , otherwise. Without loss of generality, it is assumed that parameters ai , bi , and di are all
integers. Let Ci denote the completion time of job Ji . The goal of the problem is to find a
schedule such that the total completion time, or the sum of completion times, of all jobs
is minimized. This problem is denoted as P m/pi ai or ai bi , di / Ci by adopting the
standard three-field notation.
For convenience, a job is called early if its starting time is before or at its deteriorating
date; tardy, otherwise. For schedule S, the objective value is denoted by ZS. In addition,
let M be a sufficiently large positive number. Before formulating the proposed mixed integer
programming model, the following variables have to be defined.
Based on the definitions and notation described above, the considered problem can
now be formulated as 0-1 mixed integer programming model, as shown below.
n
Minimize ZS Ci . 3.1
i1
Mathematical Problems in Engineering 5
Subject to:
ai , si ≤ di
pi ∀i 1, . . . , n, 3.2
ai bi , otherwise
n
x0ik 1 ∀k 1, . . . , m, 3.3
i1
n
xin1k 1 ∀k 1, . . . , m, 3.4
i1
n
xijk yjk ∀j 1, . . . , n, ∀k 1, . . . , m, 3.5
j
i0,i /
n1
xijk yik ∀i 1, . . . , n, ∀k 1, . . . , m, 3.6
j
j1,i /
Ci ≥ pi Mx0ik − 1 ∀i 1, . . . , n, ∀k 1, . . . , m, 3.7
Cj ≥ Ci pj M xijk − 1 ∀i 1, . . . , n, ∀j 1, . . . , n, ∀k 1, . . . , m, 3.8
si ≥ Mx0ik − 1 ∀i 1, . . . , n, ∀k 1, . . . , m, 3.9
sj ≥ Ci M xijk − 1 ∀i 1, . . . , n, ∀j 1, . . . , n, ∀k 1, . . . , m, 3.10
m
yik 1, ∀i 1, . . . , n, 3.11
k1
In the above mathematical model, the objective 3.1 minimizes the total completion
time. Constraints 3.2 depict the processing time of the step-deteriorating jobs. Constraints
3.3 and 3.4 ensure that only one job can be processed at the first and the last position
on each machine. Constraints 3.5 and 3.6 guarantee that each job is scheduled only once
and processed by at most one machine. Constraints 3.7 define the completion time of the
first job assigned to a machine. Constraints 3.8 represent that the completion time of a job
in sequence on each machine will be at least equal to the sum of the completion time of the
preceding job and the processing time of the present job, if the job is immediately schedule
after the previous job. Constraints 3.9 define the starting time of the first job assigned to a
machine. Constraints 3.10 state that the starting time of a job in sequence on each machine is
greater than or equal to the completion time of the preceding job. Constraints 3.11 confirm
that each job is only processed to exactly one machine. Constraints 3.12 specify that the
decision variable x and y is binary over all domains.
This described problem is NP-complete because the single machine problem 1/pi ai
or ai bi , di / Ci has proven to be NP-complete by 24. There exists no polynomial time
algorithm for the exact solution of the considered problem. Therefore, we have to look for
some efficient heuristic or meta-heuristic algorithms to yield near optimal solutions of large-
sized problems in a reasonable computational time.
6 Mathematical Problems in Engineering
Property 1. On each machine of an optimal solution for the considered problem, early jobs
and tardy jobs are sequenced in the nondecreasing order of ai and ai bi , respectively.
Proof. The proof is straightforward from the SPT rule on ai and ai bi , respectively.
Procedure: the modified WCSA for the parallel machine scheduling problem
Inputs: n, m, ai , di , bi for i 1,. . ., n
Output: the near optimal scheme S opt and the associated total completion time TC
Begin
let N0 {J1 , J2 , . . . , Jn } be the sequence that sorts all the jobs by ascending order of their
normal time ai ,
set TC infinity, and S optk Φ for k 1, . . . , m % initialize the T C and S opt
set var max{2, n/m} % initialize the ranges of l1 and l2
for l1 1; l1 ≤ var; l1
ω1 0.4 0.75 − 0.4 × l1 − 1/var − 1
for l2 1; l2 ≤ var; l2
ω2 0.2 0.5 − 0.2 × l2 − 1/var − 1
ω3 1 − ω1 − ω2
set N N0
set S Φ for k 1, . . . , m % initialize the scheme S
set Cmac k 0, for k 1, . . . , m % initialize the completion
time of all machines
set Cjob i 0, for i 1, . . . , n % initialize the completion
time of all jobs
for i 1; i ≤ m; i
select the machine h that has the least completion time in all machines
select the job Jh that has the least normal time from N
Sh Sh ∪ {Jh }
Cmac h Cmac h aJh
Cjob Jh Cmac h
delete job Jh from N
end for
for j m 1; j ≤ n; j
select the machine f that has the least completion time in all machines
ifCmac f > max{dJr , Jr ∈ N} % tardy jobs are sequenced in the
nondecreasing order of ai bi
select the job Jf with the smallest aJf bJf from N
Sf Sf ∪ {Jf }
Cmac f Cmac f aJf bJf
Cjob Jf Cmac f
delete job Jf from N
else % arrange the jobs in a ascending
order of the weight of combination
set N {dJr ≥ Cmac f, Jr ∈ N}
select job Jf with the smallest ω1 × aJf ω2 × dJf − ω3 × bJf from N
Sf Sf ∪ {Jf }
Cmac f Cmac f aJf
Cjob Jf Cmac f
delete job Jf from N
end if
end for
T C temp sum{Cjob i 0, for i 1, . . . , n}
if T C temp < T C % update the scheme S opt and the
assisted completion time T C
T C T C temp
S opt S
end if
end for
end for
End
Initial solution 1 4 5 7 2 6 3 8
Machine 1 1 5 2 6 3
Machine 2 4 7 8
to conceptual simplicity to understand and implement and the high flexibility and brilliant
adaptability to different problems. Consequently, we intend to use it to solve our problem.
Since the VNS was proposed in 1997, some variants of VNS are developed by Hansen
and Mladenović 25. Variable neighborhood descent VND, as one of these variants of VNS,
performs change of neighborhoods in a deterministic way and adopts the first improvement
rather than the best improvement. In the subsequent subsection, we are going to employ a
VNS with five powerful NSSs based on insertion, swap, and inversion under the framework
of VND.
Job number
Parameter
J1 J2 J3 J4 J5 J6 J7 J8
ai 10 13 28 55 63 81 90 95
di 14 48 21 52 55 60 36 7
bi 14 3 5 41 35 16 47 2
Procedure: NSS1
improvement yes
while improvement yes do
improvement no
i1
while i ≤ n do n is the number of jobs
select randomly one job p without repletion from the current solution x
x swap job i and job p on their positions.
if Zx < Zx do
xx
in
improvement yes
end if
ii1
end while
end while
on insertion or swap operations. In addition, the inversion operation has been reported
to be significantly surpassing all the modified forms of crossover of genetic algorithm
especially tailored to deal with combinational problems 48. Therefore, we define five types
of neighborhood search structures based on insertion, swap, and inversion.
To generate a neighboring solution, NSS1 makes some slight changes in the candidate
solution by swapping the positions of all jobs in the sequence with one selected job at
random and without repetition. All jobs are selected one after another without repetition
in a random order. If we observe the first improvement, the associated sequence is accepted
and the procedure restarts. According to the NSS1 , some available neighboring searches are
found and used to improve the candidate solution. The whole procedure repeats so long
as no improvement is obtained through swap all jobs with another randomly selected job.
The procedure of NSS1 is described in Algorithm 2 in detail. After implementing all swap
operations, we believe that there is little hope for further improvement just by swapping
two jobs on their positions. Hence, it necessitates consider another neighborhood structure to
escape from this local optimum of the NSS1 . Subsequently, we need to introduce another NSS
to search potential improvement based on insertion moves.
In our NSS2 , the procedure is similar to the NSS1 . The only difference between
them is their move pattern that NSS2 changes the position of one job based on insertion
neighborhood. A job is removed from the sequence at random and without repetition, and
then relocated another random selected position. The other procedure is the same as the
10 Mathematical Problems in Engineering
Procedure: NSS3
i1
while i ≤ n do
j j 1
while j ≤ n do
select randomly two jobs p and q from the current solution x
x swap the two selected jobs on their positions
if Zx < Zx do
xx
jn
in
end if
j j 1
end while
ii1
end while
NSS1 . To avoid duplicated description, it is unnecessary to go into these details here. After
relocating all jobs, we feel that the algorithm cannot rely solely on single move to improve
the quality of the current solution. Hence, we need to introduce other neighborhood search
structure to generate more complex neighbors than the above two structures.
In the NSS3 , the number of randomly selected jobs is 2. The manner of choosing these 2
jobs is all the combinations of two-out-of-n jobs. Swap the selected jobs on their positions and
observe if the solution is improved. Once observing the first improvement, the associated
solution is accepted and the procedure restarts. If not, this search strategy repeats for the
subsequent combinations. Algorithm 3 illustrated the whole procedure of the NSS3 . By the
same way of switching from NSS1 to NSS2 , the double insertions are employed in the fourth
neighborhood structure NSS4 . In the NSS4 , the number of removed jobs is set equal to 2.
To relocate them, the two jobs are reinserted into two new randomly selected positions. The
other procedure is the same to the NSS3 .
To improve further the search performance of the proposed VNS, an inversion
operator is embedded into the NSS5 when the foregoing neighborhood search methods
terminate. In the NSS5 , it randomly selects two positions, known as the points of inversion,
and inverts the sequence between these positions. The inversion procedure is repeated
ϕ times for producing a new job sequence. Owing to the impact of the parameter ϕ on
neighborhood search, it must be tuned. The inversion operation is unidirectional, that is,
only the inverted sequence is improved, the corresponding solution is accepted to replace the
incumbent one. The procedure of is described in Algorithm 4.
The general outline of the proposed VNS is shown in Algorithm 5. According to using
the above neighborhood search structures, the proposed VNS can be quick to yield the near
optimal solution. If the neighborhood search structure continues to increase the number of
selected jobs, the quality of the algorithm will be not improved according to our primary
experiments.
Mathematical Problems in Engineering 11
Procedure: NSS5
for i 1 to ϕ do
select two jobs p and q at random from the current solution x
x invert partial jobs between job p and job q
if Zx < Zx do
xx
end if
end for
5. Computational Analysis
To test the performance of our approaches, computational experiments were carried out
for the P m/pi = ai or ai bi , di / Ci problem. More specifically, we also carried out
comprehensive computational and statistical tests to assess the performance of MWCSA,
and VNS. Firstly, we gave a detailed description of the different instance sets that we have
employed. After reporting the results, we carried out comprehensive statistical analyses in
order to soundly test the significance of the reported results.
Because there is not a widely available set of benchmark instances for the problem
under study, data for the test instances are generated randomly. The normal processing
times ai are randomly generated from an integer uniform distribution on U1, 100. The
deterioration penalties bi are randomly picked from an integer uniform distribution on
f×n
U1, 100 × β, where β 0.5. Let Df i1 ai /m, 0 ≤ f ≤ 1. The deteriorating dates di are
randomly selected from three intervals U0, D0.5 , UD0.5 , D, and U0, D. In this paper, we
have divided the benchmark instances into 2 different types according to the number of jobs.
For small-sized instances, we test all combinations of n {6, 8, 10} and m {2, 3}. For large-
sized instances, we use the combinations of n {20, 40, 60, 80, 100} and m {2, 4, 6, 8, 10} to
evaluate the proposed algorithms. Since the deteriorating date may be generated from three
different intervals, we totally have 6 × 3 25 × 3 93 instances.
12 Mathematical Problems in Engineering
In order to validate the performance of the proposed approaches, attempts are made
to solve the MIP model presented in Section 3 using the software IBM-ILOG CPLEX 12.3
solvers. The software is run on a PC with Intel Pentium Dual-Core 2.60 GHz CPU and 2 GB
RAM. Because the problem under consideration is NP-complete, it is impossible to obtain
optimal solutions by using some polynomial time algorithms. Therefore, only small-sized
instances can be solved to optimality using CPLEX. For the small-sized instances, we use
relative percentage deviation RPD as a performance measure to compare those results of
the CPLEX software, the MWCSA and the VNS. RPD is obtained by given formula below:
Z Alg − ZOPL
RPD × 100, 5.1
ZOPL
where ZAlg is the total completion time of the solution obtained by a given algorithm
and instance and ZOPL is the total completion time of the optimal schedule given by
CPLEX. The SRF small rate first of Section 4 is also brought into the large-sized instances
for comparison. Our purpose to utilize SRF is to use its result as upper bound for a given
instance to evaluate the VNS. In addition, the MWCSA is used to obtain the initial solution
of the VNS for improving the quality of solutions. The hybrid algorithm, denoted by VNS
MWCSA, is employed for the large-sized instances.
All proposed algorithms in the previous sections were coded in MATLAB 7.11 and
implement on the same PC. According to the preliminary tests, the stopping criterion used
when testing all instances with the proposed algorithms is set to the maximum iterations
time fixed to 200. Thus, there is only one parameter ϕ is tuned in our proposed VNS.
The considered candidate levels are 10, 30, 50, and 70. A set of 15 large-sized instances are
randomly generated when the number of jobs is fixed to 100. All the 15 instances are solved
by these VNSs with different parameter candidate levels. The results were analyzed by the
one-way ANOVA test. The means plot and Fisher LSD intervals for the 4 levels of parameter
ϕ are shown in Figure 2. As we can see, ϕ of 50 and 70 provide the better results among
all levels. However, it needs more computational time when ϕ is 70. Therefore, the most
reasonable value of parameter ϕ is 50.
CPLEX, SRF, and MWCSA are deterministic and only one run is necessary. The VNS
and VNS MWCSA are stochastic and we need to run some replicates in order to better assess
the results. We run each algorithm five times. The listed objective value and computational
time are the means of five reported results.
Table 2 provides a comparison of the solutions to the mixed integer programming
model generated by the CPLEX and those solutions provided by execution of the proposed
MWCSA and VNS algorithms. All small-sized instances were solved to optimality by CPLEX
as shown in Table 1. The computational time of the CPLEX increases rapidly as the instance
become larger. When the number of jobs is equal to 10, solutions to MIP models require far
longer than 1 hour. From Table 1, the longest computational time of CPLEX is 144761.59
seconds 40.21 hours. It is impractical in scheduling for such a long time. In addition,
we found that CPLEX is not greatly affected by the distribution of the deteriorating dates
almost. While the deteriorating dates are randomly picked from the interval UD0.5 , D, all
corresponding instances are relatively easy. This is because the number of deteriorating jobs
is less. It is worthwhile to note that the MWCSA behaves well for all small-sized instances
and gives an average result of 0.62% deviation from the optimal solution. Specially, eleven
of these instances can be solved optimally by the proposed heuristic. Its execution time is far
Mathematical Problems in Engineering 13
Table 2: Comparison of CPLEX results with the proposed algorithms for small-sized instances.
0.9
0.88
0.86
0.84
0.82
Mean
0.8
0.78
0.76
0.74
0.72
10 30 50 70
ϕ
Figure 2: Means plot and Fisher LSD intervals for the different levels of parameter ϕ the significance level
α is 0.05.
14 Mathematical Problems in Engineering
Table 3: Test results of the algorithms on the instances with the interval U0, D0.5 .
less than 1 second so that its log is not necessary. It can also be noticed that the solutions from
the VNS algorithm are optimal for these small-sized instances. Meanwhile, the computational
time of these instances given by VNS are not more than 1 second. This means that the VNS
algorithm is effective in solving the scheduling problem under study and has significantly
better optimization performance than the others.
Tables 3, 4, and 5 show the results of large-sized instances where the deteriorating
dates are generated from three different intervals. In order to compare the performance of
the proposed algorithms, the ratios of the results of a given algorithm to the values obtained
by SRF were calculated out. For the instances with the deteriorating dates generated from
interval U0, D0.5 , VNS behaves slightly better than MWCSA. The difference in average
rate between two algorithms is only 2.26%. For other two types of instances with intervals
UD0.5 , D and U0, D, the VNS and VNS MWCSA strikingly outperformed the MWCSA
heuristic because of its larger search space. As a rule of thumb, a given algorithm takes
possession of the better initial solution and should yield better result. However, the hybrid
Mathematical Problems in Engineering 15
Table 4: Test results of the algorithms on the instances with the interval UD0.5 , D.
algorithm, with longer run times, does not perform noticeably better than the VNS. It is
chiefly because the VNS carries out enough neighborhood searches before the stopping
criterion meets. The rates of three different types of instances were plotted into smooth
curves shown in Figures 3, 4, and 5. It can be seen from those figures that the performance
of MWCSA heuristic is not particularly robust as regards the distribution of the deteriorating
date. Alternatively, VNS and VNS MWCSA are statistically better than MWCSA for three
types of instances.
6. Conclusions
This paper considers the identical parallel machines scheduling problem with step-
deteriorating jobs. The processing time of each job is a step function of its starting time and
16 Mathematical Problems in Engineering
Table 5: Test results of the algorithms on the instances with the interval U0, D.
a deteriorating date that is individual to all jobs. The problem is to determine the allocation
of jobs to machines as well as the sequence of the jobs assigned to each machine for the
criteria of minimizing the total completion time. A mathematical model for this problem
has been formulated. Since the problem under study is NP-complete, it is impossible to
solve large-sized instances to optimality. To solve the tackled problem, a heuristic MWCSA
and a VNS are proposed to obtain the near optimal solutions. In order to further improve
the quality of solution, the heuristic MWCSA has been hybridized with the VNS algorithm
and implemented to provide a good initial solution. Numerical experiments are conducted
on small- and large-sized instances. Computational results show that MWCSA produces
some good solutions compared to CPLEX, but the performance is greatly affected by the
distribution of the deteriorating date. In contrast, VNS and VNS WMCSA are robust as
regards three types of instances. Therefore, fairly good solutions can be obtained by the
proposed methods within reasonable amount of time.
Mathematical Problems in Engineering 17
0.95
0.9
Rate
0.85
0.8
0.75
0.7
0 5 10 15 20 25 30
Problem number
MWCSA
VNS
VNS + MWCSA
Figure 3: Comparison of rates between the three algorithms for the case with the interval U0, D0.5 .
1.2
1.15
1.1
1.05
Rate
0.95
0.9
0.85
0.8
0 5 10 15 20 25 30
Problem number
MWCSA
VNS
VNS + MWCSA
Figure 4: Comparison of rates between the three algorithms for the case with the interval UD0.5 , D.
For further study, it is worth considering the setup times between jobs for the
problem of scheduling jobs with piecewise-deterioration on multimachine. In addition, other
efficient constructive heuristics and neighborhood properties are worthwhile to investigate
for improving the quality of solutions.
18 Mathematical Problems in Engineering
1.1
1.05
0.95
Rate
0.9
0.85
0.8
0.75
0.7
0 5 10 15 20 25 30
Problem number
MWCSA
VNS
VNS + MWCSA
Figure 5: Comparison of rates between the three algorithms for the case with the interval U0, D.
Acknowledgments
This work is partially supported by the National Natural Science Foundation of China under
Grant no. 51175442, the Youth Foundation for Humanities and Social Sciences of Ministry
of Education of China under Grant no. 12YJCZH296, the Ph.D. Programs Foundation
of Ministry of Education of China under Grant no. 200806131014, and the Fundamental
Research Funds for the Central Universities under Grant nos. SWJTU09CX022, 2010ZT03.
References
1 S. Browne and U. Yechiali, “Scheduling deteriorating jobs on a single processor,” Operations Research,
vol. 38, no. 3, pp. 495–498, 1990.
2 J. N. D. Gupta and S. K. Gupta, “Single facility scheduling with nonlinear processing times,”
Computers and Industrial Engineering, vol. 14, no. 4, pp. 387–393, 1988.
3 A. Bachman and A. Janiak, “Minimizing maximum lateness under linear deterioration,” European
Journal of Operational Research, vol. 126, no. 3, pp. 557–566, 2000.
4 A. Bachman, A. Janiak, and M. Y. Kovalyov, “Minimizing the total weighted completion time of
deteriorating jobs,” Information Processing Letters, vol. 81, no. 2, pp. 81–84, 2002.
5 C.-C. Wu, W.-C. Lee, and Y.-R. Shiau, “Minimizing the total weighted completion time on a single
machine under linear deterioration,” International Journal of Advanced Manufacturing Technology, vol.
33, no. 11-12, pp. 1237–1243, 2007.
6 J.-B. Wang, L. Lin, and F. Shan, “Single-machine group scheduling problems with deteriorating jobs,”
International Journal of Advanced Manufacturing Technology, vol. 39, no. 7-8, pp. 808–812, 2008.
7 J.-B. Wang, C. T. Ng, and T. C. E. Cheng, “Single-machine scheduling with deteriorating jobs under a
series-parallel graph constraint,” Computers & Operations Research, vol. 35, no. 8, pp. 2684–2693, 2008.
8 M. M. Mazdeh, F. Zaerpour, A. Zareei, and A. Hajinezhad, “Parallel machines scheduling to minimize
job tardiness and machine deteriorating cost with deteriorating jobs,” Applied Mathematical Modelling,
vol. 34, no. 6, pp. 1498–1510, 2010.
Mathematical Problems in Engineering 19
9 L. Sun, L. Sun, K. Cui, and J. B. Wang, “A note on flow shop scheduling problems with deteriorating
jobs on no-idle dominant machines,” European Journal of Operational Research, vol. 200, no. 1, pp. 309–
311, 2010.
10 D. Wang and J. B. Wang, “Single-machine scheduling with simple linear deterioration to minimize
earliness penalties,” International Journal of Advanced Manufacturing Technology, vol. 46, no. 1–4, pp.
285–290, 2010.
11 J.-B. Wang, “Flow shop scheduling with deteriorating jobs under dominating machines to minimize
makespan,” International Journal of Advanced Manufacturing Technology, vol. 48, no. 5–8, pp. 719–723,
2010.
12 C.-M. Wei and J.-B. Wang, “Single machine quadratic penalty function scheduling with deteriorating
jobs and group technology,” Applied Mathematical Modelling, vol. 34, no. 11, pp. 3642–3647, 2010.
13 X. Huang and M.-Z. Wang, “Parallel identical machines scheduling with deteriorating jobs and total
absolute differences penalties,” Applied Mathematical Modelling, vol. 35, no. 3, pp. 1349–1353, 2011.
14 J.-B. Wang and C. Wang, “Single-machine due-window assignment problem with learning effect and
deteriorating jobs,” Applied Mathematical Modelling, vol. 35, no. 8, pp. 4017–4022, 2011.
15 J.-B. Wang, J.-J. Wang, and P. Ji, “Scheduling jobs with chain precedence constraints and deteriorating
jobs,” Journal of the Operational Research Society, vol. 62, no. 9, pp. 1765–1770, 2011.
16 J.-B. Wang and C.-M. Wei, “Parallel machine scheduling with a deteriorating maintenance activity
and total absolute differences penalties,” Applied Mathematics and Computation, vol. 217, no. 20, pp.
8093–8099, 2011.
17 S.-H. Yang and J.-B. Wang, “Minimizing total weighted completion time in a two-machine flow shop
scheduling under simple linear deterioration,” Applied Mathematics and Computation, vol. 217, no. 9,
pp. 4819–4826, 2011.
18 J.-B. Wang and M.-Z. Wang, “Single-machine scheduling with nonlinear deterioration,” Optimization
Letters, vol. 6, no. 1, pp. 87–98, 2012.
19 J.-B. Wang, M.-Z. Wang, and P. Ji, “Single machine total completion time minimization scheduling
with a time-dependent learning effect and deteriorating jobs,” International Journal of Systems Science,
vol. 43, no. 5, pp. 861–868, 2012.
20 B. Alidaee and N. K. Womer, “Scheduling with time dependent processing times: review and
extensions,” Journal of the Operational Research Society, vol. 50, no. 7, pp. 711–720, 1999.
21 T. C. E. Cheng, Q. Ding, and B. M. T. Lin, “A concise survey of scheduling with time-dependent
processing times,” European Journal of Operational Research, vol. 152, no. 1, pp. 1–13, 2004.
22 P. S. Sundararaghavan and A. S. Kunnathur, “Single machine scheduling with start time dependent
processing times: some solvable cases,” European Journal of Operational Research, vol. 78, no. 3, pp.
394–403, 1994.
23 W. Kubiak and S. van de Velde, “Scheduling deteriorating jobs to minimize makespan,” Naval Research
Logistics, vol. 45, no. 5, pp. 511–523, 1998.
24 T. C. E. Cheng and Q. Ding, “Single machine scheduling with step-deteriorating processing times,”
European Journal of Operational Research, vol. 134, no. 3, pp. 623–630, 2001.
25 P. Hansen and N. Mladenović, “Variable neighborhood search: principles and applications,” European
Journal of Operational Research, vol. 130, no. 3, pp. 449–467, 2001.
26 N. Mladenović and P. Hansen, “Variable neighborhood search,” Computers & Operations Research, vol.
24, no. 11, pp. 1097–1100, 1997.
27 A. S. Kunnathur and S. K. Gupta, “Minimizing the makespan with late start penalties added to
processing times in a single facility scheduling problem,” European Journal of Operational Research,
vol. 47, no. 1, pp. 56–64, 1990.
28 G. Mosheiov, “Scheduling jobs with step-deterioration; Minimizing makespan on a single- and multi-
machine,” Computers and Industrial Engineering, vol. 28, no. 4, pp. 869–879, 1995.
29 A. A. K. Jeng and B. M. T. Lin, “Makespan minimization in single-machine scheduling with step-
deterioration of processing times,” Journal of the Operational Research Society, vol. 55, no. 3, pp. 247–256,
2004.
30 A. A. K. Jeng and B. M. T. Lin, “Minimizing the total completion time in single-machine scheduling
with step-deteriorating jobs,” Computers and Operations Research, vol. 32, no. 3, pp. 521–536, 2005.
31 C.-C. He, C.-C. Wu, and W.-C. Lee, “Branch-and-bound and weight-combination search algorithms
20 Mathematical Problems in Engineering
for the total completion time problem with step-deteriorating jobs,” Journal of the Operational Research
Society, vol. 60, no. 12, pp. 1759–1766, 2009.
32 J. Layegh, F. Jolai, and M. S. Amalnik, “A memetic algorithm for minimizing the total weighted
completion time on a single machine under step-deterioration,” Advances in Engineering Software, vol.
40, no. 10, pp. 1074–1077, 2009.
33 M. Y. Kovalyov and W. Kubiak, “A fully polynomial approximation scheme for minimizing makespan
of deteriorating jobs,” Journal of Heuristics, vol. 3, no. 4, pp. 287–297, 1998.
34 T. C. E. Cheng, Q. Ding, M. Y. Kovalyov, A. Bachman, and A. Janiak, “Scheduling jobs with piecewise
linear decreasing processing times,” Naval Research Logistics, vol. 50, no. 6, pp. 531–554, 2003.
35 G. Moslehi and A. Jafari, “Minimizing the number of tardy jobs under piecewise-linear deterioration,”
Computers and Industrial Engineering, vol. 59, no. 4, pp. 573–584, 2010.
36 S. R. Gupta and J. S. Smith, “Algorithms for single machine total tardiness scheduling with sequence
dependent setups,” European Journal of Operational Research, vol. 175, no. 2, pp. 722–739, 2006.
37 M. R. D. Paula, M. G. Ravetti, G. R. Mateus, and P. M. Pardalos, “Solving parallel machines scheduling
problems with sequence-dependent setup times using variable neighbourhood search,” IMA Journal
of Management Mathematics, vol. 18, no. 2, pp. 101–115, 2007.
38 D. Anghinolfi and M. Paolucci, “Parallel machine total tardiness scheduling with a new hybrid
metaheuristic approach,” Computers & Operations Research, vol. 34, no. 11, pp. 3471–3490, 2007.
39 R. Driessel and L. Mönch, “Variable neighborhood search approaches for scheduling jobs on parallel
machines with sequence-dependent setup times, precedence constraints, and ready times,” Computers
and Industrial Engineering, vol. 61, no. 2, pp. 336–345, 2011.
40 J. Behnamian, M. Zandieh, and S. M. T. F. Ghomi, “Parallel-machine scheduling problems with
sequence-dependent setup times using an ACO, SA and VNS hybrid algorithm,” Expert Systems with
Applications, vol. 36, no. 6, pp. 9637–9644, 2009.
41 C.-L. Chen and C.-L. Chen, “Hybrid metaheuristics for unrelated parallel machine scheduling with
sequence-dependent setup times,” International Journal of Advanced Manufacturing Technology, vol. 43,
no. 1-2, pp. 161–169, 2009.
42 J. Gao, L. Sun, and M. Gen, “A hybrid genetic and variable neighborhood descent algorithm for
flexible job shop scheduling problems,” Computers & Operations Research, vol. 35, no. 9, pp. 2892–2907,
2008.
43 V. Roshanaei, B. Naderi, F. Jolaib, and M. Khalili, “A variable neighborhood search for job shop
scheduling with set-up times to minimize makespan,” Future Generation Computer Systems, vol. 25,
no. 6, pp. 654–661, 2009.
44 M. A. Adibi, M. Zandieh, and M. Amiri, “Multi-objective scheduling of dynamic job shop using
variable neighborhood search,” Expert Systems with Applications, vol. 37, no. 1, pp. 282–287, 2010.
45 M. Amiri, M. Zandieh, M. Yazdani, and A. Bagheri, “A variable neighbourhood search algorithm for
the flexible job-shop scheduling problem,” International Journal of Production Research, vol. 48, no. 19,
pp. 5671–5689, 2010.
46 P. Hansen, N. Mladenović, and J. A. M. Pérez, “Variable neighbourhood search: methods and
applications,” Annals of Operations Research, vol. 175, pp. 367–407, 2010.
47 M. L. Pinedo, Scheduling: Theory, Algorithms and Systems, Springer, New York, NY, USA, 2008.
48 C. Ferreira, “Combinatorial optimization by gene expression programming: inversion revisited,” in
Proceedings of the Argentine Symposium on Artificial Intelligence, pp. 160–174, Santa Fe, NM, USA, 2002.
Hindawi Publishing Corporation
Mathematical Problems in Engineering
Volume 2012, Article ID 871741, 16 pages
doi:10.1155/2012/871741
Research Article
Global Sufficient Optimality Conditions for
a Special Cubic Minimization Problem
Copyright q 2012 Xiaomei Zhang et al. This is an open access article distributed under the
Creative Commons Attribution License, which permits unrestricted use, distribution, and
reproduction in any medium, provided the original work is properly cited.
We present some sufficient global optimality conditions for a special cubic minimization problem
with box constraints or binary constraints by extending the global subdifferential approach
proposed by V. Jeyakumar et al. 2006. The present conditions generalize the results developed in
the work of V. Jeyakumar et al. where a quadratic minimization problem with box constraints or
binary constraints was considered. In addition, a special diagonal matrix is constructed, which is
used to provide a convenient method for justifying the proposed sufficient conditions. Then, the
reformulation of the sufficient conditions follows. It is worth noting that this reformulation is also
applicable to the quadratic minimization problem with box or binary constraints considered in the
works of V. Jeyakumar et al. 2006 and Y. Wang et al. 2010. Finally some examples demonstrate
that our optimality conditions can effectively be used for identifying global minimizers of the
certain nonconvex cubic minimization problem.
1. Introduction
Consider the following cubic minimization problem with box constraints:
1
min fx bT x3 xT Ax aT x,
2
n CP 1
s.t. x∈D ui , vi ,
i1
1 A−Q0, Qdiagα ,...,α
∂L fx bT x3 xT Qx βT xβAQxa, αi ∈R 1 n . 2.3
2
Proof. Suppose that there exists a diagonal matrix Q diagα1 , . . . , αn , such that A − Q 0.
Let
1
lx bT x3 xT Qx βT x, β A − Qx a. 2.4
2
1 T T
φx fx − lx x A − Qx a − β x. 2.5
2
Since ∇2 φx A − Q 0, for all x ∈ Rn , we know that φx is a convex function on Rn . Note
that ∇φx A−Qxa−β 0, and so x is a global minimizer of φx, that is, φx ≥ φx,
for all x ∈ Rn . This means that lx ∈ ∂L fx.
Next we prove the converse.
Let lx ∈ ∂L fx, lx bT x3 1/2xT Qx βT x. By definition,
Hence
1 T T
φx fx − lx x A − Qx a − β x ≥ fx − lx, ∀x ∈ Rn . 2.7
2
Thus, x is a global minimizer of φx. So, ∇φx 0 and ∇2 φx 0, that is,
A − Q 0, A − Qx a − β 0, 2.8
hence β A − Qx a.
⎧
⎪
⎨−1
⎪ if xi ui ,
xi 1 if xi vi ,
⎪
⎪
⎩Ax a if xi ∈ ui , vi . 2.9
i i
diagx1 , . . . , xn .
X
4 Mathematical Problems in Engineering
αi min{0, αi },
2.10
Q diagα1 , . . . , αn .
By Proposition 2.2, we obtain the following sufficient global optimality condition for
CP 1.
Theorem 2.3. For CP 1, let x x1 , . . . , xn T ∈ D and u u1 , . . . , un T , v v1 , . . . , vn T .
Suppose that there exists a diagonal matrix Q diagα1 , . . . , αn , αi ∈ R, i 1, . . . , n, such that
A − Q 0, and for all x ∈ D, bi xi − xi ≥ 0, i 1, . . . , n. If
1
XAx a − Qv − u ≤ 0, 2.11
2
1
lx bT x3 xT Qx βT x,
2 2.12
β A − Qx a.
Obviously if lx − lx ≥ 0 for each x ∈ D, then x is a global minimizer of CP 1.
Note that
n
αi
lx − lx xi − xi 2 Ax aT x − x bT x3 − x3 . 2.14
i1
2
αi
xi − xi 2 Ax ai xi − xi bi xi3 − x3i ≥ 0, i 1, . . . , n, xi ∈ ui , vi , 2.15
2
then, from 2.14, it holds that lx − lx ≥ 0. So x is a global minimizer of lx over box
constraints.
On the other hand, suppose that x is a global minimizer of lx, x ∈ D. Then it holds
that
n
αi
lx − lx xi − xi 2 Ax aT x − x bT x3 − x3 ≥ 0, ∀x ∈ D. 2.16
i1
2
Mathematical Problems in Engineering 5
we still have
αi
− lx
lx xi − xi 2 Ax aT x − x bT x3 − x3 ≥ 0, i 1, . . . , n. 2.18
2
This means that if x is a global minimizer of lx over box constraints. Then 2.15 holds.
Combining the above discussion, we can conclude that x is a global minimizer of lx
over box constraints if and only if 2.15 holds. So next, we just need to prove 2.15 in order
to show that x is a global minimizer of lx.
We first see from 2.11, for each i 1, . . . , n, that
αi
− vi − ui xi Ax ai ≤ 0. 2.19
2
αi
− xi − ui xi Ax ai ≤ 0, 2.20
2
and
αi
xi − vi xi Ax ai ≤ 0. 2.21
2
αi
− xi − ui Ax a2i ≤ 0. 2.22
2
αi
xi − xi 2 Ax ai xi − xi bi xi3 − x3i
2
αi
≥ xi − xi 2 bi xi3 − x3i
2 2.23
bi xi − xi xi2 xi xi x2i
≥ 0.
6 Mathematical Problems in Engineering
αi
xi − ui Ax ai ≥ 0. 2.24
2
So we have
αi
xi − xi 2 Ax ai xi − xi bi xi3 − x3i
2
αi
≥ xi − ui 2 Ax ai xi − ui bi xi3 − u3i
2
2.25
αi
xi − ui Ax ai xi − ui bi xi3 − u3i
2
≥ 0.
αi
xi − vi Ax ai ≤ 0. 2.26
2
Then
αi
xi − xi 2 Ax ai xi − xi bi xi3 − x3i
2
αi
≥ xi − vi 2 Ax ai xi − vi bi xi3 − vi3 2.27
2
≥ 0.
So, if condition 2.11 holds, then 2.15 holds. And, from 2.14, we can conclude that
x is a global minimizer of CP 1.
Theorem 2.3 shows that the existence of diagonal matrix Q plays a crucial role because
if this diagonal matrix Q does not exist, then we have no way to use this theorem. If the
diagonal matrix Q exists, then the key problem is how to find it. These questions also exist in
8, 12 .
The following corollary will answer the questions above.
Corollary 2.4. For CP 1, let x ∈ D. Assume that, for all x ∈ D, it holds that bi xi − xi ≥ 0 i
1, . . . , n. Then one has the following conclusion.
XAx a ≤ 0, 2.28
2 When A is not a positive semidefinite matrix, if there exists an index i0 , 1 ≤ i0 ≤ n, such
that
then there is no such diagonal matrix Q that meets the requirements of the Theorem 2.3.
3 Let
When A is not a positive semidefinite matrix and the condition XAx a ≤ 0 holds, if
A − Q 0 holds, then x is a global minimizer of CP 1. Otherwise, one can conclude that there is no
such diagonal matrix Q that meets the requirements of the Theorem 2.3.
Proof. 1 Suppose that A 0 and the condition XAx a ≤ 0 holds. Choosing Q Q 0,
by Theorem 2.3, we can conclude that x is a global minimizer of CP 1.
2 When A is not a positive semidefinite matrix, if there exists an index i0 , 1 ≤ i0 ≤ n,
such that
then
Suppose there exists a diagonal matrix Q that meets all conditions in Theorem 2.3.
Condition 2.11 can be rewritten in the following form:
1
xi Ax ai − αi vi − ui ≤ 0, i 1, . . . , n. 2.33
2
3 Next we will consider the case that A is not a positive semidefinite matrix, and
condition XAx a ≤ 0 holds.
We construct a diagonal matrix Q diagα1 , . . . , αn where αi 2xi Ax ai /vi −
ui , i 1, . . . , n, and A − Q 0. Then it suffices to show that condition 2.11 in Theorem 2.3
hold.
Note that xi Ax ai ≤ 0, i 1, . . . , n, and then
Since αi ≤ 0, we have αi αi . So
1
xi Ax ai − αi vi − ui 0, i 1, . . . , n. 2.38
2
Apparently this means that the constructed diagonal matrix Q also satisfies condition 2.11.
According to Theorem 2.3, we can conclude that x is a global minimizer of CP 1.
If the constructed diagonal matrix Q does not meet the condition A − Q 0, then
we can conclude that there is no such diagonal matrix Q that can meet the requirements of
Theorem 2.3.
To show this, suppose that there exists a diagonal matrix Q∗ diagα∗1 , . . . , α∗n , which
satisfies A − Q∗ 0 and 2.11.
From 2.11, we have
n
n
ri
n
min fx bi xi3 xi2 ai xi
i1 i1
2 i1
CP 2
n
s.t. x∈D ui , vi ,
i1
where ui , vi , ai , bi , ri ∈ R and ui ≤ vi , i 1, 2, . . . , n.
Mathematical Problems in Engineering 9
1
xi ri xi ai − ri vi − ui ≤ 0,
2 2.40
bi xi − xi ≥ 0,
Proof. For CP 2, choose Q A diagr1 , . . . , rn . If 2.40 holds, then, by Theorem 2.3, x is a
global minimizer of CP 2.
1
min fx bT x3 xT Ax aT x
2
n CP 3
s.t. x ∈ DB {ui , vi },
i1
Theorem 3.1. For CP 3, let x x1 , . . . , xn T ∈ DB . Suppose that there exists a diagonal matrix
Q diagα1 , . . . , αn , αi ∈ R, i 1, . . . , n such that A − Q 0, and for all x ∈ DB , bi xi − xi ≥
0 i 1, . . . , n. If
1
XAx a − Qv − u ≤ 0, 3.1
2
1
lx bT x3 xT Qx βT x,
2 3.2
β A − Qx a.
Then
Obviously if lx − lx ≥ 0 for each x ∈ DB , then x is a global minimizer of CP 3.
10 Mathematical Problems in Engineering
Note that
n
αi
lx − lx xi − xi 2 Ax aT x − x bT x3 − x3 . 3.4
i1
2
Thus, x is a global minimizer of lx with binary constraints if and only if, for each
i 1, . . . , n, xi ∈ {ui , vi },
αi
xi − xi 2 Ax ai xi − xi bi xi3 − x3i ≥ 0. 3.5
2
αi
− vi − ui xi Ax ai ≤ 0. 3.6
2
αi
xi − ui 2 bi xi3 − u3i Ax ai xi − ui ≥ 0. 3.8
2
So 3.5 holds.
Case 2. If xi vi , then 3.6 is equivalent to
αi
− vi − ui Ax ai ≤ 0. 3.9
2
αi
vi − xi 2 − bi vi3 − xi3 − Ax ai vi − xi ≥ 0. 3.10
2
So 3.5 holds.
From 3.5, we can conclude that x is a global minimizer of problem CP 3
Corollary 3.2. For CP 3, let x ∈ DB . Suppose that, for all x ∈ DB , bi xi − xi ≥ 0 i 1, . . . , n.
(1) When A is a positive semidefinite matrix, if
XAx a ≤ 0, 3.11
(2) Let
When A is not a positive semidefinite matrix, if A−Q 0 holds, then x is a global minimizer of CP 3.
Otherwise, one can conclude that there is no such diagonal matrix Q that meets the requirements of
Theorem 3.1.
1
xi Ax ai − αi vi − ui 0, i 1, . . . , n. 3.14
2
It obviously means that the diagonal matrix Q also satisfies condition 3.1. According to
Theorem 3.1, x is a global minimizer of CP 3.
Note that there is difference between formula 3.1 and formula 2.11. In formula
3.1, the diagonal elements αi of a diagonal matrix Q are allowed to be positive or
nonpositive. But in formula 2.11, the diagonal elements αi of a diagonal matrix Q must meet
the conditions αi ≤ 0. So we have to discuss the sign of the terms xi Ax ai i 1, . . . , n in
Corollary 3.2.
We now consider a special case of CP 3:
n
n
ri
n
min fx bi xi3 xi2 ai xi
i1 i1
2 i1
CP 4
n
s.t. x ∈ DB {ui , vi },
i1
where bi , ri , ai , ui , vi ∈ R and ui ≤ vi , i 1, . . . , n.
Proof. For CP 4, choose Q A diagr1 , . . . , rn . If conditions 3.15 hold, then, by
Theorem 3.1, x is a global minimizer of CP 4.
4. Numerical Examples
In this section, six examples are given to test the proposed global sufficient optimality
condition.
7 3 3 1 3
min x 5x23 2x33 x12 x22 x32 2x1 x2 x1 x3 x2 x3 x1 5x2 3x3
3 1 2 2 2
4.1
3
s.t. x∈D 1, 2 .
i1
Let
⎡ ⎤
3 2 1
⎢ ⎥
⎢ ⎥
A ⎢2 2 1⎥ 4.2
⎣ ⎦
1 1 1
⎡ ⎤
−15
⎢ 2 ⎥
⎢ ⎥
⎢ ⎥
XAx a ⎢ −10 ⎥ < 0. 4.3
⎢ ⎥
⎣ ⎦
−6
1 3 1
min −x13 − 3x23 x43 − x12 − x22 x32 − x42 − 2x1 x2 x1 x4 2x2 x4 − x1 − 4x2 5x4
2 2 2
4.4
4
s.t. x∈D −1, 1 .
i1
Mathematical Problems in Engineering 13
Let
⎡ ⎤
−2 −2 0 1
⎢ ⎥
⎢−2 −2 0 2 ⎥
⎢ ⎥
⎢ ⎥
A⎢ ⎥ 4.5
⎢0 0 3 0⎥
⎢ ⎥
⎣ ⎦
1 2 0 −1
and a −1, −4, 0, 5T , b −1, −3, 0, 1/2T . Obviously A not is a positive semidefinite matrix.
Considering x 1, 1, 0, −1T , obviously we have, for each xi ∈ −1, 1 , bi xi − xi ≥
0 i 1, 2, 3, 4.
Note that Ax a −6, −10, 0, 9T . Then letting X diag1, 1, 0, −1, we have XAx
T
a −6, −10, 0, −9 ≤ 0. Let α1 2x1 Ax a1 /v1 − u1 −6, α2 −10, α3 0 and α4
−9. Then Q diag−6, −10, 0, −9, which satisfies A − Q 0. According to Corollary 2.43,
x 1, 1, 0, −1 is a global minimizer.
2 3 1
min −x13 x23 − 3x33 − 2x12 − x22 − 4x32 x42 − 2x1 5x2 − 3x3
3 5 2
4.6
4
s.t. x∈D −1, 1 .
i1
Let A diag−4, −6/5, −8, 1 and a −2, 5, −3, 0T , b −1, 2/3, −3, 0T , r
−4, −6/5, −8, 1T .
Consider x 1, −1, 1, 0T .
Let X diag1, −1, 1, 0, r −4, −6/5, −8, 0T . Then
1
x1 r1 x1 a1 − r1 v1 − u1 −2 < 0,
2
1
x2 r2 x2 a2 − r2 v2 − u2 −5 < 0,
2
1 4.7
x3 r3 x3 a3 − r3 v3 − u3 −3 < 0,
2
1
x4 r4 x4 a4 − r4 v4 − u4 0,
2
xi ∈ −1, 1 , bi xi − xi ≥ 0, i 1, 2, 3, 4.
5 8 3 3
min 4x13 x23 − 3x33 − x43 − x12 − x22 3x32 − x42 2x1 x2 x1 x4 x2 x3 − x3 x4
2 3 2 2
9
5x1 x2 − 2x3 − 2x4 4.8
2
4
s.t. x ∈ DB {−1, 1}.
i1
Let
⎡ ⎤
−3 2 0 1
⎢ ⎥
⎢2 −3 1 0⎥
⎢ ⎥
⎢ ⎥
A⎢ ⎥ 4.9
⎢0 1 6 −1⎥
⎢ ⎥
⎣ ⎦
1 0 −1 −2
and a 5, 9/2, −2, −2T , b 4, 5/2, −3, −8/3T . Obviously A not is a positive semidefinite
matrix.
Considering x −1, −1, 1, 1T , it follows that, for each x ∈ DB , bi xi − xi ≥
0 i 1, 2, 3, 4. Note that Ax a 7, 13/2, 2, −6T and X diag−1, −1, 1, 1. Let α1
2x1 Ax a1 /v1 − u1 −7. Similarly we have, α2 −13/2, α3 2 and α4 −6.
Then Q diag−7, −13/2, 2, −6, which satisfies A − Q 0. According to Theorem 3.12,
x −1, −1, 1, 1 is a global minimizer.
1 5
min 3x13 − 8x23 5x33 x43 − x12 4x22 − x32 − 3x42 x1 − 2x2 3x3 4x4
2 3
4.10
4
s.t. x ∈ DB {−1, 0}.
i1
Let a 1, −2, 3, 4T , b 3, −8, 5, 1/2T , and r −2, 8, −10/3, −6T .
Consider x −1, 0, −1, −1T .
diag−1, 1, −1, −1. Then
Let X
1
x1 r1 x1 a1 − r1 v1 − u1 −2 < 0,
2
1
x2 r2 x2 a2 − r2 v2 − u2 −6 < 0,
2
1 14
x3 r3 x3 a3 − r3 v3 − u3 − < 0,
2 3
Mathematical Problems in Engineering 15
1
x4 r4 x4 a4 − r4 v4 − u4 −7 < 0,
2
x ∈ DB , bi xi − xi ≥ 0, i 1, 2, 3, 4.
4.11
1 5
min 3x13 − 8x23 5x33 x43 − x12 − 4x22 x32 − 3x42 − x1 2x2 3x3 4x4
2 3
4.12
4
s.t. x ∈ DB {−1, 0}.
i1
Let a −1, 2, 3, 4T , b 3, −8, 5, 1/2T and r −2, −8, 10/3, −6T .
Consider x −1, 0, −1, −1T .
diag−1, 1, −1, −1. Then
Let X
1
x1 r1 x1 a1 − r1 v1 − u1 0,
2
1
x2 r2 x2 a2 − r2 v2 − u2 6 > 0,
2
4.13
1 4
x3 r3 x3 a3 − r3 v3 − u3 − < 0,
2 3
1
x4 r4 x4 a4 − r4 v4 − u4 −7 < 0.
2
We can see that the conditions are not true in x −1, 0, −1, −1T in Corollary 3.3. But
x −1, 0, −1, −1 is a global minimizer. This fact exactly shows that the conditions are just
sufficient.
Acknowledgment
This research was supported by Innovation Program of Shanghai Municipal Education
Commission 12ZZ071, Shanghai Pujiang Program 11PJC059, and the National Natural
Science Foundation of China 71071113.
References
1 R. A. Canfied, “Multipoint cubic surrogate function for sequential approximate optimization,”
Structural and Multidisciplinary Optimization, vol. 27, pp. 326–336, 2004.
2 Y. Nesterov, “Accelerating the cubic regularization of Newton’s method on convex problems,”
Mathematical Programming, vol. 112, no. 1, pp. 159–181, 2008.
16 Mathematical Problems in Engineering
3 C. S. Lin, P.-R. Chang, and J. Y. S. Luh, “Formulation and optimization of cubic polynomial joint
trajectories for industrial robots,” IEEE Transactions on Automatic Control, vol. 28, no. 12, pp. 1066–
1074, 1983.
4 A. Beck and M. Teboulle, “Global optimality conditions for quadratic optimization problems with
binary constraints,” SIAM Journal on Optimization, vol. 11, no. 1, pp. 179–188, 2000.
5 J.-B. Hiriart-Urruty, “Global optimality conditions in maximizing a convex quadratic function under
convex quadratic constraints,” Journal of Global Optimization, vol. 21, no. 4, pp. 445–455, 2001.
6 J.-B. Hiriart-Urruty, “Conditions for global optimality 2,” Journal of Global Optimization, vol. 13, no. 4,
pp. 349–367, 1998.
7 A. Strekalovsky, “Global optimality conditions for nonconvex optimization,” Journal of Global Opti-
mization, vol. 12, no. 4, pp. 415–434, 1998.
8 V. Jeyakumar, A. M. Rubinov, and Z. Y. Wu, “Sufficient global optimality conditions for non-convex
quadratic minimization problems with box constraints,” Journal of Global Optimization, vol. 36, no. 3,
pp. 471–481, 2006.
9 Z. Y. Wu, V. Jeyakumar, and A. M. Rubinov, “Sufficient conditions for global optimality of bivalent
nonconvex quadratic programs with inequality constraints,” Journal of Optimization Theory and
Applications, vol. 133, no. 1, pp. 123–130, 2007.
10 V. Jeyakumar, A. M. Rubinov, and Z. Y. Wu, “Non-convex quadratic minimization problems with
quadratic constraints: global optimality conditions,” Mathematical Programming A, vol. 110, no. 3, pp.
521–541, 2007.
11 Z. Y. Wu, “Sufficient global optimality conditions for weakly convex minimization problems,” Journal
of Global Optimization, vol. 39, no. 3, pp. 427–440, 2007.
12 Y. Wang and Z. Liang, “Global optimality conditions for cubic minimization problem with box or
binary constraints,” Journal of Global Optimization, vol. 47, no. 4, pp. 583–595, 2010.
13 D. Pallaschke and S. Rolewicz, Foundations of Mathematical Optimization, Convex Analysis without
Linearity, Kluwer Academic Publishers, Dordrecht, The Netherlands, 1997.
Hindawi Publishing Corporation
Mathematical Problems in Engineering
Volume 2012, Article ID 423524, 24 pages
doi:10.1155/2012/423524
Research Article
A Two-Stage DEA to Analyze the Effect of
Entrance Deregulation on Iranian Insurers:
A Robust Approach
Copyright q 2012 Seyed Gholamreza Jalali Naini et al. This is an open access article distributed
under the Creative Commons Attribution License, which permits unrestricted use, distribution,
and reproduction in any medium, provided the original work is properly cited.
We use two-stage data envelopment analysis DEA model to analyze the effects of entrance
deregulation on the efficiency in the Iranian insurance market. In the first stage, we propose a robust
optimization approach in order to overcome the sensitivity of DEA results to any uncertainty in the
output parameters. Hence, the efficiency of each ongoing insurer is estimated using our proposed
robust DEA model. The insurers are then ranked based on their relative efficiency scores for an
eight-year period from 2003 to 2010. In the second stage, a comprehensive statistical analysis
using generalized estimating equations GEE is conducted to analyze some other factors which
could possibly affect the efficiency scores. The first results from DEA model indicate a decline
in efficiency over the entrance deregulation period while further statistical analysis confirms that
the solvency ignorance which is a widespread paradigm among state owned companies is one of
the main drivers of efficiency in the Iranian insurance market.
1. Introduction
In line with the Iran’s Third Development Plan in 1999, the Iranian parliament approved the
establishment of private insurance companies in August 2002 with the aim of improving
efficiency, increasing consumer choices through increased rivalry, and finally enhancing
transparency in the market 1. There are some differences in the literature regarding to
the deregulation impact analysis on the efficiency of insurance companies. While Rees
et al. 2 reported small improvement in German and Britain life insurance market after
deregulation, Hussels and Ward also could not find any strong evidence of deregulation
effects on insurance business efficiency between 1992 until 2002 3, Wang and her colleague
showed the whole market getting more competitive due to leaders market-share losing after
2 Mathematical Problems in Engineering
entrance deregulation in life insurance 4. Jeng and Lai’s study showed that the deregulation
and liberalization, reduction of government or other barriers to market, do not have major
adverse impact on the technical, cost, and revenue efficiency of existing firms in the long run.
The dominance of existing firms has declined but persisted throughout the sample period 5.
Cummins and Rubio-Misas found a positive effect of deregulation on the Spanish insurance
companies’ efficiency 6. Boonyasai et al. also found similar evidence in some Southeast
Asian countries stemmed from joint deregulation and liberalization policies 7. Based on
the previous studies, one could observe that in some cases the developed insurance markets
have improved slightly after deregulation, while there is the possibility of having no effect or
affecting negatively.
In this paper, we test the effects of deregulation on Iranian insurance market. After
Iran’s revolution in 1979, all private insurers were merged compulsorily into four state-
owned companies. Following the entrance deregulation, the Iranian insurance market has
witnessed considerable increase in the number of companies since 2003. A criticism of the
entrance deregulation process was however the lack of sufficient supervision in the insurance
market. To find more evidence whether the Iranian insurance companies have been able
to improve their efficiency as a result of the deregulation process, this paper analyzes the
efficiency of the Iranian insurance market using a two-stage robust DEA model. We use data
from 2003 to 2010, which include the accomplishment of the entrance deregulation. To our
best knowledge, this paper is the first research conducted to examine the relative efficiency
of all Iranian insurance companies for the period of entrance deregulation.
The academic novelty of this research is presenting comprehensive literature review
and applying a two-stage robust DEA model to the problem. We show that traditional DEA
has a limitation to caver all aspects of insurers’ behavior, and a robust DEA model can be used
to overcome the limitations of previous methods. We also use GEE model to extract the most
significant factors explaining robust CRS efficiency scores. According to our study, applying
GEE model as second stage in analyzing efficiency scores is another novelty of this research.
To the best of our knowledge, some efforts have been conducted to analyze the effects of
insurance deregulation on the efficiency of the industry throughout the world using DEA,
however, few researches, if any, considered this issue within Iranian insurance market.
The rest of the paper is arranged as follows: in Section 1, we review different
characteristics of Iranian Insurance market. Section 2 reviews the existing literature on
the topic, especially those related to mathematical programming approach in efficiency
measurement DEA, robust optimization, and generalized estimating equations GEE. Section 3
discusses applications of DEA in institutional changes of insurance industry. In order to
describe different approaches in output selection for DEA model, Section 4 investigates the
Value Added versus the Financial Intermediary approaches. Section 5 presents the results of
efficiency analysis of Iranian insurance companies from 2003 until 2010. This section presents
the first stage and compares traditional CRS DEA scores with robust CRS DEA scores and tries
to find factors explaining efficiencies. In Section 6, a GEE model is applied to capture the most
important factors that explain efficiencies. Finally, Section 7 illustrates the conclusions.
its rank based on Penetration Ratio is a little better among its neighbors, as shown in Table 2.
However, If the index is compared with the global average, 6.89% in 2010 20, then the level
of the ratio would not be satisfactory. It seems that the insurance sector has not been able
to play its primary role in the economy when compared with its counterparts in developed
countries. During last 10 years, Iranian insurance industry has witnessed two major changes
in the institutions. The first one was entrance deregulation which abolished the monopoly of
state-owned companies in 2001. Based on the new law, the private insurance companies have
been established and number of insurance companies increased to 23 from 4 since 2001. The
effects of this institutions change can be seen in Tables 1 and 2, where the premium per capita
and penetration ratio is increasing at very high clip from US$ 11.1 to US$ 34.4 and from
0.86% to 1.28%, respectively, within three years. The second institutions change was price
deregulation in 2009. Based on the new law, the insurances companies are allowed to set their
own premium rates in property and casualty lines of business P&L. It is also obvious that
following price deregulation the growth of the insurance industry has been revitalized after a
period of calm. By the way, due to the lack of available data, the main purpose of this paper
focuses on the first event.
The last issue is that the share of life insurance in Iran is very low in comparison with
its global average. Based on the yearbook of Central Insurance of IR Iran, the average of life
insurance share in total written premium of Iran’s market is around 7.5% in the last 10 years.
The analysis of this issue is beyond the scope of this paper, but this phenomenon has caused
all companies established in Iran to be general mixed insurers which mostly do business in
P&L lines. This in turn, helps us to deal with all companies in a similar way when assessing
the efficiency.
3. Theoretical Framework
3.1. The Economic Efficiency
Frontier methodologies have been used in majority of papers published in recent years. There
are two common tools in frontier methodologies: the econometric frontier analysis and the data
envelopment analysis DEA, Both have their own pros and cons. Unlike the data envelopment
analysis approach, it’s not allowed in econometric stochastic frontier approach to use various
inputs and outputs and it also requires researchers to define functional form on the data and
set assumptions about distributional form of the inefficiency term. Both of them presume
the production function is known. The economic efficiency is stemmed from production
frontier in theory of the firm. Figure 1 shows a production frontier PF for a firm with single
input/output.
If a firm is producing at point I, J in time t, it could produce more efficiently by
moving to the frontier PFt horizontally or vertically. If it moves to PFt horizontally by
reducing its excess input, this is called input-oriented and, it is called output-oriented if
the point moves vertically by producing more output. The technical efficiency of the firm
is calculated by the ratio 0H/0I for input-oriented approach, which is the reciprocal of its
distance from the frontier PFt . Defining efficiency by the concept of the distance from the
production frontier is formulated by Shephard 21 as below; assume a producer uses input
4
Table 1: Premium per Capita PPC trend of selected countries 2003–2010 US Dollar.
Year
S.C. 2001 2002 2003 2004 2005 2006 2007 2008 2009 2010
P.P.C. P.P.C. A.G. P.P.C. A.G. P.P.C. A.G. P.P.C. A.G. P.P.C. A.G. P.P.C. A.G. P.P.C. A.G. P.P.C. A.G. P.P.C. A.G.
UAE 302.4 317 5% 310.7 −2% 350.2 13% 414.2 18% 585.1 41% 811.6 39% 1114 37% 1111.8 0% 1248.1 12%
Qatar NA NA NA NA NA 444.4 NA 442.3 0% 683.5 55% 640.2 −6% NA NA 548.6 NA 618.8 13%
Bahrain 219.7 295.2 34% NA NA NA NA NA NA NA NA NA NA 588.7 NA 624.8 6% 527 −16%
Kuwait 131.5 154.1 17% 148 −4% 161.2 9% 185.5 15% 227.2 22% 257.3 13% 313 22% 153 −51% 235.5 54%
Oman 77.4 84 9% 99 18% 103.1 4% 113.7 10% 143.7 26% 159.5 11% 218.2 37% 220.4 1% 260.8 18%
Saudi Arabia 47.2 41.6 −12% 41.2 −1% 51.4 25% 57.1 11% 63.1 11% 91.7 45% 121.4 32% 151.5 25% 178.4 18%
Iran 11.1 17.3 56% 22.9 32% 29.1 27% 34.4 18% 41.1 19% 50.9 24% 59.1 16% 63.9 8% 76.8 20%
Mathematical Problems in Engineering
Table 2: Penetration ratio PR trend of selected countries 2003–2010 %.
Year
Mathematical Problems in Engineering
S.C. 2001 2002 2003 2004 2005 2006 2007 2008 2009 2010
P.R. P.R. A.G. P.R. A.G. P.R. A.G. P.R. A.G. P.R. A.G. P.R. A.G. P.R. A.G. P.R. A.G. P.R. A.G.
UAE 1.25 1.28 2% 1.12 −13% 1.65 47% 1.53 −7% 1.7 11% 1.9 12% 2 5% 2.5 25% 2.1 −16%
Qatar NA NA NA NA NA 1.23 NA 1.09 −11% 1.1 1% 0.9 −18% NA NA 0.8 NA 0.8 0%
Bahrain 1.93 2.08 8% NA NA NA NA NA NA NA NA NA NA 2 NA 2.3 15% 2.8 22%
Kuwait 0.79 0.95 20% 0.92 −3% 0.93 1% 0.79 −15% 0.7 −11% 0.6 −14% 0.6 0% 0.5 −17% 0.5 0%
Oman 0.96 1.01 5% 1.44 43% 1.28 −11% 1.14 −11% 1 −12% 1.1 10% 1.1 0% 1.2 9% 1.3 8%
Saudi Arabia 0.53 0.48 −9% 0.47 −2% 0.48 2% 0.46 −4% 0.5 9% 0.6 20% 0.6 0% 1 67% 1.1 10%
Iran 0.86 1 16% 1.16 16% 1.25 8% 1.28 2% 1.3 2% 1.29 −1% 1.35 5% 1.39 3% 1.5 8%
S.C.: selected countries; P.P.C: premium per capita; A.G.: annual growth; P.R.: penetration ratio.
Source: 8.
5
6 Mathematical Problems in Engineering
Y
P F t+1
P Ft
(xt+1 , yt+1 )
M
J (Xt , Yt )
0 L K H I X
Figure 1: Production Frontier PF for a firm with one input and one output.
x −1
D x, y sup θ : , y V y inf θ : θx, y V y , 3.1
θ
where V , D, and θ are production frontier, distance function, and distance, respectively.
The technical efficiency TEx, y is therefore defined as TEx, y 1/Dx, y. It should
be mentioned here that TEx, y for each decision making unit can be obtained by linear
programming 22. If a firm uses two or more inputs, inefficiency can be also stemmed
from the fact that it’s not deploying the cost minimizing combination of inputs. This kind
of inefficiency is called allocative inefficiency. A firm is considered to be fully cost efficient if
it operates at a point where both technical and allocative efficiency are met. Cost efficiency is
then defined as follows:
choose PFt or PFt 1 as the frontier reference. With respect to the period t frontier, an input-
oriented Malmquist productivity index can be defined as:
D t xt , y t
M t t 1 t 1 .
t
3.3
D x ,y
Similarly, the Malmquist index based on period t 1 frontier can be defined. To avoid
arbitrarily selecting one frontier to compute the index, the geometric mean could be applied
as follows:
D t xt , y t D t 1 xt , y t
M xt 1 , y t 1 , xt , y t × . 3.4
D t xt 1 , y t 1 D t 1 xt 1 , y t 1
and a1j , . . . , asj , respectively, where the ratio of the weighted sum of outputs to the weighted
sum of inputs for some given DMUo is maximized as follows:
s
xr ar0
max δ0 r1 m
i1 vi bi0
s
r1 xr arj 3.5
subject to m ≤ 1, ∀j 1, . . . , n,
i1 vi bij
xr , vi ≥ 0,
where xr and vi are the weight factors, and the δ0 , ar0 , and bi0 are the observed efficiency,
output and input values, respectively, of DMUo , the DMU to be evaluated. Model 3.5 is
a nonlinear fractional programming model which could be converted into the following LP
model 26:
max δ0
s
m
subject to xr arj − vi bij ≤ 0, ∀j 1, . . . , n, 3.6
r1 i1
ur , vi ≥ 0.
max W
s
subject to w− rj≤0
xr α
r1
s
m
rj − vi bij ≤ 0,
xr α ∀j 1, . . . , n 3.7
r1 i1
m
vi bij 1
i1
xr , vi ≥ 0,
where, x and v are input and output variables and Indices i, r, and j represent the number
of inputs, outputs and DMUs respectively. One could observe that in this formulation,
the parameter α rj is an uncertain value. We cannot solve this problem with popular
Linear Programming techniques, since the primary assumption, that is, certainty of input
parameters is violated.
Mathematical Problems in Engineering 9
min c x
subject to Ax ≥ b, 3.8
x ∈ X.
Bertsimas and Sim proposed an approach for linear optimization that provides full
control on the degree of conservatism and keeps the advantages of the linear framework of
Soyster. They defined the scaled deviation from nominal value of aij as shown here;
aij − aij
ηij ,
aij
3.9
n
ηij ≤ Γi , ∀i 1, . . . , m,
j1
where ηij has an unknown but symmetric distribution which could only take value within
−1, 1. Despite the fact that the cumulative scaled deviation of constraint i can take no value
within −n, n, but it is assumed to be confined. In this approach, Γi is some control parameter
known as the price of robustness. These parameters adjust the robustness of the method
against the level of conservatism of the solution. For Γi 0, we get nominal model and
no uncertain coefficient involved. On the other hand, Γi n means that the ith constraint
of the problem is protected against all possible realizations of uncertain coefficients. For any
value Γ within 0, n, the decision maker takes into account a tradeoff between the level of
the protection of constraint and the level of solution conservation.
Based on Bertsimas approach, we can reformulate DEA model. The reformulated DEA
model, known as robust counterpart is as follwos:
max W
s
s
subject to W − xr aro z0 Γ pr0 ≤ 0,
r1 r1
s
m
s
xr arj − vi bij zj Γ prj ≤ 0, ∀j 1, . . . , n,
r1 i1 r1
3.10
zj prj ≥ arj yr , ∀r 1, . . . , s, ∀j 1, . . . , n,
m
vi bij 1, ∀j 1, . . . , n,
i1
− yr ≤ xr ≤ yr , ∀r 1, . . . , s,
xr , zj , yr , vi , prj ≥ 0,
where arj vector of nominal value of arj , Γ degree of uncertainty within constraint
parameters, arj precision of estimation of arj , zj some auxiliary variable related to the robust
counterpart denoting the cost of robustness in each constraints, prj some auxiliary variable
related to the robust counterpart counting the number of uncertain parameters in each
constraints, and y decision variable for making the absolute term |xr | to linear one. Other
notations are defined in previous equations.
This robust counterpart is obviously a linear programming model which can be solved
with popular solver packages. Since original DEA model is linear programming problem,
Mathematical Problems in Engineering 11
incorporating uncertainty does not deteriorate solvability of the model. In other words,
applying this reformulation preserves the type of original linear problem.
N
∂gn T
g E Cov−1 gn gn , 3.11
n1
∂β
N
∂μn T
0 E Cov−1 δn δn − μn , 3.12
n1
∂β
working model of the correlation structure in the observed dependent variables, where α is a
possibly vector-valued parameter 34.
Cross-country
studies; four life Group life insurance premiums,
Salaries, wages, commissions,
insurance markets: individual life insurance
Boonyasai et al. 7 DEA and business and services
Korea, Philippines, premiums, and investment
expense.
Taiwan, and income
Thailand.
Claims incurred net change in
13
160 Ukrainian
Various types of premiums, such
insurance Equity, liabilities, and fixed and
Badunenko et al. 17 DEA as personal, property, liability,
companies current assets
and so forth
2003–2005.
German and british
Annual average number of
life insurance Net written premiums, additions
Hussels and Ward 3 DEA employees, total assets minus
companies to reserves.
total liabilities.
1991–2002.
Total operating costs, total Profit and loss account, net
Nigerian insurers number of employees, and total premiums, settled claims,
Barros et al. 18 DEA
1994–2005 investments. outstanding claims, and
investment.
Benefit payments disaggregated
into four categories: ordinary life
Taiwanese life insurance, personal accident
Home office labor, agent labor,
insurance insurance, individual health
Jeng and Lai 5 DEA business service, and equity
companies insurance, and group insurance,
capital.
Mathematical Problems in Engineering
6. Computational Results
6.1. First Stage: Efficiency Scores, Traditional DEA versus
Robust DEA Results
To calculate insurers’ efficiencies, we used panel data for the years 2003–2010, obtained from
the Central Insurance yearbook of IR Iran on 20 mixed general insurance companies for a
period of 8 years. The sample consists of 139 observations. The insurance companies that are
considered in this analysis represent almost all of the market. Keeping in mind the popular
DEA rule of thumb, the number of companies in each year is almost greater than three times
the number of inputs plus output.
To determine inputs, we followed previous works discussed in the literature review
section, and for outputs we followed value-added approach as described in Cummins and
Weiss 23 but in order to capture the intermediary function of insurance companies we
added ROE as an output. In short, we measured output by: 1 losses incurred; 2 return on
equity ROE; measured inputs by 1 number of employees; 2 general and administrative
expenses’ 3 surplus that is total asset minus total liabilities. It should be mentioned here that
all insurance companies in Iran are general mixed, and in the last 10 years the average of
life insurance share in total premium of Iran’s market is around 7.5%. In fact Iranian insurers
mostly do business in P&L line. This helps us to consider loss incurred as output for all
observations. Table 4 presents the CRS efficiency scores for the Iranian insurance companies.
Some conclusions can be stemmed from Table 4.
First, based on DEA with certain and uncertain outputs, the overall inefficiency gap for
the Iranian insurance industry is 0.18 and 0.3, respectively. However, it should be mentioned
that these gaps do not follow any particular trend. This finding implies that in overall, no
major change could be found in the market. It is also true about standard deviations in both
DEA models see Table 4.
Second, as expected, CRS efficiency of models with certain outputs is always greater
than that of with uncertain outputs, but the gaps between private and state owned insurers
is deferent. As it is shown in Figure 2, the average gap in private insurers is much higher
than that of in state owned insurers. It could be observed that the private insurers would be
much more venerable if the conditions changed. It convince us to work with Robust CRS
efficiency scores instead of traditional CRS scores because of the fact that in case of any
16
Table 4: CRS relative efficiency observed in Iranian insurance companies in percent %: 2003–2010; DEA with certain and uncertain outputs traditional
DEA versus robust DEA.
Year
DMUs 2003 2004 2005 2006 2007 2008 2009 2010
Certain Uncertain Certain Uncertain Certain Uncertain Certain Uncertain Certain Uncertain Certain Uncertain Certain Uncertain Certain Uncertain
outputs outputs outputs outputs outputs outputs outputs outputs outputs outputs outputs outputs outputs outputs outputs outputs
DMU1a 100 100 100 100 100 100 100 100 100 100 100 100 100 100 100 100
DMU2a 100 100 100 100 100 100 100 100 100 100 100 100 100 100 71 67
DMU3a 100 100 100 100 100 100 68 68 100 100 100 100 100 100 100 100
DMU4a 46 46 51 45 72 61 56 42 64 64 71 71 74 73 60 53
DMU5 41 41 69 33 26 14 27 20 56 51 35 34 70 64 97 92
DMU6 100 88 71 58 100 99 79 79 75 74 79 78 100 100 100 100
DMU7 70 60 67 61 57 57 66 64 98 98 100 90 87 77 69 68
DMU8 86 71 100 100 86 81 100 100 100 53 100 58 100 100 100 100
DMU9 32 31 49 42 55 55 51 26 51 28 54 54 53 47 44 29
DMU10 100 100 100 100 100 100 100 100 100 100 100 100 100 100 100 100
DMU11 100 94 100 78 100 44 100 62 65 31 52 38 44 33 52 21
DMU12 100 100 100 100 100 71 100 33 100 97 100 74 100 100 100 100
DMU13 100 100 63 56 54 50 79 78 95 95 100 100 100 100 100 100
DMU14 92 46 21 6 68 57 100 100 100 100 100 100 100 100
DMU15 100 100 100 100 100 100 100 88 100 92 100 76
DMU16 34 22 73 57 97 96 100 100 100 64 74 69
DMU17 100 55 61 53 43 34 58 52 72 49 51 35
DMU18 77 52 69 40 52 24 64 30 69 33
DMU19 62 48 64 46 53 30
DMU20 100 100 100 36
Average 83 79 83 73 77 65 78 66 84 76 82 74 86 79 82 70
Standard
26 26 20 27 29 32 21 27 21 29 23 26 19 26 21 30
deviation
a
State-owned insurers.
Mathematical Problems in Engineering
Mathematical Problems in Engineering 17
change in outputs such as ROE or loss incurred, the efficiency of private insurers will change
dramatically. Indeed, the right model for this uncertain data is robust model.
Third, it seems that some companies could be able to compete with their state-owned
companies counterparts. For example, DMU8, DMU10, and DMU14 which established in
2003, 2003, and 2004, respectively, could reach to the state owned insurers DMU1, DMU2,
DMU3, and DMU4. It shows that state-owned insurers are not secure. They should upgrade
the quality of their management practices, responding to the results of the present research
see Figure 3.
The reason why these insurers could be able to catch their state-owned counterparts
is out of our research scope, and can be done in separate research. But it can be briefly noted
that all these companies are captive.
14
12
10
scores
8
0
2003 2004 2005 2006 2007 2008 2009 2010
Year
State owned
Private
Figure 2: Average gap between traditional efficiency scores and robust efficiency scores for different types
of insurance companies.
100
Average of robust CRS
90
80
efficiency scores
70
60
50
40
30
20
10
0
2003 2004 2005 2006 2007 2008 2009 2010
Year
Average of robust CRS efficiency scores of state = owned insurers
Average of robust CRS efficiency scores of DMU8, DMU10, and DMU14
Figure 3: Comparison between average of robust CRS efficiency scores for state owned insurers and
selective private insurers.
institutions change. Based on Figure 5 we conclude that FLI may have a significant effect
on efficiency scores. This factor is similar to CastNew which is used in Barros et al. 19.
Based on the material before, to find the most significant factors explaining efficiency
scores we consider six factors for next stage, GEE model. The factors are: Ownership, Being in
stock exchange market, Insurer’s size, Insurer’s market share, capital structure, and solvency index.
Mathematical Problems in Engineering 19
1200 100
Efficiency (%)
60
600 50
40
400
30
200 20
10
0 0
2003 2004 2005 2006 2007 2008 2009 2010
Figure 4: Comparison between premium to surplus ratio and efficiency for different types of insurance
companies.
100
90
80
70
60
(%)
50
40
30
20
10
0
2003 2004 2005 2006 2007 2008 2009 2010
Figure 5: Comparison between capital structure and efficiency for different types of insurance company
Financial Leverage Index FLI, a ratio indicating the capital structure, is defined as the ratio between
assets to equity.
where δit represents the CRS efficiency score of insurer i at time t. Private is a dummy
variable, which is one for state-owned insurance companies. The inclusion of this variable
is based on the assumption that private companies may exhibit higher efficiency due to type
of ownership. Stock is a dummy variable which is one for insurance companies whose shares
could be traded on the stock exchange aiming to capture the effect of transparency due to
the stock market governance requirements. Big is a dummy variable which is one for big
companies, and is measured by the total value of assets. Following Barros et al. 19 we set
a limit for companies’ total assets. If a company’s total assets values more than $300 million
U.S., it is considered as a big company, and variable Big for that is equal to 1. It helps us to
summarize data without losing valuable information because there are two major categories
of companies in Iran. Some of them are old and huge, and the others are recently established
and small. Marketshare is the market share of each insurance company within the market.
Capitalstructure is the ratio of equity to total assets Equity to total assets is equal to dividing
return on asset (ROA) by return on equity (ROE) or FLI−1 . or FLI−1 , aiming to capture the effect
of capital structure. PrmtoSrp is the ratio of premium divided by surplus aiming to measure
the risk-aversion behavior of insurers and the company’s attitude toward solvency issue.
It is obvious that an insurer’s efficiency in a given year is correlated with its efficiency
in the other years. In other words, there are correlations among the yearly observations
belong to an insurer. GEE is an extension of the quasi-likelihood approach, is used to
analyze longitudinal and other correlated data. Some articles do discuss how much statistical
information is obtainable from observations on individuals in clusters such as cluster of an
insurance company’s efficiencies during 8 years.
Following Liang and Zeger 34, we employ GEE method to determine the most
important variables, with 139 observations categorized into 20 clusters. The results are
presented in Table 5. Lots of models are estimated in order to compare the results. The
results are quite stable since the variables that were significant in the Model 1, remained
significant after dropping the insignificant variables. Based on Table 5, we can conclude
that private contributes negatively to efficiency, signifying that this type of companies face
different constraints in the Iranian insurance market. Second, the variable Capitalstructure
shows that the specific capital structure of Iranian insurance companies exercises a positive
effect on efficiency. Finally the variable PrmtoSrp has a positive influence on efficiency. This
result along with the result of capitalstructure shows that the companies which do not respect
to solvency measures are more efficient than the others. Marketshare, Big, and Stock do not
have a significant effect on efficiency.
Since the GEE model makes a working correlation matrix in order to obtain the optimal
estimators, we test higher degrees of correlations between with-in-the-subjects observations.
This can be seen in Table 6, where we estimate parameter with M 5, M 6, and M 7. This
means that the insurer’s efficiency score in a given year is related to the insurer’s efficiency
in adjacent 5 years if we set M 5. It is obvious that m cannot be more than 7 due to the fact
that the maximum number of observations within a cluster or with-in-the-subject is equal to
8 a company’s efficiency scores during 8 years. Table 6 shows that Private and PrmtoSrp are
consistently significant while we increase the degree of with-in-the-subject dependency.
In GEE model, QIC is a measure used to choose best correlation structure which can
be applied to determine the best subsets of covariates for a particular model. The best model
is the one with the smallest QIC value. If we consider QIC as the model’s goodness of fit, it
gets better while we increase M, the with-in-the-subject dependency.
Mathematical Problems in Engineering 21
Table 5: Second stage—Generalized estimating equations models that capture the most important factors
explaining the efficiency score for robust DEA.
Models
Predictors Model 1 Model 2 Model 3 Model 4
Hypothesis Hypothesis Hypothesis Hypothesis
B B B B
Test-Sig. Test-Sig. Test-Sig. Test-Sig.
Intercept .681 .006 .681 .005 .779 .000 1.141 .000
Private −.367 .000 −.370 .000 −.352 .000 −.402 .000
Stock .118 .143 .118 .138 — — — —
Big .165 .193 .172 .129 .178 .093 — —
Marketshare .001 .751 — — — — — —
Capitalstructure .007 .003 .007 .003 .006 .004 .005 .005
PrmtoSrp .018 .000 .018 .000 .017 .000 .017 .000
QICa 46.367 44.537 41.933 35.142
QICCa 25.739 23.785 21.424 19.649
Residual normal testb .224
Asymp. Sig. 2-tailed
a
Computed using the full log quasilikelihood function.
b
Being normal is not rejected at 5% level if the statistic is more than 0.05.
Table 6: Sensitivity analyzing of significant factors, Private and PrmtoSrp, based on the within-subject
dependencies.
8. Concluding Remarks
In this paper, we have analyzed technical efficiency for Iranian insurance companies between
2003 and 2010, a period that insurers experienced intense volatility due to the entrance
deregulation of the market. We propose a two-stage procedure to analysis the most important
determinants affecting efficiency scores. In the first stage, we obtained the CRS efficiency
scores by robust DEA model proposed by Bertsimas and Sim in 2003. In the second stage, we
determined the most important factors that can explain the efficiency scores by using GEE
developed by Liang and Zeger in 1986. The major results of our study are that ownership type
and failure to meet the risk management rules are the main drivers of efficiency. In other words,
any state owned insurer which issued more policies without respect to the sufficient capital
22 Mathematical Problems in Engineering
provision could obtain better score. It should be mentioned here that state owned companies
issued more policies and paid more loss, in the hope that the government supports in case
of difficult financial situations. Indeed, instead of issuing insurance policies in proportion
to their capital, the insurance policies issued on behalf the government credit. This finding
shows the violation of competition rules by state, and the inadequacies of the institutions
necessary for private sector development.
What should the managers of inefficient insurance companies do to improve
efficiency? First, in order to prepare the institutions, they must pursue the new regulations
that require state-owned insurers to have sufficient capital and prevent them to issue
insurance policies unlimitedly. Also, following the Williamson if we consider culture as
the first level of institutions 40, it seems that private companies should try to change the
commonly thought that state-owned companies are more reliable. This cannot be achieved
unless their quality of services is as good as their state-owned counterparts, and it cannot
be done unless they design and manage their processes efficiently. They should establish a
benchmark management procedure in order to evaluate their relative position and to adopt
appropriate managerial procedures for catching up with the frontier of “best practices.”
It seems that some private companies have been able to compete with their state-owned
counterparts. Finally, the regulatory authority has an important role in making a fair business
environment and improving the efficiency of insurers by 1 participation in developing
new rules of business and in enforcing its regulatory duties. 2 Developing indicators to
monitor solvency and requiring state-owned companies to comply with. One of the central
insurance of IR Iran initial duties is to arrange the market so that insurers could compete fairly
while they meet at least the minimum required solvency margin. 3 Publishing information
in order to introduce greater transparency into the market especially those related to the
sufficient capital and reserves for future commitments.
References
1 Management and Planning Organization, “The law of the third economic, social, and cultural
develpoment plan,” Tech. Rep., Management and Planning Organization, Tehran, Iran, 2003.
2 R. Rees, E. Kèssner, P. Klemperer, and C. Matutes, “Regulation and efficiency in European insurance
markets,” Economic Policy, no. 29, pp. 365–397, 1999.
3 S. Hussels and D. R. Ward, “The impact of deregulation on the German and UK life insurance markets:
an analysis of efficiency and productivity between 1991–2002,” Working Paper, Cranfield Research
Paper Series, 2007.
4 J. L. Wang, L. Y. Tzeng, and E. L. Wang, “The nightmare of the leader: the impact of deregulation on
an oligopoly insurance market,” Journal of Insurance, vol. 26, no. 1, pp. 15–28, 2003.
5 V. Jeng and G. C. Lai, “The impact of deregulation on efficiency: an analysis of life insurance industry
in Taiwan from 1981 to 2004,” Risk Management and Insurance Review, vol. 11, no. 2, pp. 349–375, 2008.
6 J. D. Cummins and M. Rubio-Misas, “Deregulation, consolidation, and efficiency: evidence from the
Spanish insurance industry,” Journal of Money, Credit and Banking, vol. 38, no. 2, pp. 323–355, 2006.
7 T. Boonyasai, M. F. Grace, and H. D. Skipper Jr., The Effect of Liberalization and Deregulation on Life
Insurer Efficiency [M.S. thesis], Georgia State University, Atlanta, Ga, USA, 1999.
8 Central Insurance of IR Iran, Statistical Year Book 2009-2010, Central Insurance of IR Iran, Tehran, Iran.
9 J. D. Cummins, G. Turchetti, and M. A. Weiss, “Productivity and technical. Productivity and technical
efficiency in the Italian insurance industry,” Working Paper 96-10, Wharton School, Philadelphia, Pa,
USA, 1996.
10 H. Fukuyama, “Investigating productive efficiency and productivity changes of Japanese life
insurance companies,” Pacific Basin Finance Journal, vol. 5, no. 4, pp. 481–509, 1997.
11 J. D. Cummins, M. A. Weiss, and H. Zi, “Organizational form and efficiency: the coexistence of stock
and mutual property-liability insurers,” Management Science, vol. 45, no. 9, pp. 1254–1269, 1999.
Mathematical Problems in Engineering 23
12 B. Mahlberg and T. Url, “The transition to the single market in the German insurance industry,”
Working Paper, Austrian Institute of Economic Research, 2000.
13 A. G. Noulas, T. Hatzigayios, J. Lazaridis, and K. Lyroudi, “Non-parametric production frontier
approach to the study of efficiency of non-life insurance companies in Greece,” Journal of Financial
Management and Analysis, vol. 14, no. 1, pp. 19–26, 2001.
14 B. Mahlberg and T. Url, “Effects of the single market on the Austrian insurance industry,” Empirical
Economics, vol. 28, no. 4, pp. 813–838, 2003.
15 K. C. Ennsfellner, D. Lewis, and R. I. Anderson, “Production efficiency in the Austrian insurance
industry: a Bayesian examination,” Journal of Risk and Insurance, vol. 71, no. 1, pp. 135–159, 2004.
16 C. P. Barros, N. Barroso, and M. R. Borges, “Evaluating the efficiency and productivity of insurance
companies with a malmquist index: a case study for Portugal,” Geneva Papers on Risk and Insurance,
vol. 30, no. 2, pp. 244–267, 2005.
17 O. Badunenko, B. Grechanyuk, and O. Talavera, “Development under regulation: the way of the
Ukrainian insurance market,” Discussion Papers of DIW Berlin, German Institute for Economic
Research, Berlin, Germany, 2006.
18 C. P. Barros, A. Ibiwoye, and S. Managi, “Productivity change of Nigerian insurance companies: 1994–
2005,” African Development Review, vol. 20, no. 3, pp. 505–528, 2008.
19 C. P. Barros, M. Nektarios, and A. Assaf, “Efficiency in the Greek insurance industry,” European Journal
of Operational Research, vol. 205, no. 2, pp. 431–436, 2010.
20 D. Stalb and L. Bevere, World Insurance in 2010, Swiss Reinsurance, 2011.
21 R. W. Shephard, Theory of Cost and Production Functions, Princeton University Press, Princeton, NJ,
USA, 1970.
22 A. Charnes, W. W. Cooper, A. Y. Lewin, and L. M. Seiford, Data Envelopment Analysis: Theory,
Methodology and Applications, Kluwer Academic, New York, NY, USA, 1995.
23 J. D. Cummins and M. A. Weiss, “Analyzing firm performance in the insurance industry using frontier
efficiency methods,” Center for Financial Institutions Working Papers, Wharton School Center for
Financial Institutions, University of Pennsylvania, 1998.
24 D. Deprins, L. Simar, and H. Tulkens, “Measuring labor-efficiency in post offices,” in The Performance
of Public Enterprises: Concepts and Measurement, pp. 243–268, Elsevier Science, Amsterdam, The
Netherlands, 1984.
25 J. D. Cummins and H. Zi, “Comparison of frontier efficiency methods: an application to the U.S. life
insurance industry,” Journal of Productivity Analysis, vol. 10, no. 2, pp. 131–152, 1998.
26 A. Charnes, W. W. Cooper, and E. Rhodes, “Measuring the efficiency of decision making units,”
European Journal of Operational Research, vol. 2, no. 6, pp. 429–444, 1978.
27 S. T. Cooper and E. Cohn, “Estimation of a frontier production function for the South Carolina
educational process,” Economics of Education Review, vol. 16, no. 3, pp. 313–327, 1997.
28 D. Bertsimas and M. Sim, “Robust discrete optimization and network flows,” Mathematical
Programming B, vol. 98, no. 1–3, pp. 49–71, 2003.
29 M. A. Shafiaa, M. Aghaeeb, and A. Jamilia, “A new mathematical model for the job shop scheduling
problem with uncertain processing times,” International Journal of Industrial Engineering Computations,
vol. 2, pp. 295–306, 2010.
30 A. Soyster, “Convex programming with set-inclusive constraints and applications to inexact linear
programming,” Operations Research, vol. 21, no. 5, pp. 1154–1157, 1973.
31 A. Ben-Tal and A. Nemirovski, “Robust solutions of linear programming problems contaminated with
uncertain data,” Mathematical Programming B, vol. 88, no. 3, pp. 411–424, 2000.
32 M. Gharakhani, I. Kazemi, and H. Alizadeh Haji, “A robust DEA model for measuring the relative
efficiency of Iranian high schools,” Management Science Letter, vol. 1, no. 3, pp. 389–404, 2011.
33 J. A. Hanley, A. Negassa, M. D. D. B. Edwardes, and J. E. Forrester, “Statistical analysis of correlated
data using generalized estimating equations: an orientation,” American Journal of Epidemiology, vol.
157, no. 4, pp. 364–375, 2003.
34 K. Y. Liang and S. L. Zeger, “Longitudinal data analysis using generalized linear models,” Biometrika,
vol. 73, no. 1, pp. 13–22, 1986.
35 M. Eling and M. Luhnen, “Frontier efficiency methodologies to measure performance in the insurance
industry: overview and new empirical evidence,” University of St. Gallen Working Papers on Risk
Management and Insurance Paper 56, 2008.
36 P. L. Brockett, W. W. Cooper, L. L. Golden, J. J. Rousseau, and Y. Wang, “Financial intermediary versus
production approach to efficiency of marketing distribution systems and organizational structure of
insurance companies,” Journal of Risk and Insurance, vol. 72, no. 3, pp. 393–412, 2005.
24 Mathematical Problems in Engineering
37 A. N. Berger and D. B. Humphrey, “Measurement and efficiency issues in commercial banking,” in
Output Measurement in the Service Sectors, pp. 245–300, University of Chicago Press, Chicago, Ill, USA,
1992.
38 J. T. Leverty and M. F. Grace, “The robustness of output measures in property-liability insurance
efficiency studies,” Journal of Banking and Finance, vol. 34, no. 7, pp. 1510–1524, 2010.
39 T. J. Coelli, P. Rao, and G. E. Battese, An Introduction to Efficiency and Productivity Analysis, Kluwer
Academic Press, New York, NY, USA, 1998.
40 O. E. Williamson, “The new institutional economics: taking stock, looking ahead,” Journal of Economic
Literature, vol. 38, no. 3, pp. 595–613, 2000.
Hindawi Publishing Corporation
Mathematical Problems in Engineering
Volume 2012, Article ID 841410, 36 pages
doi:10.1155/2012/841410
Research Article
Solving Constrained Global Optimization Problems
by Using Hybrid Evolutionary Computing and
Artificial Life Approaches
Jui-Yu Wu
Department of Business Administration, Lunghwa University of Science and Technology, No. 300,
Section 1, Wanshou Road, Guishan, Taoyuan County 333, Taiwan
Copyright q 2012 Jui-Yu Wu. This is an open access article distributed under the Creative
Commons Attribution License, which permits unrestricted use, distribution, and reproduction in
any medium, provided the original work is properly cited.
This work presents a hybrid real-coded genetic algorithm with a particle swarm optimization
RGA-PSO algorithm and a hybrid artificial immune algorithm with a PSO AIA-PSO
algorithm for solving 13 constrained global optimization CGO problems, including six nonlinear
programming and seven generalized polynomial programming optimization problems. External
RGA and AIA approaches are used to optimize the constriction coefficient, cognitive parameter,
social parameter, penalty parameter, and mutation probability of an internal PSO algorithm. CGO
problems are then solved using the internal PSO algorithm. The performances of the proposed
RGA-PSO and AIA-PSO algorithms are evaluated using 13 CGO problems. Moreover, numerical
results obtained using the proposed RGA-PSO and AIA-PSO algorithms are compared with those
obtained using published individual GA and AIA approaches. Experimental results indicate that
the proposed RGA-PSO and AIA-PSO algorithms converge to a global optimum solution to a
CGO problem. Furthermore, the optimum parameter settings of the internal PSO algorithm can
be obtained using the external RGA and AIA approaches. Also, the proposed RGA-PSO and AIA-
PSO algorithms outperform some published individual GA and AIA approaches. Therefore, the
proposed RGA-PSO and AIA-PSO algorithms are highly promising stochastic global optimization
methods for solving CGO problems.
1. Introduction
Many scientific, engineering, and management problems can be expressed as constrained
global optimization CGO problems, as follows:
Minimize fx,
s.t. gm x ≤ 0, m 1, 2, . . . , M,
2 Mathematical Problems in Engineering
hk x 0, k 1, 2, . . . , K,
xnl ≤ xn ≤ xnu , n 1, 2, . . . , N,
1.1
where fx denotes an objective function; gm x represents a set of m nonlinear inequality
constraints; hk x refers to a set of k nonlinear equality constraints; x represents a vector
of decision variables which take real values, and each decision variable xn is constrained
by its lower and upper boundaries xnl , xnu ; N is the total number of decision variables
xn . For instance, generalized polynomial programming GPP belongs to the nonlinear
programming NLP method. The formulation of GPP is a nonconvex objective function
subject to nonconvex inequality constraints and possibly disjointed feasible region. The
GPP approach has been successfully used to solve problems including alkylation process
design, heat exchanger design, optimal reactor design 1, inventory decision problem
economic production quantity 2, process synthesis and the design of separations,
phase equilibrium, nonisothermal complex reactor networks, and molecular conformation
3.
Traditional local NLP optimization approaches based on a gradient algorithm are
inefficient for solving CGO problems, while an objective function is nondifferentiable.
Global optimization methods can be divided into deterministic or stochastic 4. Often
involving a sophisticated optimization process, deterministic global optimization methods
typically make assumptions regarding the problem to be solved 5. Stochastic global
optimization methods that do not require gradient information and numerous assumptions
have received considerable attention. For instance, Sun et al. 6 devised an improved
vector particle swarm optimization PSO algorithm with a constraint-preserving method
to solve CGO problems. Furthermore, Tsoulos 7 developed a real-coded genetic algorithm
RGA with a penalty function approach for solving CGO problems. Additionally, Deep and
Dipti 8 presented a self-organizing GA with a tournament selection method for solving
CGO problems. Meanwhile, Wu and Chung 9 developed a RGA with a static penalty
function approach for solving GPP optimization problems. Finally, Wu 10 introduced an
artificial immune algorithm AIA with an adaptive penalty function method to solve CGO
problems.
Zadeh 11 defined “soft computing” as the synergistic power of two or more fused
computational intelligence CI schemes, which can be divided into several branches: gran-
ular computing e.g., fuzzy sets, rough sets, and probabilistic reasoning, neurocomputing
e.g., supervised, unsupervised, and reinforcement neural learning algorithms, evolutionary
computing e.g., GAs, genetic programming, and PSO algorithms, and artificial life e.g.,
artificial immune systems 12. Besides, outperforming individual algorithms in terms of
solving certain problems, hybrid algorithms can solve general problems more efficiently
13. Therefore, hybrid CI approaches have recently attracted considerable attention as a
promising field of research. Various hybrid evolutionary computing GA and PSO methods
and artificial life such as AIA methods approaches have been developed for solving
optimization problems. These hybrid algorithms focus on developing diverse candidate
solutions such as chromosomes and particles of population/swarm to solve optimization
problems more efficiently. These hybrid algorithms use two different algorithms to create
diverse candidate solutions using their specific operations and then merge these diverse
Mathematical Problems in Engineering 3
candidate solutions to increase the diversity of the candidate population. For instance, Abd-
El-Wahed et al. 14 developed an integrated PSO algorithm and GA to solve nonlinear
optimization problems. Additionally, Kuo and Han 15 presented a hybrid GA and PSO
algorithm for bilevel linear programming to solve a supply chain distribution problem.
Furthermore, Shelokar et al. 16 presented a hybrid PSO method and ant colony optimization
method for solving continuous optimization problems. Finally, Hu et al. 17 developed an
immune cooperative PSO algorithm for solving the fault-tolerant routing problem.
Compared to the above hybrid CI algorithms, this work optimizes the parameter
settings of an individual CI method by using another individual CI algorithm. A standard
PSO algorithm has certain limitations 17, 18. For instance, a PSO algorithm includes
many parameters that must be set, such as the cognitive parameter, social parameter, and
constriction coefficient. In practice, the optimal parameter settings of a PSO algorithm are
tuned based on trial and error and prior knowledge is required to successfully manipulate
the cognitive parameter, social parameter, and constriction coefficient. The exploration and
exploitative capabilities of a PSO algorithm are limited to optimum parameter settings.
Moreover, conventional PSO methods involve premature convergence that rapidly losses
diversity during optimization.
Fortunately, optimization of parameter settings for a conventional PSO algorithm
can be considered an unconstrained global optimization UGO problem, and the diversity
of candidate solutions of the PSO method can be increased using a multi-nonuniform
mutation operation 19. Moreover, the parameter manipulation of a GA and AIA method
is easy to implement without prior knowledge. Therefore, to overcome the limitations of
a standard PSO algorithm, this work develops two hybrid CI algorithms to solve CGO
problems efficiently. The first algorithm is a hybrid RGA and PSO RGA-PSO algorithm,
while the second algorithm is a hybrid AIA and PSO AIA-PSO algorithm. The proposed
RGA-PSO and AIA-PSO algorithms are considered to optimize two optimization problems
simultaneously. The UGO problem optimization of cognitive parameter, social parameter,
constriction coefficient, penalty parameter, and mutation probability of an internal PSO
algorithm based on a penalty function approach is optimized using external RGA and AIA
approaches, respectively. A CGO problem is then solved using the internal PSO algorithm.
The performances of the proposed RGA-PSO and AIA-PSO algorithms are evaluated using a
set of CGO problems e.g., six benchmark NLP and seven GPP optimization problems.
The rest of this paper is organized as follows. Section 2 describes the RGA, PSO
algorithm, AIA, and penalty function approaches. Section 3 then introduces the proposed
RGA-PSO and AIA-PSO algorithms. Next, Section 4 compares the experimental results of
the proposed RGA-PSO and AIA-PSO algorithms with those of various published individual
GAs and AIAs 9, 10, 20–22 and hybrid algorithms 23, 24. Finally, conclusions are drawn
in Section 5.
2. Related Works
2.1. Real-Coded Genetic Algorithm
GAs are stochastic global optimization methods based on the concepts of natural selection
and use three genetic operators, that is, selection, crossover, and mutation, to explore and
exploit the solution space. RGA outperforms binary-coded GA in solving continuous function
optimization problems 19. This work thus describes operators of a RGA 25.
4 Mathematical Problems in Engineering
2.1.1. Selection
A selection operation selects strong individuals from a current population based on their
fitness function values and then reproduces these individuals into a crossover pool. The
several selection operations developed include the roulette wheel, ranking, and tournament
methods 19, 25. This work uses the normalized geometric ranking method, as follows:
r−1
pj q 1 − q , j 1, 2, . . . , psRGA , 2.1
q
q psRGA , 2.2
1− 1−q
r individual ranking based on fitness value, where 1 represents the best, r 1, 2, . . . , psRGA ,
psRGA population size of the RGA.
2.1.2. Crossover
While exploring the solution space by creating new offspring, the crossover operation
randomly selects two parents from the crossover pool and then uses these two parents to
generate two new offspring. This operation is repeated until the psRGA /2 is satisfied. The
whole arithmetic crossover is easily implemented, as follows:
v1 β × v1 1 − β × v2 ,
2.3
v2 1 − β v1 β × v2 ,
where v1 and v2 parents, v1 and v2 offspring, β uniform random number in the interval
0, 1.5.
2.1.3. Mutation
Mutation operation can increase the diversity of individuals candidate solutions. Multi-
nonuniform mutation is described as follows:
⎧
⎨xcurrent,n xnu − xcurrent,n pert gRGA if U1 0, 1 < 0.5,
xtrial,n 2.4
⎩x
current,n − xcurrent,n − xn pert gRGA if U1 0, 1 ≥ 0.5,
l
where pertgRGA U2 0, 11 − gRGA /gmax,RGA 2 , perturbed factor, U1 0, 1 and U2 0, 1
uniform random variable in the interval 0, 1, gmax,RGA maximum generation of the RGA,
gRGA current generation of the RGA, xcurrent,n current decision variable xn , xtrial,n trial
candidate solution xn .
Mathematical Problems in Engineering 5
lb
vj,n gPSO 1 vj,n gPSO c1 r1j gPSO pj,n gPSO − xj,n gPSO
gb
c2 r2j gPSO pj,n gPSO − xj,n gPSO j 1, 2, . . . , psPSO , n 1, 2, . . . , N,
2.5
xj,n gPSO 1 xj,n gPSO vj,n gPSO 1 j 1, 2, . . . , psPSO , n 1, 2, . . . , N. 2.6
Shi and Eberhart 27 developed a modified PSO algorithm by incorporating an inertia
weight ωin into 2.7 to control the exploration and exploitation capabilities of a PSO
algorithm, as follows:
lb
vj,n gPSO 1 ωin vj,n gPSO c1 r1j gPSO pj,n gPSO − xj,n gPSO
gb
c2 r2j gPSO pj,n gPSO − xj,n gPSO j 1, 2, . . . , psPSO , n 1, 2, . . . , N.
2.7
A constriction coefficient χ was inserted into 2.8 to balance the exploration and
exploitation tradeoff 28–30, as follows:
lb
vj,n gPSO 1 χ vj,n gPSO ρ1 gPSO pj,n gPSO − xj,n gPSO
gb
ρ2 gPSO pj,n gPSO − xj,n gPSO j 1, 2, . . . , psPSO , n 1, 2, . . . , N,
2.8
6 Mathematical Problems in Engineering
where
2U3 0, 1
χ , 2.9
2−τ − ττ − 4
U3 0, 1 uniform random variable in the interval 0, 1, τ τ1 τ2 , τ1 c1 r1j , τ1 c2 r2j .
This work considers parameters ωin and χ to update the particle velocities, as follows:
lb
vj,n gPSO 1 χ ωin vj,n gPSO c1 r1j gPSO pj,n gPSO − xj,n gPSO
gb
c2 r2j gPSO pj,n gPSO − xj,n gPSO j 1, 2, . . . , psPSO , n 1, 2, . . . , N,
2.10
where ωin gmax,PSO − gPSO /gmax,PSO , increased gPSO value reduces the ωin , gmax,PSO
maximum generation of the PSO algorithm.
According to 2.10, the optimal values of parameters c1 , c2 , and χ are difficult to obtain
through a trial and error. This work thus optimizes these parameter settings by using RGA
and AIA approaches.
1 N 1
prj ,
N n1 ednj
2.11
xn∗ − xnj
dnj , j 1, 2, . . . , rs, n 1, 2, . . . , N,
xn∗
where prj probability that Ab j recognizes Ab∗ the best solution, xn∗ the best Ab∗ with
the highest Ab-Ag affinity, xnj decision variables xn of Ab j, rs repertoire population
size of the AIA.
The Ab∗ is recognized by other Abj in a current Ab repertoire. Large prj implies that
Abj can effectively recognize Ab∗ . The Abj with prj that is equivalent to or larger than the
threshold degree prt is reproduced to generate an intermediate Ab repertoire.
where pertgAIA {U5 0, 11 − gAIA /gmax,AIA }2 perturbation factor, gAIA current
generation of the AIA, gmax,AIA maximum generation number of the AIA, U4 0, 1 and
U5 0, 1 uniform random number in the interval 0, 1.
This operation has two tasks, that is, a uniform search and local fine-tuning.
⎧ 2 ⎫
⎨M K ⎬
Minimize fpseudo x, ρ fx ρ max 0, gm x hk x , 2.14
⎩m1 k1
⎭
where fpseudo x, ρ pseudo-objective function obtained using an original objective function
plus a penalty term, ρ penalty parameter.
Unfortunately, the penalty function scheme is limited by the need to fine-tune the
penalty parameter ρ 8. To overcome this limitation, this work attempts to find the optimum
ρ for each CGO problem using the RGA and AIA approaches. Additionally, to obtain high-
quality RGA-PSO and AIA-PSO solutions accurate to at least five decimal places for the
violation of each constraint to a specific CGO problem, the parameter ρ is within the search
space 1 × 109 , 1 × 1011 .
3. Method
3.1. RGA-PSO Algorithm
Figure 1 shows the pseudocode of the proposed RGA-PSO algorithm. The external RGA
approach is used to optimize the best parameter settings of the internal PSO algorithm, and
the internal PSO algorithm is employed to solve CGO problems.
Mathematical Problems in Engineering 9
gRGA ← gRGA 1
end
end
end
.
.
.
External RGA
Step 1 initialize the parameter settings. Parameter settings are given such as psRGA ,
crossover probability pc , mutation probability of the external RGA approach pm,RGA , the lower
and upper boundaries of these parameters c1 , c2 , χ, ρ, and the mutation probability of the
internal PSO algorithm pm,PSO . The candidate solutions individuals of the external RGA
represent the optimized parameters of the internal PSO algorithm. Finally, Figure 2 illustrates
the candidate solution of the external RGA approach.
10 Mathematical Problems in Engineering
.
.
.
Step 2 compute the fitness function value. The fitness function value fitnessj of the external
RGA approach is the best objective function value fx∗PSO obtained from the best solution
x∗PSO of each internal PSO algorithm execution, as follows:
fitnessj f x∗PSO , j 1, 2, . . . , psRGA . 3.1
Candidate solution j of the external RGA approach is incorporated into the internal
PSO algorithm, and a CGO problem is then solved using the internal PSO algorithm, which
is executed as follows.
M
2
fpseudo,j f xPSO,j ρ× max 0, gm xPSO,j , j 1, 2, . . . , psPSO . 3.2
m1
Step 3 update the particle velocity and position. The particle position and velocity can
be updated using 2.6 and 2.10, respectively.
Step 4 implement a mutation operation. The standard PSO algorithm lacks evolution
operations of GAs such as crossover and mutation. To maintain the diversity of
particles, this work uses the multi-nonuniform mutation operator defined by 2.4.
Step 5 perform an elitist strategy. A new particle swarm is created from internal step
3. Notably, fxPSO,j of a candidate solution j particle j in the particle swarm
Mathematical Problems in Engineering 11
Internal steps 2 to 5 are repeated until the gmax,PSO value of the internal PSO algo-
rithm is satisfied.
End
Step 3 implement selection operation. The parents in a crossover pool are selected using
2.1.
Step 4 perform crossover operation. In GAs, the crossover operation performs a global
search. Thus, the crossover probability pc usually exceeds 0.5. Additionally, candidate
solutions are created using 2.3.
Step 5 conduct mutation operation. In GAs, the mutation operation implements a local
search. Additionally, a solution space is exploited using 2.4.
Step 6 implement an elitist strategy. This work updates the population using an elitist
strategy. A situation in which the fitnessj of candidate solution j in the new population is
larger than that in the current population suggests that the weak candidate solution j is
replaced. Additionally, a situation in which the fitnessj of candidate solution j in the new
population is equal to or worse than that in the current population implies that the candidate
solution j in the current population survives. In addition to maintaining the strong candidate
solutions, this strategy eliminates weak candidate solutions.
External Steps 2 to 6 are repeated until the gmax,RGA value of the external RGA approach
is met.
External AIA
Step 1 initialize the parameter settings. Several parameters must be predetermined. These
include rs and the threshold for Ab-Ab recognition prt , as well as the lower and upper
boundaries of these parameters c1 , c2 , χ, ρ, and pm,PSO . Figure 5 shows the Ab and Ag
representation.
12 Mathematical Problems in Engineering
Step 1 create an initial particle swarm. An initial particle swarm is created based on psPSO
from xnl , xnu of a CGO problem. A particle represents a candidate solution of a
CGO problem.
Step 2 calculate the objective function value. Equation 3.2 is used as the pseudo-
objective function value of the internal PSO algorithm.
Step 3 update the particle velocity and position. Equations 2.6 and 2.10 can be used
to update the particle position and velocity.
Step 4 implement a mutation operation. The diversity of the particle swarm is increased
using 2.4.
Step 5 perform an elitist strategy. A new particle swarm population is generated from
internal step 3. Notably, fxPSO,j of a candidate solution j particle j in the
particle swarm is evaluated. Here, a pairwise comparison is made between the
fxPSO,j value of candidate solutions in the new and current particle swarms. The
elitist strategy guarantees that the best candidate solution is always preserved in
the next generation. The current particle swarm is updated to the particle swarm of
the next generation.
Internal steps 2 to 5 are repeated until the gmax,PSO value of the internal PSO
algorithm is satisfied.
End
Consistent with the Ab-Ag affinity metaphor, an Ab-Ag affinity is determined using 3.3, as
follows:
max affinityj −1 × f x∗PSO j 1, 2, . . . , rs. 3.3
Following the evaluation of the Ab-Ag affinities of Abs in the current Ab repertoire, the
Ab with the highest Ab-Ag affinity Ab∗ is chosen to undergo clonal selection operation in
external Step 3.
Step 3 perform clonal selection operation. To control the number of antigen-specific Abs,
2.11 is used.
Step 4 implement Ab-Ag affinity maturation. The intermediate Ab repertoire that is created
in external Step 3 is divided into two subsets. These Abs undergo somatic hypermutation
operation by using 2.12 when the random number is 0.5 or less. Notably, these Abs suffer
receptor editing operation using 2.13 when the random number exceeds 0.5.
Mathematical Problems in Engineering 13
Step 5 introduce diverse Abs. Based on the bone marrow operation, diverse Abs are created
to recruit the Abs suppressed in external Step 3.
External Steps 2–6 are repeated until the termination criterion gmax,AIA is satisfied.
4. Results
The 13 CGO problems were taken from other studies 1, 20, 21, 23, 34. The set of CGO
problems comprises six benchmark NLP problems TPs 1–4 and 12–13, and seven GPP
problems, in which TP 5 alkylation process design in chemical engineering, TP 6 optimal
reactor design, TP 12 a tension/compression string design problem, and TP 13 a pressure
vessel design problem are constrained engineering problems, were used to evaluate the
performances of the proposed RGA-PSO and AIA-PSO algorithms. In the appendix, the
objective function, constraints, boundary conditions of decision variables, and known global
14 Mathematical Problems in Engineering
Ag
Stimulate
Suppress
Candidate solution
x1 x2 x3 x4 x5
Ab
Ab
Paratope (recognizer)
Ab RGA-PSO/AIA-PSO
optimum for TPs 1−11 are described and the problem characteristics of TPs 5, 12, and 13 are
further detailed.
The proposed RGA-PSO and AIA-PSO algorithms were coded in MATLAB software
and executed on a Pentium D 3.0 GHz personal computer. Fifty independent runs were
conducted to solve each test problem TP. Numerical results were summarized, including
the best, median, mean, and worst results, as well as the standard deviation S.D. of objective
function values obtained using RGA-PSO and AIA-PSO solutions, mean computational CPU
times MCCTs, and mean absolute percentage error MAPE, as defined by
50
fx∗ − f xsstochastic /fx∗
MAPE
s1
× 100%, s 1, 2, . . . , 50, 4.1
50
where fx∗ value of the known global solution, fxsstochastic values obtained from
solutions of stochastic global optimization approaches e.g., RGA-PSO and AIA-PSO
algorithms.
Table 1 lists the parameter settings for the RGA-PSO and AIA-PSO algorithms, as
shown in Table 1.
Mathematical Problems in Engineering 15
Table 1: The parameter settings for the RGA-PSO and AIA-PSO algorithms.
pc 1
4.2. Comparison of the Results for the Proposed RGA-PSO and AIA-PSO
Algorithms with Those Obtained Using the Published Individual GA
and AIA Approaches and Hybrid Algorithms
Table 7 compares the numerical results of the proposed RGA-PSO and AIA-PSO algorithms
with those obtained using published individual GA and AIA approaches for TPs 1–4. In this
table, GA-1 is a GA with a penalty function methods, as used by Michalewicz 20. Notably,
16
Table 2: Numerical results of the proposed RGA-PSO and AIA-PSO algorithms for TPs 1–13.
Global
TP number Methods Best Mean Median Worst MAPE % S.D. MCCT sec P value
optimum
The proposed
24.323 24.568 24.521 24.831 1.078 0.13 432.39
1 24.306 RGA-PSO 0.643
The proposed
24.358 24.579 24.564 24.809 1.125 0.11 417.63
AIA-PSO
The proposed
−30665.539 −30665.539 −30665.539 −30665.534 1.34E − 06 7.36E − 04 349.55
2 −30665.539 RGA-PSO 0.478
The proposed
−30665.539 −30665.539 −30665.539 −30665.539 9.95E − 07 3.81E − 05 349.76
AIA-PSO
The proposed
680.632 680.640 680.639 680.658 1.46E − 03 5.55E − 03 398.72
3 680.630 RGA-PSO 0.839
The proposed
680.633 680.640 680.640 680.657 1.50E − 03 0.01 398.84
AIA-PSO
The proposed
−15 −15 −15 −15 1.27E − 06 1.27E − 06 466.00
4 −15 RGA-PSO 0.294
The proposed
−15 −15 −15 −15 1.20E − 10 2.64E − 11 450.28
AIA-PSO
The proposed
1227.2321 1228.8471 1228.3935 1231.6003 1.34E − 01 1.27 481.86
5 1227.1978 RGA-PSO 0.002∗
The proposed
1227.1598 1228.1800 1227.9762 1229.6590 8.02E − 02 0.71 461.30
AIA-PSO
The proposed
3.9521 3.9577 3.9566 3.9722 1.67E − 01 5E − 03 439.16
6 3.9511 RGA-PSO 0.682
The proposed
3.9516 3.9581 3.9569 3.9735 1.77E − 01 4.79E − 03 438.55
AIA-PSO
The proposed
−5.7398 −5.7398 −5.7398 −5.7398 2.54E − 04 1.13E − 05 325.73
7 −5.7398 RGA-PSO 0.358
Mathematical Problems in Engineering
Table 2: Continued.
Global
TP number Methods Best Mean Median Worst MAPE % S.D. MCCT sec P value
optimum
The proposed
−5.7398 −5.7398 −5.7398 −5.7398 2.42E − 04 7.74E − 06 329.12
AIA-PSO
The proposed
−83.2497 −83.2497 −83.2497 −83.2497 5.13E − 03 1.26E − 06 211.91
8 −83.254 RGA-PSO 0.291
The proposed
−83.2497 −83.2497 −83.2497 −83.2497 5.13E − 03 6.85E − 07 213.01
AIA-PSO
The proposed
−6.0441 −5.9883 −5.9968 −5.8051 0.991 0.05 440.72
9 −6.0482 RGA-PSO 0.577
Mathematical Problems in Engineering
The proposed
−6.0467 −5.9828 −5.9979 −5.8274 1.082 0.05 441.42
AIA-PSO
The proposed
6299.8374 6299.8412 6299.8419 6299.8428 2.52E − 03 1.42E − 03 224.13
10 6300 RGA-PSO 0.001∗
The proposed
6299.8395 6299.8420 6299.8423 6299.8425 2.51E − 03 6.83E − 04 223.91
AIA-PSO
The proposed
10122.4732 10122.4925 10122.4925 10122.6444 2.01E − 03 2.26E − 02 340.89
11 10122.6964 RGA-PSO 0.884
The proposed
10122.4852 10122.4920 10122.4927 10122.4931 2.02E − 03 1.85E − 03 338.37
AIA-PSO
The proposed
0.012692 0.012724 0.012721 0.012784 — 1.46E − 05 293.67
12 — RGA-PSO 0.014∗
The proposed
0.012667 0.012715 0.012719 0.012778 — 2.00E − 05 296.15
AIA-PSO
The proposed
5885.3018 5895.0381 5885.3326 6005.4351 — 24.33 291.17
13 — RGA-PSO 0.019∗
The proposed
5885.3249 5886.5426 5885.3323 5906.7404 — 4.54 292.16
AIA-PSO
The “—” denotes unavailable information, and ∗ represents that the mean values obtained using the RGA-PSO and AIA-PSO algorithms are statistically different.
17
18 Mathematical Problems in Engineering
Table 3: The best solutions obtained using the RGA-PSO algorithm from TPs 1–13.
Table 4: The best solutions obtained using the AIA-PSO algorithm from TPs 1–13.
Table 5: The best parameter settings of the best solution obtained using the RGA-PSO algorithm from TPs
1–13.
TP number χ c1 c2 ρ pm,PSO
1 0.73759998 1.52939998 1.72753568 28709280994 0.3750
2 0.67057537 0.45010388 2 1000000000 0.2709
3 0.75493696 0.36925226 1.91198475 54900420223 0.3856
4 0.45438391 1.45998477 1.25508289 1000000000 0.1
5 1 0.77483233 2 1000000000 0.5
6 0.70600559 0.82360083 0.91946627 26622414511 0.1619
7 0.74584341 0.95474855 1.17537957 86557786169 0.1958
8 1 0.69725160 1.53028620 1000000000 0.1675
9 0.41638718 0.46594542 1.97807798 12976330236 0.3686
10 0.25378610 0.59619170 0.83891277 1000000000 0.1
11 0.48183123 2 2 1000000000 0.1
12 0.76190087 0.1 1.16855713 100000000000 0.5
13 0.71783704 1.39420750 1.33590124 29071187026 0.2098
Table 6: The best parameter settings of the best solution obtained using the AIA-PSO algorithm from TPs
1–13.
TP number χ c1 c2 ρ pm,PSO
1 0.44896780 0.71709211 1.93302154 72140307110 0.4058
2 1 1.36225259 1.97905466 182831007 0.5
3 0.46599492 0.90346435 1.89697456 69125199709 0.2613
4 0.98124982 0.27882671 0.87437226 85199047430 0.1
5 0.99082484 0.1 1.11371788 4231387044 0.5
6 0.82869043 0.88773247 2 1387448505 0.5
7 0.87571243 1.89936723 0.74306310 94752095153 0.2194
8 0.93583844 1.53906226 1.30374874 22520728225 0.1798
9 1 0.22556712 1.52263349 67578847151 0.4507
10 1 1.93003999 0.1 1351461763 0.5
11 1 1.51209364 1.63826995 1811017789 0.5
12 0.52068067 0.1 2 81914376144 0.1
13 0.82395890 1.60107152 0.93611204 17767111886 0.1813
GA-2 represents a GA with a penalty function, but without any penalty parameter, as used
by Deb 21. Also, GA-3 is an RGA with a static penalty function, as developed by Wu and
Chung 9. Notably, AIA-1 is an AIA method called CLONALG, as proposed by Cruz-Cortés
et al. 22. Finally, AIA-2 is an AIA approach based on an adaptive penalty function, as
developed by Wu 10. The numerical results of GA-1, GA-2, and AIA-1 methods for solving
TPs 1–4 were collected from the published literature 20–22. Furthermore, the GA-1, GA-2,
and AIA-1 approaches were executed under 350,000 objective function evaluations. To fairly
compare the performances of the proposed hybrid CI algorithms and the individual GA and
AIA approaches, the GA-3, AIA-2, the internal PSO algorithm of RGA-PSO method, and the
internal PSO algorithm of AIA-PSO method were independently executed 50 times under
350,000 objective function evaluations for solving TPs 1–4.
For solving TP 1, the median values obtained using the RGA-PSO and AIA-PSO
algorithms are smaller than those obtained using the GA-1, GA-3, and AIA-2 approaches, and
Mathematical Problems in Engineering 21
Table 7: Comparison of the results of the proposed RGA-PSO and AIA-PSO algorithms and those of the
published individual GA and AIA approaches for TPs 1–4.
TP Global
Methods Best Mean Median Worst MAPE%
number optimum
GA-1 20 24.690 — 29.258 36.060 —
GA-2 21 24.372 — 24.409 25.075 —
GA-3 9 24.935 27.314 27.194 33.160 12.377
1 24.306 AIA-1 22 24.506 25.417 — 26.422 —
AIA-2 10 24.377 24.669 24.663 24.988 1.495
The proposed
24.323 24.568 24.521 24.831 1.078
RGA-PSO
The proposed
24.358 24.579 24.564 24.809 1.125
AIA-PSO
GA-1 20 — — — — —
GA-2 21 — — — — —
GA-3 9 −30665.526 −30662.922 −30664.709 −30632.445 8.53E − 03
2 −30665.539 AIA-1 22 −30665.539 −30665.539 — −30665.539 —
AIA-2 10 −30665.539 −30665.526 −30665.527 −30665.506 4.20E − 05
The proposed
−30665.539 −30665.539 −30665.539 −30665.534 1.34E − 06
RGA-PSO
The proposed
−30665.539 −30665.539 −30665.539 −30665.539 9.95E − 07
AIA-PSO
GA-1 20 680.642 — 680.718 680.955 —
GA-2 21 680.634 — 680.642 680.651 —
GA-3 9 680.641 680.815 680.768 681.395 2.72E − 02
3 680.630 AIA-1 22 680.631 680.652 — 680.697 —
AIA-2 10 680.634 680.653 680.650 680.681 3.45E − 03
The proposed
680.632 680.640 680.639 680.658 1.46E − 03
RGA-PSO
The proposed
680.633 680.640 680.640 680.657 1.50E − 03
AIA-PSO
GA-1 20 −15 −15 −15 −15 —
GA-2 21 — — — — —
GA-3 9 −13.885 −12.331 −12.267 −10.467 17.795
4 −15 AIA-1 22 −14.987 −14.726 — −12.917 —
AIA-2 10 −14.998 −14.992 −14.992 −14.988 5.08E − 02
The proposed
−15 −15 −15 −15 1.27E − 06
RGA-PSO
The proposed
−15 −15 −15 −15 1.20E − 10
AIA-PSO
The “—” denotes unavailable information.
the worst values obtained using RGA-PSO and AIA-PSO algorithms are smaller than those
obtained using the GA-1, GA-2, GA-3, AIA-1, and AIA-2 approaches. For solving TP 2, the
median and worst values obtained using the RGA-PSO and AIA-PSO algorithms are smaller
than those obtained using the GA-3 method. For solving TP 3, the median and worst values
22 Mathematical Problems in Engineering
GA-3 versus GA-3 versus GA-3 versus AIA-2 versus AIA-2 versus RGA-PSO versus
TP number
AIA-2 RGA-PSO AIA-PSO RGA-PSO AIA-PSO AIA-PSO
1 0.000∗ 0.000∗ 0.000∗ 0.000∗ 0.001∗ 0.643
∗ ∗ ∗ ∗ ∗
2 0.003 0.003 0.003 0.000 0.000 0.478
∗ ∗ ∗ ∗ ∗
3 0.000 0.000 0.000 0.000 0.000 0.839
∗ ∗ ∗ ∗ ∗
4 0.000 0.000 0.000 0.000 0.000 0.294
∗
Represents that the mean values obtained using two algorithms are statistically different.
obtained using the RGA-PSO and AIA-PSO algorithms are smaller than those obtained using
the GA-1 and GA-3 approaches. For solving TP 4, the median and worst values obtained
using the RGA-PSO and AIA-PSO algorithms are smaller than those obtained using the GA-
3 method, and the worst values obtained using the RGA-PSO and AIA-PSO algorithms are
smaller than those obtained using the AIA-1 approach. Moreover, the GA-3 method obtained
the worst MAPE% for TP 1 and TP 4. Table 8 lists the results of the t-test for the GA-3,
AIA-2, RGA-PSO, and AIA-PSO methods. This table indicates that the mean values of the
RGA-PSO, and AIA-PSO algorithms are not statistically significant, since P values are larger
than a significant level 0.05, and the mean values between GA-3 versus AIA-2, GA-3 versus
RGA-PSO, GA-3 versus AIA-PSO, AIA-2 versus RGA-PSO, and AIA-2 versus AIA-PSO are
statistically significant. According to Tables 7 and 8, the mean values obtained using the RGA-
PSO and AIA-PSO algorithms are better than those of obtained using the GA-3 and AIA-1
methods for TPs 1–4.
Table 9 compares the numerical results obtained using the proposed RGA-PSO and
AIA-PSO algorithms and those obtained using AIA-2 and GA-3 for solving TPs 5–13.
The AIA-2, GA-3, the internal PSO algorithm of the RGA-PSO approach, and the internal
PSO algorithm of AIA-PSO approach were independently executed 50 times under 300,000
objective function evaluations. Table 9 shows that MAPE% obtained using the proposed
RGA-PSO and AIA-PSO algorithms is close to 1%, or smaller than 1% for TPs 5–11, indicating
that the proposed RGA-PSO and AIA-PSO algorithms can converge to global optimum for
TPs 5–11. Moreover, the worst values obtained using the RGA-PSO and AIA-PSO algorithms
are significantly smaller than those obtained using the GA-3 method for TPs 5, 6, 11, and
13. Additionally, the worst values obtained using the RGA-PSO and AIA-PSO algorithms are
smaller than those obtained using the AIA-2 method for TPs 5, 6, and 13.
Table 10 summarizes the results of the t-test for TPs 5–13. According to Tables 9 and 10,
the mean values of the RGA-PSO and AIA-PSO algorithms are smaller than those of the GA-3
approach for TPs 5, 6, 7, 8, 9, 10, 11, and 13. Moreover, the mean values obtained using the
RGA-PSO and AIA-PSO algorithms are smaller than those of the AIA-2 approach for TPs 6,
7, 8, 10, and 12. Totally, according to Tables 7−10, the performances of the hybrid CI methods
are superior to those of individual GA and AIA methods.
The TPs 12 and 13 have been solved by many hybrid algorithms. For instance,
Huang et al. 23 presented a coevolutionary differential evolution CDE that integrates
a coevolution mechanism and a DE approach. Zahara and Kao 24 developed a hybrid
Nelder-Mead simplex search method and a PSO algorithm NM-PSO. Table 11 compares
Table 9: Comparison of the numerical results of the proposed RGA-PSO and AIA-PSO algorithms and those of the published individual AIA and RGA for
TPs 5–13.
TP number Global optimum Methods Best Mean Median Worst MAPE%
AIA-2 10 1227.3191 1228.6097 1227.8216 1239.5205 1.15E − 01
GA-3 9 1228.5118 1279.3825 1263.2589 1481.3710 4.266
5 1227.1978
The proposed RGA-PSO 1227.2321 1228.8471 1228.3935 1231.6003 1.34E − 01
The proposed AIA-PSO 1227.1598 1228.1800 1227.9762 1229.6590 8.02E − 02
AIA-2 10 3.9518 4.1005 4.0460 4.2997 3.781
GA-3 9 3.9697 4.1440 4.1358 4.3743 4.881
6 3.9511
The proposed RGA-PSO 3.9521 3.9577 3.9566 3.9722 1.67E − 01
The proposed AIA-PSO 3.9516 3.9581 3.9569 3.9735 1.77E − 01
AIA-2 10 −5.7398 −5.7398 −5.7398 −5.7396 6.03E − 04
GA-3 9 −5.7398 −5.7375 −5.7383 −5.7316 3.98E − 02
7 −5.7398
The proposed RGA-PSO
Mathematical Problems in Engineering
Table 9: Continued.
TP number Global optimum Methods Best Mean Median Worst MAPE%
AIA-2 10 5885.5312 5900.4860 5894.7929 6014.2198 —
GA-3 9 5885.3076 5943.4714 5897.2904 6289.7314 —
13 —
The proposed RGA-PSO 5885.3018 5895.0381 5885.3326 6005.4351 —
The proposed AIA-PSO 5885.3310 5886.5426 5885.3323 5906.7404 —
The “—” denotes unavailable information.
Mathematical Problems in Engineering
Mathematical Problems in Engineering 25
GA-3 versus GA-3 versus GA-3 versus AIA-2 versus AIA-2 versus RGA-PSO
TP number
AIA-2 RGA-PSO AIA-PSO RGA-PSO AIA-PSO versus AIA-PSO
5 0.000∗ 0.000∗ 0.000∗ 0.485 0.163 0.002∗
6 0.112 0.000∗ 0.000∗ 0.000∗ 0.000∗ 0.682
7 0.000∗ 0.000∗ 0.000∗ 0.000∗ 0.000∗ 0.358
8 0.000∗ 0.000∗ 0.000∗ 0.000∗ 0.000∗ 0.291
∗ ∗ ∗
9 0.000 0.000 0.000 0.516 0.814 0.577
∗ ∗ ∗ ∗ ∗
10 0.000 0.000 0.000 0.009 0.001 0.001∗
11 0.000∗ 0.000∗ 0.000∗ 0.069 0.013∗ 0.884
∗ ∗ ∗
12 0.049 0.389 0.178 0.000 0.007 0.014∗
13 0.003∗ 0.001∗ 0.000∗ 0.240 0.000∗ 0.019∗
∗
Represents that the mean values obtained using two algorithms are statistically different.
Table 11: Comparison of the numerical results of the proposed RGA-PSO and AIA-PSO algorithms and
those of the published hybrid algorithms for TPs 12-13.
the numerical results of the CDE, NM-PSO, RGA-PSO, and AIA-PSO methods for solving TPs
12−13. The table indicates that the best, mean, and worst values obtained using the NM-PSO
method are superior to those obtained using the CDE, RGA-PSO, and AIA-PSO approaches
for TP 12. Moreover, the best, mean, and worst values obtained using the AIA-PSO algorithm
are better than those of the CDE, NM-PSO, and RGA-PSO algorithms.
According to the No Free Lunch theorem 35, if algorithm A outperforms algorithm
B on average for one class of problems, then the average performance of the former must
be worse than that of the latter over the remaining problems. Therefore, it is unlikely that
any unique stochastic global optimization approach exists that performs best for all CGO
problems.
26 Mathematical Problems in Engineering
1 Parameter manipulation of the internal PSO algorithm is based on the solved CGO
problems. Owing to their ability to efficiently solve an UGO problem, the external
RGA and AIA approaches are substituted for trial and error to manipulate the
parameters χ, c1 , c2 , ρ, and pm,PSO .
2 Besides obtaining the optimum parameter settings of the internal PSO algorithm,
the RGA-PSO and AIA-PSO algorithms can yield a global optimum for a CGO
problem.
The proposed RGA-PSO and AIA-PSO algorithms have the following limitations.
1 The proposed RGA-PSO and AIA-PSO algorithms increase the computational CPU
time, as shown in Table 2.
2 The proposed RGA-PSO and AIA-PSO algorithms are designed to solve CGO prob-
lems with continuous decision variables xn . Therefore, the proposed algorithms
cannot be applied to manufacturing problems such as job shop scheduling and
quadratic assignment problems combinatorial optimization problems.
5. Conclusions
This work presents novel RGA-PSO and AIA-PSO algorithms. The synergistic power of
the RGA with PSO algorithm and the AIA with PSO algorithm is also demonstrated by
using 13 CGO problems. Numerical results indicate that, in addition to converging to
a global minimum for each test CGO problem, the proposed RGA-PSO and AIA-PSO
algorithms obtain the optimum parameter settings of the internal PSO algorithm. Moreover,
the numerical results obtained using the RGA-PSO and AIA-PSO algorithms are superior to
those obtained using alternative stochastic global optimization methods such as individual
GA and AIA approaches. The RGA-PSO and AIA-PSO algorithms are highly promising
stochastic global optimization approaches for solving CGO problems.
Mathematical Problems in Engineering 27
Appendices
A. TP 1 [20, 21]
TP 1 has ten decision variables, eight inequality constraints, and 20 boundary conditions, as
follows:
− 10 ≤ xn ≤ 10, n 1, 2, . . . , 10.
fx∗ 24.306.
28 Mathematical Problems in Engineering
B. TP 2 [21]
TP 2 involves five decision variables, six inequality constraints, and ten boundary conditions,
as follows:
C. TP 3 [20, 21]
TP 3 has seven decision variables, four inequality constraints, and 14 boundary conditions,
as follows:
Minimize fx x1 − 102 5x2 − 122 x34 3x4 − 112 10x56
− 10 ≤ xn ≤ 10, n 1, 2, . . . , 7.
Mathematical Problems in Engineering 29
D. TP 4 [20, 21]
TP 4 involves 13 decision variables, nine inequality constraints, and 26 boundary conditions,
as follows:
4 4 13
Minimize fx 5 xn − 5 xn2 − xn
n1 n1 n5
0 ≤ xn ≤ 1, n 1, 2, . . . , 9,
0 ≤ x13 ≤ 1.
by reacting it with isobutane in the presence of acid. The decision variables xn are olefin
feed rate barrels/day x1 , acid addition rate thousands of pounds/day x2 , alkylate yield
barrels/day x3 , acid strength x4 , motor octane number x5 , external isobutane-to-olefin
ration x6 , and F-4 performance number x7 :
g4 x ω17 x5−1 ω18 x5−1 x6 ω19 x4 x5−1 − ω20 x5−1 x62 ≤ 1,
g6 x ω24 x7−1 ω25 x2 x3−1 x7−1 − ω26 x2 x3−1 x4−1 x7−1 ≤ 1,
where ωl l 1, 2, . . . , 44 denotes positive parameters given in Table 12. The global solution
to TP 5 is
l ωl l ωl l ωl
1 1.715 16 0.19120592E − 1 31 0.00061000
2 0.035 17 0.56850750E 2 32 0.0005
3 4.0565 18 1.08702000 33 0.81967200
4 10.000 19 0.32175000 34 0.81967200
5 3000.0 20 0.03762000 35 24500.0
6 0.063 21 0.00619800 36 250.0
7 0.59553571E − 2 22 0.24623121E 4 37 0.10204082E − 1
8 0.88392857 23 0.25125634E 2 38 0.12244898E − 4
9 0.11756250 24 0.16118996E 3 39 0.00006250
10 1.10880000 25 5000.0 40 0.00006250
11 0.13035330 26 0.48951000E 6 41 0.00007625
12 0.00660330 27 0.44333333E 2 42 1.22
13 0.66173269E − 3 28 0.33000000 43 1.0
14 0.17239878E − 1 29 0.02255600 44 1.0
15 0.56595559E − 2 30 0.00759500
Minimize fx 0.4x10.67 x7−0.67 0.4x20.67 x8−0.67 10 − x1 − x2
x∗ 6.4747, 2.2340, 0.6671, 0.5957, 5.9310, 5.5271, 1.0108, 0.4004, fx∗ 3.9511. F.2
32 Mathematical Problems in Engineering
G. TP 7 [1]
TP 7 has four decision variables, two nonconvex inequality constraints, and eight boundary
conditions, as follows:
Minimize fx −x1 0.4x10.67 x3−0.67
H. TP 8 [1]
TP 8 contains three decision variables subject to one nonconvex inequality constraint and six
boundary conditions, as follows:
Minimize fx 0.5x1 x2−1 − x1 − 5x2−1
H.1
s.t. g1 x 0.01x2 x3−1 0.01x1 0.0005x1 x3 ≤ 1, 1 ≤ xn ≤ 100, n 1, 2, 3.
I. TP 9 [1]
TP 9 contains eight decision variables subject to four nonconvex inequality constraints and
16 boundary conditions, as follows:
Minimize fx −x1 − x5 0.4x10.67 x3−0.67 0.4x50.67 x7−0.67
x∗ 6.4225, 0.6686, 1.0239, 5.9399, 2.2673, 0.5960, 0.4029, 5.5288, fx∗ −6.0482. I.2
J. TP 10 [34]
TP 10 contains three decision variables subject to one nonconvex inequality constraint and
six boundary conditions, as follows:
Minimize fx 5x1 50000x1−1 20x2 72000x2−1 10x3 144000x3−1
J.1
s.t. g1 x 4x1−1 32x2−1 120x3−1 ≤ 1, 1 ≤ xn ≤ 1000, n 1, 2, 3.
K. TP 11 [1, 34]
TP 11 involves five decision variables, six inequality constraints, and ten boundary condi-
tions, as follows:
Minimize g0 x 5.3578x32 0.8357x1 x5 37.2392x1
x23 x3
s.t. g1 x 1 − ≤ 0,
71785x14
4x22 − x1 x2 1
g2 x − 1 ≤ 0, L.1
12566 x2 x13 − x14 5108x12
140.45x1
g3 x 1 − ≤ 0,
x22 x3
x1 x2
g4 x − 1 ≤ 0, 0.05 ≤ x1 ≤ 2, 0.25 ≤ x2 ≤ 1.3, 2 ≤ x3 ≤ 15.
1.5
Acknowledgment
The author would like to thank the National Science Council of the Republic of China,
Taiwan, for financially supporting this research under Contract no. NSC 100-2622-E-262-006-
CC3.
References
1 C. A. Floudas, P. M. Pardalos, C. S. Adjiman, and W. R. Esposito, Handbook of Test Problems in Local and
Global Optimization, Kluwer, Boston, Mass, USA, 1999.
Mathematical Problems in Engineering 35
27 Y. Shi and R. Eberhart, “A modified particle swarm optimizer,” in Proceedings of the IEEE International
Conference on Evolutionary Computation, pp. 69–73, Anchorage, Alaska, USA, 1998.
28 M. Clerc, “The swarm and the queen: towards a deterministic and adaptive particle swarm optimiza-
tion,” in Proceedings of the IEEE Congress on Evolutionary Computation, pp. 1951–1957, Washington, DC,
USA, 1999.
29 M. Clerc and J. Kennedy, “The particle swarm-explosion, stability, and convergence in a multidimen-
sional complex space,” IEEE Transactions on Evolutionary Computation, vol. 6, no. 1, pp. 58–73, 2002.
30 A. P. Engelbrecht, Fundamentals of Computational Swarm Intelligence, John Wiley & Sons, 2005.
31 N. K. Jerne, “Idiotypic networks and other preconceived ideas,” Immunological Reviews, vol. 79, pp.
5–24, 1984.
32 L. N. de Castro and F. J. Von Zuben, “Artificial Immune Systems: Part I—Basic Theory and Appli-
cations,” FEEC/Univ. Campinas, Campinas, Brazil, 1999, ftp://ftp.dca.fee.unicamp.br/pub/docs/
vonzuben/tr dca/trdca0199.pdf.
33 C. A. Coello Coello, “Theoretical and numerical constraint-handling techniques used with evolution-
ary algorithms: a survey of the state of the art,” Computer Methods in Applied Mechanics and Engineering,
vol. 191, no. 11-12, pp. 1245–1287, 2002.
34 M. J. Rijckaert and X. M. Martens, “Comparison of generalized geometric programming algorithms,”
Journal of Optimization Theory and Applications, vol. 26, no. 2, pp. 205–242, 1978.
35 D. H. Wolpert and W. G. Macready, “No free lunch theorems for optimization,” IEEE Transactions on
Evolutionary Computation, vol. 1, no. 1, pp. 67–82, 1997.
Hindawi Publishing Corporation
Mathematical Problems in Engineering
Volume 2012, Article ID 726564, 17 pages
doi:10.1155/2012/726564
Research Article
A Hybrid Algorithm Based on ACO and PSO for
Capacitated Vehicle Routing Problems
Copyright q 2012 Yucheng Kao et al. This is an open access article distributed under the Creative
Commons Attribution License, which permits unrestricted use, distribution, and reproduction in
any medium, provided the original work is properly cited.
The vehicle routing problem VRP is a well-known combinatorial optimization problem. It has
been studied for several decades because finding effective vehicle routes is an important issue of
logistic management. This paper proposes a new hybrid algorithm based on two main swarm
intelligence SI approaches, ant colony optimization ACO and particle swarm optimization
PSO, for solving capacitated vehicle routing problems CVRPs. In the proposed algorithm, each
artificial ant, like a particle in PSO, is allowed to memorize the best solution ever found. After
solution construction, only elite ants can update pheromone according to their own best-so-far
solutions. Moreover, a pheromone disturbance method is embedded into the ACO framework to
overcome the problem of pheromone stagnation. Two sets of benchmark problems were selected to
test the performance of the proposed algorithm. The computational results show that the proposed
algorithm performs well in comparison with existing swarm intelligence approaches.
1. Introduction
The vehicle routing problem VRP is a well-known combinatorial optimization problem in
which the computational complexity is NP-hard. The capacitated vehicle routing problem
is one of the variants of VRPs. The objective of CVRPs is to minimize the total traveling
distance of vehicles which serve a set of customers. The following constraints are considered
in a typical CVRP: each route is a tour which starts from a depot, visits a subset of the
customers, and ends at the same depot; each customer must be assigned to exactly one of
the vehicles; each customer has its own demand and the total demand of customers assigned
to a vehicle must not exceed the vehicle capacity. In the past decades, researchers proposed
different strategies to solve the CVRP. One of them is to cluster customers into different routes
and then to arrange the visiting sequence for each route. The objective function value will
2 Mathematical Problems in Engineering
be apparently influenced by the results of customer clustering and sequencing. For more
detailed descriptions of vehicle routing problems, the reader may refer to the articles by
Laporte 1, Osman 2, and Cordeau et al. 3.
Because the CVRP is an NP-hard problem 4, the optimal solution of a large-size
instance cannot be found within a reasonable time. To overcome this difficulty, many classical
heuristic methods and metaheuristic methods are proposed in the past five decades. Some
metaheuristic algorithms can provide competitive solutions to the CVRP, such as Simulated
Annealing SA, Tabu Search TS, and Genetic Algorithm GA. Metaheuristic algorithms
have some advantages, for example, the abilities to escape from local optima through
stochastic search, to speed convergence using solution replacement, to guide the search
direction with the elitist strategy, and so on. The following paragraph gives a brief review
of some articles which use these metaheuristic algorithms to solve CVRPs.
Barbarosoglu and Ozgur 5 designed a TS-based algorithm using a new neighbor-
hood generation procedure for the single-depot vehicle routing problems. The neighbors
are defined by using two procedures: the first one ignores the scattering patterns of
customer locations, and the second one considers the underlying clustering of customer
locations. Baker and Ayechew 6 put forward a hybrid of GA algorithms with neighborhood
search methods. The pure GA has three specific processes: initialization, reproduction, and
replacement. The neighborhood search methods are used to accelerate the convergence
of GA. Computational results showed that this approach is competitive with published
results obtained using TS and SA. Lin et al. 7 proposed a hybrid algorithm which takes
the advantages of SA and TS. In their paper, SA is used to adjust the probability of
accepting worse solutions according to the extent of solution improvement and the annealing
temperature, while TS is embedded in the framework of SA to avoid cycling to some extent
while searching for neighborhood.
In the recent ten years, swarm intelligence, a new category of metaheuristics, has
emerged and attracted researchers’ attention. Swarm intelligence mimics the social behavior
of natural insects or animals to solve complex problems. Some commonly used swarm
intelligence algorithms for the solution of the CVRP include ant colony optimization ACO,
particle swarm optimization PSO and artificial bee colony 8.
ACO is a population-based swarm intelligence algorithm and was proposed by Dorigo
and Gambardella 9. This algorithm has been inspired by the foraging behavior of real ant
colonies and originally designed for the traveling salesman problem TSP. The artificial
ants use pheromone laid on trails as an indirect communication medium to guide them
to construct complete solution routes step by step. More pheromone deposits on better
routes attract more ants for later search. This effect is called dynamic positive feedback and
helps speed convergence of ACO. Recently, some researchers have studied vehicle routing
problems using ACO algorithms. Applying ACO to the CVRP is quite natural, since we can
view ant nests as depots, artificial ants as vehicles, foods as customers, and trails as routes.
Some of the relevant papers are briefly reviewed as follows.
Bell and McMullen 10 made modifications of the ACO algorithm in order to solve
the vehicle routing problem. They used multiple ant colonies to search vehicle routes.
Each vehicle route is marked with unique pheromone deposits by an ant colony, but the
communication among ant colonies is limited. Later, Liu and Cai 11 proposed a new
multiple ant colonies technique, which allows ant colonies to communicate with each other
in order to escape from local optima. Chen and Ting 12 developed an improved ant colony
system algorithm, in which pheromone trails will be reset to initial values for restarting
the search if the solution is not improved after a given number of iterations. Zhang and
Mathematical Problems in Engineering 3
Tang 13 hybridized the solution construction mechanism of ACO with scatter search SS.
The algorithm stores better solutions in a reference set. Some new solutions are generated
by combining solutions selected from the reference set, and some are produced by using
the conventional ACO method. Yu et al. 14 also developed an improved ant colony
optimization for vehicle routing problems. Their algorithm uses the ant-weight strategy to
update pheromone in terms of solution quality and the contribution of each edge to the
solution. Lee et al. 15 proposed an enhanced ACO algorithm for the CVRP. Their algorithm
adopts the concept of information gain to measure the variation of pheromone concentrations
and hence to dynamically adjust the value of heuristic parameter β which determines the
importance of heuristic value η at different iterations.
PSO is also a population-based swarm intelligence algorithm and was originally
proposed by Kennedy and Eberhart 16. PSO is inspired by social behavior of bird flocking.
It has been shown that PSO can solve continuous optimization problems very well. In the
PSO, solution particles try to move to better locations in the solution space. The movements
of particles are guided by the individuals’ and the swarm’s best positions. PSO can converge
very fast due to its two unique mechanisms: memorizing personal best experiences Pbest
and information sharing of global best experiences Gbest . Note that the Gbest solution of the
particle swarm is equal to the Pbest solution of the best particle. In 1998, Shi and Eberhart 17
enhanced PSO by adding the concept of inertia weight, which becomes the standard version
of PSO.
PSO being originally developed for continuous optimization problems, a special
solution representation or solution conversion should be designed first in order to solve
CVRPs. Chen et al. 18 first proposed a PSO-based algorithm to solve the CVRP. In
their approach, each iteration has two main steps: customers are first clustered by using
a discrete PSO algorithm DSPO and then sequenced by applying a SA algorithm. Due
to its long solution strings, their approach takes much computational time in solving large
scale problems. To improve Chen et al.’s work, Kao and Chen 19 addressed a new
solution representation and solved the CVRP with a combinatorial PSO algorithm. Ai and
Kachitvichyanukul 20 presented two solution representations for solving CVRPs. For
example, in their second solution representation SR-2, each vehicle is represented in three
dimensions, with two for the reference point and one for the vehicle coverage radius. SR-
2 employs these points and radius to construct vehicle routes. The particle solutions are
adjusted by using a continuous PSO. Marinakis et al. 21 proposed a hybrid PSO algorithm
to tackle large-scale vehicle routing problems. Their proposed algorithm combines a PSO
algorithm with three heuristic methods, with the first for particle initialization, the second
for solution replacement, and the third for local search.
This study proposes a new hybrid algorithm for the capacitated vehicle routing
problem, which is based on the framework of ACO and is hybridized with the merits of PSO.
The reasons why ACO and PSO, rather than SA, TS, and GA, are adopted in the proposed
algorithm are given as follows. First, SA and TS perform the so-called single-starting-point
search and thus their performance relies highly on a good initial solution. However, GA,
ACO, and PSO are all population-based algorithms and can start the search from multiple
points. Their initial solutions have little influence on their performance. Thus, we consider
adopting the population-based algorithms to solve CVRPs. Second, ACO and PSO have
memory that enables the algorithms to retain knowledge of good solutions, while the genetic
operators of GA may destroy previously learned knowledge when producing the offspring.
In view of these two considerations, we select ACO and PSO as the solution approach for this
paper.
4 Mathematical Problems in Engineering
In the past, most relevant papers adopted either ACO 10–15 or PSO 18–21 alone
without trying to use both in combination for solving CVRPs. In this paper, we try to
integrate ACO with PSO to develop a new hybrid approach which can take advantage of
both algorithms. That is, the proposed algorithm uses the solution construction approach of
ACO to cluster customers and build routes at the same time and use the short-term memory
inspired by PSO to speed convergence through laying pheromone on the routes of Gbest and
Pbest solutions.
Like most ACO-based algorithms, the proposed hybrid algorithm also faces the
limitation of pheromone stagnation, which results in premature convergence. To solve this
problem, Shuang et al. 22 employed the mechanism of PSO to modify the pheromone
updating rules of ACO. Their proposed algorithm is called PS-ACO and is used to solve
traveling salesman problems TSPs. PS-ACO can improve the performance of ACO to some
extent, but it may still be trapped in local optima due to the overaccumulation of pheromone
on some edges when solving more complicated problems like CVRPs. To attain a high degree
of search accuracy, this paper proposes a pheromone disturbance approach to overcome the
problem of pheromone stagnation. The remainder of this paper is organized as follows.
Section 2 defines the mathematical formulation of the CVRP. The proposed methodology
is described in Section 3. Section 4 presents computational results. Finally, conclusions are
drawn in the last section.
di : demand of customer i, d0 0;
Xijk : 0-1 variable, where Xijk 1 if the edge from customer i to customer j is traveled by
vehicle k; otherwise, Xijk 0. Note that i /
j;
p: penalty coefficient;
Objective function
N
N
K
Minimize Cij Xijk , 2.1
i0 j0 k1
K
N
subject to Xijk 1, j 1, 2, . . . , N, 2.2
k1 i0
K
N
Xijk 1, i 1, 2, . . . , N, 2.3
k1 j0
N
N
k
Xiu − k
Xuj 0, k 1, 2, . . . , K; u 1, 2, . . . , N, 2.4
i0 j0
N
N
Xijk di ≤ Q, k 1, 2, . . . , K, 2.5
i0 j0
N
N
Xijk Cij Si ≤ T, k 1, 2, . . . , K, 2.6
i0 j0
N
N
Xijk Xjik ≤ 1, i 0; k 1, 2, . . . , K, 2.7
j1 j1
Xijk ≤ |R| − 1, R ⊆ {1, . . . , N}, 2 ≤ |R| ≤ N − 1; k 1, 2, . . . , K, 2.8
i,j∈R
Equation 2.1 is the objective function of the CVRP. Equations 2.2 and 2.3 ensure
that each customer can be served by only one vehicle. Equation 2.4 maintains the continuity
at each node for every vehicle. Equation 2.5 ensures that the total customer demand of
a vehicle cannot exceed its maximum capacity. Similarly, 2.6 ensures that the total route
distance of a vehicle cannot exceed its route length limit. Equation 2.7 makes sure that
every vehicle can be used at most once and must start and end at the depot. The subtour
elimination constraints are given in 2.8. Equation 2.9 is the integrality constraint.
3. PACO Algorithm
This section describes the proposed solution algorithm to the capacitated vehicle routing
problem. The algorithm, called PACO, hybridizes the solution construction mechanism of
6 Mathematical Problems in Engineering
ACO and the short-term memory mechanism of PSO to find optimal or near optimal vehicle
routes.
and length constraints see 2.5 and 2.6. Two penalty functions are added to the original
objective function, as defined in 3.1, one for excess vehicle capacity and the other for
excess route length. These penalty functions increase the objective function value of infeasible
solutions so as to prefer feasible solution to be selected as elite ants which are allowed to
perform local search and to update pheromone on their routes see Sections 3.5 and 3.6.
⎡ ⎧ ⎫ ⎧ ⎫⎤
N N K K ⎨ N N ⎬ ⎨ N N
⎬
Minimize f Cij Xijk p ⎣max 0, Xijk di −Q max 0, Xijk Cij Si − T ⎦.
i0 j0 k1 k1
⎩ i0 j0 ⎭ ⎩ i0 j0 ⎭
3.1
Step 3 Local search. The top r best ants perform local search.
Step 4. Update the Gbest and Pbest solutions of ants and select r elite ants.
Step 6 Pheromone disturbance. Randomly disturb the pheromone matrix, reset the Pbest
solutions for some ants, and go to Step 8.
Step 7 Pheromone updating. Update the pheromone matrix based on the Pbest solutions of
elite ants.
Step 8. If iteration number reaches the maximum number of iterations MaxIte, go to Step 9;
otherwise, go to Step 2 for the next iteration.
Begin
Initialization
Solution construction
Local search
fGbest (t + 1) < No
fGbest (t)?
NG = NG + 1
Yes
NG = 0
No
NG = w ?
Yes
Pheromone Pheromone
updating disturbance
Pbest
resetting
No Reach
NG = 0
maxIter?
Yes
Output G best
End
PACO can cluster customers into K vehicles and arrange vehicle visiting sequences at the
same time. At each of the iterations, m ants construct individual solutions independently.
Each solution contains K vehicle routes. Each ant starts from the depot, selects customers
for the first vehicle, moves back to the depot before the capacity or distance limit is violated,
then restarts from the depot and selects customers for the second vehicle. The procedure is
repeated until all customers are selected and their sequences are arranged in used vehicles.
It is possible that the last vehicle will serve all of the remaining customers even if it violates
the capacity or distance limit. The procedure of ant solution construction ensures that the
constraints defined in 2.7 and 2.8 can be satisfied.
Mathematical Problems in Engineering 9
The customer selection follows the state transition rule defined in 3.2. Suppose ant s
is moving from customer i to the next customer, v:
⎧
⎨arg max τij α ηij β q ≤ q0 ,
v j∈Us 3.2
⎩V q > q0 ,
α β
τij ηij
V : Pij α β , 3.3
j∈Us τij ηij
where Us is the set of customers that remain to be selected by ant s positioned on customer i,
τij is the pheromone trail on edge i, j, ηij is the inverse of the distance of edge i, j, α and β
are parameters which determine the relative importance of pheromone versus distance, q is
a random number uniformly distributed in the interval of 0 and 1, q0 is a parameter ranged
between 0 and 1, and Pij is the probability that ant s moves from customer i to customer j.
If q ≤ q0 , then ant s uses the greedy method 3.2 to select the next customer; otherwise, it
uses the probabilistic rule 3.3 to determine the next customer. If vehicle k is full or reaches
its distance limit, ant s has to go back to the depot and restarts from the depot to load the
next vehicle. However, if vehicle k is the last vehicle of ant s, then it has to serve all of the
remaining customers, even if the capacity or distance limit is violated.
Δ f S − fS,
3.4
Δ
P S exp − ,
T t
10 Mathematical Problems in Engineering
where S is the current solution, S is the new solution, fS is the objective function value
of S, fS is the new objective function value, t is the current temperature, and P S is
the probability that SA accepts new solution S . After carrying out neighborhood search R
times, SA reduces the temperature. New temperature T t 1 is equal to λ × T t, where
0 < λ < 1. PACO adopts a best improvement strategy in its local search step. That is, the best-
so-far solution will be recorded during the run of the SA algorithm. When the termination
condition is met, the ant solution will be replaced with the best-so-far solution of SA if the
latter is really better.
After solution construction and local search, all ants compare their Pbest solutions with
their iteration solutions and perform Pbest replacement if the iteration solutions are better.
Then, ants are ranked in terms of the goodness of their Pbest solutions, and the first r ants are
the elite ants. Of course, the Gbest solution of all ants is equal to the Pbest solution of the best
elite ant.
r
P
τij 1 − ρ τij Δτij bests ΔτijGbest ,
s2
P 1
Δτij bests , 3.5
fPsbest
1
ΔτijGbest ,
fGbest
where ρ is the pheromone evaporation rate and ranges between 0 and 1, r is the total number
P
of elite ants, Δτij bests and ΔτijGbest , are the pheromone added by elite ant s and the best ant,
respectively, and fPsbest and fGbest are the objective function values of elite ant s and the best
ant, respectively.
j
i
1 2 3 4 5
1 0 0.1 0.8 0.2 0.5
2 0 0.2 0.1 0.2
3 0 0.3 0.5
4 0 0.8
5 0
the search convergence and takes more computational time. It is suggested that w be equal
to the number of customers.
The basic idea of pheromone disturbance is similar to arithmetic crossover in GA, but
it is used to produce new pheromone trails, rather than to generate new solutions. Pheromone
disturbance has three steps. The first step is to select edges for disturbance. The disturbance
rate μ determines the probability that an edge is selected for disturbance. If μ is set too
high, it will be difficult to retain previous search experiences; on the other hand, if it is set
too low, the effect of disturbance will not be evident. The second step is to cluster selected
edges into groups, each of which has two edges with the same customer node. The groups
with single edges will be abandoned. The third step is to use a random number q ∈ 0, 1 to
determine a disturbance type for each of the paired edges. There are three types of pheromone
disturbance: unchanging, replacement, and weighted average. As defined in 3.6, paired
edges i, j and i, u have three possible results. That is, the pheromone on edge i, j may
remain the same, be replaced by the pheromone level of edge i, u, or be replaced with the
weighted average of the pheromone levels of these two edges.
⎧
⎪
⎪
t
⎨τij q < 0.2,
τijt 1 τiu
t
0.2 ≤ q < 0.4 , 3.6
⎪
⎪
⎩δτ t 1 − δτiu
t
, j
/ u, q ≥ 0.4,
ij
where δ is a uniform random number in the range 0, 1 and determines the ratios of
pheromone on the two edges.
After pheromone disturbance, ant s positioned on customer node i has a chance to
explore different edges rather than to keep selecting the same next customer node to move to.
Hence, pheromone disturbance increases the probability of finding optimal solutions. Note
that we deal with symmetric CVRPs in this paper, where the distances between customer
nodes are independent of the direction of traversing the edges. The same situation applies to
pheromone trails. Accordingly, dij dji and τij τji for each pair of nodes.
We use an example to illustrate the idea of pheromone disturbance. Suppose we have
five customers, then there are 20 edges with pheromone deposits. The original pheromone
matrix is shown in Table 1. Since τij τji for each pair of nodes, the algorithm considers only
half the edges to be disturbed.
The first step is to select edges for disturbance. Each edge is selected with a probability
of μ 0.3. For each edge, we generate a random number from uniform distribution in the
range between 0 and 1. The generated random numbers are shown in Table 2. A edge is
selected if its random number is less than μ. Table 2 tells us that four edges are selected and
marked in bold type. The second step is to pair two edges that have the same customer node.
12 Mathematical Problems in Engineering
j
i
1 2 3 4 5
1 0 0.5 0.2 0.6 0.1
2 0 0.1 0.2 0.7
3 0 0.8 0.4
4 0 0.5
5 0
j
i
1 2 3 4 5
1 0 0.1 0.8 0.2 0.65
2 0 0.1 0.11 0.2
3 0 0.3 0.5
4 0 0.8
5 0
Scanning the matrix in Table 2 row by row, we obtain the pairing results: {edge 1, 3, edge
1, 5} and {edge 2, 3, edge 2, 4}. The third step uses 3.6 to determine new pheromone
trails for paired edges. Random number q is first generated to determine a disturbance type
for each selected edge. The disturbance results are shown in Table 3. For example, for the first
paired edges, option 1 unchanging applies to edge 1, 3 and option 3 weighted average
applies to edge 1, 5. Note that here δ is a random number in the range of 0, 1. In our
example of edge 1, 5, δ is equal to 0.5. The new pheromone matrix is presented in Table 4.
4. Computational Results
The PACO algorithm described in Section 3 was coded in Java, and all experiments were
performed on a personal computer with Intel Core 2 CPU T7500 running at 2.20 GHz. Two
sets of benchmark problems were selected to evaluate the effectiveness of our proposed
algorithm for the capacitated vehicle routing problem.
The first set has 16 test problems and can be downloaded from the website
http://www.branchandcut.org/VRP/data/. The problems in this benchmark set are subject
to capacity constraints only. The total number of customers varies from 29 to 134, and the
total number of vehicles ranges from 3 to 10. The locations of customers appear in clusters
in the problems with their names initiated with B and M, while in the remaining problems
customers are randomly scattered or semiclustered. The first benchmark set was used by
Chen et al. 18 and Ai and Kachitvichyanukul 20 to test their PSO-based algorithms.
The second benchmark set can be downloaded from the website http://
people.brunel.ac.uk/∼mastjjb/jeb/orlib/vrpinfo.html. It has been widely used in previous
studies and contains 14 classical test problems selected from Christofides et al. 23. Problems
1, 2, 3, 4, 5, 11, and 12 consider the constraint of capacity only, while the remaining problems
are subject to the capacity and distance limits. The total number of customers varies from
50 to 199, and the total number of vehicles ranges from 5 to 18. Besides, customers are
randomly distributed in the first ten problems whereas customers are clustered in the last
four problems.
The PACO parameters set as follows were found to be robust for most of the test
problems according to our pilot tests. PACO parameters are maximum iteration number
MaxIte 1000, population size pop N/2, number of elite ants r 3, penalty coefficient
p 100, q0 0.8, ρ 0.5, α 2, β 1, and τ0 1/N×Lnn , where Lnn is the tour length
found by the nearest neighbor heuristic. Local search parameters are initial temperature t0
2, final temperature tf 0.01, R max{N × K/2, 250}, λ 0.9. Pheromone disturbance
parameters are: μ 0.3, w N, Δf 5.
Table 5 lists the computational results of PACO on two sets of test problems. The best
solution, average solution, worst solution, and standard deviation Std. computed over 20
independent runs on each problem are summarized, along with their average computational
time in seconds required to reach the final best solutions. The best solutions equal to the
best-known solutions of benchmark problems are asterisked and typed in bold. Table 5
reveals that PACO is able to generate reasonable good solutions for most of CVRPs in terms
of solution quality. Twelve out of sixteen test problems can be solved successfully by the
proposed algorithm in the first benchmark set. For the second set, seven out of fourteen test
problems can be solved successfully by PACO.
To evaluate the pheromone disturbance strategy and Pbest -resetting operation, the
proposed algorithm is compared with standard ACO 9 and PS-ACO 22 in terms of
their convergence trends. Standard ACO and PS-ACO were originally proposed for solving
traveling salesman problems TSP, not CVRPs. We implemented these two algorithms
in Java and tried to apply them to solve CVRPs with the same local search method and
parameter settings used in PACO.
Benchmark problem C1 was selected to test three ACO-based algorithms. Three data
sets of iteration-best solutions are plotted in Figure 2, The curves reveal that both PACO
and PS-ACO can converge faster than standard ACO but only PACO can find the optimal
solution. It demonstrates that, with pheromone disturbance and Pbest solution resetting,
PACO can effectively escape from local optima and find better solutions. In Figure 2, a
14 Mathematical Problems in Engineering
650
600
550
500
0 100 200 300 400 500 600 700 800 900
PACO
ACO
PSACO
couple of peaks on the PACO curve indicate the effects of pheromone disturbance and Pbest -
resetting. The disadvantage of PS-ACO is that all of the local best solutions are considered
in the pheromone updating procedure. When most ants have similar Pbest solutions, the total
amount of pheromone increment on edges will become very large. It results in the state of
pheromone stagnation and the search is trapped by a local optimum.
We conducted a comparative study to compare PACO with a couple of swarm
intelligence methods available for the CVRP. The comparative study contains two parts. In the
first part, the PACO algorithm is compared with three different PSO-based algorithms, which
were all tested on the 16 problems from the first benchmark set. The smaller the objective
function value, the better the solution. Table 6 displays the computational results of PACO
and the best results found in the papers of Chen et al. 18, Ai and Kachitvichyanukul 20,
and Kao and Chen 19, denoted as DPSO-SA, SR-2, CPSO-SA, respectively. Experimental
Mathematical Problems in Engineering 15
No. Problem N K BKS DPSO-SA 18 SR-2 20 CPSO-SA 19 PACO
1 A-n33-k5 32 5 661 661∗ /32.2 661∗ /13 661∗ /0.7 661∗ /0.12
2 A-n46-k7 45 7 914 914∗ /128.9 914∗ /23 917/2.4 914∗ /0.16
3 A-n60-k9 59 9 1354 1354∗ /308.8 1355/40 1354∗ /6.5 1354∗ /14.15
4 B-n35-k5 34 5 955 955∗ /37.6 955∗ /14 955∗ /1.2 955∗ /0.06
5 B-n45-k5 44 5 751 751∗ /134.2 751∗ /20 751∗ /4.8 751∗ /1.12
6 B-n68-k9 67 9 1272 1272∗ /344.2 1274/50 1274/27.2 1275/87.19
7 B-n78-k10 77 10 1221 1239/429.4 1223/64 1237/24 1221∗ /54.38
8 E-n30-k3 29 3 534 534∗ /28.4 534∗ /16 534∗ /0.3 534∗ /0.05
9 E-n51-k5 50 5 521 528/300.5 521∗ /22 521∗ /4.6 521∗ /0.51
10 E-n76-k7 75 7 682 688/526.5 682∗ /60 692/9.5 685/18.95
11 F-n72-k4 71 4 237 244/398.3 237∗ /53 237∗ /5.3 237∗ /6.28
12 F-n135-k7 134 7 1162 1215/1526.3 1162∗ /258 1200/202.8 1170/246.85
13 M-n101-k10 100 10 820 824/874.2 820∗ /114 825/6.1 820∗ /66.02
14 M-n121-k7 120 7 1034 1038/1733.5 1036/89 1039/51.5 1034∗ /8.18
15 P-n76-K4 75 4 593 602/496.3 594/48 596/27.6 593∗ /8.61
16 P-n101-k4 100 4 681 694/977.5 683/86 691/29.4 683/25.7
Notes: x/y f best solution/shortest CPU Time s, BKS is the best-known solution provided by published papers, DPSO-
SA used Intel Pentium IV CPU 1.8 GHz with 256 M RAM, SR-2 used Intel Pentium IV CPU 3.4 GHz with 1 GB RAM, and
CPSO-SA used Intel Core 2 CPU E8400 3 GHz with 3.5 G RAM.
results show that PACO is able to obtain the same or better results compared with three PSO-
based algorithms.
In the second part of the comparative study, the computational results of PACO are
compared with three ACO-based algorithms, which were all tested on the 14 problems
from the second benchmark set. Table 7 reports the best results obtained by these various
ACO algorithms. The best results of IACS, SS ACO, and EACO are found in 12, 13, 15,
16 Mathematical Problems in Engineering
respectively. Table 7 indicates that PACO generates very good solutions for seven problems,
which are equal to the best solutions published so far. For three compared algorithms, only
EACO is better than PACO because it can find best-known solutions for eight problems. It
can be noticed that both PACO and EACO can obtain optimum solutions for six problems
in common. For problem C11, PACO can reach the best-known solution but EACO cannot,
while, for problems C3 and C8, EACO can outperform PACO. However, the best solutions
produced by PACO and EACO are nearly equal for these three problems.
Tables 6 and 7 also list the shortest computational times required to reach the final best
solution over 20 independent runs. From the data, it can be observed that the computation
time taken by PACO is reasonable in practice in comparison with existing PSO- and ACO-
based algorithms. It shows that pheromone disturbance can improve solutions but does not
increase much computational time. Note that the computation time has not been reported
in 15. Instead, the authors of 15 used the maximum execution time as the termination
condition. They stopped the EACO algorithm after one hour of running for 14 benchmark
problems. The overall result of comparative study shows that the proposed algorithm is
competitive with recent swarm intelligence approaches in terms of solution quality and CPU
time.
5. Conclusion
This paper proposes a hybrid algorithm, PACO, which takes advantage of ant colony
optimization and particle swarm optimization for capacitated vehicle problems. During the
searching process, artificial ants construct solution routes, memorize the best solution ever
found, and lay pheromone on the routes of swarm and personal best solutions. To prevent
being trapped in local optima and to increase the probability of obtaining better solutions,
PACO performs pheromone disturbance and short-term memory resetting operations to
adjust stagnated pheromone trails. Disturbed pheromone trails guide ants to find new Pbest
and Gbest solutions. The merits of PSO adopted in PACO can speed convergence during
a run, even after pheromone disturbance operations. Computational results show that the
performance of PACO is competitive in terms of solution quality when compared with
existing ACO- and PSO-based approaches. For future research, PACO can be modified to
extend its application to vehicle routing problems with time windows or multiple depots,
among others.
Acknowledgments
The authors would like to thank the referees for their valuable comments and suggestions
that have greatly improved the quality of this paper. The authors are also grateful to National
Science Council, Taiwan Grant no. NSC 99-2410-H-036-003-MY2 for the financial support.
References
1 G. Laporte, “The vehicle routing problem: an overview of exact and approximate algorithms,”
European Journal of Operational Research, vol. 59, no. 3, pp. 345–358, 1992.
2 I. H. Osman, “Metastrategy simulated annealing and tabu search algorithms for the vehicle routing
problem,” Annals of Operations Research, vol. 41, no. 4, pp. 421–451, 1993.
Mathematical Problems in Engineering 17
Research Article
New Bounds for Ternary Covering Arrays Using a
Parallel Simulated Annealing
Copyright q 2012 Himer Avila-George et al. This is an open access article distributed under
the Creative Commons Attribution License, which permits unrestricted use, distribution, and
reproduction in any medium, provided the original work is properly cited.
A covering array CA is a combinatorial structure specified as a matrix of N rows and k columns
over an alphabet on v symbols such that for each set of t columns every t-tuple of symbols is
covered at least once. Given the values of t, k, and v, the optimal covering array construction
problem CAC consists in constructing a CA N; t, k, v with the minimum possible value of N.
There are several reported methods to attend the CAC problem, among them are direct methods,
recursive methods, greedy methods, and metaheuristics methods. In this paper, There are three
parallel approaches for simulated annealing: the independent, semi-independent, and cooperative
searches are applied to the CAC problem. The empirical evidence supported by statistical analysis
indicates that cooperative approach offers the best execution times and the same bounds as the
independent and semi-independent approaches. Extensive experimentation was carried out, using
182 well-known benchmark instances of ternary covering arrays, for assessing its performance
with respect to the best-known bounds reported previously. The results show that cooperative
approach attains 134 new bounds and equals the solutions for other 29 instances.
1. Introduction
A covering array, denoted by CAN; t, k, v, is a matrix M of size N × k which takes values
from the set of symbols {0, 1, 2, . . . , v − 1} called the alphabet, and every submatrix of size
N × t contains each tuple of symbols of size t or t-tuple, at least once. The value N is the
number of rows of M, k is the number of parameters, where each parameter can take v
values, and the interaction degree between parameters is described by the strength t. Each
2 Mathematical Problems in Engineering
combination of t columns must cover all the vt t-tuples. Given that there are kt sets of t
columns in M, the total number of t-tuples in M must be vt kt . When a t-tuple is missing in
a specific set of t columns, we will refer to it as a missing t-wise combination. Then, M is a
covering array if the number of missing t-wise combinations is zero.
When a matrix has the minimum possible value of N to be a CAN; t, k, v, the value
N is known as the covering array number. The covering array number is formally defined
as CANt, k, v min{N : ∃ CAN; t, k, v}. Given the values of t, k, and v, the optimal
covering array construction problem CAC consists in constructing a CAN; t, k, v such
that the value of N is minimized.
A major application of covering arrays CAs arises in software interaction testing
1, where a covering array can be used to represent an interaction test suite as follows. In
a software test, we have k components or factors. Each of these parameters has v values or
levels. A test suite is an N × k array where each row is a test case. Each column represents a
component, and a value in the column is the particular configuration. By mapping a software
test problem to a covering array of strength t, we can guarantee that we have tested, at
least once, all t-way combinations of component values 2. Thus, software testing costs can
be substantially reduced by minimizing the number of test cases N in the covering array.
Please observe that software interaction testing is a black-box testing technique, and thus
it exhibits weaknesses that should be addressed by employing white-box testing techniques.
For a detailed example of the use of covering arrays in software interaction testing, the reader
is referred to 3.
In this paper, we aim to develop an enhanced sequential simulated annealing ESSA
algorithm for finding near-optimal covering arrays. Simulated annealing algorithm is a
general-purpose stochastic optimization technique that has proved to be an effective tool
for approximating globally optimal solutions to many optimization problems. However, one
of the major drawbacks of the technique is the time it requires to obtain good solutions
moreover, when the evaluation function requires too much time. To address this drawback,
we propose three parallel simulated annealing approaches to solve the CAC problem. The
objective is to find the best bounds to some ternary covering arrays by using parallelism. To
our knowledge, the application of parallel simulated annealing to the CAC problem has not
been reported in the literature. Some methods of parallelization of simulated annealing are
discussed in 4–8.
The remainder of this paper is organized as follows. Some techniques that have been
used for constructing covering arrays are presented in Section 2. Section 3 describes the
components of our sequential annealing algorithm. In Section 4, three parallel simulated
annealing approaches are discussed. Section 5 describes the experimental results. Finally,
Section 6 presents the conclusions derived from the research presented in this paper.
and a group. This construction builds the covering array column by column by considering
the group acting on the columns of the starter vector. Meagher and Stevens 10 extended
the idea exhibited in 9, presenting a strategy for obtaining the starter vector by local search
and the selection of a group action. Recently, Lobb et al. 11 presented a generalization of
this method to permit any number of fixed points, permit an arbitrary group acting on the
symbols, and permit an arbitrary group acting on the columns. With all these generalizations
were obtained new bounds for covering arrays of strength two.
Tang and Woo 12 used constant weight vectors to construct test suites to be applied
to logic circuit testing. Martinez-Pena et al. 13 introduced a new method for constructing
covering arrays using trinomial coefficients, and they improved the results presented in 12.
Martinez-Pena et al. used the trinomial coefficients for the representation of the search space
in the construction of ternary covering arrays. It is clear that any covering array is formed by
a row set. In this sense, a trinomial coefficient represents a particular subset of rows which
may belong to a ternary covering array.
Hartman 3 presented a recursive construction which gives a method of squaring the
number k of columns in a covering array of strength t while multiplying the rows N by a
factor dependent only on t and v, but independent of k. This factor is related to the Turan
numbers T t; v that are defined to be the number of edges in the Turan graph.
Colbourn et al. 14 presented a product construction for t 2. In general, the product
of two covering arrays where t 2 consists in obtaining a new covering array where the
number of columns is equal to the product of the columns of the ingredients, and the number
of rows is equal to the sum of the rows of each ingredient.
The majority of commercial and open-source test data generating tools use greedy
algorithms for covering arrays construction AETG, TCG, ACTS, DDA, among others.
Cohen et al. 1 presented a strategy called AETG. In AETG, covering arrays are
constructed one row at a time. To generate a row, the first t-tuple is selected based on the
one involved in most uncovered pairs. Remaining factors are assigned levels in a random
order. Levels are selected based on the one that covers the most new t-tuples. For each
row that is actually added to the covering array, there are a number of, M, candidate rows
that are generated, and only a candidate that covers the most new t-tuples is added to the
covering array. Once a covering is constructed, a number, R, of test suites are generated and
the smallest test suite generated is reported. This process continues until all pairs are covered.
Tung and Aldiwan 15 proposed a tool called TCG. In TCG, one row is added at a
time to a covering array until all pairs are covered. Before each row is added, a number of
up to M candidate rows are generated, and the best candidate covering the most new pairs
is added. M is defined to be the maximum cardinality of factors the maximum number of
levels associated with any factor. To construct each row, factors are assigned levels in an
order based on a nonascending order of the cardinality of each factor. Each level for the factor
is evaluated, and a count of the number of pairs that are covered is used to determine whether
or not to select a level for a factor.
Bryce and Colbourn 16 presented an algorithm called DDA. The DDA constructs one
row of a covering array at a time using a steepest ascent approach. Factors are dynamically
fixed one at a time in an order based on density. New rows are continually added until all
interactions have been covered.
Lei and Tai 17 introduced a new algorithm called IPO, for pairwise testing. For a
system with two or more input parameters, the IPO strategy generates a pairwise test set for
the first two parameters, extends the test set to generate a pairwise test set for the first three
parameters, and continues to do so for each additional parameter. Contrary to many other
4 Mathematical Problems in Engineering
algorithms that build covering arrays one row at a time, the IPO strategy constructs them one
column at a time. Lei et al. 18 introduced an algorithm for the efficient production of covering
arrays, called IPOG, which generalizes the IPO strategy from pairwise testing to multiway
testing. The main idea is that covering arrays of k − 1 columns can be used to efficiently build
a covering array with degree k.
Ronneseth and Colbourn 19 introduced a new algorithm for constructing covering
arrays, the BBA. The BBA’s fundamental idea is to combine smaller covering arrays by
reordering the rows and then to append additional rows for the remaining uncovered pairs.
Some stochastic algorithms in artificial intelligence, such as tabu search 20, 21,
simulated annealing 22, generic algorithms, and ant colony optimization algorithm 23, provide
an effective way to find approximated solutions. In these algorithms, the optimization focuses
on one value of N at a time, attempting to find a covering array for that size.
A simulated annealing metaheuristic henceforth called SAC has been applied by
Cohen et al. in 22 for constructing covering arrays. SAC starts with a randomly generated
initial solution M whose cost cM is measured as the number of uncovered t-tuples. In
their implementation, Cohen et al. use a simple geometric function Tn 0.9998Tn−1 with an
initial temperature fixed at Ti 0.20. At each temperature, 2000 neighboring solutions are
generated. The algorithm stops either if a valid covering array is found, or if no change in
the cost of the current solution is observed after 500 trials. The authors justify their choice
of these parameter values based on some experimental tuning. They conclude that their
simulated annealing implementation is able to produce smaller covering arrays than other
computational methods, sometimes improving upon algebraic constructions.
Torres-Jimenez and Rodriguez-Tello 24 introduced a new simulated annealing
implementation for constructing binary covering arrays.
1 if ΔC < 0,
P 3.1
e−ΔC/T otherwise.
Uphill moves enable the system to escape from the local minima; without them, the
system would be trapped into a local minimum. Too high of a probability for the occurrence
of uphill moves, however, prevents the system from converging. In simulated annealing,
the probability is controlled by temperature in such a manner that at the beginning of the
procedure the temperature is sufficiently high, in which a high probability is available, and
Mathematical Problems in Engineering 5
as the calculation proceeds, the temperature is gradually decreased, lowering the probability
25.
The following paragraphs will describe each of the components of the implementation
of our simulated annealing implementation. The description is done given the matrix rep-
resentation of a covering array.
i−1
k
mi,v ,
1 if ms,v /
gri dms,v , mi,v , where dms,v , mi,v 3.3
s1 v1 0 otherwise.
An example is shown in 3.4; the number of symbols different between rows r1 and c1
is 4 and between r2 and c1 is 3 summing up 7. Then, the Hamming distance for the candidate
row c1 is 7:
⎧ ⎧
⎪
⎪ r 2 1 0 1 ⎪
⎪
⎨ 1 ⎨dr1 , c1 4
⎪
Rows r2 1 2 0 1 , Distances dr2 , c1 3 , 3.4
⎪
⎪ ⎪
⎪
⎪
⎩c 1 0 2 1 0 ⎩gc 7
1
Equation 3.4 is an example of the Hamming distance between two rows r1 , r2 that
are already in the matrix M and a candidate row c1 .
6 Mathematical Problems in Engineering
N1 s if x ≤ P,
N3 s, x 3.5
N2 s if x > P.
i idle times can be reduced since asynchronous communications overlap a part of the
computation;
ii less communication overhead, an isolated access to the global state is needed by
each processor at the end of each Markov subchain;
iii the probability of being trapped in a local optimum can be smaller. This is because
not all the processors start from the same state in each Markov subchain.
5. Experimental Results
This section presents an experimental design and results derived from testing the parallel
ISA, SSA, and CSA algorithms described in the Section 4. In order to show the performance
of these approaches, three experiments were developed. The first experiment had as purpose
to fine-tune the probabilities of the neighborhood functions to be selected. The second
experiment had as purpose to compare the three approaches in terms of parallel execution
time. Among the three approaches, the CSA approach was the fastest. The third experiment
evaluated the quality of the solutions of the CSA approach over a new benchmark proposed
in this paper. The results were compared against the best-known solutions reported in the
literature to construct covering arrays 30.
The three parallel approaches were implemented using C language and the message
passing interface MPI library. The implementations were run on the Tirant supercom-
puter The Tirant supercomputer: http://www.uv.es/siuv/cas/zcalculo/res/informa.wiki.
Tirant comprises 256 JS20 compute nodes blades and 5 p515 servers. Every blade has two
IBM Power4 processors at 2.0 GHz running Linux operating system with 4 GB of memory
RAM and 36 GB of local disk storage. All the servers provide a total of nearly 10 TB of
disk storage accessible from every blade through GPFS global parallel file system. Tirant
has in total 512 processors, 1 TB of memory RAM, and 19 TB of disk storage. The following
parameters were used for all simulated annealing implementations:
12 12
11 11
10 10
9 9
8 8
Best times
7 7
6 6
5 5
4 4
3 3
2 2
1 1
0 0
[0, 1]
[0.1, 0.9]
[0.2, 0.8]
[0.3, 0.7]
[0.4, 0.6]
[0.5, 0.5]
[0.6, 0.4]
[0.7, 0.3]
[0.8, 0.2]
[0.9, 0.1]
[1, 0]
Probabilities
ca1 ca4 ca7
ca2 ca5
ca3 ca6
P1 x1 P2 x2 q. 5.1
A solution to the given linear Diophantine equation must satisfy 5.2. This equation
has 11 solutions, each solution is an experiment that tests the degree of participation of each
neighborhood function in ESSA to accomplish the construction of a covering array. Every
combination of the probabilities was applied by ESSA to construct the set of covering arrays
shows in Table Ia, and each experiment was run 31 times; with the data obtained for each
experiment, we calculate the median. A summary of the performance of ESSA with the
probabilities that solved the 100% of the runs is shown in Table Ib:
2
Pi xi 1.0 5.2
i1
Finally, given the results shown in Figure 1, the best configuration of probabilities was
P1 0.3 and P2 0.7 because it found the covering arrays in smaller time median value.
The values P1 0.3 and P2 0.7 were kept fixed in the second experiment.
Mathematical Problems in Engineering 11
Table 1: a A set of 7 covering arrays configurations. b Performance of ESSA. Columns 3–9 show the
time taken to construct each of the covering arrays.
a
Id CA description
ca1 CA19; 2, 30, 3
ca2 CA35; 3, 5, 3
ca3 CA58; 3, 10, 3
ca4 CA86; 4, 5, 3
ca5 CA204; 4, 10, 3
ca6 CA243; 5, 5, 3
ca7 CA1040; 5, 15, 3
b
Covering arrays
P1 P2
ca1 ca2 ca3 ca4 ca5 ca6 ca7
0 1 4789.763 3.072 46.989 12.544 3700.038 167.901 0.102
0.1 0.9 1024.635 0.098 0.299 0.236 344.341 3.583 0.008
0.2 0.8 182.479 0.254 0.184 0.241 173.752 1.904 0.016
0.3 0.7 224.786 0.137 0.119 0.222 42.950 1.713 0.020
0.4 0.6 563.857 0.177 0.123 0.186 92.616 3.351 0.020
0.5 0.5 378.399 0.115 0.233 0.260 40.443 1.258 0.035
0.6 0.4 272.056 0.153 0.136 0.178 69.311 2.524 0.033
0.7 0.3 651.585 0.124 0.188 0.238 94.553 2.127 0.033
0.8 0.2 103.399 0.156 0.267 0.314 81.611 5.469 0.042
0.9 0.1 131.483 0.274 0.353 0.549 76.379 4.967 0.110
1 0 7623.546 15.905 18.285 23.927 1507.369 289.104 2.297
32
30
28
26
24
22
20
18
Speedup
16
14
12
10
8
6
4
2
0
0 2 4 6 8 10 12 14 16 18 20 22 24 26 28 30 32
P
ISA approach
SSA approach
CSA approach
Figure 2: Median speedup for the ISA, SSA, and CSA approaches.
the literature 30, which were produced using the following state-of-the-art procedures:
orthogonal array construction, Roux type constructions, doubling constructions, algebraic
constructions, deterministic density algorithm DDA, Tabu search, and IPOG-F.
For this experiment, we have fixed the maximum computational time expended by
our PSA for constructing a covering array to 72 hours and 50 processors. We create a new
benchmark composed of 182 covering arrays distributed as follows:
i 47 covering arrays with strength t 3, degree 4 ≤ k < 50, and order v 3,
ii 46 covering arrays with strength t 4, degree 5 ≤ k < 50, and order v 3,
iii 45 covering arrays with strength t 5, degree 6 ≤ k < 50, and order v 3,
iv 44 covering arrays with strength t 6, degree 7 ≤ k < 50, and order v 3,
The detailed results produced by this experiment are listed in Table 2. The first two
columns in each subtable indicate the strength t and the degree k of the selected instances.
Next two columns show, in terms of the size N of the covering arrays, the best-known
solution reported in the literature and the improved bounds produced by the CSA approach.
Last column depicts the difference between the best result produced by CSA approach and
the best-known solution Δ β − ϑ.
Figure 3 compares the results shown in Table 2 involving the CSA algorithm and the
best-known solutions. The analysis of the data presented led us to the following observation.
The solutions quality attained by the CSA approach is very competitive with respect to that
produced by the state-of-the-art procedures summarized in column 3 ϑ. In fact, it is able to
improve on 134 previous best-known solutions.
6. Conclusions
The long execution time of simulated annealing due to its sequential nature hinders its
application to realistically sized problems, in this case, the CAC problem when t > 3,
Mathematical Problems in Engineering 13
Table 2: It shows the improved bounds produced by CSA approach. Column ϑ represents the best-known
solution reported in the literature 30. Column β represents the best solution in terms of N produced by
CSA approach. Last column Δ depicts the difference between the best result produced by CSA approach
and the best-known solution Δ β − ϑ. a Improved bounds on CAN3, k, 3; b improved bounds on
CAN4, k, 3; c improved bounds on CAN5, k, 3; d improved bounds on CAN6, k, 3.
a
t k ϑ β Δ
3 4 27 27 0
3 5 33 33 0
3 6 33 33 0
3 7 40 39 −1
3 8 42 42 0
3 9 45 45 0
3 10 45 45 0
3 11 45 45 0
3 12 45 45 0
3 13 50 49 −1
3 14 51 50 −1
3 15 57 57 0
3 16 60 59 −1
3 17 60 59 −1
3 18 60 59 −1
3 19 60 59 −1
3 20 60 59 −1
3 21 66 67 1
3 22 66 71 5
3 23 69 71 2
3 24 72 71 −1
3 25 75 72 −3
3 26 78 72 −6
3 27 81 79 −2
3 28 81 79 −2
3 29 87 84 −3
3 30 87 84 −3
3 31 90 88 −2
3 32 90 89 −1
3 33 90 89 −1
3 34 90 89 −1
3 35 90 89 −1
3 36 90 89 −1
3 37 90 89 −1
3 38 90 89 −1
3 39 90 89 −1
3 40 90 89 −1
3 41 98 94 −4
3 42 98 94 −4
3 43 100 99 −1
3 44 100 99 −1
3 45 103 99 −4
3 46 103 101 −2
3 47 106 101 −5
3 48 106 101 −5
3 49 108 101 −7
3 50 108 102 −6
14 Mathematical Problems in Engineering
b
t k ϑ β Δ
4 5 81 81 0
4 6 111 111 0
4 7 123 123 0
4 8 141 135 −6
4 9 159 135 −24
4 10 159 164 5
4 11 183 183 0
4 12 201 201 0
4 13 219 219 0
4 14 237 249 12
4 15 237 277 40
4 16 237 277 40
4 17 300 287 −13
4 18 307 300 −7
4 19 313 313 0
4 20 315 321 6
4 21 315 338 23
4 22 315 347 32
4 23 315 359 44
4 24 377 375 −2
4 25 384 375 −9
4 26 393 387 −6
4 27 393 387 −6
4 28 393 392 −1
4 29 393 406 13
4 30 393 401 8
4 31 440 424 −16
4 32 445 431 −14
4 33 454 438 −16
4 34 462 447 −15
4 35 471 440 −31
4 36 471 456 −15
4 37 471 460 −11
4 38 471 465 −6
4 39 471 468 −3
4 40 499 472 −27
4 41 506 484 −22
4 42 509 488 −21
4 43 518 494 −24
4 44 522 497 −25
4 45 526 497 −29
4 46 530 506 −24
4 47 534 510 −24
4 48 542 516 −26
4 49 549 523 −26
4 50 549 525 −24
Mathematical Problems in Engineering 15
c
t k ϑ β Δ
5 6 243 243 0
5 7 351 351 0
5 8 405 405 0
5 9 483 405 −78
5 10 483 405 −78
5 11 705 550 −155
5 12 723 600 −123
5 13 723 828 105
5 14 922 890 −32
5 15 963 944 −19
5 16 963 1025 62
5 17 1117 1117 0
5 18 1167 1165 −2
5 19 1197 1190 −7
5 20 1266 1257 −9
5 21 1317 1312 −5
5 22 1346 1319 −27
5 23 1405 1387 −18
5 24 1447 1420 −27
5 25 1486 1440 −46
5 26 1521 1493 −28
5 27 1538 1527 −11
5 28 1579 1555 −24
5 29 1615 1585 −30
5 30 1647 1616 −31
5 31 1681 1643 −38
5 32 1724 1671 −53
5 33 1783 1702 −81
5 34 1783 1724 −59
5 35 1851 1748 −103
5 36 1882 1778 −104
5 37 1909 1800 −109
5 38 1937 1829 −108
5 39 1960 1851 −109
5 40 1986 1866 −120
5 41 2023 1896 −127
5 42 2046 1923 −123
5 43 2069 1940 −129
5 44 2091 2089 −2
5 45 2112 2111 −1
5 46 2130 2129 −1
5 47 2150 2149 −1
5 48 2174 2168 −6
5 49 2191 2189 −2
5 50 2213 2211 −2
16 Mathematical Problems in Engineering
d
t k ϑ β Δ
6 7 729 729 0
6 8 1152 1152 0
6 9 1431 1600 169
6 10 1449 1849 400
6 11 1449 2136 687
6 12 2181 2482 301
6 13 2734 2744 10
6 14 2907 3220 313
6 15 3234 3338 104
6 16 3443 3672 229
6 17 3658 3882 224
6 18 3846 4098 252
6 19 4054 4256 202
6 20 4486 4400 −86
6 21 4678 4600 −78
6 22 4853 4732 −121
6 23 4942 4941 −1
6 24 5193 5100 −93
6 25 5257 5238 −19
6 26 5709 5380 −329
6 27 5853 5810 −43
6 28 6003 5965 −38
6 29 6150 6110 −40
6 30 6281 6250 −31
6 31 6413 6393 −20
6 32 6535 6518 −17
6 33 6656 6642 −14
6 34 6772 6760 −12
6 35 6877 6871 −6
6 36 6989 6978 −11
6 37 7092 7086 −6
6 38 7194 7187 −7
6 39 7293 7284 −9
6 40 7391 7385 −6
6 41 7490 7478 −12
6 42 7574 7569 −5
6 43 7672 7661 −11
6 44 7757 7748 −9
6 45 7845 7836 −9
6 46 7938 7928 −10
6 47 8013 8005 −8
6 48 8092 8089 −3
6 49 8179 8176 −3
6 50 8256 8253 −3
Mathematical Problems in Engineering 17
110 550
100 500
90 450
400
80
350
70
300
N
N
60
250
50 200
40 150
30 100
20 50
5 10 15 20 25 30 35 40 45 50 5 10 15 20 25 30 35 40 45 50
k k
1200 4000
1000 3200
800 2400
600 1600
400 800
200 0
5 10 15 20 25 30 35 40 45 50 5 10 15 20 25 30 35 40 45 50
k k
Figure 3: Graphical comparison of the quality solutions between CSA and the state of the art 30.
5 < k ≤ 100, and v 3. A more efficient way to reduce execution time and make the simulated
annealing a more promising method is to parallelize sequential simulated annealing. It is a
challenging task. In fact, there are many approaches that may be considered in parallelizing
simulated annealing. However, an inappropriate strategy used will likely result in poor
performance.
In this paper, we have used three different approaches to do the parallelization of
the simulated annealing algorithm. From the experimental results, we found that the ISA
approach has the worst performance, and it does not scale well. The SSA approach offers
reasonable execution times; compared to the ISA, communication overhead in the SSA
approach would be increased when the size of the problem grows. However, each processor
can utilize the information from other processors such that the decrease in computations and
idle times can be greater than the increase in communication overhead. For instance, a certain
processor which is trapped in an inferior solution can recognize its state by comparing it
18 Mathematical Problems in Engineering
with others and may accelerate the annealing procedure. That is, processors may collectively
converge to a better solution.
The CSA approach is the one which offers the best results; it significantly reduces the
execution time of the SSA approach by employing asynchronous information exchange. The
quality solutions attained by the CSA approach are very competitive with respect to that
produced by the state-of-the-art procedures; in fact, it is able to improve on 134 previous
best-known solutions and equals the solutions for other 29 instances.
These experimental results confirm the practical advantages of using CSA algorithm
in the software testing area. It is a robust algorithm yielding smaller test suites than other
representative state-of-the-art algorithms, which allows reducing software testing costs.
The new bounds are available in Cinvestav covering array repository CAR, which
is available under request at http://www.tamps.cinvestav.mx/∼jtj/authentication.php. We
have verified all covering arrays described in this paper using the tool described in 31.
Acknowledgments
The authors thankfully acknowledge the computer resources and assistance provided by
Spanish Supercomputing Network TIRANT-UV. This research work was partially funded
by the following projects: CONACyT 58554; Cálculo de Covering Arrays; 51623-Fondo Mixto
CONACyT; Gobierno del Estado de Tamaulipas.
References
1 D. M. Cohen, S. R. Dalal, J. Parelius, and G. C. Patton, “The combinatorial design approach to
automatic test generation,” IEEE Software, vol. 13, no. 5, pp. 83–88, 1996.
2 K. Z. Zamli, M. F. J. Klaib, M. I. Younis, N. A. M. Isa, and R. Abdullah, “Design and implementation
of a t-way test data generation strategy with automated execution tool support,” Information Sciences,
vol. 181, no. 9, pp. 1741–1758, 2011.
3 A. Hartman, “Software and hardware testing using combinatorial covering suites,” in Graph Theory,
Combinatorics and Algorithms, M. C. Golumbic and I. B. A. Hartman, Eds., vol. 34 of Operations
Research/Computer Science Interfaces, pp. 237–266, Springer, New York, NY, USA, 2005.
4 E. H. L. Aarts and P. J. M. van Laarhoven, “Statistical cooling: a general approach to combinatorial
optimization problems,” Philips Journal of Research, vol. 40, no. 4, pp. 193–226, 1985.
5 D. Janaki Ram, T. H. Sreenivas, and K. G. Subramaniam, “Parallel simulated annealing algorithms,”
Journal of Parallel and Distributed Computing, vol. 37, no. 2, pp. 207–212, 1996.
6 S. Y. Lee and K. G. Lee, “Synchronous and asynchronous parallel simulated annealing with multiple
Markov chains,” IEEE Transactions on Parallel and Distributed Systems, vol. 7, no. 10, pp. 993–1008, 1996.
7 D. J. Chen, C. Y. Lee, C. H. Park, and P. Mendes, “Parallelizing simulated annealing algorithms based
on high-performance computer,” Journal of Global Optimization, vol. 39, no. 2, pp. 261–289, 2007.
8 Z. J. Czech, W. Mikanik, and R. Skinderowicz, “Implementing a parallel simulated annealing
algorithm,” in Proceedings of the 8th International Conference on Parallel Processing and Applied
Mathematics (PPAM’10), vol. 6067 of Lecture Notes in Computer Science, pp. 146–155, Springer, Berlin,
Germany, 2010.
9 M. Chateauneuf and D. L. Kreher, “On the state of strength-three covering arrays,” Journal of
Combinatorial Designs, vol. 10, no. 4, pp. 217–238, 2002.
10 K. Meagher and B. Stevens, “Group construction of covering arrays,” Journal of Combinatorial Designs,
vol. 13, no. 1, pp. 70–77, 2005.
11 J. R. Lobb, C. J. Colbourn, P. Danziger, B. Stevens, and J. Torres-Jimenez, “Cover starters for covering
arrays of strength two,” Discrete Mathematics, vol. 312, no. 5, pp. 943–956, 2012.
12 D. T. Tang and L. S. Woo, “Exhaustive test pattern generation with constant weight vectors,” IEEE
Transactions on Computers, vol. C-32, no. 12, pp. 1145–1150, 1983.
13 J. Martinez-Pena, J. Torres-Jimenez, N. Rangel-Valdez, and H. Avila-George, “A heuristic approach
Mathematical Problems in Engineering 19
for constructing ternary covering arrays using trinomial coefficients,” in Proceedings of the 12th Ibero-
American Conference on Artificial Intel-Ligence—IBERAMIA, vol. 6433 of Lecture Notes in Computer
Science, pp. 572–581, Springer, Berlin, Germany, 2010.
14 C. J. Colbourn, S. S. Martirosyan, G. L. Mullen, D. Shasha, G. B. Sherwood, and J. L. Yucas, “Products
of mixed covering arrays of strength two,” Journal of Combinatorial Designs, vol. 14, no. 2, pp. 124–138,
2006.
15 Y. W. Tung and W. S. Aldiwan, “Automating test case generation for the new generation mission
software system,” in Proceedings of the IEEE Aerospace Conference, vol. 1, pp. 431–437, IEEE Press, March
2000.
16 R. C. Bryce and C. J. Colbourn, “The density algorithm for pairwise interaction testing,” Software
Testing Verification and Reliability, vol. 17, no. 3, pp. 159–182, 2007.
17 Y. Lei and K. C. Tai, “In-parameter-order: a test generation strategy for pair-wise testing,” in
Proceedings of the 3rd IEEE International Symposium on High-Assurance Systems Engineering (HASE’98),
pp. 254–261, IEEE Computer Society, 1998.
18 Y. Lei, R. Kacker, D. R. Kuhn, V. Okun, and J. Lawrence, “IPOG: a general strategy for T-way software
testing,” in Proceedings of the 14th Annual IEEE International Conference and Workshops on the Engineering
of Computer-Based Systems (ECBS’07), pp. 549–556, IEEE Computer Society, Tucson, Ariz, USA, March
2007.
19 A. H. Ronneseth and C. J. Colbourn, “Merging covering arrays and compressing multiple sequence
alignments,” Discrete Applied Mathematics, vol. 157, no. 9, pp. 2177–2190, 2009.
20 K. J. Nurmela, “Upper bounds for covering arrays by tabu search,” Discrete Applied Mathematics, vol.
138, no. 1-2, pp. 143–152, 2004.
21 L. Gonzalez-Hernandez, N. Rangel-Valdez, and J. Torres-Jimenez, “Construction of mixed covering
arrays of variable strength using a tabu search approach,” in Proceedings of the 4th International
Conference on Combinatorial Optimization and Applications (COCOA’10), vol. 6508 of Lecture Notes in
Computer Science, pp. 51–64, Springer, Berlin, Germany, 2010.
22 M. B. Cohen, C. J. Colbourn, and A. C. H. Ling, “Augmenting simulated annealing to build interaction
test suites,” in Proceedings of the 14th International Symposium on Software Reliability Engineering
(ISSRE’03), pp. 394–405, IEEE Computer Society, 2003.
23 T. Shiba, T. Tsuchiya, and T. Kikuno, “Using artificial life techniques to generate test cases for
combinatorial testing,” in Proceedings of the 28th Annual International Computer Software and Applications
Conference (COMPSAC’04), vol. 1, pp. 72–77, IEEE Computer Society, September 2004.
24 J. Torres-Jimenez and E. Rodriguez-Tello, “New bounds for binary covering arrays using simulated
annealing,” Information Sciences, vol. 185, no. 1, pp. 137–152, 2012.
25 Y. Jun and S. Mizuta, “Detailed analysis of uphill moves in temperature parallel simulated annealing
and enhancement of exchange probabilities,” Complex Systems, vol. 15, no. 4, pp. 349–358, 2005.
26 L. Cavique, C. Rego, and I. Themido, “Subgraph ejection chains and tabu search for the crew
scheduling problem,” Journal of the Operational Research Society, vol. 50, no. 6, pp. 608–616, 1999.
27 M. M. Atiqullah, “An efficient simple cooling schedule for simulated annealing,” in Proceedings of the
International Conference on Computational Science and its Applications (ICCSA’04), vol. 3045 of Lecture
Notes in Computer Science, pp. 396–404, Springer, Berlin, Germany, 2004.
28 S. Kirkpatrick, C. D. Gelatt, and M. P. Vecchi, “Optimization by simulated annealing,” Science, vol.
220, no. 4598, pp. 671–680, 1983.
29 Z. Czech, “Three parallel algorithms for simulated annealing,” in Parallel Processing and Applied
Mathematics, vol. 2328 of Lecture Notes in Computer Science, pp. 210–217, Springer, Berlin, Germany,
2006.
30 C. J. Colbourn, “Covering array tables for t2, 3, 4, 5, 6,” 2012, http://www.public.asu.edu/∼ccolb-
ou/src/tabby/catable.html.
31 H. Avila-George, J. Torres-Jimenez, N. Rangel-Valdez, A. Carrión, and V. Hernández, “Supercomput-
ing and a Grid computing on the verification of coveringarrays,” The Journal of Supercomputing. In
press.
Hindawi Publishing Corporation
Mathematical Problems in Engineering
Volume 2012, Article ID 475018, 16 pages
doi:10.1155/2012/475018
Research Article
A VNS Metaheuristic with Stochastic Steps for
Max 3-Cut and Max 3-Section
Ai-fan Ling1, 2
1
Research Center of Security and Future, School of Finance, Jiangxi University of Finance and Economics,
Nanchang 330013, China
2
Key Laboratory of Management, Decision and Information Systems, Academy of Mathematics and Systems
Science, CAS, Beijing 100190, China
Copyright q 2012 Ai-fan Ling. This is an open access article distributed under the Creative
Commons Attribution License, which permits unrestricted use, distribution, and reproduction in
any medium, provided the original work is properly cited.
A heuristic algorithm based on VNS is proposed to solve the Max 3-cut and Max 3-section
problems. By establishing a neighborhood structure of the Max 3-cut problem, we propose a
local search algorithm and a variable neighborhood global search algorithm with two stochastic
search steps to obtain the global solution. We give some numerical results and comparisons
with the well-known 0.836-approximate algorithm. Numerical results show that the proposed
heuristic algorithm can obtain efficiently the high-quality solutions and has the better numerical
performance than the 0.836-approximate algorithm for the NP-Hard Max 3-cut and Max 3-section
problems.
1. Introduction
Given a graph GV ; E, with nodes set V and edges set E, the Max 3-cut problem is to
find a partition S0 ⊂ V , S1 ⊂ V and S2 ⊂ V , of the set V , such that S0 S1 S2 V ,
Si Sj ∅i / j and the sum of the weights on the edges connecting the different parts is
maximized. Similar to the Max cut problem, the Max 3-cut problem has long been known
to be NP complete 1, even for any un-weighted graphs 2, and has also applications in
circuit layout design, statistical physics, and so on 3. However, due to the complexity of this
problem, its research progresses is much lower than that of the Max cut problem. Based on
the semidefinite programming relaxation proposed by Goemans and Williamson 4, Frieze
and Jerrum 5 obtained a 0.800217-approximation algorithm for the Max 3-Cut problem.
Recently, Goemans and Williamson 6 and Zhang and Huang 7 improved Frieze and
2 Mathematical Problems in Engineering
2 1 1 2
− yi · yj − yj · yi 1 − Re yi · yj . 1.1
3 3 3 3
2
M3C : max fy wij 1 − Re yi · yj
3 i<j
1.2
2 n
s.t. y ∈ {1, ω, ω } ,
2
CSDP : max wij 1 − Re yi · yj
3 i<j
s.t. yi 1, i 1, 2, . . . , n, 1.3
Akij · Y ≥ −1, i, j 1, 2, . . . , n, k 0, 1, 2
Y 0,
Mathematical Problems in Engineering 3
where Yij yi · yj , Akij ωk ei eTj ω−k ej eTi and ei denotes the vector with zeros everywhere
except for an unit in the ith component. It is easily to verify that constraints Akij · Y ≥ −1 can
be expressed as
1
Re ωk Yij ≥ − , k 0, 1, 2. 1.4
2
To get an approximate solution of M3C, Goemans and Williamson 6 do not directly solve
the CSDP, but solve an equivalent real SDP with following form Although some softwares,
such as SeDuMi 8 and the earlier version of SDPT3-4.0 9, can deal with SDPs with complex
data, this does not reduce the dimensions of problems:
1 Q O
RSDP : max ·X
2 O Q
T
ei ei O
s.t. · X 2, i 1, 2, . . . , n,
O ei eTi
⎡ ⎤
Re Akij − Im Akij
⎣ ⎦ · X ≥ −2, 1 ≤ i < j ≤ n, k 0, 1, 2
Im Akij Re Akij
A0ij O 1.5
· X 0, 1 ≤ i < j ≤ n,
O −A0ij
O A0ij
· X 0, 1 ≤ i < j ≤ n,
A0ij O
O ei eTi
· X 0, i 1, 2, . . . , n,
ei eTi O
X ∈ S 2n ,
R −S
X . 1.6
S R
The final two classes of equality constraints ensure that Sii 0 i 1, . . . , n and S is a skew-
symmetric matrix.
4 Mathematical Problems in Engineering
⎧
⎪
⎪ 2π
⎪
⎪ 1, if Argξi ∈ 0, ,
⎪
⎪ 3
⎪
⎪
⎨ 2π 4π
yi ω, if Argξi ∈ , , 1.7
⎪
⎪ 3 3
⎪
⎪
⎪
⎪
⎪
⎪ 4π
⎩ω2 , if Argξi ∈ , 2π ,
3
where Arg· ∈ 0, 2π means the complex angle principal value of a complex number.
Goemans and Williamson 6 had verified that, see also Zhang and Huang 7,
y ≥ 0.836 · Q · Y .
f 1.8
The algorithm proposed by Goemans and Williamson 6 can obtain a very good
approximate ratio, and RSDP can be solved by interior point algorithm, but the 0.836-
approximate algorithm will be not practical in numerical study for the Max 3-cut problem.
From RSDP, one can see that for a graph with n nodes, RSDP has 2n 5nn − 1/2 constraints
and 3nn − 1/2 slack variables via the inequality constraints. That is to say, RSDP has a 2n
dimensional unknown symmetrical semidefinite positive matrix variable and a 3nn − 1/2
dimensional unknown vector variable, and 2n 5nn − 1/2 constraints, and has also many
matrices without an explicit block diagonal structure although they are sparse. For instance,
when n 100, RSDP becomes a very-high-dimensional semidefinite programming problem
with 14850 slack variables and 24950 constraints. Further, as we known, it is only a class of
universal and medium-scale instances for Max 3-cut problems with 50 to 100 nodes. Hence,
it will be very time consuming to solve such a RSDP relaxation of M3C using the current
existing any SDP softwares. This leads that 0.836-approximate algorithm is not suitable for
computational study of the Max 3-cut problem. This limitation for solving M3C based on
CSDP or RSDP relaxation motivates us to find a new efficient and fast algorithm for the
practical purpose for the Max 3-cut problem.
In the current paper, we first establish a definition of K-neighborhood structure of the
Max 3-cut problem and design a local search algorithm to find the local minimizer. And then,
we propose a variable neighborhood search VNS metaheuristic with stochastic steps which
is originally considered by Mladenović and Hansen 10, by which we can find efficiently
a high-quality global approximate solution of the Max 3-cut problem. Further, combining a
greed algorithm, we extend the proposed algorithm to the Max 3-section problem. To the
best of our knowledge, it is first time to consider the computational study of the Max 3-
cut problem. In order to test the performance of the proposed algorithm, we compare the
numerical results with Goemans and Williamson’s 0.836-approximate algorithm.
This paper is organized as follows. In Section 2, we give some definitions and lemmas.
In Section 3, we present the VNS metaheuristic for solving the Max 3-cut problem. The VNS
is extended to the Max 3-section problem in Section 4. In Section 5, we give some numerical
results and comparisons.
Mathematical Problems in Engineering 5
2. Preliminaries
In this section, we will establish some definitions and give some facts for our sequel purpose.
For the third roots of unity, 1, ω, ω2 , we can get the following fact:
2 2
2 2.1
1−ω ω − ω2 1 − ω2 3.
Denote S {1, ω, ω2 }n . Then based on 2.1, for any y ∈ S, we may definite a K-neighborhood
of y as follows.
Definition 2.1. For any y ∈ S and any positive integer number K 1 ≤ K ≤ n, one defines the
K-neighborhood, denoted by NK y, of y as the set
n
2
NK y z∈S: z−y 2
zi − yi ≤ 3K . 2.2
i1
Definition 2.3. For any u ∈ {0, 1, 2}, define two maps from {1, ω, ω2 } to itself as follows:
τi ωu ωu i ∈ {1, ω, ω2 }, i 1, 2.
Clearly, for any u ∈ {0, 1, 2}, τi ωu / ωu , i 1, 2 and τ1 ωu / τ2 ωu . Applying
Definition 2.3, for any z ∈ Ny there exists an unique component, zk say, of z, such that
yk and either zk τ1 yk or zk τ2 yk , and other components of z and y are the same.
zk /
For simplicity, for any z ∈ Ny with zk / yk and zi yi i 1, . . . , n, i /
k, we denote by
z τ1k y or z τ2k y corresponding to zk τ1 yk or zk τ2 yk . By Definitions 2.1
and 2.3, for any y ∈ S, we can structure its 1-neighborhood points using maps defined by
Definition 2.3; that is, we have the following result.
Lemma 2.4. Let τi · i 1, 2 be defined by Definition 2.3. Then, for any y ∈ S and any fixed positive
integer number k 1 ≤ k ≤ n, one has
that is, τ1k y and τ2k y are two 1-neighborhood points of y.
maximizer of f over S. A 1-local maximizer of the function f is also called a local maximizer
of the function f over S.
Let y0 y10 , . . . , yn0 T ∈ S be a feasible solution of problem M3C. If y0 is not a local maximizer
of f, then for all y ∈ Ny0 , we may find a y ∈ Ny0 , such that f
y max{fy : y ∈ Ny0 }.
It is clear that f y ≥ fy . If y
0
is not still a local maximizer of f, then replacing y0 with y
and repeating the process until a point y satisfying fy max{fy : y ∈ N y} is found,
which indicates that y is a local maximizer of f.
For any positive integer number k 1 ≤ k ≤ n, let yk y1k , . . . , ynk T τik y0 ∈
Ny0 i 1, 2; that is,
yik yi0 , i 1, 2, . . . , k − 1, k 1, . . . , n;
3.1
ykk 0
/ yk .
Denote
δk f y0 − f yk . 3.2
Based on Lemma 3.1, if we know the value of fy0 , then we can obtain the value
function fyk at next iterative point yk by calculating δk by 3.3, instead of calculating
directly the values fyk , which reduces sharply the computational cost. By Definition 2.1,
there exist two points satisfying 3.1 for fixed k; that is, when yk ∈ Ny0 and 3.1 is
satisfied, then either ykk τ1 yk0 or ykk τ2 yk0 . For our convenience, we denote δk by
δ1 k when ykk τ1 yk0 and by δ2 k when ykk τ2 yk0 . In what follows, we describe the local
search algorithm for the Max 3-cut problem denoted by LSM3C; by this algorithm, we can
get a local maximizer of functionfy over S.
For LSM3C, one has the following.
δi∗ k∗ min{δ1 1, δ2 1, . . . , δ1 k, δ2 k, . . . , δ1 n, δ2 n}. 3.4
1 Arbitrary choose a point y0 ∈ S, implement LSM3C starting from y0 ∈ S and denote
the obtained local maximizer by y . Set K 1.
2 Randomly take a point y ∈ ∂NIK y and implement again LSM3C from y, and
.
denote the obtained new local maximizer by y
> f
3 If fy y, then set y and K 1; go to Step 2.
y
4 If K < Kmax ≤ n, set K K 1; go to Step 2. Otherwise, return y
as an approximate
global solution of problem M3C and stop.
The subscript IK in Step 2 is a function of K and is also a positive integer number
not greater than n. IK reflects the main skill of converting the current neighborhood of
local maximizer y into another neighborhood of y . For a given Kmax , let m n/Kmax and
K0 n − mKmax , where a means the integral part of a. We divide the n neighborhoods of y ,
Ny, N2
y, . . . , NK y, . . . , Nn
y into Kmax neighborhood blocks NI1 y, . . . , NIKmax
y,
such that, for K 1, 2, . . . , Kmax − K0 ,
y ⊆ NIK
NK−1m 1 y ⊆ NKm 1
y, 3.5
NK−1m y
1 1 ⊆ NIK
y ⊆ NKm 1 y.
3.6
has m integers and each subset of the last K0 subsets has m 1 integers. For any integer
K≤ Kmax , let
or
Then we can randomly choose a point y in ∂NIK y, where c ∈ 0, 1 is a random number
from uniformly distribution U0, 1, such that NIK
y satisfies 3.5 or 3.6.
VNS-k stops when the maximum K neighborhood is reached. Additionally, we also
consider another termination criterion of VNS based on the maximum CPU-time and denoted
by VNS-t. VNS-t can obtain a better solution than VNS-k since VNS-t actually runs several
times VNS-k in the maximum allowing time tmax , but it generally has to spend more
computational time. The VNS-t can be stated as follows.
For VNS-t, one has the following.
1 Set tCPU 0, running VNS-k for an arbitrary initial point y0 ∈ S, and let a local
optimal solution y
be obtained.
3 If tCPU < tmax , then set K 1; go to Step 2 in VNS-k. Otherwise, return y
as an
approximate global solution of problem M3C and stop.
We mention that it differs from the classical variable neighborhood search meta-
heuristic that is originally proposed by Mladenović and Hansen 10. In order to obtain a
global optimal solution or a high-quality approximate solution of problem M3C, we use
two stochastic steps in VNS. First, for a fixed K, a K-neighbor of y is chosen randomly.
Second, by the definition of IK, when we change the neighborhood of y from NIK−1 to
NIK , NIK may take any a neighborhood among NK−1m j , j 1, 2, . . . , m of y
, which is
decided by random number c. In VNS, positive integer Kmax decides the maximum search
neighborhood block of y , which also decides directly the CPU-time of VNS. Based on the
second stochastic step, we may choose a relative small Kmax comparing with n. This can
decrease our computational time.
0, then the Max 3-section problem can be formulated as the following programming problem
M3S:
2
M3S : max wij 1 − Re yi · yj
3 i<j
n
4.1
s.t. yi 0,
i1
y ∈ S,
s.t. eeT · Y 0,
4.2
yi 1, i 1, 2, . . . , n,
Akij · Y ≥ −1, i, j 1, 2, . . . , n, k 0, 1, 2
Y 0,
where e is the column vector of all ones. Andersson 12 extended Frieze and Jerrum’s
random rounding method to M3S and obtained a 2/3 O1/n3 -approximate algorithm,
which is the current best approximate ratio for M3S; also see the recent research of Gaur et
al. 13. The author of the current paper considers a special the Max 3-Section problem and
obtains a 0.6733-approximate algorithm; see Ling 2009 14.
Clearly, the feasible region of problem M3S is a subset of S, and the optimal value of
problem M3S is not greater than that of problem M3C. Assume that we have get a global
optimal solution or a high-quality approximate solution y of problem M3C. It is clear that y
may not satisfy the condition ni1 yi 0. But we may adjust y to get a new feasible solution
ys using a greedy algorithm, such that ys satisfies ni1 yis 0. This is the motivation that we
propose the greedy algorithm for the Max 3-section problem.
For the sake of our analysis, without loss of generality, we assume that the local
maximizer y satisfies |S0 | max{|S0 |, |S1 |, |S2 |}. This means that S0 {i : yi 1} is the
subset of V with maximum cardinal number. If |Sk | max{|S0 |, |S1 |, |S2 |}k / 0, k 1, 2,
then we may set yiN wk yi , i 1, . . . , n. The resulted new solution yN y1N , . . . , yiN
will not change the objective value since f y fwk y k
/ 0, k 1, 2; moreover, the new
partition {S0 , S1 , S2 } based on y satisfies |SN
N N N N
0 | max{|S 0 |, |S1 |, |S2 |}. By our assumption,
N N N
The sizes adjusting greedy algorithm of Cases 3 and 4 are similar to Cases 1 and 2.
Hence, we mainly consider Cases 1 and 2 for adjusting the partition of V from S {S0 , S1 , S2 }
to S {S0 , S1 , S2 } such that |Sk | n/3, k 0, 1, 2. Denote
δ0 i wij , i ∈ S0 ,
j∈S1 ∪S2
δ01 i wij , i ∈ S0 , δ10 i wij , i ∈ S1 ,
j∈S1 j∈S0
4.3
δ02 i wij , i ∈ S0 , δ20 i wij , i ∈ S2 ,
j∈S2 j∈S0
δ12 i wij , i ∈ S1 , δ21 i wij , i ∈ S2 .
j∈S2 j∈S1
where d01 i∈S0 δ01 i, d02 i∈S0 δ02 i, d12 i∈S1 δ12 i.
In what follows, we describe the size adjusting greedy algorithms SAGAs for Cases
1 and 2, and denote the greedy algorithms for the two cases by SAGA1 and SAGA2,
respectively.
For SAGA1, one has the following.
1 Calculate
i∈S0 δ02 i i∈S1 δ12 i
m02 , m12 . 4.5
|S0 | |S1 |
2 If m02 ≥ m12 , let S1 {j1 , j2 , . . . , j|S1 | }, where δ12 jl ≥ δ12 jl 1 , l 1, 2, . . . , |S1 |. Set
S1 {j1 , j2 , . . . , jn/3 }, S2 S2 S1 \ S1 and renew to calculate
δ02 i wij ,
4.6
j∈S2
for each i ∈ S0 . Let S0 {i1 , i2 , . . . , i|S0 | }, where δ02
ik ≥ δ02 ik 1 . Set S0
{i1 , i2 , . . . , in/3 } and S2 S2 S0 \ S0 .
Mathematical Problems in Engineering 11
3 If m02 < m12 , let S0 {i1 , i2 , . . . , i|S0 | }, where δ02 ik ≥ δ02 ik 1 , k 1, 2, . . . , |S0 |, set
S0 {i1 , i2 , . . . , in/3 }, S2 S2 S0 \ S0 , and then renew to calculate
δ12 i wij ,
4.7
j∈S2
for each i ∈ S1 . Set S1 {j1 , j2 , . . . , jn/3 } and S2 S2 S1 \ S1 , where δ12
jk ≥
δ12 jk 1 here.
4 Return the current partition S {S0 , S1 , S2 }; stop.
d01 d02
m01 , m02 . 4.8
|S0 | |S0 |
S0 i1 , i2 , . . . , i|S0 | , 4.9
S0 i1 , i2 , . . . , i|S0 |−q1 , S1 S1 S0 \ S0 , 4.10
and let
S0 i1 , i2 , . . . , i S , 4.12
| 0|
where δ02 ik ≥ δ02
ik 1 , k 1, 2, . . . , |S0 |. Set
! "
S0 i1 , i2 , . . . , in/3 , S2 S2 S0 \ S0 . 4.13
S0 i1 , i2 , . . . , i|S0 | , 4.14
12 Mathematical Problems in Engineering
S0 i1 , i2 , . . . , i|S0 |−q2 , S2 S2 S0 \ S0 , 4.15
and let
S0 i1 , i2 , . . . , i S , 4.17
| 0|
where δ01 ik ≥ δ01
ik 1 , k 1, 2, . . . , |S0 |. Set
! "
S0 i1 , i2 , . . . , in/3 , S1 S1 S0 \ S0 . 4.18
5. Numerical Results
This section describes the obtained experimental results for some instances of Max 3-
cut and Max 3-Section problems using the proposed VNS metaheuristic. We also show a
quantitative comparison with 0.836-approximate algorithm. The computational experiments
are performed in an Intel Pentium 4 processor at 2.0 GHz, with 512 MB of RAM, and all
algorithms are coded in Matlab. Because RSDP relaxation of M3C includes many slack
variables, many constraints, and matrices variables without a block diagonal structures, in
our numerical comparisons, we choose SDPT3-4.0 9, one of the best and well-known solvers
of semidefinite programming, to solve RSDP relaxation of M3C.
All our test problems are generated randomly by the following way. Let p ∈ 0, 1 be
a constant and r ∈ 0, 1 a random number. If r ≤ p, then there is an edge between nodes i
and j with weight wij , that is, a random integer between 1 and 10. Otherwise, wij 0; that
is, there is no edge between nodes i and j. Because of the limits of memory of SDPT3, when
n > 200, RSDP becomes a huge semidefinite programming problem with not less than 59700
slack variables and 99900 constraints and is out of memory of SDPT3. Hence, in the numerical
experiments, we consider 30 instances with p 0.1, 0.3, 0.6, and n varying from 20 to 200.
Firstly, we check the influence of Kmax on the quality of solution obtained by VNS-k.
For a given graph, we take Kmax 3, 5, 10, 15, 30; Table 1 presents the results, where Wnp in the
first column of this table and the following tables means that a graph is randomly generated
with nodes n and density p; for instance, W30.6 presents a graph generated randomly with
n 30 and p 0.6. We find from Table 1 that the influence of Kmax to objective value
denoted by Obj in Table 1 is slight when Kmax > 5, but the CPU time increases sharply as
Kmax increases. This result is actually not surprising. Indeed, because IK > K, we choose
Mathematical Problems in Engineering 13
Table 1: The objective value obtained by VNS for M3C with different Kmax .
randomly a point y in ∂NIK y, instead of ∂NK y. This avoids to choose too large Kmax
which leads to more CPU-time cost. Hence, in sequel numerical comparisons, we fix Kmax 5
for all test problems.
Secondly, we compare VNS VNS-k, VNS-t metaheuristic with 0.836-approximate
algorithm for all test problems. To avoid the effect of initial points, for each test problem,
after RSDP is solved, we run the round procedure of 0.836-approximate algorithm and VNS
metaheuristic ten times, respectively.
Table 2 gives the result of numerical comparisons. In the numerical presentations of
Table 2, Objrsdp is the optimal value of problem RSDP; that is, it is an upper bound of M3C.
ObjGM is the largest value obtained by 0.836-approximate algorithm in the ten tests. Objvns
stands for the largest value obtained by VNS for M3C in the ten tests, respectively. m and s.v.
are the number of constraints and slack variables s.v., respectively. tGM and tvns−k are the
average time second associated with the two algorithms in the ten tests. For the maximum
CPU time of VNS-t, we take tmax 2tvns−k , but the real CPU time of VNS-t will be greater than
tmax . Additionally, for measuring the performance of solutions, we take
for M3S. Clearly, ρ can reflect how close to the solution obtained by VNS from the optimal
solution of RSDP. One can see from Table 2 that 1 the VNS metaheuristic not only can
obtain a better solution than 0.836-approximate algorithm for all test problems, but also
that the elapsed CPU-time of VNS metaheuristic is much less than that of 0.836-approximate
algorithm for all test problems, 2 the performance of solution can be improved by VNS-t for
most of test problems when the termination criterion of VNS is based on the maximum CPU-
time, but VNS-t spends more computational time than VNS-k. The improved performance
can be reflected by ∇ρ ρt − ρk in the final column of Table 2. Average speaking, VNS-t
improves 0.91 percentage point.
Finally, we consider the solution of M3S by combining VNS-k and greedy sizes-
adjusted algorithm SAGA stated in Section 4. Let y be an approximate solution of M3C
obtained by VNS; we can obtain an approximate solution of M3S from SAGA. The numerical
results are reported by Table 3 in which Objvns saga stands for the largest value obtained
by VNS-k plus SAGA for M3S. Although our sizes-adjusted algorithm may decrease the
objective value obtained by VNS, the changes of objective values are very slight from
14 Mathematical Problems in Engineering
m 0.836-algorithm VNS
Wnp Objrsdp
s.v. ObjGM tGM Objvns-k tvns-k ρk % Objvns−t ρt % ∇ρ%
W20.1 990 121 117 22.93 119 0.28 −1.65 119 −1.65 0
W20.3 570 232 225 25.58 228 0.29 −1.72 228 −1.72 0
W20.6 492 464 14.27 479 0.40 −2.65 479 −2.24 0.41
W30.1 2235 147 139 161.56 144 0.81 −2.04 144 −2.04 0
W30.3 1305 566 521 140.12 550 0.97 −2.83 550 −2.30 0.53
W30.6 1181 1101 94.93 1151 0.92 −2.54 1151 −2.03 0.51
W45.1 5040 675 594 202.48 605 0.92 −10.38 612 −9.33 1.05
W45.3 2970 1300 1147 211.72 1192 1.02 −8.31 1300 −7.15 1.16
W45.6 2441 2272 217.57 2313 1.02 −5.25 2350 −3.73 1.52
W60.1 8970 1069 918 256.35 965 2.14 −9.73 981 −8.23 1.50
W60.3 5310 2494 2228 256.57 2301 3.13 −7.74 2332 −6.50 1.24
W60.6 4403 4054 266.68 4213 2.48 −4.32 4307 −2.18 2.14
W80.1 15960 1567 1357 537.58 1415 5.82 −9.71 1438 −8.23 1.48
W80.3 9480 4094 3666 546.23 3801 7.35 −7.16 3912 −4.45 2.71
W80.6 7746 7328 495.34 7423 6.31 −4.17 7481 −3.42 0.75
W100.1 24950 2526 2149 893.52 2262 12.73 −10.46 2302 −8.87 1.60
W100.3 14850 6418 5814 882.73 6010 12.66 −6.36 6113 −4.75 1.61
W100.6 11956 11212 865.28 11391 11.99 −4.73 11422 −4.47 0.26
W120.1 33990 3746 3222 1476.54 3335 34.07 −10.98 3392 −9.45 1.53
W120.3 21420 9173 8266 1498.18 8575 35.41 −6.52 8623 −6.00 0.52
W120.6 16748 15596 1567.43 16056 43.80 −4.14 16114 −3.79 0.35
W150.1 56175 5821 5020 2618.11 5208 69.20 −10.54 5257 −9.69 0.85
W150.3 33525 14432 13209 3021.87 13543 70.60 −6.15 13607 −5.72 0.43
W150.6 26310 25076 3172.22 25405 74.46 −3.44 25517 −3.01 0.43
W180.05 80910 4303 3578 9043.25 3726 133.10 −13.41 3812 −11.41 1.20
W180.1 48330 7236 6328 10225.54 6481 130.07 −10.44 6547 −9.52 0.92
W180.3 20147 18436 9887.36 19031 132.14 −5.54 19213 −4.64 0.90
W180.6 37292 35386 10004.11 35949 102.77 −3.61 36124 −3.13 0.48
W200.05 99900 5174 4306 25872.33 4484 71.90 −13.40 4509 −12.85 0.55
W200.1 59700 9271 8092 29003.28 8799 149.50 −5.10 8853 −4.51 0.59
W200.5 38831 36481 28774.17 37477 200.16 −3.49 37552 −3.29 0.20
Table 3. Particular, objective values of some problems do not decrease, instead increase,
such as W150.3. We do not compare the obtained results with Andersson’s 2/3-approximate
algorithm. Because we find that all approximate solutions of M3S obtained by VNS plus
SAGA still are better than that of 0.836-approximate algorithm with the exception of only
W30.1 and W30.3.
6. Conclusions
A variable neighborhood stochastic metaheuristic was proposed to solve the Max 3-cut and
Max 3-section problems in this paper. Our algorithms can solve Max 3-cut and Max 3-
section problems with different sizes and densities. Although 0.836-approximate algorithm
Mathematical Problems in Engineering 15
Table 3: The numerical results of combining VNS-k metaheuristic with SAGA for M3S.
has the very good theoretic results, in numerical aspects, our comparisons indicate that the
proposed VNS metaheuristic is superior to the well-known 0.836-approximate algorithm
and can efficiently obtain very high-quality solutions of the Max 3-cut and Max 3-section
problems.
We mention that the proposed algorithm in fact can deal with higher dimensional G-set
graphs problems created by Pro. Rinaldi using a graph generator, rudy. But, we cannot give
numerical comparisons with 0.836-approximate algorithm since RSDP relaxations of these
problems are out of memory of the current all SDP software. In additionally, if we increase
Kmax or tmax in numerical implementing, then the quality of solution of M3C will further be
improved by VNS.
Funding
This work is supported by the National Natural Science Foundations of China no. 71001045,
10971162, the China Postdoctoral Science Foundation no. 20100480491, the Natural Science
Foundation of Jiangxi Province of China no. 20114BAB211008, and the Jiangxi University
of Finance and Economics Support Program Funds for Outstanding Youths.
Acknowledgment
The authors would like to thank the editor and an anonymous referee for their numerous
suggestions for improving the paper.
16 Mathematical Problems in Engineering
References
1 R. M. Karp, “Reducibility among combinatorial problems,” in Complexity of Computer Computations,
R. Miller and J. Thatcher, Eds., pp. 85–103, Plenum Press, New York, NY, USA, 1972.
2 M. R. Garey, D. S. Johnson, and L. Stockmeyer, “Some simplified NP-complete graph problems,”
Theoretical Computer Science, vol. 1, no. 3, pp. 237–267, 1976.
3 F. Barahona, M. Gröetschel, G. Reinelt, and M. Juenger, “An application of combinotiorial optimiza-
tion to statistical physics and circuit layout design,” Operations Research, vol. 36, no. 3, pp. 493–513,
1988.
4 M. X. Goemans and D. P. Williamson, “Improved approximation algorithms for maximum cut
and satisfiability problems using semidefinite programming,” Journal of the Association for Comput-
ing Machinery, vol. 42, no. 6, pp. 1115–1145, 1995.
5 A. Frieze and M. Jerrum, “Improved approximation algorithms for MAX k-CUT and MAX
BISECTION,” in Integer Programming and Combinatorial Optimization, E. Balas and J. Clausen, Eds.,
vol. 920, pp. 1–13, 1995.
6 M. X. Goemans and D. P. Williamson, “Approximation algorithms for MAX-3-CUT and other
problems via complex semidefinite programming,” Journal of Computer and System Sciences, vol. 68,
no. 2, pp. 442–470, 2004.
7 S. Zhang and Y. Huang, “Complex quadratic optimization and semidefinite programming,” SIAM
Journal on Optimization, vol. 16, no. 3, pp. 871–890, 2006.
8 J. F. Sturm, “Using SeDuMi 1.02, ‘a MATLAB toolbox for optimization over symmetric cones’,”
Optimization Methods and Software, vol. 11, no. 1–4, pp. 625–653, 1999.
9 K.-C. Toh, M. J. Todd, and R. H. Tütüncü, “SDPT3 version 4.0 beta—a MATLAB software for
semidefinite-quadratic-linear programming,” 2004, http://www.math.nus.edu.sg/∼mattohkc/sdpt3
.html.
10 N. Mladenović and P. Hansen, “Variable neighborhood search,” Computers & Operations Research, vol.
24, no. 11, pp. 1097–1100, 1997.
11 P. Hansen, N. Mladenović, and J. A. Moreno Pérez, “Variable neighbourhood search: methods and
applications,” Annals of Operations Research, vol. 175, pp. 367–407, 2010.
12 G. Andersson, “An approximation algorithm for Max p-Section,” Lecture Notes in Computer Science,
vol. 1563, pp. 237–247, 1999.
13 D. R. Gaur, R. Krishnamurti, and R. Kohli, “The capacitated max k-cut problem,” Mathematical
Programming, vol. 115, no. 1, pp. 65–72, 2008.
14 A.-F. Ling, “Approximation algorithms for Max 3-section using complex semidefinite programming
relaxation,” in Combinatorial Optimization and Applications, vol. 5573 of Lecture Notes in Computer
Science, pp. 219–230, 2009.
Hindawi Publishing Corporation
Mathematical Problems in Engineering
Volume 2012, Article ID 761708, 12 pages
doi:10.1155/2012/761708
Research Article
Opposition-Based Barebones Particle Swarm for
Constrained Nonlinear Optimization Problems
Hui Wang
School of Information Engineering, Nanchang Institute of Technology, Nanchang 330099, China
Copyright q 2012 Hui Wang. This is an open access article distributed under the Creative
Commons Attribution License, which permits unrestricted use, distribution, and reproduction in
any medium, provided the original work is properly cited.
This paper presents a modified barebones particle swarm optimization OBPSO to solve con-
strained nonlinear optimization problems. The proposed approach OBPSO combines barebones
particle swarm optimization BPSO and opposition-based learning OBL to improve the quality
of solutions. A novel boundary search strategy is used to approach the boundary between the
feasible and infeasible search region. Moreover, an adaptive penalty method is employed to handle
constraints. To verify the performance of OBPSO, a set of well-known constrained benchmark
functions is used in the experiments. Simulation results show that our approach achieves a
promising performance.
1. Introduction
Many engineering problems can be converted to constrained optimization problems. The aim
of constrained optimization is to find a feasible solution with minimized cost In this paper,
we only consider minimization problems. A general constrained minimization optimization
problem can be defined as follows.
subject to
gj x ≤ 0, j 1, 2, . . . , q,
1.2
hj x 0, j q 1, 2, . . . , m,
2 Mathematical Problems in Engineering
where gx is inequality constraint, hx is the equality constraint, m is the number of
constraints, q is the number of inequality constraints, and m − q is the number of equality
constraints.
Particle swarm optimization PSO is a population-based stochastic search algorithm
developed by Kennedy and Eberhart 1. Although PSO shares many similarities with
evolutionary algorithms EAs, the standard PSO does not use evolution operators such as
crossover and mutation. For PSO’s simple concept, easy implementation, and effectiveness,
it has been widely applied to many optimization areas.
Although PSO has shown a good performance over many optimization problems, it
does not work well when solving complex problems. Especially for constrained optimization
problems, the standard PSO could hardly search promising solutions. The possible reason is
that constrained optimization problems are usually multimodal and having some constraints.
PSO could easily fall into local minima and hardly search feasible solutions. To enhance the
performance of PSO on constrained optimization problems, many improved PSO variants
have been proposed in the past several years.
Meng et al. 2 used a quantum-inspired PSO QPSO to solve constrained economic
load dispatch. The QPSO shows stronger search ability and quicker convergence speed,
not only because of the introduction of quantum computing theory, but also due to two-
novel strategies: self-adaptive probability selection and chaotic sequences mutation. Coelho
3 presented a novel quantum-behaved PSO QPSO to solve constrained engineering
design problems. The proposed approach embedded a Gaussian mutation into QPSO to
prevent premature convergence to local minima. Sun et al. 4 proposed an improved vector
PSO IVPSO based on multidimensional search, in which a simple constraint-preserving
method is used to handle constraints. Liu et al. 5 used a hybrid PSO called PSO-DE to
solve constrained numerical and engineering optimization problems. The PSO-DE integrates
PSO with differential evolution DE to obtain a good performance. Venter and Haftka 6
proposed a new method to solve constrained optimization problems. The constrained, single
objective optimization problem is converted into an unconstrained, biobjective optimization
problem that is solved using a multiobjective PSO algorithm. Lu and Chen 7 presented
an enhanced PSO by employing a dynamic inertia weight to avoid premature convergence.
The inertia weight of every individual is dynamically controlled by the Euclidean distance
between individual and the global best individual. Daneshyari and Yen 8 proposed a
cultural-based constrained PSO to incorporate the information of the objective function and
constraint violation. The archived information facilitates communication among swarms in
the population space and assists in selecting the leading particles in three different levels:
personal, swarm, and global levels.
There have been many modifications to the original PSO algorithm to improve the
efficiency and robustness of the search. Although these modified PSO variants have shown
good search abilities, their performance greatly depends on the control parameters in the
velocity updating model, such as inertia weight w and acceleration coefficients c1 and c2 .
Recently, a parameter-free PSO, known barebones PSO BPSO 9, used Gaussian normal
distribution to update the particles in the population. It does not involve inertia weight,
acceleration coefficients, and velocity. Its performance has been found to be competitive,
and a number of BPSO algorithms have been proposed in the past several years. Omran
et al. 10 incorporate the idea of BPSO into DE. Krohling and Mendel 11 employed a jump
strategy in BPSO to avoid premature convergence. Motivated by the idea of BPSO, this paper
presents an improved BPSO, namely OBPSO, to solve constrained nonlinear optimization
problems. In OBPSO, opposition-based learning OBL concept 12 is used for population
Mathematical Problems in Engineering 3
initialization and generation jumping. To verify the performance of OBPSO, a set of well-
known constrained benchmark problems are used in the experiments. Results obtained by
the proposed OBPSO are compared with those in the literature and discussed.
The rest of the paper is organized as follows. In Section 2, the standard PSO and
barebones PSO are briefly introduced. Section 3 describes our proposed approach. Section 4
presents experimental simulations, results, and discussions. Finally, the work is concluded in
Section 5.
vij t 1 w · vij t c1 · rand1ij · pbestij t − xij t
c2 · rand2ij · gbestj t − xij t , 2.1
where i 1, 2, . . . , N is the particle’s index, N is the population size, Xi xi1 , xi2 , . . . , xiD is
the position of the ith particle; Vi vi1 , vi2 , . . . , viD represents the velocity of the ith particle;
the pbesti pbesti1 , pbesti2 , . . . , pbestiD is the best previous position yielding the best fitness
value for the ith particle; gbest gbest1 , gbest2 , . . . , gbestD is the global best particle found
by all particles so far. The parameter w, called inertia factor, which is used to balance the
global and local search abilities of particles 13, rand1ij and rand2ij are two random numbers
generated independently within the range of 0, 1, c1 and c2 are two learning factors which
control the influence of the social and cognitive components, and t 1, 2, . . . indicates the
iteration number.
Recently, Kennedy 9 developed the barebones PSO BPSO. This new version of PSO
eliminates the velocity term, and the position is updated as follows.
gbestj t pbestij t
xij t 1 G , gbestj t − pbestij t , 2.2
2
where xij t1 is the position of the ith particle in the population, and G represents a Gaussian
distribution with mean gbestj tpbestij t/2 and standard deviation |gbestj t−pbestij t|.
Note that the particle positions are sampled by the above Gaussian distribution. The
BPSO facilitates initial exploration, due to large deviation initially, pbest will be far from the
gbest. As the number of generation increases, the deviation approaches to zero, by focussing
on exploitation of the pbest and gbest 14.
4 Mathematical Problems in Engineering
xj∗ aj bj − xj . 3.1
By staying within variables’ interval static boundaries, we would jump outside of the
already shrunken search space and the knowledge of the current converged search space
would be lost. Hence, we calculate opposite particles by using dynamically updated interval
boundaries aj t, bj t as follows 15.
where xij is the jth position element of the ith particle in the population, xij∗ is the opposite
particle of xij , aj t and bj t are the minimum and maximum values of the jth dimension in
current search space, respectively, randaj t, bj t are random numbers within aj t, bj t,
xmin , xmax is the box-constraint of the problem, and N is the population size, and t 1, 2, . . .,
indicates the generations.
hj x
cvj x 3.5
min 0, gj x ,
where hj x is the jth equality constraint, and gj x is the jth inequality constraint.
Mathematical Problems in Engineering 5
⎧
⎪
⎨fx, if x is feasible
Fx
m
3.6
⎩fx
⎪ kj · cvj x , otherwise,
j1
where m is the number of constraints, kj is a penalty coefficient for each constraint, and fx
is defined by
where fx is the average objective function values in the current swarm, and it is computed
as
1 N
fx · fxi , 3.8
N i1
cvj x
kj fx · m 2
, 3.9
l1 cvl x
where cvl x is the violation of the lth constraint averaged over the current swarm, and it is
computed by
1 N
cvj x · cvij x, 3.10
N i1
where cvij x is the violation of ith particle on the jth constraint.
Infeasible region
New solution
Boundaries
Feasible region
where j 1, 2, . . . , D.
Note that the boundary search strategy works when the current population contains
feasible and infeasible solutions. Figure 1 clearly illustrates the boundary search. As seen, if
the new solution z is infeasible, then replace y with z. Because z is nearer to the boundary
than y. If z is feasible, then replace x with z. Because z is nearer to the boundary than x.
4. Experimental Verifications
4.1. Test Problems
To verify the performance our proposed approach, we employ a set of 13 benchmark
functions from the literature 21, 22. The main characteristics of these benchmark functions
are summarized in Table 1. For specific definitions of these functions, please refer to 23.
Mathematical Problems in Engineering 7
Table 1: Summary of main characteristics of benchmark problems, where F means the feasible region, S
is the whole search region, NE represents nonlinear equality, NI indicates nonlinear inequality, LI means
linear inequality, and α is the number of active constraints at optimum.
problem G05. BPSO achieves better results than PSO on 7 problems. Both of them can find
the global optimum on 5 problems. It demonstrates that the barebones PSO is better than
standard PSO for these problems.
For the comparison of OBPSO with BPSO, both of them obtain the same results on 6
problems. For the rest 7 problems, OBPSO performs better than BPSO. It demonstrates that
the opposition-based learning is helpful to improve the quality of solutions.
To verify the effects of the boundary search strategy, we compare the performance of
OBPSO with OBPSO-1. For the OBPSO-1 algorithm, it does not use the proposed boundary
search strategy. As seen, OBPSO outperforms OBPSO-1 on 6 problems, while they obtain
the same results for the rest 7 problems. These results demonstrate the effectiveness of the
boundary search strategy.
Figure 2 shows the evolutionary processes on four representative problems. It can be
seen that OBPSO converges faster than other 3 PSO algorithms. The OBPSO-1 shows faster
convergence rate than PSO and BPSO. This confirms that the opposition-based learning is
beneficial for accelerating the evolution 15.
Table 2: Mean best function values achieved by PSO, BPSO, OBPSO-1, and OBPSO, and the best results
among the four algorithms are shown in bold.
Functions Optimum PSO mean BPSO mean OBPSO-1 mean OBPSO mean
G01 −15 −13.013 −14.46 −15 −15
G02 −0.803619 −0.47191 −0.58944 −0.70536 −0.79973
G03 −1.0 −1.00468 −1.00339 −1.00287 −1.00126
G04 −30665.539 −30665.5 −30665.5 −30665.5 −30665.5
G05 5126.4981 5213.14 5228.32 5154.76 5126.68
G06 −6961.814 −6961.81 −6961.81 −6961.81 −6961.81
G07 24.306 25.9185 25.3492 24.8576 24.4196
G08 −0.095825 −0.095825 −0.095825 −0.095825 −0.095825
G09 680.630 680.679 680.630 680.630 680.630
G10 7049.248 7639.4 7474.5 7292.82 7049.2605
G11 0.750 0.749 0.749 0.749 0.749
G12 −1.0 −1.0 −1.0 −1.0 −1.0
G13 0.05395 0.819146 0.61415 0.52312 0.33837
−2 0
Mean best function value
−4
Mean best function value
−0.2
−6
−8 −0.4
−10
−12 −0.6
−14
−0.8
−16
0 200 400 600 800 1000 0 200 400 600 800 1000
Iterations (G1) Iterations (G2)
10000 1.4
Mean best function value
1.2
Mean best function value
9000 1
0.8
8000 0.6
0.4
7000 0.2
0 200 400 600 800 1000 0 200 400 600 800 1000
Iterations (G10) Iterations (G13)
Figure 2: The evolutionary processes of PSO, BPSO, OBPSO-1, and OBPSO on four selected problems.
10 Mathematical Problems in Engineering
Table 3: Comparison results of OBPSO with other three PSO algorithms, where “w/t/l” means that OBPSO
wins in w functions, ties in t functions, and loses in l functions, compared with its competitors. The best
results among the four algorithms are shown in bold.
Functions Optimum NVPSO 4 mean RVPSO 21 mean SAVPSO 22 mean OBPSO mean
G01 −15 −13.871875 −14.7151 −14.4187 −15
G02 −0.803619 −0.336263 −0.74057 −0.413257 −0.79973
G03 −1.0 −1.00484 −1.0034 −1.0025 −1.00126
G04 −30665.539 −30665.5 −30665.5 −30665.5 −30665.5
G05 5126.4981 5126.4957 5202.3627 5241.0549 5126.68
G06 −6961.814 −6961.81 −6961.81 −6961.81 −6961.81
G07 24.306 25.1301 24.989 24.317 24.4196
G08 −0.095825 −0.095825 −0.095825 −0.095825 −0.095825
G09 680.630 680.634430 680.653 680.630 680.630
G10 7049.248 7409.065752 7173.2661 7049.2701 7049.2606
G11 0.750 0.749 0.749 0.749 0.749
G12 −1.0 −1.0 −1.0 −1.0 −1.0
G13 0.05395 0.465217 0.552753 0.681123 0.33837
w/t/l — 7/5/1 8/5/0 6/6/1 —
The parameter settings of NVPSO are described in 4. For RVPSO and SAVPSO,
their parameter settings are given in 22. For OBPSO, we use the same parameter values as
described in the previous section. For each test functions, all algorithms stop running when
the number of iterations reaches to the maximum value 1,000.
Table 3 presents average results of NVPSO, RVPSO, SAVPSO, and OBPSO over 30
runs, where Mean represents the mean best function values. The comparison results among
OBPSO and other algorithms are summarized as w/t/l in the last row of the table, which
means that OBPSO wins in w functions, ties in t functions, and loses in l functions, compared
with its competitors.
From the results of Table 3, OBPSO outperforms NVPSO on 7 problems, while NVPSO
only achieves better results on a single problem. For the rest 5 problems, both OBPSO and
NVPSO can find the global optimum. OBPSO performs better than RVPSO on 8 problems,
while both of them obtain the same results for the rest 5 problems. For the comparison of
SAVPSO and OBPSO, both of them achieve the same results on 6 problems. For the rest 7
problems, OBPSO wins 6, while SAVPSO wins only 1.
5. Conclusion
This paper proposes a modified barebones particle swarm optimization to solve constrained
nonlinear optimization problems. The proposed approach is called OBPSO which employs
two novel strategies including opposition-based learning and boundary search. Compared
to other improved PSO variants, OBPSO is almost a parameter-free algorithm except for
the probability of opposition. Moreover, an adaptive penalty method is used to handle
constraints. Experimental verifications on a set of constrained benchmark functions show
that OBPSO achieves a promising performance compared to four other PSO variants. The
parameter po may affect the performance of OBPSO. To determine the best choice of po ,
different values of po will be investigated. This will be conducted in the future work.
Mathematical Problems in Engineering 11
Acknowledgments
The authors would like to thank the editor and anonymous reviewers for their detailed and
constructive comments that helped them to increase the quality of this work. This work
is supported by the Science and Technology Plan Projects of Jiangxi Provincial Education
Department nos. GJJ12641, GJJ12633, and GJJ12307, and the National Natural Science
Foundation of China nos. 61070008, 61165004.
References
1 J. Kennedy and R. Eberhart, “Particle swarm optimization,” in Proceedings of the IEEE International
Conference on Neural Networks. Part 1, pp. 1942–1948, December 1995.
2 K. Meng, H. G. Wang, Z. Y. Dong, and K. P. Wong, “Quantum-inspired particle swarm optimization
for valve-point economic load dispatch,” IEEE Transactions on Power Systems, vol. 25, no. 1, pp. 215–
222, 2010.
3 L. D. S. Coelho, “Gaussian quantum-behaved particle swarm optimization approaches for con-
strained engineering design problems,” Expert Systems with Applications, vol. 37, no. 2, pp. 1676–1683,
2010.
4 C. L. Sun, J. C. Zeng, and J. S. Pan, “An new vector particle swarm optimization for constrained
optimization problems,” in Proceedings of the International Joint Conference on Computational Sciences
and Optimization (CSO ’09), pp. 485–488, April 2009.
5 H. Liu, Z. Cai, and Y. Wang, “Hybridizing particle swarm optimization with differential evolution for
constrained numerical and engineering optimization,” Applied Soft Computing Journal, vol. 10, no. 2,
pp. 629–640, 2010.
6 G. Venter and R. T. Haftka, “Constrained particle swarm optimization using a bi-objective
formulation,” Structural and Multidisciplinary Optimization, vol. 40, no. 1–6, pp. 65–76, 2010.
7 H. Lu and X. Chen, “A new particle swarm optimization with a dynamic inertia weight for solving
constrained optimization problems,” Information Technology Journal, vol. 10, no. 8, pp. 1536–1544, 2011.
8 M. Daneshyari and G. G. Yen, “Constrained multiple-swarm particle swarm optimization within a
cultural framework,” IEEE Transactions on Systems, Man, and Cybernetics, Part A, vol. 18, pp. 1–16,
2011.
9 J. Kennedy, “Bare bones particle swarms,” in Proceedings of the IEEE Swarm Intelligence Symposium (SIS
’03), pp. 80–87, 2003.
10 M. G. H. Omran, A. P. Engelbrecht, and A. Salman, “Bare bones differential evolution,” European
Journal of Operational Research, vol. 196, no. 1, pp. 128–139, 2009.
11 R. A. Krohling and E. Mendel, “Bare bones particle swarm optimization with Gaussian or cauchy
jumps,” in Proceedings of the IEEE Congress on Evolutionary Computation (CEC ’09), pp. 3285–3291, May
2009.
12 H. R. Tizhoosh, “Opposition-based learning: a new scheme for machine intelligence,” in Proceedings of
the International Conference on Computational Intelligence for Modelling, Control and Automation (CIMCA
’05) and International Conference on Intelligent Agents, Web Technologies and Internet Commerce (IAWTIC
’05), pp. 695–701, November 2005.
13 Y. Shi and R. C. Eberhart, “A modified particle swarm optimizer,” in Proceedings of the IEEE
International Conference on Evolutionary Computation (ICEC ’98), pp. 69–73, May 1998.
14 A. P. Engelbrecht, “Heterogeneous particle swarm optimization,” in Proceedings of the International
Conference on Swarm Intelligence, pp. 191–202, 2010.
15 R. S. Rahnamayan, H. R. Tizhoosh, and M. M. A. Salama, “Opposition-based differential evolution,”
IEEE Transactions on Evolutionary Computation, vol. 12, no. 1, pp. 64–79, 2008.
16 H. Wang, Z. Wu, S. Rahnamayan, Y. Liu, and M. Ventresca, “Enhancing particle swarm optimization
using generalized opposition-based learning,” Information Sciences, vol. 181, no. 20, pp. 4699–4714,
2011.
17 H. Wang, Z. Wu, and S. Rahnamayan, “Enhanced opposition-based differential evolution for solving
high-dimensional continuous optimization problems,” Soft Computing, vol. 15, no. 11, pp. 2127–2140,
2011.
18 A. C. C. Lemonge and H. J. C. Barbosa, “An adaptive penalty scheme for genetic algorithms in
structural optimization,” International Journal for Numerical Methods in Engineering, vol. 59, no. 5, pp.
703–736, 2004.
12 Mathematical Problems in Engineering
19 E. K. da Silva, H. J. C. Barbosa, and A. C. C. Lemonge, “An adaptive constraint handling technique
for differential evolution with dynamic use of variants in engineering optimization,” Optimization and
Engineering, vol. 12, no. 1-2, pp. 31–54, 2011.
20 X. Pan, Y. Cao, and Q. Pu, “Improved particle swarm optimization with adaptive constraint handling
for engineering optimization,” Journal of Information and Computational Science, vol. 8, no. 15, pp. 3507–
3514, 2011.
21 H. Y. Lu and W. Q. Chen, “Dynamic-objective particle swarm optimization for constrained
optimization problems,” Journal of Combinatorial Optimization, vol. 12, no. 4, pp. 409–419, 2006.
22 H. Y. Lu and W. Q. Chen, “Self-adaptive velocity particle swarm optimization for solving constrained
optimization problems,” Journal of Global Optimization, vol. 41, no. 3, pp. 427–445, 2008.
23 J. J. Liang, T. P. Runarsson, E. Mezura-Montes et al., “Problem definitions and evaluation criteria
for the CEC 2006, special session on constrained real-parameter optimization,” Tech. Rep., Nanyang
Technological University, Singapore, 2006.
Hindawi Publishing Corporation
Mathematical Problems in Engineering
Volume 2012, Article ID 456047, 18 pages
doi:10.1155/2012/456047
Research Article
A New Hybrid Nelder-Mead Particle Swarm
Optimization for Coordination Optimization of
Directional Overcurrent Relays
Copyright q 2012 A. Liu and M.-T. Yang. This is an open access article distributed under the
Creative Commons Attribution License, which permits unrestricted use, distribution, and
reproduction in any medium, provided the original work is properly cited.
1. Introduction
Transmission lines are exposed to the environment and stretch long distances, which
increases the probability of failure far beyond that of other components of the network,
including generators, transformers, and switchgear equipment.
2 Mathematical Problems in Engineering
The results show that the proposed method, comprising a linear programming LP problem
and a mixed-integer nonlinear programming MINLP problem, is capable of overcoming the
relay coordination problem of a power system. The combined approach effectively increases
the convergence rate of calculation and enhances the capability of the PSO when processing
under constraints. Finally, we discuss the results and draw conclusions.
n
min J wi tik , 2.1
i1
where tjk reveals the operating time of relay j, and the relay is the backup relay of relay i. CTI
is the minimum coordination time interval; its value ranges from 0.2 to 0.5 s. In this study, a
CTI of 0.2 s was chosen.
The function for the nonlinear relay characteristics is based on IEEE standard C37.112-
1996 and is represented as follows:
28.2
ti TDSi × 2 0.1217 , 2.3
Ifi /Ipi − 1
where TDSi and Ipi are the time dial setting and the pickup current setting of the ith relay,
respectively. Ifi is the short-circuit fault current passing through the ith relay.
The constants and exponent in 2.3 define the shape of the extremely inverse trip
characteristics.
The results of this research not only describe the methodology of DOCR coordination
optimization but also demonstrate the feasibility of the TDS and Ip settings of the relays. In
general, DOCR allows for a continuous TDS value, but a discrete Ip setting. To satisfy this
requirement, this study explored both linear and nonlinear programming for DOCR coor-
dination optimization. The variable TDS is optimized according to a predefined Ip for each
DOCR, and this optimization problem can be viewed as a linear programming LP problem.
4 Mathematical Problems in Engineering
For the nonlinear programming NLP problem, variables TDS and Ip are optimized for each
DOCR. In the LP or NLP problem of the DOCR coordination optimization, the TDS values
can range continuously from 0.1000 to 1.1000, and the Ip values can range discretely between
10 and 1000 with a step size of 1, depending on the close-in fault current for each relay.
1 For a random solution, determine the degree of constraint violation ΔV using the
following equation:
V g m×1 ⇒ ΔV −gj x k×1 when gj x > 0, j 1, . . . , m, 3.1
where V is the vector of inequality constraints g, and k is the degree of constraint
violation ΔV .
2 Compute Δx g, where Δx g are the derivatives of these constraints with respect to
the solution vector n decision variables
Δx V Δx g k×n , x 1, . . . , k. 3.2
Mathematical Problems in Engineering 5
Pworst Pbest
Pworst Pbest Pworst Pbest
Pcent
Pcent
Pcent
Prefl Pexp
a b c
Pcont
Pcent Pcent Pcent
Pcont
3 The relationship between changes in the constraint violation ΔV and the solution
vector Δx is expressed as
ΔV Δx V × Δx ⇒ Δx Δx V −1 × ΔV. 3.3
xt 1
xt Δx xt Δx V −1 × ΔV. 3.4
On the other hand, if Prefl has a comparatively low fitness, the simplex is contracted.
Contraction can either be outward or inward, depending upon whether Prefl is better or worse
than Pworst , respectively. The contraction operations i.e., Pworst is replaced with Pcont are
shown in Figures 1d and 1e. If neither contraction improves the worst solution in the
simplex, the best point in the simplex is computed, and a shrinkage is then performed; all the
points of the simplex are moved a little closer towards the best solution Pbest , as shown in
Figure 1f.
rand
co 0.5 , 3.7
3
where c1 and c2 are two positive constants; c0 is an inertia weight, and rand is a random
value between 0, 1. Zahara and Hu suggested c1 c2 2 and c0 0.5 rand /2 15.
However, many experiments have shown that using c0 0.5 rand /3 provides better
results. Equation 3.5 illustrates the calculation of a new velocity for each individual. The
velocity of each particle is updated according to its previous velocity Vid , the particle’s
previous best location pid , and the global best location pgd . Particle velocities for each
dimension are clamped to a maximum velocity Vmax . Equation 3.6 shows how each
particle’s position is updated in the search space.
it easily falls into a local best solution. This drawback is improved by integrating the two
algorithms. Combining the two algorithms and the gradient-based repair methods enables
feasible optimal solutions to be found that satisfy the constraint conditions 18.
Using the advantages mentioned above, the NM-PSO method clearly overcomes the
drawbacks of low convergence speed, the need for more particles, and the inability to deal
with constraint conditions to accurately find optimal solutions.
4. Case Study
The appearance and parameters of the relevant line equipment of two typical test systems
are introduced. We discuss the fault current and the corresponding DOCR relationship of the
coordination pairs of P/B when a close-in three-phase short fault occurs in transmission lines.
Taking the above two test systems as examples, this study validated the feasibility
of the proposed NM-PSO optimization algorithm to solve the DOCR optimal coordination
problem. The results were compared with PSO and original NM-PSO algorithm. The results
of the comparison demonstrate that the proposed NM-PSO algorithm is clearly better than
PSO and original NM-PSO in terms of the objective function, the rate of convergence, and
computation speed.
In this study, the multiples 2 × n 1, 5 × n 1, 10 × n 1, and 20 × n 1 were adopted
as the number of populations to demonstrate the influence of the number of particles on
the proposed algorithm. To observe the process and changes of convergence in the objective
function, the number of iterations was set at 300, to highlight the superior performance of the
proposed system.
8 Mathematical Problems in Engineering
Formulation of the
coordination
problem
NM-PSO parameters
Generate initial
population
Evaluate solutions
PSO method
Yes Sorting
Yes
1 Line 13 3 4
Line 34
8 14 2 9 3 10 4
Line 16
Line 12 Line 45
7
13 Line 26 6 12 Line 65 5 11
1
2 6 5
3.5
Objective function
2.5
1.5
0 50 100 150 200 250 300
Iteration number
Proposed NM-PSO
Original NM-PSO
PSO
Figure 4: Convergence of PSO, original NM-PSO, and proposed NM-PSO to the optimal solution for an
IEEE 8-bus test system LP problem.
Table 1: P/B relays and the close-in fault currents for an IEEE 8-bus test system.
3
2.9
2.8
2.7
Objective function
2.6
2.5
2.4
2.3
2.2
2.1
2
0 50 100 150 200 250 300
Iteration number
Figure 5: Convergence of the NM-PSO for five different random initial populations for an IEEE 8-bus test
system LP problem.
Table 2: Optimal settings of the relays for an IEEE 8-bus test system LP problem.
Table 3: Operating time of P/B relays for an IEEE 8-bus test system LP problem.
11 12
2 3
2 3 13
7
1
1
5
6 4 9 10 8 14
4
5
15 26 29
28 27
25
8 7
24
16 17 18 19 20 21 22
23 30
6 11 10 9
39 38 31
37
12
36
34 33
40 35 32
13 14
the proposed method satisfies all constraints i.e., CTI 0.2, the best coordination setting for
DOCR can be efficiently completed.
As expected, the proposed NM-PSO yields better objective function results than PSO
and original NM-PSO. Figure 4 shows that proposed NM-PSO nearly reached the global
optimum after 189 iterations. The results of this proposed NM-PSO algorithm reveal better
convergence.
To analyze the convergence consistency of the proposed NM-PSO algorithm when
solving the LP problem in the case of different initial values, this study randomly performed
the proposed method five times. As seen in Figure 5, the proposed NM-PSO algorithm can
reduce the objective function to the same value after nearly 200 iterations. The convergence
of the proposed NM-PSO is evidently not affected by different initial values.
Table 4: P/B relays and the close-in fault currents for an IEEE 14-bus test system.
Primary relay Backup relay Primary relay Backup relay Primary relay Backup relay
No. Current No. Current No. Current No. Current No. Current No. Current
1 11650 6 654 8 3880 30 188 19 955 17 955
5 12400 2 1980 29 4720 7 1220 18 725 20 725
2 4260 4 750 29 4720 9 1990 22 1930 29 499
2 4260 12 875 29 4720 13 1070 22 1930 24 1160
2 4260 8 723 29 4720 25 449 22 1930 32 280
3 7310 1 3920 6 3830 3 1280 23 1200 21 434
3 7310 12 848 6 3830 10 1990 23 1200 29 499
3 7310 8 689 6 3830 16 560 23 1200 32 281
11 7180 4 725 4 3920 5 1370 30 1810 21 424
11 7180 1 3920 4 3920 16 562 30 1810 24 1130
11 7180 8 695 4 3920 10 1990 30 1810 32 275
7 7330 1 3920 15 4610 5 1360 31 2060 21 428
7 7330 4 716 15 4610 3 1280 31 2060 24 1150
7 7330 12 845 15 4610 10 1970 31 2060 29 494
13 3280 11 1380 9 3260 5 1390 27 2030 26 1230
12 3130 14 1250 9 3260 3 1310 27 2030 23 808
26 4640 9 2080 9 3260 16 569 25 1430 28 633
26 4640 13 1120 16 1490 40 201 25 1430 23 806
26 4640 7 1270 16 1490 18 388 24 1870 26 1230
26 4640 30 179 16 1490 37 51 24 1870 28 634
10 3110 13 1140 17 2210 15 1110 37 572 35 572
10 3110 7 1290 17 2210 37 51 36 781 38 781
10 3110 25 495 17 2210 40 199 35 1480 33 368
10 3110 30 190 39 2400 15 1120 35 1480 39 1110
14 4030 9 2090 39 2400 18 389 34 1390 36 284
14 4030 7 1270 39 2400 37 47 34 1390 39 1110
14 4030 25 489 38 2530 15 1110 40 654 36 285
14 4030 30 188 38 2530 40 191 40 654 33 370
8 3880 9 2090 38 2530 18 386 32 547 34 547
8 3880 25 489 21 564 19 564 33 783 31 783
8 3880 13 1120 20 1310 22 1310 x x x x
currents passing through them for a close-in fault in this network. There are 92 inequality
constraints corresponding to each relay pair.
This study also used PSO, original NM-PSO, and proposed NM-PSO to solve the
DOCR coordination optimization of the MINLP problem, which required obtaining the TDS
and Ip values of each DOCR. The results after 300 iterations with a population size of 1601
20 × n 1, where n is the number of TDS and Ip variables of the 40 relays are shown in
Table 5. The integer Ip can be directly applied in the current intelligent electronic device IED
settings.
Figure 7 shows the results of the comparison demonstrate that the proposed NM-PSO
algorithm is clearly better than PSO and original NM-PSO in terms of the objective function,
the rate of convergence, and computation speed.
14 Mathematical Problems in Engineering
Table 5: Optimal settings of the relays for an IEEE 14-bus test system MINLP problem.
13
12
11
10
Objective function
9
3
0 50 100 150 200 250 300
Iteration number
Proposed NM-PSO
Original NM-PSO
PSO
Figure 7: Convergence of PSO, original NM-PSO, and proposed NM-PSO to the optimal solution for an
IEEE 14-bus test system MINLP problem.
14
12
Objective function
10
2
0 50 100 150 200 250 300
Iteration number
Figure 8: Convergence of the proposed NM-PSO for five different random initial populations for an IEEE
14-bus test system MINLP problem.
2.5
Objective function
2
1.5
0.5
0
particle = 29 Particle = 71 Particle = 141 Particle = 281
PSO 2.5016 2.2912 2.0468 2.0148
Original NM-PSO 2.3864 2.1569 1.9782 1.9682
Proposed NM-PSO 2.363 2.1021 1.9783 1.9682
20
Objective function
15
10
0
Particle = 81 Particle = 201 Particle = 401 Particle = 801
PSO 19.3419 15.0223 14.441 13.5383
Original NM-PSO 14.9341 13.2828 13.4241 13.1195
Proposed NM-PSO 13.7139 13.2828 13.1982 13.049
b IEEE 14-bus test system LP problem
Figure 9: Objective function values by PSO, original NM-PSO, and proposed NM-PSO algorithms for the
LP problem with n 40 for four different populations.
problems, for the same number of iterations 300 and for four different populations 2 × n
1, 5 × n 1, 10 × n 1, and 20 × n 1.
It can clearly be seen that the proposed NM-PSO method results in better Obj-Fun
values than the PSO and original NM-PSO algorithm. For the more complicated MINLP
problem shown in Figures 10a and 10b, the proposed NM-PSO after 300 iterations results
in even better Obj-Fun values in the case of particles 2 × n 1 than the Obj-Fun values of PSO
algorithm in the case of particles 20 × n 1. Hence, the proposed method produces better
results than the PSO algorithm using fewer particles less computation time.
5. Conclusions
In this paper, the DOCR coordination problem is formulated as a constrained optimization
problem. It can be concluded that the proposed NM-PSO optimization algorithm is applicable
Mathematical Problems in Engineering 17
2.5
Objective function
1.5
0.5
0
Particle = 57 Particle = 141 Particle = 281 Particle = 561
PSO 2.2577 2.1932 2.0096 1.9066
Original NM-PSO 2.1577 2.0232 1.8572 1.7606
Proposed NM-PSO 2.1504 2.0079 1.8289 1.7301
a IEEE 8-bus test system MINLP problem
10
9
8
Objective function
7
6
5
4
3
2
1
0
Particle = 161 Particle = 401 Particle = 801 Particle = 1601
PSO 8.9499 7.6434 4.8378 4.2233
Original NM-PSO 5.1933 7.6434 4.8378 3.6362
Proposed NM-PSO 4.1063 3.7405 3.6522 3.1829
b IEEE 14-bus test system MINLP problem
Figure 10: Objective function values by PSO, original NM-PSO, and proposed NM-PSO algorithms for the
MINLP problem with n 80 for four different populations.
References
1 A. J. Urdaneta, R. Nadira, and L. G. Perez Jimenez, “Optimal coordination of directional overcurrent
relays in interconnected power systems,” IEEE Transactions on Power Delivery, vol. 3, pp. 903–911, 1988.
2 H. A. Abyaneh, M. Al-Dabbagh, H. K. Karegar, S. H. H. Sadeghi, and R. A. J. Khan, “A new optimal
approach for coordination of overcurrent relays in interconnected power systems,” IEEE Transactions
on Power Delivery, vol. 18, no. 2, pp. 430–435, 2003.
3 D. Birla, R. P. Maheshwari, and H. O. Gupta, “A new nonlinear directional overcurrent relay
coordination technique, and banes and boons of near-end faults based approach,” IEEE Transactions
on Power Delivery, vol. 21, no. 3, pp. 1176–1182, 2006.
4 A. S. Noghabi, J. Sadeh, and H. R. Mashhadi, “Considering different network topologies in optimal
overcurrent relay coordination using a hybrid GA,” IEEE Transactions on Power Delivery, vol. 24, no. 4,
pp. 1857–1863, 2009.
5 A. S. Noghabi, H. R. Mashhadi, and J. Sadeh, “Optimal coordination of directional overcurrent relays
considering different network topologies using interval linear programming,” IEEE Transactions on
Power Delivery, vol. 25, no. 3, pp. 1348–1354, 2010.
6 P. P. Bedekar and S. R. Bhide, “Optimum coordination of directional overcurrent relays using the
hybrid GA-NLP approach,” IEEE Transactions on Power Delivery, vol. 26, no. 1, pp. 109–119, 2011.
7 H. H. Zeineldin, E. F. El-Saadany, and M. M. A. Salama, “Optimal coordination of overcurrent relays
using a modified particle swarm optimization,” Electric Power Systems Research, vol. 76, no. 11, pp.
988–995, 2006.
8 M. M. Mansour, S. F. Mekhamer, and N. E. S. El-Kharbawe, “A modified particle swarm optimizer for
the coordination of directional overcurrent relays,” IEEE Transactions on Power Delivery, vol. 22, no. 3,
pp. 1400–1410, 2007.
9 R. C. Eberhart and J. Kennedy, “A new optimizer using particle swarm theory,” in Proceedings of the
6th International Symposium on Micro Machine and Human Science, pp. 39–43, Nagoya, Japan, October
1995.
10 P. Chootinan and A. Chen, “Constraint handling in genetic algorithms using a gradient-based repair
method,” Computers and Operations Research, vol. 33, no. 8, pp. 2263–2281, 2006.
11 Y. Dong, J. F. Tang, B. D. Xu, and D. Wang, “An application of swarm optimization to nonlinear
programming,” Computers and Mathematics with Applications, vol. 49, no. 11-12, pp. 1655–1668, 2005.
12 J. H. Mathews, Numerical Methods Using Matlab, Prentice-Hall, Upper Saddle River, NJ, USA, 2004.
13 T. Niknam, H. D. Mojarrad, and M. Nayeripour, “A new hybrid fuzzy adaptive particle swarm
optimization for non-convex economic dispatch,” International Journal of Innovative Computing,
Information and Control, vol. 7, no. 1, pp. 189–202, 2011.
14 C. H. Lin, J. L. Chen, and Z. L. Gaing, “Combining biometric fractal pattern and particle swarm
optimization-based classifier for fingerprint recognition,” Mathematical Problems in Engineering, vol.
2010, Article ID 328676, 14 pages, 2010.
15 E. Zahara and C. H. Hu, “Solving constrained optimization problems with hybrid particle swarm
optimization,” Engineering Optimization, vol. 40, no. 11, pp. 1031–1049, 2008.
16 E. Zahara and A. Liu, “Solving parameter identification problem by hybrid particle swarm
optimization,” in Proceedings of the International MultiConference of Engineers and Computer Scientists
2010 (IMECS ’10), pp. 36–38, March 2010.
17 A. Liu and E. Zahara, “Parameter identification problem using particle swarm optimization,” in
Proceedings of the 5th International Conference on Natural Computation (ICNC ’09), pp. 275–278, Tianjin,
China, August 2009.
18 Y. T. Kao, E. Zahara, and I. W. Kao, “A hybridized approach to data clustering,” Expert Systems with
Applications, vol. 34, no. 3, pp. 1754–1762, 2008.
19 S. Kamel and M. Kodsi, “Modeling and simulation of IEEE 14 bus system with facts controllers,”
Tech. Rep., 2009.
Hindawi Publishing Corporation
Mathematical Problems in Engineering
Volume 2012, Article ID 827193, 24 pages
doi:10.1155/2012/827193
Research Article
Evaluating the Performance of Taiwan Homestay
Using Analytic Network Process
Copyright q 2012 Yi-Chung Hu et al. This is an open access article distributed under the Creative
Commons Attribution License, which permits unrestricted use, distribution, and reproduction in
any medium, provided the original work is properly cited.
Homestay industry in Taiwan is not only thriving, but also its operation is moving gradually
toward elaboration strategy and in a specialized-operation manner these years. Nevertheless,
the evaluation frameworks of the earlier studies were sporadically constructed from an overall
perspective of homestays. Moreover, the functions, operational model, and natures of homestays
are dissimilar to those of hotels; therefore, if the evaluation criteria of homestays employ the
ones of hotels, it would appear to be incoherent and incompatible. This study has accordingly
developed and constructed a set of evaluation indicators tailor-made for homestay sector through
discussion of literatures and interviewing experts so that the evaluation framework would be
more comprehensive and more practical. In the process of interviewing experts, it was discovered
that dependences lay on the aspects and criteria. Consequently, this research chose the ANP
analytic network process to get the weights and, further, to acquire the homestay business
performance through fuzzy theory. The result reveals, as regards key aspects, homestay proprietors
and customer groups both weight the surroundings of the building and features, service quality,
operation, and management most. In respect to overall homestay performance, customer groups
consider it has reached the satisfactory level.
1. Introduction
The growing work pressure has made travelling as the best way to relieve stress in a society
pursuing high efficiency 1. In the wake of increased incomes, better living, change of
traditional values, and convenient traffic network as well as more leisure time, the demand
of recreational activities are swelling in Taiwan. Recreational activities and tourism has
steadily become part of modern life 2, which has driven a multiaspect development of
Taiwan domestic tourism and recreational market. The most considerable development
and evolvement lie in the accommodation business at tourism attractions, also known as
homestay 3. Homestays are the best Taiwan promoters. If any tourists from all over the
2 Mathematical Problems in Engineering
world visit Taiwan and stay in homestays, hosts of homestays not only can introduce them
the local cultures and past, but also can act as tour guides as well as good public relation
practitioners. Homestays, therefore, play a very essential role. Homestays accordingly are
with crucial marketing functions and can drive local economic development. On these
grounds, the authors have deemed homestay is an important and discuss-worthy issue.
Based on the World Tourism Organization data, concerning global tourism market,
there are approximately seven hundred million people travelling abroad every year. The scale
of global tourism sector including both domestic and international tour takes up over 10% of
the global GDP gross domestic product, which has reached an amount of 3.5 trillion US
dollars. Tourism 2020 Vision, which was published by the World Tourism Organization in
2001, also indicated that the number of tourists in 2020 would reach 1.56 billion persons,
showing a growth of 1.76 times comparing to the 565 million persons in 1995 4. In addition,
according to the Taiwan Tourism Bureau 5, the total tourism revenue in Taiwan had reached
TWD$514 billion in 2010, and the number of visitors has been growing significantly for a
decade. In 2011, the number of in-bound visitors to Taiwan was 6.09 millions, which was
an all-time peak and showed a growth of 9.34% compared to the number in 2010. Visitors
with “pleasure” purpose were 3.63 million persons, which was up by 11.95%. The ones with
business purpose were 984.85 thousand persons, showing an increase of 5.02%. In view of
the data above, the number of inbound visitors has kept rising, which would benefit the
development of Taiwan industries of tourism, recreation, hospitality, and accommodation.
On top of that, the Taiwan Council for Economic Planning and Development revealed
that Taiwan residents’ expenditures in recreational and cultural consumption were increasing
significantly in 2010 with a real growth rate at 9.88%, up by 0.98 percentage point compared
to 2011. The annual growth rate of consumer expenditures in restaurants and lodging has
reached 8.11%, up by 0.22 percentage point compared to 2001 6. Regarding the expected
utility of the development plans of six emerging industries biotech, tourism, green energy,
medical and caring, boutique farming, culture, and creativeness, the Taiwan Council for
Economic Planning and Development 7 proposed that the entire revenue of tourism will
increase from TWD$402.2 billion in 2008 to TWD$550 billion in 2012, including TWD$
300 billion from tourism foreign exchange earning. Further, in 2012, it will attract private
investment with TWD$200 billion. In view of the above, it not only demonstrates leisure
and entertainment activities certainly carrying weight in Taiwan people’s life, but also
highlighting immense business potentials in tourism market.
Taiwan is promoting itself to be a “Green Silicon-Valley Island” for sustainable
tourism; thus, tourism has become one of the major development targets. Homestay sector
has shot up among the trends of recreational farming and ecotourism. According to Taiwan
Tourism Bureau 8, the number of approved homestays has swelled from 1,237 in January
2006 to 3,397 in January 2012, with an increase number of 2,160. The homestay sector, on
this ground, is not only prosperous, but also with immeasurable business opportunities in
tourism sector. In recent years, the operation of homestays has developed into an elaborate
and specialized manner. With marketing and media promotion, it has become an important
option when people engage in recreational activities and has further turned into an emerging
sector with great potential 9. The Regulations for the Management of Home Stay Facilities
10 have been implemented for years since December 12, 2001, and approved homestays
would receive validation certificates. But good ones always mingling with black sheep, some
opportunists have entered this sector and often left negative images, which would impact
consumers’ image about homestays 3. The Chief Secretary of Taiwan Tourism Bureau,
Wayne Liu the present Deputy Director General, said that there would be a classified
Mathematical Problems in Engineering 3
evaluation for Taiwan’s approved homestays in the future. The preliminary plan was to
categorize them into fourteen groups, which are architecture, room features, landscape,
ecology, culture, aboriginal, experiential activities, sport activities, bicycling, hot spring, tour
guiding service, language, industrial heritage, and miscellaneous so that people could clearly
know each homestay’s features 11.
In order to safeguard visitors’ lodging interests, choices of recreational commodities
and its safety, as well as maintaining the quality of recreational activities while enjoying the
vocation, appropriate evaluation items for Taiwan homestay quality validation should be
set up by weighing up multi-aspect evaluating items so that applicable, impartial validation
assessment and criteria could be accordingly established 12. The earlier homestay studies
were mainly about customer satisfaction 13–15, marketing strategies 16–18, experiential
marketing 19–21, operation and management 22–26, and consumer behaviors 27–30,
but few were from an overall perspective to construct homestay evaluation framework.
Furthermore, the homestay sector is showing a thriving trend and it is hoped that an
evaluation mechanism could be formed through this research; notwithstanding, the main
distinguishable features of homestay operation different from those of hotels are that
homestays lay more stress on 1 inexpensive price and help-your-self service, 2 not
emphasizing luxury facilities, but aware of safety issue and hygiene facilities, 3 its service
might not be full-scaled, but is with hospitality, local color, and homeyness. What is more
important is that homestays make use of nature resources and local cultures in order to let
visitors be able to experience the local social customs in person. On top of these, homestays
can provide functions like sport, recreation, amusement, and so on to make visitors fully
enjoy their leisure time 31. On these accounts, it is obvious that the homestays and
hotels vary in their functions, business model, and natures 3, 12, 32–34; therefore, if the
hotel evaluation criteria are applied indiscriminately to homestays, it would appear to be
unconnected and illogical. This research hence deems that it is vital and crucial to develop
and construct applicable homestay evaluation indicators, which is one of the motives.
On the other hand, beside homestay validation, such as “hospitable homestays”, there
is sporadic research into homestay evaluation indicators. For example, Ou and Gian 31
made researches into leisure farming homestay visitor characteristics and needs, whose
measure aspects were facilities, service, landscape, operation, and management. Yen et al.
34 studied the homestay evaluation indicators, whose measure aspects were infrastructures,
service quality, features of resources, and the association with communities. Chen 35 has
constructed evaluation indicators for Taiwan green homestays, whose measure aspects were
green building, sustainable landscape engineering, organic farming, environment education,
community co-prosperity, and so forth. Shi 12 researched Taiwan homestay quality
validation, whose measure aspects were surroundings and facilities, business features,
service quality, and community participation. To conclude, it is found that most underlined
the hardware, service quality, or exterior settings. It was discovered in the interview process
that the forementioned measure aspects are essential to the current external context; however,
with respect to evaluation, it should be brought back to the homestay’s natures and geist,
such as the interaction between the homestay hosts and guests. This study has thus added
the aspect of “homestay geist and community co-prosperity” into the final framework and
discussed its nuances, which is another motive.
At the end, it found that there is dependence among aspects and criteria of the
homestay evaluation. Saaty 36 suggested employing ANP to solve the decision-making
issues of dependence. Zadeh 37 introduced fuzzy set theory in 1965, which is to address
the uncertainties arising from subjective belief by quantitative methods, and by fuzzifying, it
4 Mathematical Problems in Engineering
would provide a better promotability, error tolerance, and is better for applying to the real-
world nonlinear system 38.
The following structures of this paper are: Section 2 is the constructing process of the
research framework, mainly by literature review and interviewing experts; Section 3 is about
methodology, including majority rule, ANP analytic network process and Fuzzy Theory;
Section 4 is about empirical results; Section 5 is conclusion and suggestions.
3. Methodology
3.1. Majority Rule
Majority rule is a decision rule that selects alternatives which are the major consensus among
most experts. Majority rule can be the more-than-half-vote rule or two-thirds rule 40. On
Mathematical Problems in Engineering 5
this basis, this research has set that all must agree have 100 degree full influence as the
screening threshold when determining criteria relevancy.
6 Mathematical Problems in Engineering
Aspect Definition
The entire exterior and interior design of the homestay. Local
Surroundings of the building and
cultural features are demonstrated and incorporated into the
features A
homestay.
Customer satisfaction at services provided by homestay
Service quality B
proprietors.
Homestay facilities C Hardware of the homestay.
This section is about how the homestay proprietors manage, plan,
Homestay operation and
and design its rooms and surroundings as well as how they guard
management D
roomers’ lodging privacy and safety.
The homestay proprietors carry out the homestay operation in
Homestay geist and community person and frequently interact with their guests and run the
co-prosperity E homestay business with a concept of incorporating into its local
community in order to help local economic prosper.
In the real environment, there are many decision-making problems that cannot just take the
pure hierarchical relationship to construct the framework, because the high- and the low-
level element may exist the dependence relationship and interaction 41. As a result, Saaty
41 advanced ANP which has dependence and feedback, and in 2001 36, he recommended
using ANP to solve the problem of interdependent relationships among the criteria or
alternatives. In ANP, when nodes correspond to levels or components, that means there
exists the network feedback in a system 42. The elements in the nodes may influence
some or all elements in other nodes. In the network, all the nodes can be source nodes,
intermediate nodes, or sink nodes. The relationship in the network is represented by the arc,
and the direction of arrow means dependence relationship 41. When the two nodes have the
external dependence, it will be represented by the two-way arrow; the nodes in the elements
have the internal dependence, it will be represented by the circle arc 43.
ANP has four steps 41, 44 as follows.
Step 1. Establish the model and the framework. The question should be clearly described and
decomposed a rational system.
Step 2. Do the pairwise comparison to get the priority vector. The decision makers are asked to
answer the series of pairwise comparison of two random elements or criteria.
Step 3. Construct the supermatrix. The concept of super matrix is similar to the Markov chain
Saaty, 1996. To obtain the global priority vector of systematic dependence, we must put the
local priority vector into each corresponding line element to get the super matrix.
Step 4. Choose the best alternative. The alternative getting the highest weight is the best
alternative. In our study, we do not have the alternatives; therefore, we just finish from Step 1
to Step 3.
8
Table 5: Linguistic values representations for Likert’s scale and fuzzy set.
4. Empirical Result
4.1. Validity and Reliability
The framework of this research is based on literature review and expert interviews; therefore,
it is with content validity and expert validity. This research has employed Cronbach’s α
10 Mathematical Problems in Engineering
Aspect Cronbach’s α
Surroundings of the building and features A 0.76
Service quality B 0.83
Homestay facilities C 0.75
Homestay operation and management D 0.90
Homestay geist and community co-prosperity E 0.93
coefficient to assess the scale reliability. Cronbach 49 suggested that an α coefficient less than
0.35 would be deemed as low reliability, the one between 0.35 and 0.7 would be considered
as mediocre reliability, and the one above 0.7 is high reliability. The result reveals that
Cronbach’s α coefficients of all aspects are above 0.7 as in Table 6, which indicates the scales
of this research are with good reliability.
retrieved in Homestay B, two for homestay hosts and 28 for lodgers. 80 copies are all valid
samples.
A B C D E
A 1 1 1 1
B 1 1 1 1
C 1 1 1
D 1 1 1 1 1
E 1 1 1 1
A B C D E
A 0.4286 0.25 0.2727 0.25
B 0.375 0.25 0.2727 0.25
C 0.125 0.25 0.0909
D 0.4286 0.375 0.25 0.2727 0.25
E 0.1429 0.125 0.0909 0.25
sector is the largest and is followed by the one working in electronic manufacturing sector
at Homestay B. About education level, most are holding college degree. Concerning average
monthly earning, the group earning NT$ 20,001–40,000 is the major one and is followed by
the one earning NT$ 40,001–60,000 at Homestay A, and most of the Homestay B’s lodgers are
from the one earning NT$ 40,001–60,000.
4.4.1. Aspects
This research has verified that the CI consistency index value of each pairwise comparison’s
matrix is less than 0.1, and the values of limiting supermatrix were obtained by arithmetic
average in order to get the rank of aspects as in Tables 11 and 12. According to Table 11,
Homestay A’s key aspects are homestay operation and management, surroundings of
the building and features, and service quality. Key aspects of its customer group are
homestay operation and management, service quality, and surroundings of the building and
features. Both do not attach great importance to homestay facility and homestay geist and
community co-prosperity. Based on Table 12, key aspects of Homestay B are service quality,
homestay operation and management, and surroundings of the building and features. Its
customer group’s key aspects are homestay operation and management, service quality, and
surroundings of the building and features. Both Homestay B and its customer group do
not pay much attention to homestay facility as well as homestay geist and community co-
prosperity.
Mathematical Problems in Engineering 13
A B C D E
A 0.2452 0.2452 0.2452 0.2452 0.2452
B 0.2242 0.2242 0.2242 0.2242 0.2242
C 0.0773 0.0773 0.0773 0.0773 0.0773
D 0.3293 0.3293 0.3293 0.3293 0.3293
E 0.1240 0.1240 0.1240 0.1240 0.1240
Table 11: The rank of aspect importance of homestay A limiting supermatrix value.
4.4.2. Criteria
This research has verified the CI consistency index value of each pairwise comparison’s
matrix is less than 0.1, and the values of limiting supermatrix were obtained by arithmetic
average in order to get the rank of criteria as in Tables 13 and 14. According to Table 13,
the top 5 key criteria of the Homestay A are to utilize natural ventilation sufficiently, room
tidiness, to utilize plenty natural light, safety e.g., lodger insurance and room safety, and
room coziness, and the top 5 key criteria of lodgers are room coziness, to utilize plenty natural
light, to utilize natural ventilation sufficiently, safety e.g., lodger insurance and room safety,
service attitude e.g., reception service and treating lodgers with voice of the customers. In
addition, the bottom 5 criteria that Homestay A values are room themes and features e.g.,
oceanic themes, promoting and preserving local cultural resources, cooking facilities e.g.,
kitchen, contribution for living quality of local community, and medical aid e.g., first-aid
box. The bottom 5 criteria that Homestay A lodgers value are room themes and features
e.g., oceanic themes, the beautification and uniqueness of the interior design, to incorporate
the local heritage and landscape elements into design, contribution for living quality of local
community, and medical aid e.g., first-aid box.
According to Table 14, the top 5 key criteria of Homestay B are catering service and
quality e.g., the hosts prepare diversified breakfast in person, freshness of ingredients,
service attitude e.g., reception service, to treat lodgers with voice of the customers,
homestay business features e.g., aboriginal culture, room coziness, and pick-up service
e.g., offering free pick-up service. The ones of its lodgers are safety e.g., lodger insurance
and room safety, service attitude e.g., reception service, to treat lodgers with voice of the
customers, room coziness, pick-up service e.g., offering free pick-up service, and catering
service and quality e.g., the hosts prepare diversified breakfast in person, freshness of
ingredients. Furthermore, the bottom 5 criteria that the Homestay B value are to maintain the
land’s vitality and good condition in the process of design and construction, the beautification
and uniqueness of the interior design, room themes and features e.g., oceanic themes,
medical aid e.g., first-aid box, and contribution for living quality of local community. The
bottom 5 criteria that Homestay B lodgers value are greenization and uniqueness of the
14 Mathematical Problems in Engineering
Table 12: The rank of aspect importance of Homestay B limiting supermatrix value.
Table 13: The rank of criteria importance of Homestay A limiting supermatrix value.
garden design, to maintain the land’s vitality and good condition in the process of design
and construction, to incorporate the local heritage and landscape elements into design,
contribution for living quality of local community, and medical aid e.g., first-aid box.
In conclusion, both homestays set store by room coziness, but do not attach much
weight to the medical aid e.g., first-aid box, room themes and features e.g., oceanic
themes as well as contribution for living quality of local community. Customer groups of
Mathematical Problems in Engineering 15
Table 14: The rank of criteria importance of homestay B limiting supermatrix value.
both homestays regard service attitude e.g., reception service, to treat lodgers with voice
of the customers, room coziness, and safety e.g., lodger insurance and room safety, but
do not attach much importance to incorporating the local heritage and landscape elements
into design, medical aid e.g., first-aid box, and contribution for living quality of local
community.
Besides, the room coziness is the more outstanding one in the performance, of which the
weight rank is 7 and the performance is 5. The bottom 5 criteria that customer does not attach
much importance to are room themes and features e.g., oceanic themes, beautification and
uniqueness of the interior design, to incorporate the local heritage and landscape elements
into design, contribution for living quality of local community, and medical aid e.g., first-
aid box. The Homestay A customer group felt Homestay A did not perform well on these
criteria.
Based on Table 16, the top 5 key criteria of the Homestay B customer group are safety
e.g., lodger insurance and room safety, service attitude e.g., reception service, to treat
lodgers with voice of the customers, room coziness, pick-up service e.g., offering free pick-
up service, and catering service and quality e.g., the hosts prepare diversified breakfast
in person, freshness of ingredients. The customer group deemed Homestay B performed
well in most of these criteria, but not the pick-up service e.g., offering free pick-up service,
whose weight rank is 4 and performance is 7. Besides, the performance in utilizing natural
Mathematical Problems in Engineering 17
lodger insurance and room safety, but not well in incorporating the local heritage and
landscape elements into design, medical aid e.g., first-aid box, and contribution for living
quality of local community.
83, which is not far from the option of satisfactory 84; that means it is close up to satisfactory,
and it will show as satisfactory. According to Table 18, for Homestay A, 5 customers felt
highly satisfactory, 25 customers considered satisfactory, the number of slightly satisfactory is
14 and the number of feeling slightly unsatisfactory is 5. Regarding Homestay B, there were
7 customers feeling highly satisfactory, 18 thought it was a satisfactory experience, and 3 felt
slightly satisfactory. Overall, both of them are satisfactory.
With regard to homestay overall performance, five customers felt highly satisfactory
with Homestay A, 25 customers considered satisfactory, 14 were slightly satisfactory, and 5
are slightly unsatisfactory. For Homestay B, 7 customers felt highly satisfactory, 18 thought
it was a satisfactory experience, and 3 were slightly satisfactory. To conclude, the overall
performance of both homestays met the expectation of their customer groups.
5.2. Suggestion
Regarding aspects, both homestays and their customer groups attach greater importance to
surroundings of the building and features, service quality, as well as homestay operation
and management, but not to the homestay facilities and homestay geist and community
co-prosperity, which is consistent with the real situation in Taiwan. Homestay operation in
Taiwan is working toward a specialized model. Homestays tend to attract people with their
magnificent structures and gorgeous settings, among which some even spent hundreds of
millions to set up the facilities. Nevertheless, expert group suggested that is working against
homestay’s natures and geist. Homestay should be a sideline business, and the interaction
between hosts and lodgers should be very tight, but not merely about lodging. On this
ground, this study suggest it should start from educating consumers in order to make them
understand what homestays are and what they should value is not merely about the exterior
condition. Moreover, the interaction between homestay hosts and lodgers should be built up,
such as taking the lodgers to have an in-depth travel in order to give a great impression to
the customers and they would have a wanna-come-again idea. On top of those, government
could work with local communities to form “homestay villages,” which would be helpful for
driving local economic development. Smangus tribe and Tao-Mi Eco Village, for instance, are
both very successful examples of building homestay community in Taiwan.
Regarding customer groups’ criteria importance rank and their lodging homestay’s
performance, the top five key criteria of Homestays A’s customer group are room coziness,
to utilize plenty natural light, to utilize natural ventilation sufficiently, safety e.g., lodger
insurance and room safety, and service attitude e.g., reception service, to treat lodgers with
voice of the customers. The leading four performance rank of Homestay A is consistent with
its importance rank, but service attitude e.g., reception service, to treat lodgers with voice of
the customers is not performing very well, whose weight rank is 5, but performance is 7. This
study therefore suggests Homestay A can improve its service skills by on-the-job trainings or
special courses. In addition, its pick-up service offering free pick-up service also did not
perform well, whose weight rank is 6 and performance is 14. This study deems that it is
because Homestay A does not offer this service, but its customer group apparently attaches
great importance to it. Therefore, it is suggested that Homestay A could buy a shuttle bus or
outsource this service to offer this service and improve the satisfaction. The customer group
of Homestay B values safety e.g., lodger insurance and room safety most, and Homestay
B performs best in this criterion as well. It also performs great in other key criteria, such
as service attitude e.g., reception service, to treat lodgers with voice of the customers,
whose weight rank is 2 and performance is 3, and room coziness, whose weight rank is 3
and performance is 2, and catering service and quality e.g., the hosts prepare diversified
breakfast in person, freshness of ingredients, whose weigh rank is 5 and performance is
4. But Homestay B did not perform well in pick-up service offering free pick-up service,
whose weight rank is 4 and performance is 7. This study believes the main cause is same
as Homestay A, which does not have this service; therefore, the improving measure is
as forementioned. To conclude, the overall performance of both homestays is satisfactory;
Mathematical Problems in Engineering 21
Table 19
No. Date Expert opinions Response and actions
1 This study has integrated and
Both Expert A and C thought the
altered the aspects and criteria with
aspects and criteria in the prelimi-
similar implication as shown in
nary framework are appropriate. It
Table 2.
should not stress the exterior set-
2 A new aspect was added and was
tings and facilities, but the home-
named “Homestay geist and
stay geist and natures in order to
1 2011/10/27 community co-prosperity”, in which
make the framework more distributive
there are five criteria: “interaction
and comprehensive. On top of those,
between hosts and guests”, “guiding
it should underline the connection
service”, “arranging local experiential
between homestays and local commu-
activities and food”, “contribution for
nities. At the end, Expert A indicated
living quality of local community”
that aspects and criteria with similar
and “initiating preserving actions
implications could be merged.
toward local resources”
Expert B pointed out that hardware
of homestays are not on par with the
one of hotels. Therefore, they should
rely heavily on their natures and geist
as well as the interaction with guests,
and he also suggested underlining a
1 A new criteria of “promoting and
marketing concept of “homestay vil-
2 2011/11/17 preserving local cultural resources”
lage.” At the end, the expert consid-
was added into the aspect of homestay
ered the framework after integrating
geist and community co-prosperity
and altering is appropriate, but if a new
criterion of “promoting and preserv-
ing local cultural resources” is added
under the aspect of homestay geist and
community co-prosperity, it would be
more comprehensive.
1 This research was carried out
by emails to ask Expert A about 1 This research has added remarks to
3 2011/11/24 the applicability of the after-integrate the criteria under the service quality
framework. He reckoned it is distribu- aspect as demonstrated in Table 2.
tive and comprehensive.
2 Expert A and this research both
deemed that the rooms of homestays
are limited to 5 to 15 and homestay
should be a sideline business;
therefore, having employees is not an
2 Expert C thought some remarks
essential option. As a result, it would
should be included in each of the four
not be included in the framework.
criteria of the aspect of service qual-
3 Expert A and this research both
ity. In addition, she suggested adding
deemed that homestay is a sideline
“employee” and “outdoor facilities” as
business and its nature is that only if
new evaluation criteria.
there are surplus rooms, it would
become a homestay. Therefore, having
outdoor facilities is not an essential
option, which was not going to be
included in the framework.
22 Mathematical Problems in Engineering
however, if they can improve the criteria that did not have good performance, they would be
able to achieve the level of highly satisfactory.
5.3. Contribution
Researches on homestays most were about the customer satisfaction 13–15, marketing
strategies 16–18, experiential marketing 19–21, operation management 22–26, or con-
sumer behaviors 27–30, but few constructed the evaluation frameworks from perspectives
of the entire homestay business operation. Moreover, the functions, operation models, and
natures of homestays are different from hotels 3, 12, 32–34; therefore, if the evaluation
standards for hotels are applied to homestays, it would be incompatible and at variance. On
these grounds, this research has been carried out through literature review and interviewing
experts in order to develop and construct a set of evaluation indicators applicable to homestay
business and to draw up an evaluation framework, which would be more comprehensive and
tailor-made for the real practice. The expert group pointed out in the interviewing process
that aspects like hardware facilities, service quality, and exterior settings are essential to
the current demand, but they all stressed that the evaluation should be brought back to
the natures and geist of homestay business, such as the interaction between homestay hosts
and lodgers instead of only pursuing the exterior setting relentlessly. Based on this ground,
the value of this study and the evaluation frameworks’ differences between this study and
the earlier ones 12, 31, 34, 35 are to add the criterion of homestay geist and community
co-prosperity and to make a further discussion. At the end, it is hoped that the results of
this study could be a reference for homestay proprietors to improve their business and for
government to promote the homestay business in the future.
Appendix
See Table 19.
References
1 J. Y. Rong, Travel Agency: Principles, Practices, and Operations, Yang-Chih, Taipei City, Taiwan, 2002.
2 T. L. Liou and J. W. Chen, “The study of influence of perceptions in tourism developing impact
to tourism developing attitudes of residents of dahshi’s old streets,” NPUST Humanities and Social
Sciences Research, vol. 3, no. 1, pp. 20–36, 2009.
3 S. Y. Lin, M. C. Lin, and G. Y. Liu, “Taiwan host B&B quality acreditation,” Quality Magazine, vol. 46,
no. 12, pp. 32–34, 2010.
4 Z. X. Chen and R. X. Chang, Introduction to Leisure and Recreation Industry, Yang-Chih, Taipei City,
Taiwan, 2008.
5 Tourism Bureau M.O.T.C., 2012, http://admin.taiwan.net.tw/public/public.aspx?no315.
6 Now News, 2011, http://www.nownews.com/2011/05/10/91-2711236.htm.
Mathematical Problems in Engineering 23
32 Z. Q. Gao, The study on evaluation indicators for special B&B Inn—the example of seashore type [Unpublished
M.S. thesis], Kai Nan University, Lujhu Township, Taiwan, 2011.
33 S. Y. Lin, “Analysis of product levels of leisure industry—a case study of Taiwan B&B service quality,”
Journal of Customer Satisfaction, vol. 1, no. 1, pp. 145–168, 2005.
34 C. H. Yen, C. J. Chang, and L. L. Jian, “A study of evaluation indicators for taiwan home stay facility,”
Journal of Farmers’ Organizations, vol. 8, pp. 133–175, 2006.
35 Y. J. Chen, An investigation on evaluation indicators for Eco-inn selecting in Taiwan [Unpublished M.S.
thesis], National Taipei College of Nursing, Taipei City, Taiwan, 2008.
36 T. L. Saaty, Decision Making with Dependence and Feedback: The Analytic Network Process, RWS
Publications, Pittsburgh, Pa, USA, 2nd edition, 2001.
37 L. A. Zadeh, “Fuzzy Sets,” Information and Computation, vol. 8, pp. 338–353, 1965.
38 M. C. Su and X. D Chang, Machine Learning: Neural Networks, Fuzzy Systems, and Genetic Algorithms,
Quanhua, Taipei City, Taiwan, 3rd edition, 2004.
39 General Chamber of Commerce of the R.O.C., 2006, http://60.244.127.66/cgi-bin/big5/p1215/ko04.
40 J. Y. Teng, Project Evaluation: Methods and Applications, The Center of Planning and Management in
National Taiwan Ocean University, Keelung City, Taiwan, 2nd edition, 2005.
41 T. L. Saaty, Decision Making with Dependence and Feedback: The Analytic Network Process, Pittsburgh, Pa,
USA, RWS Publications, 1996.
42 T. L. Saaty, The Analytic Hierarchy Process, McGraw-Hill, New York, NY, USA, 1980.
43 J. Sarkis, “Quantitative models for performance measurement systems—alternate considerations,”
International Journal of Production Economics, vol. 86, no. 1, pp. 81–90, 2003.
44 L. M. Meade and J. Sarkis, “Analyzing organizational project alternatives for agile manufacturing
processes: an analytical network approach,” International Journal of Production Research, vol. 37, no. 2,
pp. 241–261, 1999.
45 H. J. Zimmermann, Fuzzy Set Theory and Its Applications, Kluwer, Boston, Mass, USA, 1996.
46 W. Pedrycz and F. Gomide, An Introduction to Fuzzy Sets: Analysis and Design, Complex Adaptive
Systems, MIT Press, Cambridge, Mass, USA, 1998.
47 D. Dubois and H. Prade, Fuzzy Sets and Systems: Theory and Application, vol. 144 of Mathematics in
Science and Engineering, Academic Press, New York, NY, USA, 1980.
48 M. Delgado, F. Herrera, E. Herrera-Viedma, and L. Martı́nez, “Combining numerical and linguistic
information in group decision making,” Information Sciences, vol. 107, no. 1–4, pp. 177–194, 1998.
49 L. J. Cronbach, “Coefficient alpha and the internal structure of tests,” Psychometrika, vol. 16, no. 3, pp.
297–334, 1951.
Hindawi Publishing Corporation
Mathematical Problems in Engineering
Volume 2012, Article ID 979307, 18 pages
doi:10.1155/2012/979307
Research Article
Mixed Mortar Element Method for
P1NC /P0 Element and Its Multigrid Method for
the Incompressible Stokes Problem
Copyright q 2012 Y. Jiang and J. Chen. This is an open access article distributed under the Creative
Commons Attribution License, which permits unrestricted use, distribution, and reproduction in
any medium, provided the original work is properly cited.
We discuss a mortar-type P1NC /P0 element method for the incompressible Stokes problem. We
prove the inf-sup condition and obtain the optimal error estimate. Meanwhile, we propose a
W-cycle multigrid for solving this discrete problem and prove the optimal convergence of the
multigrid method, that is, the convergence rate is independent of the mesh size and mesh level.
Finally, numerical experiments are presented to confirm our theoretical results.
1. Introduction
As we all know, the application of viscous incompressible flows is of considerable interest. For
example, the design of hydraulic turbines, or rheologically complex flows appears in many
processes which are involved in plastics and molten metals. Therefore, in recent decades,
many engineers and mathematicians have concentrated their efforts on the Stokes problem,
especially the problem that can be handled by the finite element methods. In 1, Girault
and Raviart provided a fairly comprehensive treatment of the most recent development
in the finite-element method. Some new divergence-free elements were proposed to solve
Stokes problem recently see 2, 3 and others. Due to this development in the finite-element
theory, many numerical algorithms were established to solve the Stokes equations. Among
these algorithms, multigrid methods and domain decomposition methods for the Stokes
equations are very prevalent. In 4, the authors constructed an efficient smoother. Based on
the smoother, the multigrid methods have been greatly developed see 5, 6. Meanwhile,
2 Mathematical Problems in Engineering
a FETI-DP method was extended to the incompressible Stokes equations in 7, 8, a BDDC
algorithm for this problem was developed too in 9 and others.
In the last twenty years, mortar element methods have attracted much attention and it
was first introduced in 10. This method is a nonconforming domain decomposition method
with nonoverlapping subdomains. In mortar finite-element methods, the meshes on adjacent
subdomains may not match with each other across the interfaces of the subdomains. The
coupling of the finite-element functions on adjacent meshes is done by enforcing the so-called
mortar condition across the interfaces see 10 for details. There have been considerable
researches on the mortar element methods see 11–13 and others.
In 12, the author discussed the mortar-type conforming element P2 /P1 element
method for the Stokes problem, and then Chen and Huang proposed the mortar-type
nonconforming element Q1rot /Q0 element method for the problem in 5. It is well known
that the rotated Q1 element is a rectangle element, and it is not a flexible finite element
since it is only suitable for the rectangular or L-shape-bounded domain. Moreover, the
rotated Q1 element is a quadratic element and is not as convenient as the linear elements in
calculating.
In this paper, we apply the mortar element method coupling with P1 nonconforming
finite element to the incompressible Stokes problem. The P1 nonconforming finite element is
a triangular element and it is suitable for more extensive polygonal domain than the rotated
Q1 element. Moreover, owing to its linearity, the computational work is less than the rotated
Q1 element. We prove the so-called inf-sup condition and obtain the optimal error estimate.
When solving the discrete problem, we also present a W-cycle multigrid algorithm, but the
analysis about the convergence of the multigrid is different from 5. We only prove that the
prolong operator satisfies the criterion which proposed in 14 and we obtain the optional
convergence with simpler analysis than that in 5. Meanwhile, we do some numerical
experiments which were realized in 5. From numerical results, we note that the number
of iterations is less than the rotated Q1 element method when achieving the same relative
error.
The rest of this paper is organized as follows. In Section 2, we review the Stokes
problem and introduce the mortar element method for P1 nonconforming element. Section
3 gives verification of the inf-sup condition and error estimate. The multigrid algorithm and
the convergence analysis are given in Sections 4 and 5, respectively. The last section presents
some numerical experiments. Throughout this paper, we denote by “C” a universal constant
which is independent from the mesh size and level, whose values can differ from place to
place.
2. Preliminaries
We only consider the incompressible flow problem, the steady-state Stokes problem, so that
we can compare the results with those in 5.
The partial differential equations of the model problem is
−Δu ∇p f in Ω,
div u 0 in Ω, 2.1
u 0 on ∂Ω,
Mathematical Problems in Engineering 3
The mixed variational formulation of problem 2.1 is to find u, p ∈ H01 Ω2 × L20 Ω such
that
2
au, v b v, p f, v, ∀v ∈ H01 Ω ,
2.3
b u, q 0, ∀q ∈ L20 Ω,
where the bilinear formulations a·, · on H01 Ω2 × H01 Ω2 , b·, · on H01 Ω2 × L20 Ω
and the dual parity ·, · on L2 Ω2 × L2 Ω2 are given, respectively, by
au, v ∇u · ∇v dx, b v, q − div vq dx, f, v f · v dx. 2.4
Ω Ω Ω
It is well known that the bilinear form b·, · satisfies the inf-sup condition, that is, there exists
a positive constant β for any q ∈ L20 Ω such that
b v, q
sup ≥β q L2 Ω
. 2.5
2 vH 1 Ω2
v∈H0 Ω
1
According to the assumption on Ω and the saddle point theory in 15, we know that if f ∈
L2 Ω2 , then there exists a unique solution u, p ∈ H01 Ω H 2 Ω2 × L20 Ω H 1 Ω
satisfying
We now introduce a mortar finite-element method for solving problem 2.3. First, we
partition Ω into nonoverlapping polygonal subdomains such that
N
Ω Ωi , Ωi Ωj φ if i / j. 2.7
i 1
They are arranged, so that the intersection of Ωi Ωj for i / j is an empty set or an edge, or
a vertex; that is, the partition is geometrically conforming. Denote by γm the common open
N
edge to Ωi and Ωj , then the interface Γ i 1 ∂Ωi \ ∂Ω is broken into a set of disjoint open
straight segments γm 1 ≤ m ≤ M, that is,
M
Γ γ m, γm γn φ if m / n. 2.8
m 1
4 Mathematical Problems in Engineering
By γmi we denote an edge of Ωi called mortar and by δmj an edge of Ωj that geometrically
occupies the same place called nonmortar.
With each Ωi , we associate a quasiuniform triangulation Th Ωi made of elements
that are triangles. The mesh size hi is the diameter of largest element in Th Ωi . We define
N
h max1≤i≤N hi , Th i 1 Th Ωi . Let CR nodal points be the nonconforming nodal points,
that is, the midpoints of the edges of the elements in Th Ωi . Denote the set of CR nodal
points belonging to Ωi , ∂Ωi and ∂Ω by ΩCR ih
, ∂ΩCR
ih
and ∂ΩCR
h
, respectively.
For each triangulation Th Ωi on Ωi , the P1 nonconforming element velocity space and
piecewise constant pressure space are defined, respectively, as follows:
2
Xh Ωi vi ∈ L2 Ωi |vi |τ is linear ∀τ ∈ Th Ωi ,
vi is continuous at midpoint of τ, vmi 0 ∀mi ∈ ∂ΩCR , 2.9
h
Qh Ωi qi ∈ L2 Ωi qi τ is a constant for τ ∈ Th Ωi .
Then the product space X h Ω N i 1 Xh Ωi is a global P1 nonconforming element space for
Th on Ω.
For any interface γm γmi δmj 1 ≤ m ≤ M, there are two different and
independent triangulations Th γmi and Th δmj , which produce two sets of CR nodes
belonging to γm : the midpoints of the elements belonging to Th γmi and Th δmj denoted
CR CR
by γmi and δmj , respectively.
In order to introduce the mortar condition across the interfaces γm , we need the
auxiliary test space Sh δmj which is defined by
Sh δmj
2
v ∈ L δmj
2
| v is piecewise constant on elements of triangulation Th δmj .
2.10
For each nonmortar edge δmj , define the L2 -projection operator: Qh,δmj : L2 γm 2 →
Sh δmj by
Qh,δmj v, w v, wL2 δmj , ∀w ∈ Sh δmj . 2.11
L2 δ mj
2.12
∀γm γmi δmj ⊂ Γ ,
the condition of the equality in 2.12 which the velocity function v satisfies is called mortar
condition.
Mathematical Problems in Engineering 5
Qh Ω q ∈ L20 Ω | qΩi ∈ Qh Ωi . 2.13
We now establish the discrete system for problem 2.3 based on the mixed finite-
element spaces Xh Ω × Qh Ω.
We first define the following formulations:
ahi uih , vih ∇uih · ∇vih dx, ∀uih , vih ∈ Xh Ωi ,
τ∈Th Ωi τ
2.14
bhi vih , phi − div vih · phi dx, ∀vih ∈ Xh Ωi , ∀phi ∈ Qh Ωi .
τ∈Th Ωi τ
Let
N
N
ah uh , vh ahi uh , vh , bh vh , ph bhi vh , ph . 2.15
i 1 i 1
Then the discrete approximation of problem 2.3 is to find uh , ph ∈ Xh Ω × Qh Ω such
that
ah uh , vh bh vh , ph f, vh , ∀vh ∈ Xh Ω,
2.16
bh uh , qh 0, ∀qh ∈ Qh Ω.
In the next section, we prove that the discrete problem 2.16 has a unique solution and we
obtain error estimate.
N
v2h : v2h,i , v2h,i : ahi v, v. 3.1
i 1
6 Mathematical Problems in Engineering
We can find in 1 that the local space family {Xh0 Ωi , Qh0 Ωi } is div-stable; that is,
there exists a constant β independent of hi such that
bh v h , qh
sup qh L2 Ω ,
≥ β ∀qh ∈ Qh0 Ωi , 3.2
v
h ∈X 0 Ωi
vh h,i
i
N
N
Q̆h Ω q̆ q̆i ∈ R , q̌, 1
N
q̌i |Ωi | 0 . 3.3
i 1 i 1
We first prove that the family {Xh Ω, Q̌h Ω} is div-stable.
bh vh , q̌
sup ≥ β̌ q̌ L2 Ω
∀q̌ ∈ Q̌h Ω, 3.4
vh ∈Xh Ω vh h
1
πi vmi v ds, 3.5
|ei | ei
2
πv π1 v1 , π2 v2 , . . . , πN vN , vi v|Ωi , ∀v ∈ H01 Ω . 3.6
h Ω → X
Define the operator Ξh,δmj : X h Ω by
⎧
⎨Qh,δ v|γmi − v|δmj mi , mi ∈ δm
CR
,
Ξh,δmj v mi
mj
j 3.7
⎩0, otherwise.
Mathematical Problems in Engineering 7
We can deduce that for any v ∈ H01 Ω2 , there exists a v∗h ∈ Xh Ω satisfying
b v − v∗h , q̌ 0. 3.8
M
In fact, we can set v∗h πv m 1 Ξh,δmj πv. Obviously v∗h ∈ Xh Ω and
N
N
b v − v∗h , q̌ − div v − v∗h q̌ dx − v − v∗h · nq̌ ds
i 1 τ∈Th Ωi τ i 1 τ∈Th Ωi ∂τ
M
− v − πv · nq̌ds Ξh,δmj πv · nq̌ ds
τ∈Th ∂τ τ∈Th ∂τ m 1
M
Qh,δmj πv|γmi − πv|δmj · n q̌j ds
j 1 δmj 3.9
M
πv|γmi − πv|δmj · nq̌j ds
j 1 δmj
M
v|γmi − v|δmj · nq̌j ds
j 1 δmj
0.
2
πv2h |πv|2H 1 τ ≤ C πvmi − πv mj
τ τ
2
1 1
C vds − vds
|ei | ei
e j
ej
τ
2 3.11
1 1
C v − vds − v − vds
|ei | ei ej e
τ j
⎛ 2 2 ⎞
1 1
≤C ⎝ 2 v − vds 2 v − vds ⎠,
τ |e i | ei ej ej
8 Mathematical Problems in Engineering
where mi , mj are the midpoints of the edges of τ, and v is the integral average of v in τ, by
Hölder inequality, trace theorem, and Friedrichs’ inequality we can get
2
1 1
v − vds ≤ v − v2 ds ≤ Ch−1 v − v2 ds
|ei |2 ei |ei | ei ∂τ
! "
3.12
≤ C h−2 v − v2 dx |v − v|2H 1 τ
τ
≤ C|v|2H 1 τ ,
2 2
Ξh,δmj πv ≤C Ξh,δmj πvmi
h
mi ∈δCR
mj
2
C Qh,δmj πv|γmi − πv|δmj mi
mi ∈δCR
mj
2
≤ Ch−1 Qh,δmj πv|γmi − πv|δmj mi 3.14
0,γm
2
≤ Ch−1 πv|γmi − πv|δmj
0,γm
! "
2 2
≤ Ch−1 πv|γmi − v|δmj v|δmj − πv|δmj
0,γm 0,γm
: Ch−1 K1 K2 .
K2 ≤ Chv2h,j . 3.15
2 2
πv|γmi − v|δmj πv|γmi − v|γmi ≤ Chv2h,i . 3.16
0,γm 0,γm
2
Ξh,δmj πv ≤ C v2h,i v2h,j , 3.17
h
Mathematical Problems in Engineering 9
Since {H01 Ω2 , L20 Ω} is div-stable, following 3.8 and 3.18, by Fortin rules, we have
completed the proof of Lemma 3.1
Now we recall the following Brezzi theory about the existence, uniqueness, and error
estimate for the discrete solution.
Theorem 3.2. The bilinear forms ah ·, · and bh ·, · have the following properties:
i ah ·, · is continuous and uniformly elliptic on the mortar-type P1 nonconforming space
Xh Ω, that is,
ii bh ·, · is also continuous on the space family Xh Ω × Qh Ω, that is,
bh vh , q ≤ vh h q L2 Ω
, ∀vh ∈ Xh Ω, q ∈ Qh Ω; 3.20
iii the family {Xh Ω, Qh Ω} satisfies the inf-sup condition, that is, there exists a constant β
that does not depend on h of triangulation such that
bh v, q
sup ≥β q L2 Ω
, ∀q ∈ Qh Ω, 3.21
v∈Xh Ω vh
so the problem 2.16 has a unique solution, and if one lets u, p, uh , ph be the solution of 2.3 and
2.16, respectively, where u, p ∈ H01 Ω2 × L20 Ω, u|Ωk ∈ H 2 Ωk 2 , p|Ωk ∈ H 1 Ωk , then
N
u − uh h p − ph L2 Ω
≤C hk uH 2 Ωk 2 p H 1 Ω k
. 3.22
k 1
Proof. The statements of Brezzi theory are that the properties 3.19–3.21 lead to the
existence, uniqueness, and error estimate of the discrete solution. In 16, it is proven
that ah ·, · is continuous on Xh Ω and is elliptic with a constant uniformly bounded.
Furthermore, it is straightforward that bh ·, · is continuous on Xh Ω × Qh Ω. The point
that needs verification is a uniform inf-sup condition 3.21, or equivalently that the family
{Xh Ω × Qh Ω} is div-stable.
Using local inf-sup condition 3.2 and the above lemma, arguing as the proof in
Proposition 5.1 of 12, we have the global inf-sup condition 3.21.
10 Mathematical Problems in Engineering
4. Numerical Algorithm
In this section, we present a numerical algorithm, that is, the W-cycle multigrid method for
the discrete system 2.16, and we prove the optional convergence of the multigrid method.
We use a simpler and more convenient analysis method than that in 5.
In order to set the multigrid algorithm, we need only to change the index h of the
partition Th in Section 2 to be k, and let T1 be the coarsest partition. By connecting the
opposite midpoints of the edges of the triangle, we split each triangle of T1 into four triangles
and we refine the partition T1 into T2 . The partition T2 is quasi-uniform of size h2 h1 /2.
Repeating this process, we get a sequence of the partition Tk k 1, 2, . . . , L, each quasi-
uniform of size hk h1 /2k−1 .
As in Section 2, with the partition Tk , we define the mortar P1 nonconforming element
velocity space and P0 element pressure space as Xk and Qk , respectively. We can see that
Xk k 1, 2, . . . , L are nonnested, and Qk k 1, 2, . . . , L are nested. Furthermore, we
denote the P1 nonconforming element product space on Ω by X k.
Let {ϕk } be the basis of Xk , and let {ψk } be the basis of Qk . For any vk ∈ Xk , qk ∈ Qk ,
i i
! "
Ak BkT uk f
k . 4.1
Bk 0 p 0
k
In the following of this section, we introduce our multigrid method; the key of this
method is the intergrid transfer operator.
k−1 → X
We first define the intergrid transfer operator on the product space, Lkk−1 : X k
⎧
⎪
⎪ vmi , mi ∈ κ, κ ∈ Tk−1 ,
⎪
⎪
⎪
⎨1
Lkk−1 vmi v|κ1 mi v|κ2 mi , mi ∈ ∂κ1 ∂κ2 , κ1 , κ2 ∈ Tk , 4.2
⎪2
⎪
⎪
⎪
⎪
⎩0, mi ∈ ∂Ω,
M
Rkk−1 v Lkk−1 v Ξk,δmj Lkk−1 v, 4.3
m 1
k
Jk−1 I. 4.4
Mathematical Problems in Engineering 11
Therefore, our prolongation operator on velocity space and pressure space can be
written as
$ %
k
Ik−1 Rkk−1 , Jk−1
k
. 4.5
Multigrid Algorithm
If k 1, compute the u1 , p1 directly. If k ≥ 2, do the following three steps.
where αk is a real number which is not smaller than the maximal eigenvalue of Ak .
um
k
1 1
um
k
1
Rkk−1 u∗k−1 , pkm1 1
pkm1 ∗
pk−1 . 4.8
then, um
k
1 m2 1
, pkm1 m2 1 is the result of one iteration step.
For convenience, at level k the problem 2.16 can be written as followes: find uk , pk ∈
Xk × Qk such that
Lh,k uk , pk ; vk , qk Fk vk , qk , ∀ vk , qk ∈ Xk × Qk . 4.10
12 Mathematical Problems in Engineering
Since Lh,k uk , pk ; vk , qk is a symmetric bilinear form on Xk ×Qk , there is a complete
j j
set of eigenfunctions φk , ψk , which satisfy
$ j
j %
Lh,k uk , pk ; vk , qk λj φk , vk h2 ψk , qk , ∀ vk , qk ∈ Xk × Qk ,
0 0
j j 4.11
vk , qk cj φk , ψk .
j
In order to verify that our multigrid algorithm is optimal, we need to define a set of
mesh-dependent norms. For each k ≥ 0 we equip Xk × Qk with the norm
1/2 1/2
v, q v, q v2L2 Ω h2k q
2
v, vk h2k q, q k ,
0,k 0,k L2 Ω
4.12
and define
⎧ ⎫1/2
⎨ ⎬
vk , qk λ s cj 2 , v2k ∇v, ∇vk . 4.13
s,k ⎩ j j ⎭ τ
For our multigrid algorithm, we have the following optional convergence conclusion.
Theorem 4.1. If u, p and uih , phi 0 ≤ i ≤ m 1 are the solutions of problems 2.16 and 4.10,
respectively, then there exists a constant 0 < γ < 1 and positive integer m, all are independent of the
level number k, such that
1 m 1 ≤ γ u, p − u0k , pk0
u, p − um
k , pk 0,k
. 4.14
0,k
To prove this theorem, we give in the next section two basic properties for convergence
analysis of the multigrid, that is, the smoothing property and approximation property.
Lemma 5.1 smoothing property. Assume that λmax Ak ≤ αk ≤ Cλmax Ak , if the number of
smoothing steps is m, then
m Ch−2 0
u − uh , pm − ph ≤ uh − uh . 5.1
h h 2,k m L2 Ω
Next, we prove the approximation property. We just apply the following conclusion in
14, which can simplify the complexity of theoretical analysis.
A.2 Jk−1
k
qL2 Ω ≤ CqL2 Ω , ∀q ∈ Qk−1 ,
where u, p ∈ H01 Ω ∩ H 2 Ω2 × L20 Ω ∩ H 1 Ω is the solution of 2.3 with the force term
f ∈ L2 Ω2 and uk−1 , pk−1 , uk , pk are the mixed finite element approximation of u, p at levels
k − 1 and k, respectively.
∗
Lk−1 k
Ik−1 vk , qk , vk−1 , qk−1 ,
5.2
k
Lk vk , qk , Ik−1 vk−1 , qk−1 , ∀ vk−1 , qk−1 ∈ Xk−1 × Qk−1 , vk , qk ∈ Xk × Qk .
∗
v, q − Ik−1
k k
Ik−1 v, q ≤ Ch2k v, q 2,k
, ∀ v, q ∈ Xk × Qk . 5.3
0,k
k
Proof. By Lemma 5.2, we only need to prove our prolongation operator Ik−1 that satisfies
A.1, A.2, and A.3.
For any v ∈ Xk−1 , the inequality A.1 holds. In fact
M
v − Rkk−1 v ≤ v − Lkk−1 v Ξk,δmj Lkk−1 v , 5.4
L2 Ω L2 Ω
m 1 L2 Ω
2 2
Ξk,δmj Lkk−1 v ≤ h2k Ξk,δmj Lkk−1 v mki
L2 Ω
mki ∈δCR
k,mj
2
h2k Qk,δmj Lkk−1 v |γmi − Lkk−1 v |δmj mki
mki ∈δCR
k,mj
2
≤ Chk Qk,δmj Lkk−1 v |γmi − Lkk−1 v |δmj 5.6
0,γm
2
≤ Chk Lkk−1 v |γmi − Lkk−1 v |δmj
0,γm
! 2 2.0
"
≤ Chk Lkk−1 v |γmi − v|δmj v |δmj − Lkk−1 v |δmj
0,γm 0,γm
Chk K1 K2 .
2 2
Lkk−1 v |γmi − v |δmj ≤2 Lkk−1 v |γmi − Qk−1,δmj v |γmi
0,δmj 0,γmi
5.8
2
2 Qk−1,δmj v |δmj − v |δmj .
0,δmj
2
Qk−1,δmj v |δmj − v |δmj v − Qe v2 ds, 5.9
0,γm
e∈Tk−1 δmj e
2
v − Qe v ds ≤ 2
v − c ds ≤ Chk v − c2 d+
+ v − c21,E+
s ≤ Chk +
e e e+ 5.10
≤ v|21,E+
Chk |+ ≤ Chk |v|21,E ,
Mathematical Problems in Engineering 15
Qk−1,δmj v |δmj − v |δmj ≤ Ch1/2
k
vk,j . 5.11
0,γm
2
Lkk−1 v |γmi − Qk−1,δmj v |γmi
0,γm
2
Lkk−1 v |γmi − v |γmi v |γmi Qk−1,δmj v |γmi
0,γm
5.12
2 2
≤2 Lkk−1 v − v |γmi 2 v |γmi − Qk−1,δmj v |γmi
0,γmi 0,γmi
F1 F2 .
2
F1 ≤ Chk Lkk−1 v ≤ Chk v2k−1,i . 5.13
k,i
For F2 , by trace theorem and the approximation of the operator Qk−1,δmj , we have
uk , pk − Ik−1
k
uk−1 , pk−1
0,k
M
≤ uk − Lkk−1 uk−1 Ξk,δmj Lkk−1 uk−1 h2k pk − Jk−1
k
pk−1
0,k
m 1
0,k 5.15
0,k
M
≤ Ch2k uH 2 Ω p H 1 Ω
Ξk,δmj Lkk−1 uk−1 .
m 1 0,k
16 Mathematical Problems in Engineering
2
Ξk,δmj Lkk−1 uk−1
0,k
2
≤ h2k Ξk,δmj Lkk−1 uk−1 mki
mki ∈δCR
k,mj
2
h2k Qk,δmj Lkk−1 uk−1 |γmi − Lkk−1 uk−1 |δmj mki
mki ∈δCR
k,mj
5.16
2
≤ Chk Qk,δmj Lkk−1 uk−1 |γmi − uk |γmi uk |δmj − Lkk−1 uk−1 |δmj
0,k
! 2 2
"
≤ Chk Lkk−1 uk−1 |γmi − uk |γmi Lkk−1 uk−1 |δmj − uk |δmj
0,γm 0,γm
Chk K1 K2 .
together with 5.15, A.3 has been proved, and we have completed the proof of Lemma
5.3.
6. Numerical Results
In this section, we present some numerical results to illustrate the theory developed in the
earlier sections. The examples are as same as those in 5, so that we can compare the
conclusion with the mortar rotated Q1 element method.
Here we deal with Ω 0, 12 . We choose the exact solution of 2.1 as
2
u1 2x2 1 − x2 y 1 − y 1 − 2y , u2 −2x1 − x1 − 2xy2 1 − y , 6.1
Figure 1: The coarsest mesh with h1,1 1/4 and h1,2 1/6.
Table 1: Error estimate for the mortar element method with h1,1 1/4 and h1,2 1/6.
Table 2: Iterative numbers for the W-cycle with h1,1 1/4 and h1,2 1/6.
k 2 3 4 5
iter4,4 9 8 8 9
iter5,5 8 8 7 7
From Table 2, we can see that the convergence for the W-cycle multigrid algorithm is
optimal; that is, the number of iterations is independent of the level number k. Meanwhile,
we note that the number of iterations is less than the rotated Q1 element method in 5 when
achieving the same relative error.
Acknowledgment
The authors would like to express their sincere thanks to the referees for many detailed and
constructive comments. The work is supported by the National Natural Science Foundation
of China under Grant 11071124.
References
1 V. Girault and P.-A. Raviart, Finite Element Methods for Navier-Stokes Equations, vol. 5 of Springer Series
in Computational Mathematics, Springer, Berlin, Germany, 1986.
18 Mathematical Problems in Engineering
2 S. Zhang, “On the P1 Powell-Sabin divergence-free finite element for the Stokes equations,” Journal of
Computational Mathematics, vol. 26, no. 3, pp. 456–470, 2008.
3 T. Neckel, M. Mehl, and C. Zenger, “Enhanced divergence-free elements for efficient incompressible
flow simulations in the PDE framework peano,” in Proceedings of the 5th European Conference on
Computational Fluid Dynamics (ECCOMAS CFD ’10), pp. 14–17, Lisbon, Portugal, June 2010.
4 D. Braess and R. Sarazin, “An efficient smoother for the Stokes problem,” Applied Numerical
Mathematics, vol. 23, no. 1, pp. 3–20, 1997.
5 J. Chen and P. Huang, “Analysis of mortar-type Q1rot /Q0 element and multigrid methods for the
incompressible Stokes problem,” Applied Numerical Mathematics, vol. 57, no. 5-7, pp. 562–576, 2007.
6 V. John and L. Tobiska, “A coupled multigrid method for nonconforming finite element discretiza-
tions of the 2D-Stokes equation,” Computing, vol. 64, no. 4, pp. 307–321, 2000.
7 J. Li, “A dual-primal FETI method for incompressible Stokes equations,” Numerische Mathematik, vol.
102, no. 2, pp. 257–275, 2005.
8 H. H. Kim, C.-O. Lee, and E.-H. Park, “A FETI-DP formulation for the Stokes problem without primal
pressure components,” SIAM Journal on Numerical Analysis, vol. 47, no. 6, pp. 4142–4162, 2010.
9 J. Li and O. Widlund, “BDDC algorithms for incompressible Stokes equations,” SIAM Journal on
Numerical Analysis, vol. 44, no. 6, pp. 2432–2455, 2006.
10 C. Bernardi, Y. Maday, and A. T. Patera, “A new nonconforming approach to domain decomposition:
the mortar element method,” in Nonlinear Partial Differential Equations and Their Applications, H. Brezis
and J. Lions, Eds., vol. 11 of Pitman Research Notes in Mathematics 299, pp. 13–51, Longman Scientific
Technical, Harlow, UK, 1994.
11 D. Braess, W. Dahmen, and C. Wieners, “A multigrid algorithm for the mortar finite element method,”
SIAM Journal on Numerical Analysis, vol. 37, no. 1, pp. 48–69, 1999.
12 F. Ben Belgacem, “The mixed mortar finite element method for the incompressible Stokes problem:
convergence analysis,” SIAM Journal on Numerical Analysis, vol. 37, no. 4, pp. 1085–1100, 2000.
13 F. Wang, J. Chen, W. Xu, and Z. Li, “An additive Schwarz preconditioner for the mortar-type rotated
Q1 FEM for elliptic problems with discontinuous coefficients,” Applied Numerical Mathematics, vol. 59,
no. 7, pp. 1657–1667, 2009.
14 Q.-p. Deng and X.-p. Feng, “Multigrid methods for the generalized Stokes equations based on mixed
finite element methods,” Journal of Computational Mathematics, vol. 20, no. 2, pp. 129–152, 2002.
15 F. Brezzi, “On the existence, uniqueness and approximation of saddle-point problems arising from
Lagrangian multipliers,” RAIRO Analyse Numérique, vol. 8, pp. 129–151, 1974.
16 L. Marcinkowski, “The mortar element method with locally nonconforming elements,” BIT, vol. 39,
no. 4, pp. 716–739, 1999.
17 P. G. Ciarlet, The Finite Element Method for Elliptic Problems, vol. 4, North-Holland, Amsterdam, The
Netherlands, 1978.
Hindawi Publishing Corporation
Mathematical Problems in Engineering
Volume 2012, Article ID 959040, 12 pages
doi:10.1155/2012/959040
Research Article
Applying Neural Networks to Prices Prediction of
Crude Oil Futures
Copyright q 2012 John Wei-Shan Hu et al. This is an open access article distributed under the
Creative Commons Attribution License, which permits unrestricted use, distribution, and
reproduction in any medium, provided the original work is properly cited.
The global economy experienced turbulent uneasiness for the past five years owing to large
increases in oil prices and terrorist’s attacks. While accurate prediction of oil price is important but
extremely difficult, this study attempts to accurately forecast prices of crude oil futures by adopting
three popular neural networks methods including the multilayer perceptron, the Elman recurrent
neural network ERNN, and recurrent fuzzy neural network RFNN. Experimental results indi-
cate that the use of neural networks to forecast the crude oil futures prices is appropriate and
consistent learning is achieved by employing different training times. Our results further demon-
strate that, in most situations, learning performance can be improved by increasing the training
time. Moreover, the RFNN has the best predictive power and the MLP has the worst one among
the three underlying neural networks. This finding shows that, under ERNNs and RFNNs, the pre-
dictive power improves when increasing the training time. The exceptional case involved BPNs,
suggesting that the predictive power improves when reducing the training time. To sum up, we
conclude that the RFNN outperformed the other two neural networks in forecasting crude oil
futures prices.
1. Introduction
During the past three years, the global economy has experienced dramatic turbulence owing
to uneasinessbecause of terrorists’ attacks and rapidly rising oil prices. For example, the
US light crude oil futures price rapidly climbed to the all-time peak about US$80 recently.
Simultaneously, the US Federal Reserve continuously increased its benchmark short-term
interest rates by seventeen times to prevent inflation till August 2006. Consequently, many
governments and corporate managers attempted to seek a method of accurately forecasting
the crude oil prices.
2 Mathematical Problems in Engineering
Accurate prediction of crude oil price is important yet extremely complicated and
difficult. For example, Kumar 1 found that the traditional model-based forecasts had larger
errors compared to forecast crude oil prices using futures price. However, Pham and Liu 2
and Refenes et al. 3 showed that neural networks had significant performance of forecast-
ing. According to Chen and Pham 4, numerous real-world application problems cannot
be fully described and handled via classical set theory. Meanwhile, fuzzy set theory can
deal with partial membership. Although Omlin et al. 5 argued that fuzzy neural networks
FNNs combine the advantages of both fuzzy systems and neural networks, whereas Omlin
et al. 5 proposed that most of the FNNs could only process state input-output relationships,
FNNs were unable to process temporal input sequences with arbitrary length. Since recurrent
neural networks RNNs are dynamic systems involving temporal state representation,
RNNs are computationally powerful. Although the Elman recurrent neural network ERNN
is a special case of RNN and is less efficient than standard RNN, Pham and Liu 2 posited
that ERNN can model a very large class of linear and nonlinear dynamic systems. Addi-
tionally, Lee and Teng 6 argued that the RFNNs have the same advantages as RNNs and
extended the application domain of the FNNs to temporal problems. These findings motivate
us to apply three neural network models namely, traditional backpropagation neural net-
works BPNs, Elman recurrent neural networks ERNNs, and recurrent fuzzy neural
networks RFNNs to forecast crude oil futures prices.
The focus of this paper is to apply neural networks for predicting crude oil futures
prices. This work has the following objectives: forecast crude oil futures prices using BPNs,
ERNNs, and RFNNs; compare the learning and predictive performance among BPNs,
ERNNs, and RFNNs, and explore how training time affects prediction accuracy.
This study classifies the previous literature into three main groups: 1 the studies
that compared artificial neural networks ANNs with other methods to forecast futures
prices, 2 the works that combined fuzzy systems with recurrent neural networks, and
3 the researches that examined the evolution or forecasting accuracy of energy futures
prices.
The following studies have applied various ANNs to predict futures prices. Refenes
et al. 3, Castillo and Melin 7, Giles et al. 8, Donaldson and Kamstra 9, and Sharma
et al. 10 all demonstrated that neural networks outperformed classical statistical techniques
in forecasting ability. Although Kamstra and Boyd 11 also found that ANNs outperformed
the naive model for most commodities in forecasting ability, yet Kamstra and Boyd found
that ANNs have less predictive power than linear model for barley and rye.
The following works combined fuzzy system with recurrent neural networks. Omlin
et al. 5, Juang and Lin 12, Nürnberger et al. 13, 14, Zhang and Morris 15, Giles et al.
8, Lee and Teng 6, Mastorocostas and Theocharis 16, 17, Juang 18, Yu and Ferreyra
19, Hong and Lee 20, and Lin and Chen 21 all designed to combine recurrent neural
networks RNNs with fuzzy system for identification and prediction.
The following researches examined the evolution or forecasting accuracy of energy
prices. Hirshfeld 22, Ma 23, Serletis 24, Kumar 1, Pindyck 25, Adrangi et al. 26, and
Krehbiel and Adkins 27 reviewed or examined the energy futures prices and the price risk.
Among these literatures, Adrangi et al. 26 found strong evidence of nonlinear dependencies
in crude oil futures prices, but the evidence is not consistent with chaos.
The paper is organized as follows. Three kinds of artificial neural networks are des-
cribed in Section 2. The performance of the proposed learning algorithms is examined by the
computer simulations are described in Section 4. Conclusion in presented in Section 5. The
performance valuation method is presented in Section 3.
Mathematical Problems in Engineering 3
y1 · · · yk Output
layer
Hidden
··· layer
··· Input
layer
x1 · · · x2 · · · xn
1
y , 2.1
1 e−α xi
where α ∈ R. An activation function can be differentiated since the steepest descent method is
employed to derive the weight updating rule. The response of the hidden layer is the input of
the output layer. In the second step, an overall error, ET , which is the difference between the
actual and the desired output, is minimized employing a supervised learning task performed
by MLP:
1 2
ET Ep dpk − ypk , 2.2
p 2 p k
4 Mathematical Problems in Engineering
where ET denotes the total error for a neural network across the entire training set, Ep
represents the network error for the pth pattern, dpk denotes the desired output of the kth unit
in the output layer for pattern p, and ypk is the actual output of the kth unit in the output layer
for the pth pattern.
Then the gradient method is applied to optimize the weight vector of ET to minimize
the summed square error between the actual and the desired network outputs throughout
the training period. The network weight is adjusted whenever a training data is inputted. The
size of the adjustment is positively related to the sensitivity of the error function to weight
connections. The general weight updating rule for the connection weight between the ith
input node and the jth output node is as follows:
∂ET
Δwji −η , 2.3
∂ωji
Output layer
Hidden layer
using discrete time steps. The activations of the hidden units at time t are fed backwards and
serve as inputs to “context layer” at time t 1 and thus represent a form of short-term memory
that enables limited recurrence. Moreover, the feedback links run from the hidden layer to the
context layer and produce both temporal and spatial patterns. As depicted in Figure 2, this
network is a two-layer network involving feedback in the first layer 32.
⎛ ⎞
N
xi κ f ⎝ wijx xj κ − 1 wiu uκ − 1⎠, 2.5
j1
n
y
yκ wi xi κ, 2.6
i1
x
where wi,j demonstrates weight linking the ith hidden-layer neuron and the jth context-
μ
layer neuron, wi indicates weight linking the input neuron uκ − 1 and the ith hidden-layer
y
neuron, wi refers to weight linking the output neuron yκ and the ith hidden-layer neuron,
f· expresses nonlinear activation function in the hidden-layer node, and n is the number of
hidden-layer nodes. Since it is difficult to interpret the network functions of RNNs, this study
further incorporates the fuzzy logic into RNNs.
6 Mathematical Problems in Engineering
y1 · · · y2 · · · yk
x1 · · · x 2 · · · x n
1 Input layer: the input nodes in this layer represent input variables. The input layer
only transmits the input value to the next layer directly and no computation is
Mathematical Problems in Engineering 7
conducted in this layer. From 2.5, the connection weight at the input layer wi1 is
unity:
1 1
gi xi . 2.7
2 Membership layer: the membership layer is also known as a fuzzification layer
and contains several different types of neurons, each neuron performs membership
function. The membership nodes in this layer correspond to the linguistic label of
the input variables in the input layer and serve as a unit of memory. Each of these
variables is transformed into several fuzzy sets in the membership layer where each
neuron corresponds to a particular fuzzy set, with the actual membership function
being provided by the neuron output. Each neuron in this layer represents charac-
teristics of each membership function, and Gaussian function serves as the mem-
bership function. The jth neuron in this layer has the following input and output:
⎡ 2 ⎤
2
xij − mij
2 ⎢ ⎥ 2.8
git exp⎣− ⎦,
σij2
where
mij denotes the mean value of a Gaussian membership function of the jth term
with respect to the ith input variable, σij represents the standard derivation of the
Gaussian type membership function of the jth term with respect to the ith input,
2 2
xij t is the input of this layer at the discrete time t, gij t − 1 denotes the feedback
unit of memory which stores the past network Information and represents the main
difference between FNN and RFNN, and θij indicates the connection weight of the
feedback unit.
Each node in the membership layer possesses three adjustable parameters: mij , σij ,
and θij .
3 Fuzzy rule layer: The fuzzy rule layer comprises numerous nodes, each node
corresponds to a fuzzy operating region of the process being modeled. This layer
constructs the entire fuzzy rule data set. The nodes in this layer equal the number
of fuzzy sets corresponding to each external linguistic input variable and receive
the one-dimensional membership degree of the associated rule from the nodes of
a set in the membership layer. The output of each neuron in the fuzzy rule layer
is obtained by using a multiplication operation. The input and output for the kth
neuron in the fuzzy rule layer are as follows:
T
3 3 2 2
gi Πi xi exp −Di xi − mi Di xi − mi , 2.10
8 Mathematical Problems in Engineering
4 Output layer: the output layer performs the defuzzification operation. Nodes in this
layer are called output linguistic nodes, where each node is for an individual output
of the system. The links between the fuzzy rule layer and the output layer are con-
nected by the weighting values wjk .
4
m
3
yk gk wjk gj , 2.11
i
where wjk is the output action strength of the kth output associated with the jth
4
fuzzy rule and serves as the tuning factor of this layer, gk is the final inferred result,
and yk represents the kth output of the FRNN.
3. Valuation Performance
This work uses the mean square error MSE method to assess the performance of three neu-
ral networks. The MSE is calculated as the average of the sum of the square of the error, which
is given by the difference between the actual and the designed output. MSE thus is computed
as
1 T
MSE rt − rt 2 , 3.1
T − T1 − 1 tT1
where T indicates the total number of samples, T1 refers to the number of estimated samples,
rt represents the actual output, and rt denotes the desired output.
4. Computer Simulations
4.1. Data Description
This study focuses on energy futures for the near-month. Daily oil prices for Brent, WTI,
DUBAI, and IPE are used in this investigation. The data sources are obtained from the Energy
Bureau of USA and International Petroleum Exchange IPT of the Great Britain. This work
explores the influence of training times on prediction performance so that it classifies the
training period from January 1, 1990 to April 30, 2005 into three five-year sections. Table 1
shows the different training periods.
Mathematical Problems in Engineering 9
Data set Backpropagation Elman RNN RNN with fuzzy Times for each ANN
I BPN1 ERNN1 RFNN1 1244
II BPN2 ERNN2 RFNN2 2499
III BPN3 ERNN3 RFNN3 3758
5. Conclusion
This study uses multi-layer perception MLP, Elman recurrent neural networks ERNNs
and recurrent fuzzy neural networks RFNNs to forecast the crude oil prices and compare
10 Mathematical Problems in Engineering
Table 3: The comparison of the learning ability MSE for the three neural networks.
Data set
ANNs MSE
I II III
BPN 6.84589E − 05 4.13072E − 05 3.75245E − 05 4.90969E − 05
ERNN 6.53719E − 05 3.71397E − 05 3.42490E − 05 4.355769E − 05
RFNN 2.43727E − 05 1.66990E − 05 1.78329E − 05 1.96349E − 05
Table 4: The comparison of the predictive power of the three neural networks.
the predictive power of the above three neural network models. Results of this work are
summarized as follows.
All of the MSE values obtained under different training times through MLP, ERNNs,
and RFNNs are below 0.0026768, suggesting that the use of the neural networks to forecast
the crude oil futures prices is appropriate, and consistent learning ability can be obtained by
using different training times. This investigation confirms that, under most circumstances,
the more training times the neural networks take, the more the learning performance of the
neural networks improves. The only exceptional case occurs at part 2 under the RFNN model,
where MSE is slightly less than that obtained from part 3.
Regarding the predictive power of the three neural networks, this study finds that
RFNN has the best predictive power and MLP has the least predictive power among the three
neural networks. This work also finds that, under ERNNs and RFNNs, the predictive power
improves when increasing the training time. However, the results are different from those
obtained under MLP, indicating that the predictive power improves when decreasing the
training time. Possible explanation for this phenomenon is the existence of a large difference
between the predictive value and the actual value during a 9-day period. To summarize, this
study concludes that the recurrent fuzzy neural network is the best among the three neural
networks.
Acknowledgments
The authors would like to thank Dr. Oleg Smirnov for his insightful comments at the 81st
WEA Annual Conference on July 1, 2006. Also, the authors would like to thank the anon-
ymous referees for their valuable comments. This research is partially supported by the
National Science Council of Taiwan under grant NSC 99-2410-H-033-026-MY3.
References
1 M. S. Kumar, “The forecasting accuracy of crude oil futures prices,” IMF Staff Papers, vol. 39, no. 2,
pp. 432–461, 1992.
2 D. T. Pham and X. Liu, Neural Networks for Identification Prediction and Control, Springer, London, UK,
1995.
Mathematical Problems in Engineering 11
3 A. Nicholas Refenes, A. Zapranis, and G. Francis, “Stock performance modeling using neural net-
works: a comparative study with regression models,” Neural Networks, vol. 7, no. 2, pp. 375–388, 1994.
4 Chen and Pham, “Time series prediction combining dynamic system, theory and statistical methods,”
in Proceedings of the IEEE/IAFE 1995 Computational Intelligence for Financial Engineering, pp. 151–155,
2001.
5 C. W. Omlin, K. K. Thornber, and C. L. Giles, “Fuzzy finite-state automata can be deterministically
encoded into recurrent neural networks,” IEEE Transactions on Fuzzy Systems, vol. 6, no. 1, pp. 76–89,
1998.
6 C.-H. Lee and C.-C. Teng, “Identification and control of dynamic systems using recurrent fuzzy neural
networks,” IEEE Transactions on Fuzzy Systems, vol. 8, no. 4, pp. 349–366, 2000.
7 O. Castillo and P. Melin, “Intelligent system for financial time series prediction combining dynamical
systems theory, fractal theory, and statistical methods,” in Proceedings of the IEEE/IAFE Computational
Intelligence for Financial Engineering (CIFEr ’95), pp. 151–155, April 1995.
8 C. L. Giles, D. Chen, G. Z. Sun, H. H. Chen, Y. C. Lee, and M. W. Goudreau, “Constructive learning of
recurrent neural networks: limitations of recurrent cascade correlation and a simple solution,” IEEE
Transactions on Neural Networks, vol. 6, no. 4, pp. 829–836, 1995.
9 R. G. Donaldson and M. Kamstra, “Forecast combining with neural networks,” Journal of Forecasting,
vol. 15, no. 1, pp. 49–61, 1996.
10 J. Sharma, R. Kamath, and S. Tuluca, “Determinants of corporate borrowing: an application of a neural
network approach,” American Business Review, vol. 21, no. 2, pp. 63–73, 2003.
11 I. Kamstra and M. S. Boyd, “Forecasting futures trading volume using neural networks,” The Journal
of Futures Markets, vol. 15, no. 8, pp. 935–970, 1995.
12 C. F. Juang and C. T. Lin, “A recurrent self-organizing neural fuzzy inference network,” IEEE
Transactions on Neural Networks, vol. 10, no. 4, pp. 828–845, 1999.
13 A. Nurnberger, A. Radetzky, and R. Kruse, “Determination of elastodynamic model parameters using
a recurrent neuro-fuzzy system,” in Proceedings of the 7th European Congress on Intelligent Techniques and
Soft Computing, Verlag Mainz, Aachen, Germany, 1999.
14 A. Nürnberger, A. Radetzky, and R. Kruse, “Using recurrent neuro-fuzzy techniques for the identifi-
cation and simulation of dynamic systems,” Neurocomputing, vol. 36, pp. 123–147, 2001.
15 J. Zhang and A. J Morris, “Recurrent neuro-fuzzy networks for nonlinear process modeling,” IEEE
Transactions on Neural Networks, vol. 10, no. 2, pp. 313–326, 1999.
16 P. Mastorocostas and J. Theocharis, “FUNCOM: a constrained learning algorithm for fuzzy neural
networks,” Fuzzy Sets and Systems, vol. 112, no. 1, pp. 1–26, 2000.
17 P. A. Mastorocostas and J. B. Theocharis, “A recurrent fuzzy-neural model for dynamic system
identification,” IEEE Transactions on Systems, Man, and Cybernetics B, vol. 32, no. 2, pp. 176–190, 2002.
18 C. F. Juang, “A TSK-type recurrent fuzzy network for dynamic systems processing by neural network
and genetic algorithms,” IEEE Transactions on Fuzzy Systems, vol. 10, no. 2, pp. 155–170, 2002.
19 W. Yu and A. Ferreyra, “System identification with state-space recurrent fuzzy neural networks,” in
Proceedings of the 43rd IEEE Conference on Decision and Control, Atlantis, Paradise Island, Bahamas,
December 2004.
20 Y.-Y. Hong and C.-F. Lee, “A neuro-fuzzy price forecasting approach in deregulated electricity
markets,” Electric Power Systems Research, vol. 73, no. 2, pp. 151–157, 2005.
21 C. J. Lin and C. H. Chen, “Identification and prediction using recurrent compensatory neuro-fuzzy
systems,” Fuzzy Sets and Systems, vol. 150, no. 2, pp. 307–330, 2005.
22 D. J. Hirshfeld, “A fundamental overview of the energy futures market,” The Journal of Futures Market,
vol. 3, no. 1, pp. 75–100, 1983.
23 C. W. Ma, “Forecasting efficiency of energy futures prices,” The Journal of Futures Markets, vol. 9, pp.
373–419, 1989.
24 A. Serletis, “Rational expectations, risk and efficiency in energy futures markets,” Energy Economics,
vol. 13, no. 2, pp. 111–115, 1991.
25 R. S. Pindyck, “Long-run evolution of energy prices,” Energy Journal, vol. 20, no. 2, pp. 1–24, 1999.
26 B. Adrangi, A. Chatrath, K. K. Dhanda, and K. Raffiee, “Chaos in oil prices? Evidence from futures
markets,” Energy Economics, vol. 23, no. 4, pp. 405–425, 2001.
27 T. Krehbiel and L. C. Adkins, “Price risk in the NYMEX energy complex: an extreme value approach,”
Journal of Futures Markets, vol. 25, no. 4, pp. 309–337, 2005.
28 J. J. Hopefield, “Networks and physical systems with emergent collective,” Proceedings of National
Academy Science, vol. 79, pp. 2554–2558, 1982.
12 Mathematical Problems in Engineering
29 S. Haykin, Neural Networks: A Comprehensive Fundation, Prentice Hall, Upper Saddle Rever, NJ, USA,
2nd edition, 1999.
30 J. A. Hertz, A. Krogh, and R. G. Palmer, Introduction to the Theory of Neural Computation, Addison-
Wesley Longman, Boston, Mass, USA, 1991.
31 J. L. Elman, “Finding structure in time,” Cognitive Science, vol. 14, no. 2, pp. 179–211, 1990.
32 C.-T. Lin and C. S. G. Lee, Neural Fuzzy System: A Neuro-Fuzzy Synergism to Intelligent Systems, Prentical
Hall, Upper Saddle River, NJ, USA, 1996.
33 B. Hammer and J. K. Nørskov, “Theoretical surface science and catalysis-calculations and concepts,”
Advances in Catalysis, vol. 45, pp. 71–129, 2000.
34 X. Li, Z. Chen, Z. Yuan, and G. Chen, “Generating chaos by an Elman network,” IEEE Transactions on
Circuits and Systems I, vol. 48, no. 9, pp. 1126–1131, 2001.
35 L. Zadeh, “Fuzzy Sets Yulia,” Information and Control, vol. 8, no. 3, pp. 338–353, 1965.
36 L. Zadeh, “Fuzzy sets and systems,” in Proceedings of the Symposium on Systems Theory, Polytechnic
Institute of Brooklyn, April 1965.
37 H.-J. Zimmermann, Fuzzy Set Theory and Its Applications, Kluwer Academic Publishers, Boston, Mass,
USA, 1985.
38 J. J. Buckley and Y. Hayashi, “Fuzzy neural networks: a survey,” Fuzzy Sets and Systems, vol. 66, no. 1,
pp. 1–13, 1994.
39 J. W.-S. Hu and C.-H. Chang, “Euro/USD currency option valuation combining two artificial neural
network models and various volatilities,” in Proceedings of the 80th Western Economic Association
Annual Conference (WEA ’05), San Francisco, Calif, USA, July 2005.
Hindawi Publishing Corporation
Mathematical Problems in Engineering
Volume 2012, Article ID 273530, 24 pages
doi:10.1155/2012/273530
Research Article
A Novel Method for Technology Forecasting and
Developing R&D Strategy of Building Integrated
Photovoltaic Technology Industry
Copyright q 2012 Y.-J. Chiu and T.-M. Ying. This is an open access article distributed under
the Creative Commons Attribution License, which permits unrestricted use, distribution, and
reproduction in any medium, provided the original work is properly cited.
Because of global warming, renewable energy technologies have become more essential currently,
with solar energy technology advancing worldwide. Therefore, interdisciplinary integration is
an important trend, and building-integrated photovoltaic BIPV is an emerging technology
involving the photovoltaic and building fields. The purpose of this study is to understand the
technology evolution of BIPV and to determine the R&D planning direction. This paper proposes
a hybrid approach to explore the life cycle of BIPV technology and develop the R&D strategy
of related industries. The proposed approach comprises the following submodules. First, patent
analysis is employed to transform patent documents into structured data. Second, the logistic
growth model is used to explore the life cycle of BIPV technology. Third, a patent matrix map
analysis is used to develop the R&D strategy of the BIPV industry. Through the analysis by the
logistic model, the BIPV technology is transformed from the emerging stage to the growth stage
of a long-term life cycle. The other important result is created by the three-dimensional matrix for
R&D strategies in this paper.
1. Introduction
Issues of energy conservation and prevention of global warming have been of great concern
in recent years. For energy shortages, the development of renewable energy technologies is an
inevitable trend and includes solar, wind, hydro- and bio-mass energy. There are advantages
to solar energy, including that it is natural, nonpolluting, inexhaustible, noiseless, and not
subject to geographical constraints. Solar energy contains few moving parts, and therefore,
needs almost no maintenance. The cost of solar energy is fixed at the initial investment stage.
2 Mathematical Problems in Engineering
It is the parity technology in the long-term operation. Solar energy technology can be divided
into the ground-based system, rooftop system, and building-integrated photovoltaic BIPV
system. Since the ground-based system needs considerable land and the rooftop system is
restricted to a limited number of building roofs, this study focuses on the BIPV system, which
is used to construct three-dimensional space to create the largest possible photovoltaic power
generation with less area. In addition, solar energy technology can be embedded in various
ways of designed buildings, which not only has the ability to improve the use of renewable
energy but is also by far the oldest and the most cleanly green energy worldwide. BIPV will
gain increasing attention of the building energy efficiency market in the twenty-first century.
Consequently, the ability to manage intellectual property plays a critical role in the
R&D process or new product development. The report of the World Intellectual Property
Organization WIPO shows that 90%–95% of the world’s inventions can be found in patent
documents, and 80% of these technologies do not appear in other resources, such as journals,
magazines, and encyclopedias 1. Moreover, a WIPO investigation showed that a company
can save up to 60% of time in R&D and 40% of research budget if it can use patent
documents efficiently. Patents contain substantial information including practical and specific
technological content, technical reporting, technological trends of firms, and technological
development trends 2. Patent information is a sufficient source of technical knowledge, and
the patent contains characteristics of the technology and market because it fulfills a clear
standard and originality, technical feasibility, and commercial value 3. It can also be the
basis of technical competitive analysis and technology trends analysis 4, and it can be
regarded as an important indicator of technological changes 5. Therefore, to master the
technology development trend and competitive advantage of enterprises, we must rely on
detailed patent intelligence to fully achieve maximum value from patent analysis 6.
The analysis of patent databases can provide the latest information technology and
product. The growth of the number of patents may represent a national or industrial
development of technology. It is a leading indicator of competitiveness, allowing one to grasp
the message of patent data and analysis of industry trends to trace an evasive technology or
technological innovation development strategy.
Many experts and scholars used patent analysis to investigate different levels as
follows:
1 discussion of national economic growth or innovation activities: Barroso et al. 7
studied 21 Latin American countries, including analysis of patents as a source
of technical information to improve scientific and technological research and
development and innovation in their own countries; Abraham and Moitra 8 used
patent analysis to accurately assess the technical progress and technical innovation;
2 discussion of enterprise technology ability: Liu and Shyu 1 used patent analysis of
Taiwan LED and TFT industry technology development, strategy, and technology
planning and forecasting. Ernst 9 analyzed the German mechanical engineering
industry and four patent strategies of performance associated with the company;
3 discussion of distribution of industries or enterprises: Guo 10 analyzed Taiwan
national patent database TWPAT for the number of applications and development
trend. Chen and Cheng 11 analyzed the US patent database USPTO in patent
classification and UPC solar patent analysis research;
4 an analysis of patent citation data: Hu 12 studied the USPTO citation data
seeking knowledge of the East Asian regional traffic. Tijssen 13 related to patent
Mathematical Problems in Engineering 3
citation analysis of global and domestic industry of interaction between science and
technology and knowledge flows.
Solar cells
Multiple cells
Encapsulation-lamination
Photovoltaic
modules industry
Building structure
Structural support unit, connecting bolt unit,
Construction industry
construction frame unit, turn unit . . . etc.
Electromechanical system
Balance system BOS, DC/AC control panel,
PV monitoring system, electric engineering . . . etc. Electromechanical industry
BIPV
solar energy into electricity. Therefore, BIPV can be called a static generator as well as the new
generation of photovoltaic building 21.
BIPV is a new concept of converting solar energy into electrical energy. In addition
to developing self-efficient zero-energy buildings ZEB with the use of BIPV, heat load
within the building can be reduced via the thermal resistance of the photovoltaic module
and then achieve the dual effect of energy saving and carbon reduction. At the same time,
the excess of clean energy can be sold back to the commercial electricity providers to sell to
other customers, thus creating a multiple effect of energy saving and carbon reduction. In the
future, the integrated distribution of BIPV electricity network will form the solar power grid
of a metropolitan area that we might call a photovoltaic city.
The methods of integrating photovoltaic and building can be distinguished into two
kinds of technology: installed PV module above a building and PV module integrated with
the building. BIPV’s technology industry chain can be divided into four flows, including solar
cells, photovoltaic modules, building structure, and electromechanical system Figure 1. The
industry chain goes from the most upstream raw material of the silicon, a seed crystal puller
rod for growing silicon boule, and then cuts into a silicon wafer, which is the main source
of solar cells. After being made into solar cells that can be converted into energy, the solar
cells are linked together in modules. Later, these solar modules are combined with a balance
system balance system BOS into a form such as the converter, current breaker, building
support structures, power cables, and monitoring instrumentation and then finally installed.
Figure 1 shows technology and market chain relationships for the BIPV industry. It is
possible to clearly understand technology and market relationships in the BIPV industry. For
example, PV manufacturers must consider and consistently comply with national building
codes when developing products; the construction and electromechanical industries in the
wiring and installation of PV modules; electrical and mechanical piping interface design,
integration, and coordination when building a high-performance, high-security system;
visual appeal. Finally, BIPV building systems are handed to market users after completion
and achieve effective energy savings and carbon reduction, furthering the realization of zero-
energy buildings.
Mathematical Problems in Engineering 5
R&D technology
decision-making
Patent analysis
Choose USPTO
database WPO
Patent matrix map Patent number SIPO
.
.
.
Since BIPV is integrated into the three-dimensional space with the building structure
and does not take up any additional ground space, BIPV is the best model of applying
renewable energy in the highly urbanized area. However, its installation cost is high,
which affects the development of market demand. Therefore, when developed countries are
promoting renewable energy, they rely on the government’s four policy factors, including
incentive policy, subsidies, low installation cost, quick approval process, and lending policies
for financial institutions, which sustain the promotion of the BIPV industry and market
development. In the future, the key factor in reducing the impact on market development
is the continuity of technology development. The installation cost, meaning the economies of
scale of the BIPV industry, could flourish unimaginably when the cost of solar photovoltaic
electricity and cost of commercial electricity are balanced.
3. Methodology
For this study, we intend to explore completed lifecycle evolution of BIPV technology. The
quantity and quality of data acquisition are priorities and explicitly defining the industry
technology is necessary to avoid forecasting bias. For the above reasons, we adapt the expert
opinions and growth curve to develop the forecasting models. Therefore, we combined patent
analysis, growth curve, and expert opinions to analyze the development of BIPV technology
and construct the R&D strategy. The framework of this study is presented in Figure 2. These
are summarized below.
Patent data
Patent searching
Search result
To merge
BIPV patent
research. We must understand its retrieval application purpose, scope, search method, and
search tools before patent searching. This research through the patent analysis is to predict
technology trends and use the BIPV technologies as scope, and then to group Boolean
searches through the USPTO patent database of keywords Boolean Search, its use of the
retrieval tools patented analysis software developed for Lian Ying Technology Patent and
Patent Guider 2008 Tech tools for analysis. The patent searching strategies for this study are
the main access point to keywords and to supplement, patent classification international
patent classification IPC, to achieve the accuracy and integrity of the retrieval processes in
Figure 3.
Literatures of Literatures of
building material building application
Cell Structures
Cable control (controller)
Single Crystal Tray duct machinery
poly Component construction
Amorphous silicon Framing mechanism
Thin film Building Structure
Dye-sensitized support
Copper indium gallium
diselenide
Cadmium telluride
Application
Shingle
Tile
BIPV photovoltaic
Roof
Solar Curtain
Encapsulation-lamination Wall
Panel
Rain
Roll-to-roll
Install Shelter
Flexible Galleries
Glass Method waterproof Canopy
TCO Install shockproof Sunshade
Skylight
Constitute system
Grate
Assembly Window
Patent
documents
between the x-axis items with the y-axis items and the trends. Therefore, we can understand
the relationship of its elements and functions in a two-dimensional array matrix notation
as shown in Table 1. This study uses this two-dimensional array matrix to represent the
technology function matrix, as shown in Table 2. A matrix map shows patent intelligence with
the correlation between technical information, including technology sort and purpose sort.
The matrix can be used to teach important technology and development trends. Therefore,
R&D managers can use patent matrix maps to develop related technologies and preventive
strategies.
k
P t , 3.1
1 e−αt−β
where P t means the patent numbers change with the time t; α means the slope growth
rate of S-curve; β means the inflection point of growth, which is the turning point of time
spent into technique; k means the saturation level of growth; e means nature log. Based on
the practice consideration, k ∗ 10%, k ∗ 90% is defined as the growth interval. The period
time from 10% of technology with maximum utility value to 90% presents by Δt. To proper
Mathematical Problems in Engineering 9
Row-Major
n
m R1 R2 R3 ··· ··· Rn
C1 Q11 Q12 Q13 ··· ··· Q1n
C2 Q21 Q22 Q23 ··· ··· Q2n
Column-Major C3 Q31 Q32 Q33 ··· ··· Q3n
.. .. .. .. .. .. ..
. . . . . . .
Cm Qm1 Qm2 Qm3 ··· ··· Qmn
Purpose sort
P1 P2 P3 ··· ··· Pn
T1 PDQ 11 PDQ 12 PDQ 13 ··· ··· PDQ 1n
T2 PDQ 21 PDQ 22 PDQ 23 ··· ··· PDQ 2n
Technology sort T3 PDQ 31 PDQ 32 PDQ 33 ··· ··· PDQ 3n
.. .. .. .. .. .. ..
. . . . . . .
Tm PDQ m1 PDQ m2 PDQ m3 ··· ··· PDQ mn
Notes: PDQ patent document quantity.
represent the growth process and make it with more fitness, α is replaced by 3.2 in practice
23. Therefore, the P t is proposed as 3.3 as follows:
ln 81
Δt , 3.2
α
k
P t . 3.3
1 e−ln 81/Δt−t−tm
This research is through the BIPV technology growth model calculation and describe
technical development trend.
40 300
35
250
1515
15 14
12 100
11
10 9 9 10
9
7 7 7 7
6 66 6 50
5 5
5 4
3 3 3
2 2 2 1
1 1
0 0
1977
1979
1980
1981
1982
1983
1985
1986
1987
1988
1989
1990
1992
1993
1994
1995
1996
1997
1998
1999
2000
2001
2002
2003
2004
2005
2006
2007
2008
2009
2010
2011
(Time)
numbers are shown in Figure 7. Global warming and near-future energy crisis results show
an increase in related patents since early 2000.
Analyzing the number of patents for technology, purpose, and application Figure 8
can determine the BIPV technology development level of a country. The patent output
of United States US is the first; Japan JP is the second. Germany DE accounted for
approximately 5% of BIPV technology in Europe. Results show that countries most focused
on R&D of BIPV technology are the United States, Japan, and Germany.
Results of the firm-level analysis show major manufacturers and product applications.
Table 3 shows PV module technology directions and strategies in the BIPV industry. The table
shows that most manufacturers continue developing roof technology and its applications.
However, some manufacturers have changed R&D strategies to focus on glass curtain walls
and external walls.
34
32
30
28
26
24
22
Patent number
20
18
16 US
14 TW
LI
12 JP
10 IT
GB
8 FR
DE
6 CH
4 CA
BM
2 BE
0 AU
T0101
T0102
T0103
T0104
T0201
T0202
T0203
T0204
T0205
T0206
T0207
T0301
T0302
T0303
T0304
T0401
T0402
T0403
T0404
Tech
46
44
42
40
38
36
34
32
30
28
Patent number
26
24
22
20 US
18 TW
16 LI
14 JP
12 IT
GB
10 FR
8 DE
6 CH
4 CA
BM
2 BE
0 AU
E0101
E0103
E0104
E0105
E0202
E0203
E0204
E0205
E0206
E0301
E0304
E0305
E0306
E0307
E0401
E0402
E0403
E0404
E0501
E0502
E0504
Effect
AU CA FR JP US
BE CH GB LI
BM DE IT TW
Figure 8: Continued.
12 Mathematical Problems in Engineering
80
70
60
50
Patent number
US
40 TW
LI
JP
30 IT
GB
FR
20 DE
CH
10 CA
BM
BE
0 AU
E0101
E0102
E0103
E0104
E0201
E0202
E0203
E0204
E0205
E0301
E0302
E0303
E0304
E0305
Effect
AU CA FR JP US
BE CH GB LI
BM DE IT TW
Purpose Building
Generate Quality Install Cost
materials to and and
electric down
Technology diversify reliability maintain
Figure 9: Main purpose matrix map of BIPV. Y Q: patent number of years in accordance with this
technology and the effectiveness of patent. The pie chart represents the number of patents. Patent number
of the more means larger the pie chart, and vice versa smaller.
Apply
Outside wall Sole function
Roof of building
Technology of building of structure
24(52) 3(3)
Engineering technology
Figure 10: Main apply matrix map of BIPV. Y Q: patent number of years in accordance with this
technology and the effectiveness of patent. The pie chart represents the number of patents. Patent number
of the more means larger the pie chart, and vice versa smaller.
technology development from 1977 to 2011. The technical function matrix gives a more
precise formulation of the future direction of development policy.
We can understand the technology development trends by analyzing the effect of the
technology and the application of BIPV. After further study through a combination of two
oriented analyses, we will be better able to judge technology trends and liquidity phenomena.
According to technology mapping from 1977 to 2011, the main applications of BIPV are in the
roof of the building Figure 10, especially in USPTO patent pieces in the United States, Japan,
and Germany.
14 Mathematical Problems in Engineering
Large-scale billboard
Telephone booth
Roof of slanted
Technology
Farmhouse
Roof of flat
Photovoltaic building materials 9(10) 2(2)
12(23) 7(8) 1(2) 2(3) 2(2) 3(4)
Module
Drainage 3(3)
Figure 11: Subapplication matrix map of BIPV. YQ: patent number of years in accordance with this
technology and the effectiveness of patent. The pie chart represents the number of patents. Patent number
of the more means larger the pie chart, and vice versa smaller.
Due to high latitudes, the technology development is mainly to oblique roof. We also
can find out technology application and purpose in the field of BIPV technology development
by Figures 11 and 12. The subtechnology matrix can provide more accurate positioning and
development of the strategic direction of future research and development.
We can use dynamic trend analysis, technology research, and development activities
to watch the technology development of the whole picture through patent maps. The cross-
analysis by purpose and applications of technology can help us to grasp the accurate R&D
direction.
After analyzing the technology purpose and application matrix map, it was combined
with time to show the evolution of BIPV purposes and applications over time. This study
proposes a patent number weight over time concept. In Figure 13, there are five major
purposes of BIPV technology on the x-axis and three major applications of BIPV technology
Mathematical Problems in Engineering 15
Waterproof ability
Low-cost module
Easy to maintain
Clean properties
Pervious to light
Technology
Scale economy
Easy to install
Vicariousness
Lightweight
Pure ability
Shade
Photovoltaic building materials 12(25) 3(4) 3(3) 1(2) 3(3) 2(3) 9(10)
Module
Photovoltaic module 5(7) 4(5) 2(2) 2(5) 1(1) 1(1) 3(3) 2(3) 1(1) 1(1) 1(1) 5(6) 1(1) 4(4) 1(2) 1(3)
Glass packaging 4(4) 1(1) 1(1) 1(1)
Test and verify 1(1)
Structure 2(2) 1(1) 1(1) 7(13) 4(4)
Building structure
Framework 2(2) 1(1) 1(1) 1(1) 1(1) 3(3) 16(34) 1(1) 1(1) 2(2) 3(3)
Adjustment 2(2) 1(1) 2(2) 1(1)
Installs the method 1(1) 1(1) 4(6) 1(1) 5(5) 8(10) 1(2) 2(2)
Clean of surface glass 2(2)
Purge for snow 3(3)
Figure 12: Subpurpose matrix map of BIPV. Y Q: patent number of years in accordance with this
technology and the effectiveness of patent. The pie chart represents the number of patents. Patent number
of the more means larger the pie chart, and vice versa smaller.
on the y-axis. Over three periods s1 − s3 , the growth in patent number weights of outside
wall building materials and the installation and maintenance of solar panels can be identified.
The patent number weights include technical application and purpose over three periods.
Technical purposes include electricity generation, diversified building materials, quality
and reliability, installation and maintenance, and cost reduction. Technical applications are
structure function, outside building wall, and building roof. The weights for the technical
purpose and application of the patent number are calculated using 4.1:
pxi ,sj
wxi ,sj m × 100%
i1 pxi ,sj
4.1
pyi ,sj
wyi ,sj n × 100%,
i1 pyi ,sj
where wxi ,sj is the patent number weight for different technical purposes xi in a different
stage sj . pxi ,sj is the patent number for different technical purposes xi in a different stage
sj . wyi ,sj is the patent number weight for different technical applications yi in a different
stage sj . pyi ,sj is the patent number for different technical purposes yi in a different stage
sj .
16 Mathematical Problems in Engineering
80
S1
(y3 ) Roof of
S2 90
building
S3 100
10 20 30 40 50 60 70 80 90 100
Weights for the technical application of the patent number (%)
10 20 30 40 50 60 70 80 90 100
Weights for the technical application of the patent number (%)
Light penetration
in the diurnally
BIPV
Key factors of BIPV R&D R&D
technology development. Therefore, this field of research for BIPV technology is in line with
the number of patents in the outside wall of building development trend forecast in this study
Figure 17. Through the logistic growth, the curve graph shows the technical saturation point
for the 449 patents. Defined according to the value of k, k × 10%, k × 90% corresponds to
the growth point in time, in February 2010. The turning point from growth to maturity is in
January 2022. To close to 90% saturation point in time through the system to calculate the
growth time of 23.6 years in August 2033, its BIPV technology patent saturation point about
cumulative to 449 patents.
18 Mathematical Problems in Engineering
2000
Patent number of total by USPTO
700
Patent number of total by USPTO
300
200
100
90% CR varying saturation
0
1980 1990 2000 2010 2020 2030 2040 2050
(Year)
Figure 17: Technology growth curve of BIPV for outside wall of building.
Mathematical Problems in Engineering 19
Life cycle Introductory stage Growth stage Maturity stage Saturation stage
Beginning time 1977.07 2008.02 2032.01 2057.06
Life cycle Introductory stage Growth stage Maturity stage Saturation stage
Beginning time 1987.05 2010.02 2022.01 2033.08
2000
A: Growth saturation point
Patent number of total by USPTO
1400
1200
Figure 18: S-curve of BIPV. S-curve A: BIPV technology develop forecasting. S-curve B: BIPV on outside
wall of building technology develop forecasting.
By Loglet Lab statistics describing logistic growth curve diagram out BIPV technology
life cycle in Table 5 and BIPV technology application in the outside wall of building
development of the technology life cycle in Table 6.
Figure 18 shows the BIPV technology S-curve. S-curve A refers to BIPV technology
developing forecasting, and S-curve B refers to BIPV outside wall building technology
developing forecasting. S-curve A shows that the introductory BIPV stage was from 1977
to 2008; the growth stage is from 2009 to 2032; the maturity stage is from 2033 to 2057; the
growth saturation point is near 2057. S-curve B shows that the introductory stage of BIPV
outside wall building technology was from 1987 to 2010; the growth stage is from 2011 to
2022; the maturity stage is from 2023 to 2033; the growth saturation point is near 2033.
5. R&D Strategy
BIPV technology functions through the research and technical application of matrix state
space dimensions, combined with the logistic growth curve that takes on a timeline surface.
It shows BIPV technology patents in the field focused on the building roof assembly and
installation techniques and applications. In the vertical plane of buildings, the technology
and application of patents clearly show the phenomenon of loose or blank. This one-block
technology for three-dimensional space of the buildings shows the use of maximum power
generation area in the world’s major countries to actively develop the future generation of
20 Mathematical Problems in Engineering
Patent
R&D
Degree of R&D Benefit strategy Description
protection capability creation
The patent with high-strength and high-product ben-
H H H Leader efits usually adopt a cross-licensing model and other
competitors to obtain the balance of terror
Actively to the goal of Sheep to obtain royalty patent
H H L Attacker
portfolio through intensive construction
Through the acquisition of key patents, in addition to
H L H Stalker protection products and the other can sue the other
party infringement strategy
As a means through to obtain and hold patents and
H L L Patent Troll patent infringement litigation as a threat to force
payment of royalties
Revenue up to the scale of the degree will become the
L H H The target main subject matter of the patent litigation with the
attacks to pay high royalties
L H L Problem Uncertainty region
Technologies and strategies for competitive advan-
L L H Responder tage, and only passively follow the changes in the
environment
L L L Give up
Mathematical Problems in Engineering 21
security and reliability, unique color, and pattern change while maintaining high
photoelectric conversion efficiency. This enhances market appeal;
3 amorphous silicon solar-cell efficiency: amorphous silicon thin-film solar-cell
modules can be bent and absorb diffuse light to produce electrical energy
characteristics. These advantages are not matched by silicon solar-cell modules.
Amorphous silicon thin-film solar-cell module use in the BIPV industry is small.
However, power generation efficiency of amorphous silicon thin-film solar-cell
modules is less than that of silicon solar-cell modules. In the future solar-cell
market, silicon solar-cell modules will remain the main market, but amorphous
silicon thin-film solar-cell modules will continue to grow. Therefore, enhancing
the photoelectric conversion efficiency of amorphous silicon thin-film solar-cell
modules should be a BIPV industry priority;
4 strategic integration of industry groups: to continually improve BIPV market
attractiveness and reduce manufacturing costs, the BIPV industry must integrate
the PV, construction, and electromechanical industries to form an industrial group.
Active industrial and technical cooperation can promote and enhance the BIPV
industry for future development;
5 import green building materials: as government subsidy policies gradually
decrease, subsidies will no longer affect BIPV market development. Future BIPV
technology development must be combined with the concept of green buildings
and comply with building material specifications. Making BIPV products a part of
green building materials is necessary for the future.
think that the proposed model is useful for forecasting the trend of technology, especially
in the field of BIPV. In the future, we hope it will also be applied in other technologies and
industries.
References
1 S. J. Liu and J. Shyu, “Strategic planning for technology development with patent analysis,”
International Journal of Technology Management, vol. 13, no. 5-6, pp. 661–680, 1997.
2 J. L. Chen, S. J. Liu, and C. H. Tseng, “Technological innovation and strategy adaptation in the product
life cycle,” Technology Management: Strategies and Application, vol. 5, no. 3, pp. 183–202, 2000.
3 S. Kuznets, “Innovative activity: problems of definition and measurement,” in The Rate and Direction
of Inventive Activity, R. Nelson, Ed., Princeton University Press, New Jersey, NY, USA, 1962.
4 B. H. Hall, Z. Griliches, and J. A. Hausman, “Patents and R&D: is there a lag?” International Economic
Review, vol. 27, no. 2, pp. 265–283, 1986.
5 H. Ernst, “The use of patent data for technological forecasting: the diffusion of CNC-technology in
the machine tool industry,” Small Business Economics, vol. 9, no. 4, pp. 361–381, 1997.
6 S. B. Chang, Establishing two stages technological positions model: derived from the patent analysis of business
method [Doctoral dissertation], National Yunlin University of Science and Technology, Yunlin, Taiwan,
2004.
7 W. Barroso, L. Quoniam, and E. Pacheco, “Patents as technological information in Latin America,”
World Patent Information, vol. 31, no. 3, pp. 207–215, 2009.
8 B. P. Abraham and S. D. Moitra, “Innovation assessment through patent analysis,” Technovation, vol.
21, no. 4, pp. 245–252, 2001.
9 H. Ernst, “Patenting strategies in the German mechanical engineering industry and their relationship
to company performance,” Technovation, vol. 15, no. 4, pp. 225–240, 1995.
10 C. H. Guo, “Taiwan solar cell patent application for the current situation,” Guang Lian: Guang Dian
Chan Ye Ji Ji Shu Xue Bao, vol. 76, pp. 50–56, 2008 Chinese.
11 Y. B. Chen and X. Y. Cheng, “Solar cell patent analysis and development trends,” in Proceedings of the
ITRI Innovation and Technology Management Conference, pp. 1–8, 2006.
12 A. G. Hu, “The regionalization of knowledge flows in East Asia: evidence from patent citations data,”
World Development, vol. 37, no. 9, pp. 1465–1477, 2009.
13 R. J. W. Tijssen, “Global and domestic utilization of industrial relevant science: patent citation analysis
of science—technology interactions and knowledge flows,” Research Policy, vol. 30, no. 1, pp. 35–54,
2001.
14 C. V. Trappey, H. Y. Wu, F. Taghaboni-Dutta, and A. J. C. Trappey, “Using patent data for technology
forecasting: China RFID patent analysis,” Advanced Engineering Informatics, vol. 25, no. 1, pp. 53–64,
2011.
15 C. Choi and Y. Park, “Monitoring the organic structure of technology based on the patent
development paths,” Technological Forecasting and Social Change, vol. 76, no. 6, pp. 754–768, 2009.
16 Y. H. Chen, C. Y. Chen, and S. C. Lee, “Technology forecasting and patent strategy of hydrogen energy
and fuel cell technologies,” International Journal of Hydrogen Energy, vol. 36, no. 12, pp. 6957–6969, 2011
Chinese.
17 A. C. Cheng, C. J. Chen, and C. Y. Chen, “A fuzzy multiple criteria comparison of technology
forecasting methods for predicting the new materials development,” Technological Forecasting and
Social Change, vol. 75, no. 1, pp. 131–141, 2008.
18 W. B. Ashton and R. K. Sen, “Using patent information in technology business planning-II,” Research
Technology Management, vol. 32, no. 1, pp. 36–42, 1989.
19 S. M. Millett and E. J. Honton, A Manager’s Guide to Technology Forecasting and Strategy Analysis
Methods, Battelle Press, Ohio, USA, 1991.
20 A. L. Porter, Forecasting and Management of Technology, John Wiley & Sons, New York, NY, USA, 1991.
21 S. T. Yeh, “Design application and case study of BIPV in Taiwan,” in Proceedings of the 4th Architectural
Forum in Taiwan—The Wisdom of Building and Eco-City, National Federation of the Architects
Association of the Republic of China, 2007.
24 Mathematical Problems in Engineering
22 T. Liour, C. W. Chen, C. B. Chen, and C. C. Chen, “White-light LED lighting technology life cycle
forecasting and its National and company-wide competitiveness,” in Proceedings of the International
Conference on Asia Pacific Business Innovation & Technology Management, pp. 1–14, 2011.
23 W. T. Chang, Technology development and corporation competence in magneto-resistive random access
memory: a study base on USPTO patent [Master’s dissertation], Frng Chia University, Taiwan, 2007.
Hindawi Publishing Corporation
Mathematical Problems in Engineering
Volume 2012, Article ID 182584, 12 pages
doi:10.1155/2012/182584
Research Article
Solving the Tractor and Semi-Trailer Routing
Problem Based on a Heuristic Approach
Copyright q 2012 Hongqi Li et al. This is an open access article distributed under the Creative
Commons Attribution License, which permits unrestricted use, distribution, and reproduction in
any medium, provided the original work is properly cited.
We study the tractor and semi-trailer routing problem TSRP, a variant of the vehicle routing
problem VRP. In the TSRP model for this paper, vehicles are dispatched on a trailer-flow network
where there is only one main depot, and all tractors originate and terminate in the main depot. Two
types of decisions are involved: the number of tractors and the route of each tractor. Heuristic
algorithms have seen widespread application to various extensions of the VRP. However, this
approach has not been applied to the TSRP. We propose a heuristic algorithm to solve the TSRP.
The proposed heuristic algorithm first constructs the initial route set by the limitation of a driver’s
on-duty time. The candidate routes in the initial set are then filtered by a two-phase approach. The
computational study shows that our algorithm is feasible for the TSRP. Moreover, the algorithm
takes relatively little time to obtain satisfactory solutions. The results suggest that our heuristic
algorithm is competitive in solving the TSRP.
1. Introduction
In this paper, we consider the tractor and semi-trailer routing problem TSRP, a variant of the
vehicle routing problem VRP. The VRP is one of the most significant problems in the fields
of transportation, distribution, and logistics. The basic VRP consists of some geographically
dispersed customers, each requiring a certain weight of goods to be delivered or picked up.
A fleet of identical vehicles dispatched from a depot is used to deliver the goods, and the
vehicles must terminate at the depot. Each vehicle can carry a limited weight and only one
vehicle is allowed to visit each customer. It is assumed that some parameters e.g., customer
demands and travel times are known with certainty. The solution of the problem consists
2 Mathematical Problems in Engineering
of finding a set of routes that satisfy the freight demand at minimal total cost. In practice,
additional operational requirements and restrictions, as in the case of the truck and trailer
routing problem TTRP, may be imposed on the VRP 1. The TTRP was first studied by
Semet and Taillard et al. 2 and Gerdessen 3 in the 1990s, and it was subsequently studied
by Chao 4, Scheuerer 5, and others. In the TTRP, the use of trailers a commonly neglected
feature in the VRP is considered. Some customers can be served by a combination vehicle
i.e., a truck pulling a trailer, as in type II in Figure 1, while other customers can only
be served by a truck as type I in Figure 1 due to some limitations such as government
regulations, limited maneuvering space at customer sites, road conditions, and so forth. These
constraints exist in many practical situations 1.
The VRP and its various extensions have long been one of the most studied
combinatorial optimization problems due to the problem’s complexity and extensive
applications in practice 6–10. The truck and trailer combination is employed widely by
enterprises around the world, but there additional features introduced by trailers that have
attracted some research. A number of studies have concentrated on applications of the TTRP.
For instance, Semet and Taillard et al. 2 and Caramia and Guerriero et al. 11 gave some
real-world TTRP applications in collection and delivery operations in rural areas or crowded
cities with accessibility constraints. Theoretically, being an extension of the VRP, the TTRP is
NP-Hard. The TTRP is computationally more difficult to solve than the VRP 1. Because the
VRP is usually tackled by heuristic methods 6–9, 12–15, it is feasible to develop heuristic
approaches for the TTRP.
Gerdessen 3 extended the VRP to the vehicle routing problem with trailers and
investigated the optimal deployment of a fleet of truck-trailer combinations by a construction
and improvement heuristic. Scheuerer 5 proposed construction heuristics called T-Cluster
and T-Sweep along with a tabu search algorithm for the TTRP. Tan et al. 16 proposed
a hybrid multiobjective evolutionary algorithm featuring specialized genetic operators,
variable-length representation; and local search heuristics to solve the TTRP. Lin et al. 1
proposed a simulated annealing SA heuristic for the TTRP and suggested that SA is
competitive with tabu search TS for solving the TTRP. Villegas et al. 17 solved the TTRP
by using a hybrid metaheuristic based on a greedy randomized adaptive search procedure
GRASP, variable neighborhood search VNS, and path relinking PR.
Villegas et al. 18 proposed two metaheuristics based on GRASP, VND, and
evolutionary local search ELS to solve the single truck and trailer routing problem with
satellite depots STTRPSD. Considering the number of available trucks and trailers to be
limited in the TTRP, Lin et al. 19 relaxed the fleet size constraint and developed a SA
heuristic for solving the relaxed truck and trailer routing problem RTTRP. Lin et al. 20
proposed a SA heuristic for solving the truck and trailer routing problem with time windows
TTRPTW.
Research to date has considered most types of road vehicles, especially trucks and
truck and trailer combinations. However, there has been little research on the types of tractor
and semi-trailer combinations. Hall and Sabnani et al. 21 studied routes that consisted of
two or more segments and two or more stops in the tour for a tractor. At each stop, the tractor
could drop off one or more trailers and pick up one or more trailers. Control rules based on
predicted route productivity were developed to determine when to release a tractor. Derigs et
al. 22 presented two approaches to solve the vehicle routing problem with multiple uses of
tractors and trailers. The primary objective was to minimize the number of required tractors.
Cheng et al. 23 proposed a model for a steel plant to find the tractor and semi-trailer
equipment and running routes for the purpose of minimizing transport distance. Liang 24
Mathematical Problems in Engineering 3
I II
Tractor Tractor
Semi-trailer Trailer Semi-trailer
III IV
Figure 1: The basic types of vehicles. Note: In practice, many vehicle types are used in road freight
transportation. This figure only lists four basic types. A large number of other types can be derived from
the four basic types by the number of axles, tires and the combination style. Enterprises in most of the
countries in the world employ various types.
established a dispatching model of tractors and semi-trailers in a large steel plant and used a
tabu search algorithm to find the optimal driving path and the cycle program.
We aim to propose a heuristic for the TSRP. This aim is based on the practical
knowledge that tractor and semi-trailer combinations are popular in some countries,
particularly China. The remainder of this paper is organized as follows. Section 2 compares
the TTRP and the TSRP and defines the TSRP. Section 3 proposes a heuristic algorithm to solve
the TSRP. Section 4 employs the heuristic algorithm to solve some experimental networks of
the TSR. Section 5 draws conclusions and gives future research directions.
2. Problem Definition
2.1. The TTRP and the TSRP
Although there is little literature devoted to the definition and solution of the TSRP in the
fields of transportation or logistics, plenty of research has been done on the TTRP, providing
important references for the TSRP. In the TTRP, a heterogeneous fleet composed of mtu trucks
and mtr trailers mtu > mtr serves a set of customers from a main depot. Each customer
has a certain demand, and the distances between any two points including customers and
depots are known. The capacities of the trucks and trailers are determinate. Some customers
must be served only by a truck, while other customers can be served either by a truck or
by a combination vehicle. The objective of the TTRP is to find a set of routes with minimum
total distance or cost so that each customer is visited in a route performed by a compatible
vehicle, the total demand of the customers visited on a route does not exceed the capacity of
the allocated vehicle, and the numbers of required trucks and trailers are not greater than mtu
and mtr , respectively 1, 17. There are three types of routes in a TTRP solution, as illustrated
in Figure 2: 1 a pure truck route traveled by a single truck; 2 a pure vehicle route without
any subtour traveled by a combination vehicle; 3 a combination vehicle route consisting of
a main tour traveled by a combination vehicle and at least one subtour traveled by the truck
alone.
The vehicle types in the TSRP are different from those in the TTRP. The TTRP focuses
on trucks and trailers, both of which can carry cargo. The TSRP involves mta tractors and
mst semi-trailers mta < mst . Although a tractor cannot carry cargo, it has more flexible
dispatching options, and it can pull different semi-trailers on various segments of its route
by the pick-up and drop-off operation at depots.
4 Mathematical Problems in Engineering
T4
C2
C1 C2
T3
C1 C1 C2
T5
C2
C3
C2
C4
1
C
D
C7 D C4 D C2 C4
C7 D C4 D C3 C4
C4 C3
C7
C6
C6
C5
D
C
5C
T2
6
C5 T1
D Depot
Loaded carriages of trucks T1 , T2 , . . . , Tn Customers for trucks visiting
The TSRP can be formally defined on a directed graph G V, A, where V
{0, 1, 2, . . . , n} is the set of vertices and A {i, j : i, j ∈ V } is the set of arcs. Each arc
i, j is generally associated with a transportation distance decided by road infrastructure.
The freight flow from i to j is regarded as certain weight of arc i, j. Vertex 0, . . . , v v < n
represents the main depots, in which many tractors and semi-trailers park. Some loaded
semi-trailers wait for visiting customers, and other unloaded semi-trailers wait for visiting
or maintenance. The remaining vertices si in V i.e., V \ {0, . . . , v} correspond to customers
who have m m ≥ 1 loaded semi-trailers waiting to visit l 1 ≤ l ≤ n and l ≤ m orientations
at the beginning of the simulation. Customers may have other unloaded semi-trailers waiting
for loading.
There are various tractor-driving modes on graph G during one daily period. For
example, 1 tractor Trj , pulling one loaded semi-trailer, goes from its main depot to a
customer in one-day period, and the customer has tractor-parking available; 2 tractor Trj ,
pulling one loaded semi-trailer, goes from its main depot to c1 . After the dropping and pulling
operations at c1 , the tractor goes on to another customer, c2 . The tractor Trj terminates at a
customer who has tractor-parking available. 3 It is similar to the running course listed in 2,
but Tractor Trj terminates at its main depot. The most basic elements of tractor-driving modes
include the following: how many semi-trailers can be pulled synchronously by a tractor, how
many vertexes are passed by the tractor, how many times per day the tractor can drop off one
or more trailers and pick up one or more trailers, whether a tractor terminates at its original
main depot, whether the semi-trailer pulled by a tractor loads cargo, and if a tractor runs
alone. In addition, a time window constraint is probably required.
Mathematical Problems in Engineering 5
8 1 4
1 1
4 2 0
6 − Four =
5 7 1 1 1 3
5 4 1 1 0
1
4 1 0
− = 1
2 1
1 3 1 1 0 2
1 1 0
3 1 2
−
1 1 =
0
2 1 1
···
Figure 3: An example of selecting unit-flow networks. Note: Black points denote depots. Lines denote the
distribution of freight flows. Numbers near lines denote freight flow volume unit: one semi-trailer.
C9
C8 C3 D
C6
C5
C2
C3 C3
D
C2 C4
C1 C
2
C1
DC
1
C7
D Depot
C1 , C2 , . . . , Cn Customers
Loaded semi-trailers
The TSRP model in this paper uses the unit-flow network. Vertex 0 represents the main
depot where some loaded semi-trailers are waiting to be delivered to customers. The vertices
in V \ {0} correspond to customers who have m 1 ≤ m ≤ n loaded semi-trailers waiting for
going to m customers. The n customers have no parked tractors. The tractor-driving modes
must satisfy some constraints, including the following: tractors terminate at the main depot,
a tractor can pull one loaded semi-trailer and can also run alone, and the working time of a
tractor is decided by its driver team. All tractors or vehicles a vehicle is one tractor pulling
one semi-trailer originate and terminate at the main depot. Whenever a tractor passes by a
customer, the tractor picks up a semi-trailer. Whenever a vehicle passes by a customer, the
vehicle drops off its semi-trailer and picks up another one. After one-day period, the number
of semi-trailers parked in every customer point is not less than a minimum Figure 4.
The TSRP model consists of determining the number of tractors to be used and the
route of each tractor so that the variable costs and service level are balanced, while each
route starts and ends at the main depot. Variable costs are reduced by decreasing the overlap
distance of tractors running alone. The service level is based on the percentage of customer
demand that is satisfied.
The elements of the initial solution set are tractor routes. To give the form of a route,
we suggest the following procedure. The number of drivers assigned to each tractor is k. The
on-duty time of each driver is T hours per person · day. The distance between depots i and
j is dij . The depot sequence on each route is denoted by H − s1 − · · · − sf − H, which is the
form of a route. The same customer is visited only once on a certain tractor’s route, and each
si i 1, 2, . . . f is unique. The on-duty time T is a constraint on the route. That is,
dij
ρ1 · kT ≤ f · ts 2 · tH ≤ ρ2 · kT, 3.1
i j
v
where ts and tH are the temporary rest time and the residence time, respectively. ρ1 0 < ρ1 <
1 and ρ2 1 ≤ ρ2 ≤ τ, τ is a limited number are the lower and upper limits of the utilization
ratio of the on-duty time, respectively. v is the average velocity of the tractor.
We suggest the steps below to construct elements of the initial solution set.
Step 1. Transform the distance matrix into a running time matrix. Use v as the parameter in
the transformation. Factors affecting v in practice include the tractor condition, the driver’s
skill, and traffic conditions. We estimate v by enterprise experience.
Step 2. Search the running time matrix. Let f be the sum of customers on a route. If f is very
large, there are too many customers on the route to allow too much temporary rest time.
Therefore, f has a maximum, and the maximum is certainly less than kT − 2 · tH /ts . Once
f is found, we implement an entire search on the running time matrix to find all routes that
satisfy the constraint 3.1.
Step 3. Compare the routes with freight flow demand. Every route in Step 2, which contains
many segments, is constructed by the segment running time of all customer pairs. In fact, not
all pairs of customers require freight exchange. There are segments on which no freight flows,
and tractors run alone on such segments. To save variable costs, the time of tractors running
alone is limited. Therefore, we obtain the initial solution set after the elimination of routes
based on the freight flow demand and the cost-saving requirement.
1 The first step is for the same type. An overlap arc i, j where the tractor running
time is less than a maximum tij max can be accepted. If the tractor needs more time
8 Mathematical Problems in Engineering
than tij max on arc i, j, only one of the routes containing arc i, j is permitted to
be chosen.
2 The second step is for different types. A “tractor route—overlap arc” matrix A0 ,
with elements aik is constructed. The rows of matrix A0 are serial numbers sni of
routes and the columns are various overlap arcs. The elements of matrix A0 are 1
or 0. If the element on row i and column k is “1”, then the sni route has an overlap
arc. When aik 1, any elements of the matrix that satisfy aij 1 are considered.
The column which contains aij has a sum i aij . If the route with serial number sni
is chosen, j i aij should be at a minimum. Once the sni route is chosen, other
routes that have the same overlap arcs with the sni route are eliminated from A0 .
Consequently, a row in A0 changes, and a new matrix A1 appears. The operation
is repeated until there is no row available in the last “tractor route—overlap arc”
matrix.
In the first phase, a transitional solution set that contains such elements as the sni
routes is constructed by improving the initial solution set.
Step 1. Construct a zero matrix O its elements are oij whose rows and columns are depots
i.e., the main depot and customer depots. Because of transportation demand, there are
freight flows between two particular depots. Because the segments of tractor routes in the
transitional solution set are defined by depots, we add 1 to oij when there is a route containing
arc i, j. We call such an operation a “fill”. By a “fill” operation, we mark all segments of
routes in the transitional solution set on matrix O. A new matrix B0 is thus formed.
Step 2. All route segments in the transitional solution set actually have corresponding
elements in matrix B0 . If a certain percentage e.g., 80∼100% of all corresponding elements of
the sni route are greater than 1, the sni route is eliminated. When the sni route is eliminated,
all of the corresponding elements subtract 1. We call such an operation a “cut”. Repeat the
“cut” operation, and a new matrix B1 finally forms. The routes corresponding to B1 make up
the satisfactory solution set.
In some cases, the number of nonzero elements of B1 is less than that of the freight flow
demands . Therefore, routes corresponding to B1 cannot satisfy all transportation demand. In
order to satisfy more customers’ demands , we can add some routes that contain overlap
arcs to increase the market adaptability of the satisfactory solution. However, too many
overlap arcs can exist because of uneven freight flows. Therefore, to balance the service level
and costs, meeting a certain percentage e.g., 80% of all transportation demand can be the
objective.
4. Computational Study
We abstract the transportation network on an N × N grid, where the nodes denote the main
depot and customer depots. In our computational study, the “RANDOM” function in Matlab,
Mathematical Problems in Engineering 9
To
From
H 1 2 3 4 5 6 7 8 9 10 11 12 13
H 0 5.5 4.0 4.5 1.5 1.0 1.0 4.0 3.0 4.5 1.0 3.5 3.5 3.0
1 5.5 0 1.5 1.0 4.0 4.5 4.5 2.5 2.5 3.0 5.5 4.0 5.0 6.5
2 4.0 1.5 0 2.5 2.5 3.0 3.0 4.0 3.0 4.5 4.0 3.5 3.5 5.0
3 4.5 1.0 2.5 0 3.0 3.5 3.5 1.5 1.5 2.0 4.5 3.0 4.0 5.5
4 1.5 4.0 2.5 3.0 0 0.5 0.5 4.5 3.5 5.0 1.5 4.0 4.0 3.5
5 1.0 4.5 3.0 3.5 0.5 0 1.0 5.0 4.0 5.5 2.0 4.5 4.5 4.0
6 1.0 4.5 3.0 3.5 0.5 1.0 0 4.0 3.0 4.5 1.0 3.5 3.5 3.0
7 4.0 2.5 4.0 1.5 4.5 5.0 4.0 0 1.0 0.5 4.0 2.5 3.5 5.0
8 3.0 2.5 3.0 1.5 3.5 4.0 3.0 1.0 0 1.5 3.0 1.5 2.5 4.0
9 4.5 3.0 4.5 2.0 5.0 5.5 4.5 0.5 1.5 0 3.5 2.0 3.0 4.5
10 1.0 5.5 4.0 4.5 1.5 2.0 1.0 4.0 3.0 3.5 0 2.5 2.5 2.0
11 3.5 4.0 3.5 3.0 4.0 4.5 3.5 2.5 1.5 2.0 2.5 0 1.0 2.5
12 3.5 5.0 3.5 4.0 4.0 4.5 3.5 3.5 2.5 3.0 2.5 1.0 0 1.5
13 3.0 6.5 5.0 5.5 3.5 4.0 3.0 5.0 4.0 4.5 2.0 2.5 1.5 0
To
From
H 1 2 3 4 5 6 7 8 9 10 11 12 13
b a
H 0 1 1 1 1 1 1 1 1 1 1 1 1 1
1 1 0 1 0 1 1 0 0 0 0 0 0 0 1
2 1 1 0 0 0 0 0 0 1 0 1 1 0 0
3 1 0 0 0 1 0 0 0 0 0 1 0 1 1
4 1 0 0 1 0 0 0 0 1 1 0 0 0 1
5 1 0 0 1 1 0 0 0 0 0 1 0 0 1
6 1 0 0 0 1 1 0 0 0 0 1 1 0 0
7 1 0 0 0 1 0 0 0 0 0 1 1 1 0
8 1 0 1 0 0 0 1 0 0 1 0 1 0 0
9 1 0 1 0 1 0 1 0 0 0 0 1 0 0
10 1 1 0 0 0 1 0 1 1 0 0 0 0 0
11 1 0 1 0 1 0 0 0 1 0 0 0 0 1
12 1 0 1 0 0 1 0 0 0 0 0 1 0 1
13 1 0 0 0 1 0 1 0 1 1 0 0 0 0
a
“1” denotes that there is one semi-trailer flow between two depots.
b
“0” denotes that there is no freight flow.
which can generate random arrays from a specified distribution, is used. By RANDOM
“norm”,1,1,10,10, a random array is generated. We select the negative positions of the array
as nodes and the minimum position as the main depot. The distance between any two nodes
is calculated by the gaps of rows and columns. The “RANDOM” function in MATLAB is also
used to determine the freight flow between two depots. The network expressed by Table 1
and the flow expressed by Table 2 make up example No. 1. By the above generation method,
we produce some transportation networks that are used to test the heuristic algorithm.
10 Mathematical Problems in Engineering
Table 3: The satisfactory solution of the TSRP achieved by the heuristic algorithm.
Working time
Types of the routes Form of route needed by
routes hours
A tractor with two drivers. Two different
customer depots are passed by. H—1—13—H 17
The route repeats once a day.
H—2—8—6—10—5—H;
A tractor with two drivers. Five different H—6—4—13—8—11—H;
customer depots are passed by. H—6—5—3—12—11—H; 17.5
The route repeats once a day. H—6—11—8—9—4—H;
H—10—8—2—1—5—H
A tractor with two drivers. Six different
customer depots are passed by. H—6—5—4—8—9—11—H 17
The route repeats once a day.
According to some enterprise experience, a driver’s on-duty time is 8.5 hours per
day. One or two drivers are assigned to a tractor. A tractor with two drivers can work
consecutively for no more than 17 hours in a 24-consecutive-hour period. The temporary
rest time in customer depots is 0.5 hour, and the residence time in H is 1 hour. By using the
approach mentioned in Section 3.1, we attain different types of tractor routes for the No. 1
example. There are 175 elements in the initial solution set. By using the 2-phase approach, we
attain the satisfactory solution of the No. 1 example Table 3. When the enterprise employs
tractor routes as the satisfactory solution, it can satisfy 80 percent of all transportation
demand. Sixteen tractors and thirty-two drivers are needed during a 24-consecutive-hour
period. The total running time of 16 tractors is 230 hours per day. In 15 percent of the total
running time, tractors run alone.
It is feasible to propose exact algorithms e.g., integer programming for the TSRP
when the initial solution set is constructed. We proposed a 0-1 integer programming for
the No. 1 example and attained the exact solution. The exact solution can satisfy 84
percent of all transportation demand. Sixteen tractors and thirty-two drivers are needed
during a 24-consecutive-hour period. In 10 percent of the total running time, tractors
run alone. We implemented the proposed heuristic algorithm using Matlab and the 0-1
integer programming with QS. Although the exact algorithm can attain a slightly better
solution, it requires more calculating time. For the No. 1 example, the solving time using
the heuristic algorithm was approximately 80 seconds while that for the exact algorithm was
approximately 2000 seconds. We suggest that the heuristic algorithm has an advantage for
solving the TSRP.
Mathematical Problems in Engineering 11
We have repeated the generation of random arrays over 50 times to obtain some typical
computational networks. The heuristic algorithm was employed on these networks. We ran
the experiments of this section on a computer with an AMD Athlontm X2 Dual-Core QL-
65 running at 2.10 GHz under Windows 7 ultimate 32 bits with 2 GB of RAM. Table 4
summarizes the characteristics of each solution in the 12-instance testbed.
Acknowledgments
This work was partially funded by the Science and Technology Plan of Transportation
of Shandong Province 2009R58 and the Fundamental Research Funds for the Central
Universities YWF-10-02-059. This support is gratefully acknowledged.
12 Mathematical Problems in Engineering
References
1 S. W. Lin, V. F. Yu, and S. Y. Chou, “Solving the truck and trailer routing problem based on a simulated
annealing heuristic,” Computers and Operations Research, vol. 36, no. 5, pp. 1683–1692, 2009.
2 F. Semet and E. Taillard, “Solving real-life vehicle routing problems efficiently using tabu search,”
Annals of Operations Research, vol. 41, no. 4, pp. 469–488, 1993.
3 J. C. Gerdessen, “Vehicle routing problem with trailers,” European Journal of Operational Research, vol.
93, no. 1, pp. 135–147, 1996.
4 I. M. Chao, “A tabu search method for the truck and trailer routing problem,” Computers and
Operations Research, vol. 29, no. 1, pp. 33–51, 2002.
5 S. Scheuerer, “A tabu search heuristic for the truck and trailer routing problem,” Computers and
Operations Research, vol. 33, no. 4, pp. 894–909, 2006.
6 J. Renaud and F. F. Boctor, “A sweep-based algorithm for the fleet size and mix vehicle routing
problem,” European Journal of Operational Research, vol. 140, no. 3, pp. 618–628, 2002.
7 H. C. Lau, M. Sim, and K. M. Teo, “Vehicle routing problem with time windows and a limited number
of vehicles,” European Journal of Operational Research, vol. 148, no. 3, pp. 559–569, 2003.
8 F. Li, B. Golden, and E. Wasil, “A record-to-record travel algorithm for solving the heterogeneous fleet
vehicle routing problem,” Computers and Operations Research, vol. 34, no. 9, pp. 2734–2742, 2007.
9 S. Liu, W. Huang, and H. Ma, “An effective genetic algorithm for the fleet size and mix vehicle routing
problems,” Transportation Research Part E, vol. 45, no. 3, pp. 434–445, 2009.
10 E. Cao and M. Lai, “The open vehicle routing problem with fuzzy demands,” Expert Systems with
Applications, vol. 37, no. 3, pp. 2405–2411, 2010.
11 M. Caramia and F. Guerriero, “A milk collection problem with incompatibility constraints,” Interfaces,
vol. 40, no. 2, pp. 130–143, 2010.
12 D. Pisinger and S. Ropke, “A general heuristic for vehicle routing problems,” Computers & Operations
Research, vol. 34, no. 8, pp. 2403–2435, 2007.
13 Y. Marinakis and M. Marinaki, “A hybrid multi-swarm particle swarm optimization algorithm for the
probabilistic traveling salesman problem,” Computers & Operations Research, vol. 37, no. 3, pp. 432–442,
2010.
14 A. Garcia-Najera and J. A. Bullinaria, “An improved multi-objective evolutionary algorithm for the
vehicle routing problem with time windows,” Computers & Operations Research, vol. 38, no. 1, pp.
287–300, 2011.
15 L. Hong, “An improved LNS algorithm for real-time vehicle routing problem with time windows,”
Computers and Operations Research, vol. 39, no. 2, pp. 151–163, 2012.
16 K. C. Tan, Y. H. Chew, and L. H. Lee, “A hybrid multi-objective evolutionary algorithm for solving
truck and trailer vehicle routing problems,” European Journal of Operational Research, vol. 172, no. 3,
pp. 855–885, 2006.
17 J. G. Villegas, C. Prins, C. Prodhon, A. L. Medaglia, and N. Velasco, “A GRASP with evolutionary
path relinking for the truck and trailer routing problem,” Computers and Operations Research, vol. 38,
no. 9, pp. 1319–1334, 2011.
18 J. G. Villegas, C. Prins, C. Prodhon, A. L. Medaglia, and N. Velasco, “GRASP/VND and multi-start
evolutionary local search for the single truck and trailer routing problem with satellite depots,”
Engineering Applications of Artificial Intelligence, vol. 23, pp. 780–794, 2010.
19 S. W. Lin, V. F. Yu, and S. Y. Chou, “A note on the truck and trailer routing problem,” Expert Systems
with Applications, vol. 37, no. 1, pp. 899–903, 2010.
20 S. W. Lin, V. F. Yu, and C. C. Lu, “A simulated annealing heuristic for the truck and trailer routing
problem with time windows,” Expert Systems with Applications, vol. 38, pp. 15244–15252, 2011.
21 R. W. Hall and V. C. Sabnani, “Control of vehicle dispatching on a cyclic route serving trucking
terminals,” Transportation Research Part A, vol. 36, no. 3, pp. 257–276, 2002.
22 U. Derigs, R. Kurowsky, and U. Vogel, “Solving a real-world vehicle routing problem with multiple
use of tractors and trailers and EU-regulations for drivers arising in air cargo road feeder services,”
European Journal of Operational Research, vol. 213, no. 1, pp. 309–319, 2011.
23 Y. R. Cheng, B. Liang, and M. H. Zhou, “Optimization for vehicle scheduling in iron and steel works
based on semi-trailer swap transport,” Journal of Central South University of Technology, vol. 17, no. 4,
pp. 873–879, 2010.
24 B. Liang, Research on semi-trailer loop swap transportation applied in large-scale iron and steel works [M.S.
thesis], Central South University, 2009.
Hindawi Publishing Corporation
Mathematical Problems in Engineering
Volume 2012, Article ID 849634, 10 pages
doi:10.1155/2012/849634
Research Article
Combining Diffusion and Grey Models Based on
Evolutionary Optimization Algorithms to Forecast
Motherboard Shipments
Copyright q 2012 Fu-Kwun Wang et al. This is an open access article distributed under the
Creative Commons Attribution License, which permits unrestricted use, distribution, and
reproduction in any medium, provided the original work is properly cited.
It is important for executives to predict the future trends. Otherwise, their companies cannot
make profitable decisions and investments. The Bass diffusion model can describe the empirical
adoption curve for new products and technological innovations. The Grey model provides short-
term forecasts using four data points. This study develops a combined model based on the rolling
Grey model RGM and the Bass diffusion model to forecast motherboard shipments. In addition,
we investigate evolutionary optimization algorithms to determine the optimal parameters. Our
results indicate that the combined model using a hybrid algorithm outperforms other methods for
the fitting and forecasting processes in terms of mean absolute percentage error.
1. Introduction
Taiwanese motherboard manufacturers create 98.5% of the worldwide desktop motherboards
and dominate the global desktop motherboard market 1. However, this industry’s growth
rate has slowed because of the trend of replacing desktop PCs with notebooks or netbooks.
In addition, the aggressive pricing by notebook/netbook manufacturers has diminished
desktop motherboard sales. It is important to develop a new forecasting model for this
rapidly changing market and to compare its results with other forecasting models. These
results can assist manufactures in making decisions on future expansion and investment.
Several studies have proposed time-series models for industrial production, demon-
strating the applicability of time-series models to industrial production forecasting. These
2 Mathematical Problems in Engineering
models typically require large amounts of data. However, Hsu 2 proved that the Grey
model GM, developed by Deng 3, requires minimal data and is the best model for limited
data prediction. Furthermore, GM can forecast in a completive environment where decision
makers have limited historical data. Chang et al. 4 applied a variable p value to a rolling
Grey model RGM to forecast semiconductor production in Taiwan. Akay and Atak 5
used the Grey prediction model with a rolling mechanism to forecast electricity demand in
Turkey. Hsu and Wang 6, 7 used the Bayesian method to improve GM1,1 for forecasting
the integrated circuit industry.
The Bass diffusion model 8 has been used to develop product life cycle curves and
to forecast the sales of the initial purchases of new products. Tseng and Hu 9 combined
fuzzy regression with the Bass model to develop a quadratic interval Bass diffusion model.
Tsaur 10 used the fuzzy grey regression model to predict the liquid critical display LCD
television market. Based on empirical data analysis, fuzzy grey regression is capable of
accurate forecasting and can give decision makers various scenarios. Wu and Chu 11 used
Gompertz, Logistic, Bass, and time-series autoregressive moving average ARMA models to
forecast mobile telephone subscription in Taiwan. Hsiao and Wang 12 applied the GM1,1,
RGM1,1, and Bass diffusion models to predict trends in the global copper clad laminate
industry.
This study proposes a combined model based on the rolling Grey and Bass diffusion
models to forecast the sale of Taiwanese motherboards more accurately. This study is
organized as follows: Section 2 reviews the Bass diffusion model and the RGM1,1; Section 3
introduces a combined model featuring an evolutionary optimization algorithm; in Section 4
we use the combined model to forecast motherboard shipments, comparing its results with
those of the RGM1,1 and Bass diffusion models; finally, we provide a conclusion.
where Ft 1 − e−pqt /1 q/p × e−pqt , nt sales at time t, m is the number
of eventual adopters, Ft is the cumulative distribution of adoptions at time t, p is the
coefficient of innovation, q is the coefficient of imitation, and ε is the normally distributed
random error term with mean zero and variance σ 2 . The adopter’s probability density
function ft for adoption at time t is given by
2
p q /p × e−pqt
ft . 2.2
2
1 q/p × e−pqt
Bass 13 used the ordinary least squares OLS method to estimate the parameters.
Schmittlein and Mahajan 14 used the maximum likelihood estimation MLE method to
improve the estimation. Srinivsan and Mason 15 used a nonlinear least square estimation
NLS method to obtain the valid error estimates. Nonlinear models are more difficult to fit
Mathematical Problems in Engineering 3
than linear models. Venkatesan and Kumar 16 used genetic algorithms GAs to estimate
the parameters, and these were consistent with the NLS method.
Grey theory is used for systems that have uncertain and imperfect information 3. It
requires only four data points to construct a prediction model. Grey prediction has three basic
operations: accumulated generating operator AGO, the inverse accumulating operator
IAGO, and the GM. The steps of the RGM1,1 model are given as follows.
An AGO operator is used to convert the original series into monotonically increasing series:
X 1 x1 1, x1 2, x1 3, . . . , x1 n , 2.4
Step 2. The first-order differential equation for the GM1,1 model is given by
dX 1 2.5
a × X 1 b,
dt
where t denotes the independent variables in the system, a represents the developed
coefficient, and b is the Grey controlled variable. The parameters of a and b can be obtained
using the OLS method. Thus, we have
a −1
u BT B × B T × YN , 2.6
b
T
P is equal 0.5 in the original model, and YN x0 i 1, x0 i 2, . . . , x0 k .
b b
x1 t 1 x0 1 − × e−at . 2.8
a a
4 Mathematical Problems in Engineering
Motherboard
shipments
Form a new
data set
(S3 and S2)
Supposing that x1 1 x0 1, the sequence one-order IAGO is acquired. Thereafter, the
sequence can be obtained as x0 t 1 x1 t 1 − x1 t.
Given t 1, 2, . . . , k, the sequence of reduction is as follows:
X 0 i; k x0 1, x0 2, . . . , x0 k 1 , 2.9
where x0 k 1 is the Grey elementary forecasting value for x0 k 1.
3. Combined Model
Combining forecasts minimizes errors 18, and many studies have demonstrated their value
19–21. In this section, we present a combined model based on the Bass diffusion model and
the RGM1,1 see Figure 1. The three major steps of the proposed combined model are as
follows.
Mathematical Problems in Engineering 5
Step 1. A new data set should be formed. The RGM1,1 usually provides better fitted values
for early periods than does the Bass diffusion model. Next, the original data points are
replaced by some predicted values by the RGM1,1 to form a new data set.
Step 2. The optimal parameters of the Bass diffusion model should be found for the original
data and a new data set. A hybrid algorithm based on the GA with PSO has been successfully
applied to real-world engineering design problems 22, 23. This study used a hybrid
algorithm that couples the GA with PSO to optimize the estimates of the parameters for the
Bass diffusion model. We then investigated three algorithms to obtain the parameters of the
Bass diffusion model. First, we used a nonlinear algorithm to obtain the initialized estimates
of the model. In addition, the confidence intervals for the parameters were used to determine
the range of the parameters for PSO and the GA. The descriptions of these three algorithms
are as follows.
1 GA: the estimated parameters can be obtained easily using Evolver Software 24.
The minimized function is defined as MAPE. Here, the population size is set by 50,
and the crossover rate and the mutation rate are set as 0.6 and 0.2, respectively.
2 PSO algorithm: we used PSO operators velocity and position updates to update
the individual with the worst fitness. Clerc and Kennedy 25 created constriction
factor k, improving the ability of PSO to constrain and control velocities. These
equations are given by
VidNew k × Vidold c1 × r1 × pid − xid
old
c2 × r2 × pgd − xid
old
,
3.1
New
xid xid
old
VidNew ,
where k is an inertia weight, c1 and c2 are two positive constants called acceleration
coefficients, and r1 and r2 are random, uniformly√distributed numbers 0, 1. The
inertia weight can be obtained by k 2/|2 − c − c2 − 4c|, where c c1 c2 > 4.
For example, if c 4.1, then w 0.729. As c increases above 4.0, k becomes smaller.
The input parameters are particle size 20, max iterations 100, w 0.729, and
c1 c2 2.05.
3 A hybrid algorithm: this hybrid algorithm couples the GA with PSO. First, we used
the PSO algorithm to obtain the estimated parameters. Thereafter, the estimated
parameters were improved using the GA.
Step 3. The fitted Bass diffusion model can be used to generate forecast values.
year 2010. First, the estimated parameters of the Bass diffusion model were obtained using
the GA. Second, the rolling interval for the RGM1,1 was set to five. Thereafter, we obtained
eight parameter settings from 1998 to 2009. Finally, the estimated parameters of the combined
model were obtained using three evolutionary algorithms GA, PSO, and a hybrid. All five
methods were used to compare the fitted values for the historical periods and the 1-year-
ahead forecast.
The parameter values of all five methods are shown in Table 3, and their forecasting
results are listed in Table 4. The MAPE values of these five methods from 1998 to 2009 are
3.692% for the Bass diffusion model using the GA, 3.395% for the RGM1,1, 2.736% for the
combined model using the GA, 2.480% for the combined model using PSO, and 2.476% for the
combined model using a hybrid algorithm. For the 1-year-ahead forecast for 2010, the MAPE
values of the five methods were 10.96% for the Bass diffusion model using the GA, 15.94%
for the RGM1,1, 9.06% for the combined model using the GA, 4.71% for the combined
model using PSO, and 4.69% for the combined model using a hybrid algorithm. Our results
demonstrate that the proposed combined model using a hybrid algorithm outperformed
the other methods for the fitting and forecasting processes in terms of MAPE, as shown in
Figure 2.
Mathematical Problems in Engineering 7
140000
120000
100000
80000
Value
60000
40000
20000
0
1996 1998 2000 2002 2004 2006 2008 2010 2012
Year
5. Conclusions
This study presented a combined model that combined the Bass diffusion model with the
RGM1,1 to forecast motherboard shipments. In addition, we investigated evolutionary
optimization algorithms to determine the optimal parameters. The results indicate that the
combined model using a hybrid algorithm provides excellent MAPE improvement. We
conclude that the combined model, using a hybrid algorithm, is suitable for forecasting in
the motherboard industry.
Future research will include a modeling comparison of our model and the Support
Vector Machine SVM model using a DNA optimization method 26.
8
Acknowledgments
The authors gratefully acknowledge the referees of this paper who helped clarify and
improve this paper. They are also thankful for the financial support from NSC in Taiwan.
References
1 Market Intelligence Center MIC, Taiwan Motherboard Statistics Data, MIC, Taipei, Taiwan.
2 L. C. Hsu, “Applying the grey prediction model to the global integrated circuit industry,” Technological
Forecasting and Social Change, vol. 70, no. 6, pp. 563–574, 2003.
3 J. L. Deng, “Introduction to grey system theory,” Journal of Grey System, vol. 1, no. 1, pp. 1–24, 1989.
4 S. C. Chang, H. C. Lai, and H. C. Yu, “A variable P value rolling grey forecasting model for Taiwan
semiconductor industry production,” Technological Forecasting and Social Change, vol. 72, no. 5, pp.
623–640, 2005.
5 D. Akay and M. Atak, “Grey prediction with rolling mechanism for electricity demand forecasting of
Turkey,” Energy, vol. 32, no. 9, pp. 1670–1675, 2007.
6 L. C. Hsu and C. H. Wang, “Forecasting the output of integrated circuit industry using a grey model
improved by the Bayesian analysis,” Technological Forecasting and Social Change, vol. 74, no. 6, pp. 843–
853, 2007.
7 L. C. Hsu and C. H. Wang, “Forecasting integrated circuit output using multivariate grey model and
grey relational analysis,” Expert Systems with Applications, vol. 36, no. 2, pp. 1403–1409, 2009.
8 F. M. Bass, “Comments on: a new product growth for model consumer durables,” Management Science,
vol. 50, no. 12, pp. 1833–1840, 2004.
9 F. M. Tseng and Y. C. Hu, “Quadratic-interval Bass model for new product sales diffusion,” Expert
Systems with Applications, vol. 36, no. 4, pp. 8496–8502, 2009.
10 R. C. Tsaur, “Forecasting analysis by using fuzzy grey regression model for solving limited time series
data,” Soft Computing, vol. 12, no. 11, pp. 1105–1113, 2008.
11 F. S. Wu and W. L. Chu, “Diffusion models of mobile telephony,” Journal of Business Research, vol. 63,
no. 5, pp. 497–501, 2010.
12 Y. Y. Hsiao and F. K. Wang, “Forecasting analysis for global copper clad laminate market,” Journal of
Convergence Information Technology, vol. 7, no. 3, pp. 250–258, 2012.
13 F. M. Bass, “A new product growth for model consumer durables,” Management Science, vol. 50, no.
12, pp. 1825–1832, 2004.
14 D. C. Schmittlein and V. Mahajan, “Maximum likelihood estimation for an innovation diffusion model
of new product acceptance,” Marketing Science, vol. 1, pp. 57–78, 1982.
15 V. Srinivasan and C. H. Mason, “Nonlinear least squares estimation of new product diffusion
models,” Marketing Science, vol. 5, pp. 169–178, 1986.
16 R. Venkatesan and V. Kumar, “A genetic algorithms approach to growth phase forecasting of wireless
subscribers,” International Journal of Forecasting, vol. 18, no. 4, pp. 625–646, 2002.
17 C. D. Lewis, Industrial and Business Forecasting Methods, Butterworth Scientific, London, UK, 1982.
18 J. S. Armstrong, “Combining forecasts: the end of the beginning or the beginning of the end?”
International Journal of Forecasting, vol. 5, no. 4, pp. 585–588, 1989.
19 D. W. Bunn, “Combining forecasts,” European Journal of Operational Research, vol. 33, no. 3, pp. 223–229,
1988.
20 F. K. Wang and K. K. Chang, “Adaptive neuro-fuzzy inference system for combined forecasts in a
panel manufacturer,” Expert Systems with Applications, vol. 37, no. 12, pp. 8119–8126, 2010.
21 C. Christodoulos, C. Michalakelis, and D. Varoutas, “On the combination of exponential smoothing
and diffusion forecasts: an application to broadband diffusion in the OECD area,” Technological
Forecasting and Social Change, vol. 78, no. 1, pp. 163–170, 2011.
22 C. F. Juang, “A hybrid of genetic algorithm and particle swarm optimization for recurrent network
design,” IEEE Transactions on Systems, Man, and Cybernetics B, vol. 34, no. 2, pp. 997–1006, 2004.
23 S. Jeong, S. Obayashi, and Y. Minemura, “Application of hybrid evolutionary algorithms to low
exhaust emission diesel engine design,” Engineering Optimization, vol. 40, no. 1, pp. 1–16, 2008.
24 Evolver Software, Version 4.0.2, Palisade Corporation, Newfield, NY, USA, 2000.
10 Mathematical Problems in Engineering
25 M. Clerc and J. Kennedy, “The particle swarm-explosion, stability, and convergence in a multidimen-
sional complex space,” IEEE Transactions on Evolutionary Computation, vol. 6, no. 1, pp. 58–73, 2002.
26 X. H. Yang and Y. Q. Li, “DNA optimization threshold autoregressive prediction model and its
application in ice condition time series,” Mathematical Problems in Engineering, vol. 2012, pp. 1–10,
2012.
Hindawi Publishing Corporation
Mathematical Problems in Engineering
Volume 2012, Article ID 209329, 15 pages
doi:10.1155/2012/209329
Research Article
Adaptive Method for Solving Optimal Control
Problem with State and Control Variables
Copyright q 2012 L. Kahina and A. Mohamed. This is an open access article distributed under
the Creative Commons Attribution License, which permits unrestricted use, distribution, and
reproduction in any medium, provided the original work is properly cited.
The problem of optimal control with state and control variables is studied. The variables are: a
scalar vector x and the control ut; these variables are bonded, that is, the right-hand side of
the ordinary differential equation contains both state and control variables in a mixed form. For
solution of this problem, we used adaptive method and technology of linear programming.
1. Introduction
Problems of optimal control have been intensively investigated in the world literature for
over forty years. During this period, a series of fundamental results have been obtained,
among which should be noted the maximum principle 1 and dynamic programming 2, 3.
Results of the theory were taken up in various fields of science, engineering, and economics.
The optimal control problem with mixed variables and free terminal time is
considered. This problem is among the most difficult problems in the mathematical theory
of control processes 4–7. An algorithm based on the concept of simplex method 4, 5, 8, 9
so called support control is proposed to solve this problem.
The aim of the paper is to realize the adaptive method of linear programming 8. In
our opinion the numerical solution is impossible without using the computers of discrete
controls defined on the quantized axes as accessible controls. This made, it possible to
eliminate some analytical problems and reduce the optimal control problem to a linear
programming problem. The obtained results show that the adequate consideration of the
dynamic structure of the problem in question makes it possible to construct very fast
algorithms of their solution.
2 Mathematical Problems in Engineering
The work has the following structure. In Section 2, The terminal optimal control
problem with mixed variables is formulated. In Section 3, we give some definitions needed
in this paper. In Section 4, the definition of support is introduced. Primal and dual ways of
its dynamical identification are given. In Section 5, we calculate a value of suboptimality. In
Section 6, optimality and ε-optimality criteria are defined. In Section 7, there is a numerical
algorithm for solving the problem; the iteration consists in two procedures: change of control
and change of a support to find a solution of discrete problem; at the end, we used a final
procedure to find a solution in the class of piecewise continuous functions. In Section 8, the
results are illustrated with a numerical example.
2. Problem Statement
We consider linear optimal control problem with control and state constraints:
tf
Jx, u g x tf Cxt Dutdt −→ max, 2.1
0 x,u
subject to
tf
Jx, u c x ctutdt −→ max, 2.3
0 x,u
subject to
t∗
Ax htutdt, 0 ≤ t ≤ tf , 2.4
0
x0 x0 , 2.5
xmin ≤ xt ≤ xmax , umin ≤ ut ≤ umax , t ∈ T 0, tf , 2.6
3. Essentials Definitions
Definition 3.1. A pair v x, u· formed of an n-vector x and a piecewise continuous
function u· is called a generalized control.
Definition 3.2. The constraint 2.4 is assumed to be controllable, that is for any m-vector g,
there exists a pair v, for which the equality 2.4 is fulfilled.
A generalized control v x, u· is said to be an admissible control if it satisfies
constraints 2.4–2.6.
Definition 3.3. An admissible control v0 x0 , u0 · is said to be an optimal open-loop control
if a control criterion reaches its maximal value
J v0 maxJv. 3.1
v
Definition 3.4. For a given ε ≥ 0, an ε-optimal control vε xε , uε · is defined by the
inequality
J v0 − Jvε ≤ ε. 3.2
First, we describe a method of computing the solution of problem 2.3–2.6 in the class of
discrete control, and then we present the final procedure which uses this solution as an initial
approximation for solving problem 2.3–2.6 in the class of piecewise continuous functions.
Definitions of admissible, optimal, ε-optimal controls for discrete functions are given
in a standard form.
Choose an arbitrary subset TB ⊂ Th of l ≤ m elements and an arbitrary subset JB ⊂ J of
m-l elements.
Form the matrix,
PB aj A I, j , j ∈ JB ; dt, t ∈ TB , 4.2
t h
where dt t hsds, t ∈ Th .
A set SB {TB , JB } is said to be a support of problem 2.3–2.6 if det PB /
0.
A pair {v, SB } of an admissible control vt x, ut, t ∈ T and a support SB is said
to be a support control.
A support control {v, SB } is said to be primally nonsingular if d∗j < xj < dj∗ , j ∈ JB ; f∗ <
ut < f ∗ , t ∈ TB .
4 Mathematical Problems in Engineering
tf
4.3
ΔJv Jv − Jv c Δx ctΔutdt.
0
Since
z
AΔx htΔutdt 0, 4.4
0
tf
ΔJv c − ν A Δx ct − ν ht Δutdt, 4.5
0
and is equal to
β βv, SB
Δj xj − xminj Δj xj − xmaxj Δtut − umin Δtut − umax ,
j∈JH − t∈T t∈T −
j∈JH
5.3
where
The number βv, SB is called a value of suboptimality of the support control {v, SB }.
From there, Jv − Jv ≤ βv, SB . Of this last inequality, the following result is deduced.
xj xmaxj if Δj < 0,
xminj ≤ xj ≤ xmaxj if Δj 0, j ∈ J,
are sufficient, and in the case of non degeneracy, they are necessary for the optimality of control v.
Theorem 6.2. For any ε ≥ 0, the admissible control v is ε-optimal if and only if there exists a support
SB such that βv, SB ≤ ε.
The iteration of the method is a change of an “old” control {v, SB } for the “new” one
{v, SB } so that β{v, SB } ≤ β{v, SB }. The iteration consists of two procedures:
1 change of an admissible control v → v,
2 change of support SB → SB .
Construction of the initial support control concerns with the first phase of the method and
can be performed with the use of the algorithm described below.
At the beginning of each iteration the following information is stored:
1 an admissible control v,
2 a support SB {TB , JB },
3 a value of suboptimality β βv, SB .
Before the beginning of the iteration, we make sure that a support control {v, SB } does not
satisfy the criterion of ε-optimality.
xj xj θ0 lj , j ∈ J,
7.1
ut ut θ0 lt, t ∈ Th ,
Support values of a pseudocontrol {xj , j ∈ JB ; ut, t ∈ TB } are computed from the equation
A I, j xj dtut g − A I, j xj dtut. 7.3
j∈JB t∈TB j∈JH t∈TH
that is,
xmin ≤ xj θ0 lj ≤ xmax , j ∈ J,
7.5
umin ≤ ut θ0 lt ≤ umax , t ∈ Th .
⎧
⎪ xmaxj − xj
⎪
⎪ if lj > 0,
⎪
⎪ lj
⎨
θj xminj − xj if lj < 0, 7.6
⎪
⎪
⎪
⎪ lj
⎪
⎩ ∞ if lj 0, j ∈ JB ,
⎧
⎪ umax − ut
⎪
⎪ if lt > 0,
⎪
⎨ lt
θt umin − ut 7.7
⎪
⎪ if lt < 0,
⎪
⎪ lt
⎩
∞ if lt 0, t ∈ TB .
Let us calculate the value of suboptimality of the support control {v, SB } with v computed
according to 7.1: βv, SB 1 − θ0 βv, SB . Consequently,
a θ0 θj0 , j0 ∈ JB ,
b θ0 θt0 , t0 ∈ TB .
where δj , j ∈ J, δt, t ∈ Th is an admissible direction of change Δ, Δ· and σ 0 is a maximal
step along this direction.
δt 0 if t ∈ TB ,
δj 0 if j
/ j0 , j ∈ JB ,
7.9
δj 0 1 if xj0 xminj0 ,
δj0 −1 if xj0 xmaxj0 ,
δj 0 if j ∈ JB ,
TB
δt 0 if t ∈ ,
t0 7.10
δt0 1 if ut0 umin ,
δt0 −1 if ut0 umax .
δj Δν A I, j , j ∈ JH ,
7.11
δt Δν dt, t ∈ TH .
Mathematical Problems in Engineering 9
TB
TB ∪ {t1 }, J B JB , 7.13
{t0 }
TB
TB , J B JB ∪ j1 , 7.14
{t0 }
JB
T B TB , JB ∪ j1 , 7.15
j0
JB
T B TB ∪ {t1 }, JB , 7.16
j0
β v, SB 1 − θ0 βv, SB − ασ 0 , 7.17
10 Mathematical Problems in Engineering
where
|ut0 − ut0 | if θ0 θt0 ,
α 7.18
xj − xj if θ0 θj0 .
0 0
1 If βv, SB > ε, then we perform the next iteration starting from the support control
{v, SB }.
2 If βv, SB 0, then the control v is optimal for problem 2.3–2.6 in the class of
discrete controls.
3 If βv, SB < ε, then the control v is ε-optimal for problem 2.3–2.6 in the class of
discrete controls.
If we would like to get the solution of problem 2.3–2.6 in the class of piecewise continuous
control, we pass to the final procedure when case 2 or 3 takes place.
⎧ 7.19
⎪
⎪ if Δt < 0
⎨umin ,
t umax ,
u if Δt > 0,
⎪
⎪
⎩∈ u , u , if Δt 0, t ∈ T .
min max h
If
tf
AI, Jx utdt g,
ht 7.20
0
tf
AI, Jx utdt
ht / g, 7.21
0
then denote T 0 {ti ∈ T, Δti 0}, where ti are zeros of the optimal cocontrol, that is,
Δti 0, i 1, s, with s ≤ m. Suppose that
Δ̇ti /
0, i 1, s. 7.22
Mathematical Problems in Engineering 11
7.23
where
Θ0 t0i , i 1, s; xj0 , j ∈ JB 7.25
fΘ 0. 7.26
We solve this system by the Newton method using as an initial approximation of the vector
Θ0 ti , i 1, s; xj , j ∈ JB . 7.27
∂f −1 Θk
Θk 1
Θ k
ΔΘ k
, ΔΘ k
− · f Θk . 7.28
∂Θk
∂f Θk k k
AI, JB ; umin − umax sign Δ̇ ti h ti , i 1, s 7.29
∂Θk
∂f Θ0
det 0.
/ 7.30
∂Θ0
12 Mathematical Problems in Engineering
For instants t ∈ TB , there exists a small μ > 0 that for any ti ∈ ti − μ, ti μ, i 1, s,
the matrix hti , i 1, s is nonsingular and the matrix ∂fΘk /∂Θk is also nonsingular if
k
elements ti , i 1, s, k 1, 2, . . . do not leave the μ-vicinity of ti , i 1, s.
∗
Vector Θk is taken as a solution of 4.6 if
∗
f Θk ≤ η, 7.31
∗
for a given η > 0. So we put θ0 θk .
The suboptimal control for problem 2.3–2.6 is computed as
x0j , j ∈ JB ,
xj0
xj , j ∈ JH
7.32
umax umin umax − umin
u0 t − sign Δ̇ t0i , t ∈ t0i , t0i 1 , i 1, s.
2 2
If the Newton method does not converge, we decrease the parameter h > 0 and perform the
iterative process again.
8. Example
We illustrate the results obtained in this paper using the following example:
25
utdt −→ min,
0
ẋ1 x3 ,
ẋ2 x4 ,
1.5
0.5
0
−0.5
−1
−1.5
−2
−2.5
0 5 10 15 20 25
“σ · dat”
“u · dat”
ψ1 −ψ3 0.1ψ4 ,
ψ2 ψ3 − 1.01ψ4 ,
ψ3 ψ1 , 8.3
ψ 4 ψ2 ,
ψ1 tf 0, ψ2 tf 0, ψ3 tf 0, ψ4 tf 0.
Problem 8.1 is reduced to canonical form 2.3–2.6 by introducing the new variable
ẋ5 u, x5 0 0. Then, the control criterion takes the form −x5 tf → max. In the class of
discrete controls with quantization period h 25/1000 0.0025, problem 8.1 is equivalent
to LP problem of dimension 4 × 1000.
To construct the optimal open-loop control of problem 8.1, as an initial support, a set
TB {5, 10, 15, 20} was selected. This support corresponds to the set of nonsupport zeroes of
the cocontrol Tn0 {2.956, 5.4863, 9.55148, 12.205, 17.6190, 19.0372}. The problem was solved
in 26 iterations, that is, to construct the optimal open-loop control, a support 4 × 4 matrix was
14 Mathematical Problems in Engineering
1.5
0.5
x3 −0.5
−1
−1.5
−2
−2.5
−2 −1 0 1 2 3
x1
2.5
1.5
0.5
0
−0.5
−1
−1.5
−2
0 5 10 15 20 25
“x1 · dat”
“x2 · dat”
Table 1
changed 26 times. The optimal value of the control criterion was found to be equal to 6.602054
in time 2.92.
Table 1 contains some information on the solution of problem 8.1 for other
quantization periods.
Mathematical Problems in Engineering 15
1.5
0.5
−0.5
−1
−1.5
−2
−2.5
0 5 10 15 20 25
“x3 · dat”
“x4 · dat”
Of course, one can solve problem 8.1 by LP methods, transforming the problem
4.6–7.8. In doing so, one integration of the system is sufficient to form the matrix of the
LP problem. However, such “static” approach is concerned with a large volume of required
operative memory, and it is fundamentally different from the traditional “dynamical”
approaches based on dynamical models 2.3–2.6. Then, problem 2.3–2.6 was solved.
In Figure 1, there are control ut and switching function for minimum principle. In
Figure 2, there is phaseportrait x1 , x3 for a system 8.1. In Figure 3, there are state variables
x1 t, x2 t for a system 8.1. In Figure 3, state variables x3 t, x4 t for a system 8.1. In
Figure 4, state variables x1 t, x2 t for a system 8.1.
References
1 L. S. Pontryagin, V. G. Boltyanski, R. Gamkrelidze, and E. F. Mischenko, The Mathematical Theory of
Optimal Processes, Interscience Publishers, New York, NY, USA, 1962.
2 R. E. Bellman, Dynamic Programming, Princeton University Press, Princeton, NJ, USA, 1963.
3 R. E. Bellman, I. Glicksberg, and O. A. Gross, Some Aspects of the Mathematical Theory of Control Processes,
Rand Corporation, Santa Monica, Calif, USA, 1958, Report R-313.
4 N. V. Balashevich, R. Gabasov, and F. M. Kirillova, “Calculating an optimal program and an optimal
control in a linear problem with a state constraint,” Computational Mathematics and Mathematical Physics,
vol. 45, no. 12, pp. 2030–2048, 2005.
5 R. Gabasov, F. M. Kirillova, and N. V. Balashevich, “On the synthesis problem for optimal control
systems,” SIAM Journal on Control and Optimization, vol. 39, no. 4, pp. 1008–1042, 2000.
6 R. Gabasov, N. M. Dmitruk, and F. M. Kirillova, “Decentralized optimal control of a group of dynamical
objects,” Computational Mathematics and Mathematical Physics, vol. 48, no. 4, pp. 561–576, 2008.
7 N. V. Balashevich, R. Gabasov, and F. M. Kirillova, “Numerical methods of program and positional
optimization of the linear control systems,” Zh Vychisl Mat Mat Fiz, vol. 40, no. 6, pp. 838–859, 2000.
8 R. Gabasov, Adaptive Method of Linear Programming, University of Karsruhe, Institute of Statistics and
Mathematics, Karsruhe, Germany, 1993.
9 L. Kahina and A. Mohamed, “Optimization of a problem of optimal control with free initial state,”
Applied Mathematical Sciences, vol. 4, no. 5–8, pp. 201–216, 2010.
Hindawi Publishing Corporation
Mathematical Problems in Engineering
Volume 2012, Article ID 971303, 14 pages
doi:10.1155/2012/971303
Research Article
A Hybrid Network Model to Extract Key Criteria
and Its Application for Brand Equity Evaluation
Copyright q 2012 C.-Y. Chen and C.-W. Li. This is an open access article distributed under the
Creative Commons Attribution License, which permits unrestricted use, distribution, and
reproduction in any medium, provided the original work is properly cited.
Making a decision implies that there are alternative choices to be considered, and a major challenge
of decision-making is to identify the adequate criteria for program planning or problem evaluation.
The decision-makers’ criteria consists of the characteristics or requirements each alternative must
possess and the alternatives are rated on how well they possess each criterion. We often use criteria
developed and used by different researchers and institutions, and these criteria have similar means
and can be substituted for one another. Choosing from existing criteria offers a practical method to
engineers hoping to derive a set of criteria for evaluating objects or programs. We have developed a
hybrid model for extracting evaluation criteria which considers substitutions between the criteria.
The model is developed based on Social Network Analysis and Maximum Mean De-Entropy
algorithms. In this paper, the introduced methodology will also be applied to analyze the criteria
for assessing brand equity as an application example. The proposed model demonstrates that it is
useful in planning feasibility criteria and has applications in other evaluation-planning purposes.
1. Introduction
System engineering is an interdisciplinary field of engineering focusing on how complex
engineering projects should be designed and managed over their life cycles, and it overlaps
with both technical and human-centered disciplines such as control engineering, industrial
engineering, organizational studies, and project management 1. The principles of system
engineering provided system thinking and have been applied to most projects and industry
fields. In the system life-cycle process, as Figure 1 illustrates, system assessment-analysis and
evaluation have to be undertaken in distinct phases.
When we want to execute an evaluation project, determining adequate criteria is
critical to achieve the evaluation. Traditionally, criteria are derived through discussion with
engineers, experts, researchers and, especially in the field of business administration, by
2 Mathematical Problems in Engineering
finding out about an object through consultation with customers. In this process, deciding
upon consistent evaluation criteria is time-consuming work. In many situations when the
assessing objects are similar, existing criteria could be referred to. If the criteria which
can be referred to are numerous, it is necessary to choose the most appropriate basis of
the benchmark assessment. After an in-depth literature review about evaluation criteria, a
researcher established an adequacy criteria algorithm that reasonably satisfied the request of
“face validity” or “content validity.”
As an example application, brand equity was chosen as it is one of the important
intangible assets that can bring competitive advantage for many enterprises. However, until
now, measuring the value of a brand has been relatively abstractive and subjective and there
are many different methods to explore the potential value of the brand in academia and
industry. Many criteria have similar meanings and can be partially or totally substituted for
one another. Based on the assumption that “a criterion can be substituted partially or totally
by another criterion”, in this paper, we develop a hybrid model for choosing the adequate
evaluation criteria. By using the methods of the Social Network Analysis SNA and the
Maximum Mean De-Entropy MMDE algorithm, the degree of substitutability of existing
reference criteria will be judged to derive a criteria list for evaluation. In this paper, the
issue of the assessment of brand equity will be addressed as an example to demonstrate
the application of the proposed research model in the planning of other evaluation
projects.
The rest of this paper is organized as follows: Section 2 describes the issues associated
with choosing the criteria, the theories of SNA and MMDE methods, and explains how
the model for this study was constructed. We also use an example, choosing the criteria to
evaluate the value of brand equity, to illustrate the steps of our model and the applications
of the model are discussed in Section 3. Finally, in Section 4, we will discuss the advantages
of the proposed model and the feasibility of its application. We draw conclusions and offer
some discussion related to future work in Section 5.
the assessment of intangible assets such as technology, patents, or brand equity or concerning
the characteristics of invisibility and abstractness, and as such, those characteristics also have
a great influence on the assessment framework itself.
There are many diverse models for brand equity available in the academic world and
in practice. These two sectors have already proposed many different assessment frameworks
which constitute various dimensions and criteria, and various analytical perspectives
accompanied by various dimensions and criteria in the assessment framework. However,
aside from the complexity of the implementation of the assessment process in reality,
the naming or definitions of dimensions and criteria are sometimes so similar making
distinguishing them with certainty very difficult. Moreover, the existence of such ambiguity
in the framework may bring about the issue of double counting resulting in bias of the
assessment result. Therefore, it can be seen that developing a series of processes to arrange
and select proper dimensions and criteria within the established framework to cover the
full meaning of the evaluation structure and to prevent measurement bias will be crucial
to building an appropriate evaluation model.
Hence, we determine the full dimensions to be applied to well-known brand equity
assessment models and use the research methodology addressed in this paper to extract
appropriate dimensions, to propose a reasonable and practical assessment framework of
brand equity and hope to provide a reference for establishing relative models in the
future.
A B A B A B
C C C
Figure 2
A and B have a relationship, between them but C is an isolated point with no relationship
between C and the other two. If the relationship is not directed, then the relation is symmetric.
In Figure 2b, there is a relationship between A and B, and a relationship between B and C,
but there is no relationship between A and C. If the relationship is not directed, the attribute
of transitive is not suitable for analysis. A directed relation, such as Figure 2c, will have the
attribute of transitive and asymmetric.
A walk is an alternating sequence of incident nodes and lines. A walk begins and ends
with nodes. The length of a walk is the number of occurrences of lines in it. Because some
nodes and some lines may be included more than once, the definition of a path as a walk
in which all nodes and all lines are distinct is necessary for our paper. If a node x can walk
to another node y, we can say that node x and node y are connected. It is likely that there
are several different connections between a given pair of nodes and that these connections
differ in length. A connection with the shortest distance between two nodes is referred to as
a geodesic, or the distance between two nodes.
“Centrality” is a concept which is used in social network analysis to identify the
“importance” of nodes in a social network. There are some indices, such as “degree,”
“closeness,” and “betweenness,” which have been commonly used as the indices of centrality
11. The definitions of these indices attempt to quantify the importance of an individual
actor/node embedded in a network. Now, many indices have been established for identifying
the important of a node. In this paper, we use the index “Bonacich Power Centrality” as the
centrality scores of network nodes.
The original degree centrality approach argues that actors who have more connections
are more likely to be powerful because they can directly affect a greater number of other
actors. Bonacich proposed a modification of the degree centrality approach to include the
concept that the same degree does not necessarily make actors equally important 12, 13.
A node is likely to be more influential if it is connected to another central node which
can quickly reach a lot of other nodes with its message. But if the nodes that you are
connected to are well connected, they are not dependent on you and that means you are
not so important although there are a lot nodes that in fact connect with you. Connected
to others makes a node central, but not powerful. One node being connected to others that
are not well connected makes one powerful, because these other actors are dependent on
you.
Bonacich proposed that both centrality and power were a function of the connections
of the nodes in one’s neighborhood. The more connections the nodes in your neighborhood
have, the more central you are. The fewer the connections the actors in your neighborhood
have, the more powerful you are. Bonacich’s power centrality measure, or BP score, is
defined as 2.1 where A is the graph adjacency matrix, 1 is column vector of 1. The β is the
Mathematical Problems in Engineering 5
“attenuation parameter,” and the α is a scaling parameter which is set as the sum of squared
scores equal to the number of nodes 14:
−1
CBP v α I − βA A 1. 2.1
For any criterion with influence from node x to node y through node z, the influence
from x to node z includes direct and indirect influence recursively. The BP score is the
notion that the power of a node is recursively defined by the sum of the power of its
alter. The nature of the recursion involved is then controlled by the power exponent:
positive values imply that vertices become more powerful as their alters become more
powerful as occurs in cooperative relations, while negative values imply that nodes
become more powerful only as their alters become weaker as occurs in competitive or
antagonistic relations. The magnitude of the exponent indicates the tendency of the effect
to decay across long walks; higher magnitudes imply slower decay. The BP score is suitable
for this paper because of the purpose of finding out the degree of substitution between
criteria.
Definition 2.1. Let a random variable with n elements denoted as X {x1 , x2 , . . . , xn }, with
a corresponding probability P {p1 , p2 , . . . , pn }, then one defines the entropy, H, of X as
follows:
H p1 , p2 , . . . , pn − pi lg pi , 2.2
subject to constraints
n
pi 1,
i1 2.3
pi lg pi 0 if pi 0.
The function “lg” means the logarithms which are taken to an arbitrary but fixed base.
The value of Hp1 , p2 , . . . , pn is the largest when p1 p2 · · · pn , and we denote this
largest entropy value as H1/n, 1/n, . . . , 1/n. Now we will define another measure for the
decreased level of entropy: de-entropy.
6 Mathematical Problems in Engineering
Definition 2.2. For a given finite discrete scheme of X, the de-entropy of X is denoted as HnD
and defined as
1 1 1
HnD H , ,..., − H p1 , p2 , . . . , pn . 2.4
n n n
By Definition 2.2, the value of H D is equal to or larger than 0. Unlike entropy, which
is used for the measure of uncertainty, the HnD can explain the amount of useful information
derived from a specific dataset, which reduces the “uncertainty” of the information. We define
the de-entropy for the purpose of searching the threshold value in order to assess the effect
of information content when adding a new node to an existing impact-relations map. By
Definition 2.1, Formula 2.5 can be proven the proof can be found in 15:
1 1 1 1 1 1
Hn H , ,..., ≤H , ,..., Hn 1 . 2.5
n n n n 1 n 1 n 1
Based on Definitions 2.1 and 2.2, the MMDE algorithm is developed to obtain a
threshold value for delineating a network diagram. This algorithm can be used to derive
a set of dispatch-nodes, the factors that strongly dispatch influences to others, and a set of
receive-nodes, which are easily influenced by another factor. According to these two sets and
a unique threshold value, we can obtain an influence network of criteria.
We propose the maximum mean de-entropy MMDE algorithm to find a threshold
value for delineating the impact relations between criteria. The MMDE algorithm has some
properties that differ from the traditional method of determining a network. First, the MMDE
algorithm serves mainly to decide the “node” rather than the “network,” and this is helpful
for understanding a problem in that it decreases the uncertainty of information. Second,
the MMDE algorithm considers the properties of the dispatch influence and the received
influence of a node, and this is useful to the analyst in determining the “nodes” or criteria,
which are easily influenced by other factors. The MMDE algorithm can also obtain a unique
threshold value, which is helpful in solving the problems that a researcher may confront
regarding the selection of a consistent threshold value and is decided by searching a suitable
criteria set. The theory and steps of this algorithm are described In Li and Tzeng 2009
16, 17, and we summarize the steps of the MMDE as follows.
Step 2. Taking the second element from the ordered triplets set to establish a dispatch-node
set.
Step 5. Similar to Steps 2 to 4, an ordered receive-node set and a maximum mean de-entropy
receive-node set can be derived.
Can the
node be replaced by The criteria meet the
Is the node a another one based
Yes No objectives of the proposed
component of the powerful on the result from hybrid model
set? MMDE algorithm?
the potential value of brand, both from academic or practical viewpoints. Generally, from
an analytical point of view, those methods can be divided into financial, marketing, and a
combination of perspectives.
Briefly, the financial perspective applies the real data from the financial statements of
the company to evaluate the brand equity; the marketing perspective tends to extend the
meaningfulness of a brand to customers, that is, consumers’ feelings toward a brand and
its potential influences on a company’s profit, is usually divided into several dimensions
to explore a brand’s potential value to a company; the combination perspective integrates
the two formers’ concepts and is viewed both from financial and marketing perspectives
which are important but to a different extent. The two methods should be considered
simultaneously in the measurement process.
In practice, there are many different models to evaluate brand equity, the most famous
and most reliable models named Interbrand, BBDO, and HIORSE. The model developed
by Interbrand, a British global branding consultancy company, covers market segmentation,
financial analysis, demand, and competition analysis and takes both marketing and financial
perspectives into consideration. It can be viewed as a conscientious measurement structure
of brand equity, but the calculating process of the model has not been completely disclosed,
preventing its widespread adoption 19.
The BBDO model, developed by a worldwide advertising agency belonging to
the Omnicom group, puts market quality, dominance of relevant market, international
orientation, brand status, and monetary basis into the measurement process. The data for the
model are relatively easy to collect, but the calculation is comparatively complex, especially
the repetition of meanings between dimensions, which could result in the possibility of
double counting 20, 21, since these three models all reflect the marketing and financial
aspects that previous researchers consider important when it comes to evaluating the
brand equity. Moreover, these three models are widely applied in practice and accepted
by many stock exchanges in the U.S. and Europe. Therefore, considering the theoretical
integrity and representativeness, we chose these three models as the empirical cases to
analyze.
HIROSE is constructed by the Japanese Government to reflect the development
background and put more emphasis on industrial competiveness 22. Therefore, this model
includes the three drivers of prestige, loyalty, and expansion and is inclined to include
Mathematical Problems in Engineering 9
Table 1: Fifteen criteria used for assessing the value of brand equity.
financial statements and can be regarded as a reliable value measurement index. Even so,
focus on the financial statements also limits its application scope in practice.
To conduct our research, a list of 15 criteria was created after comparison of these three
models, shown as Table 1. After we interviewed senior technical personnel and marketing
managers to determine the degree of substitution between each other, we can obtain the
matrix as shown in Table 2. In Table 2, aij is the element located at row i and column j. If aij
is 0, it means that the criterion i has no interrelationship with criterion j. If aij is 1, it means
that the criterion i can almost totally substitute/replace criterion j. If aij is 2, it means that the
criterion i can partially substitute/replace criterion j. The values 0, 1, and 2 are ordinal scale
by the means of distance.
Our hypothesis was analyzed using the SNA method. With the directed line that
implies the degree of substitution from one node to another, we divided the support measures
into a group that dispatches influence and a second group that receives influence so that
we could understand the influence relationships better. The purposes of the SNA enquiry
in this research, with the expert’s knowledge for contributing to a deeper comprehension
of the criteria, are the analysis of the structure and interrelationships of the criteria and the
identification of the key feasible and efficient criteria for evaluating the value of brand equity.
x3 x12 x13
x1
x14 x10
x9 x15
x2 x6 x5
4. Discussion
In this section, we will discuss the advantages of the proposed model and its application in
the brand equity evaluation.
x11
x1
x14
x9
x2
Substitution network
Negative Bonacich power score criteria Negative Bonacich power score criteria
x2 x6 x5 x2 x6 x5
Figure 6: Steps to extract nine criteria from the original fifteen criteria.
The MMDE algorithm was used to set an appropriate threshold value and obtain
adequate information to delineate the substitution network for further analysis. The results
of the MMDE algorithm can be tracked and evaluated easily because of the obvious casual
relationship. It is useful for a researcher if some specific criteria are considered as necessary
criteria for evaluation. The researcher can easily find out which criteria should also be
included in evaluating if the specific criterion has to be included.
After we obtained the results from the SNA and MMDE algorithms, we can exclude
the weaker criteria and replaceable criteria and obtain the criteria which meet the objectives
of the proposed hybrid model. In this paper, we demonstrated the effect and feasibility of
the hybrid model, the first model to reduce criteria and consider the power, centrality, and
substitution of criteria, for extracting key criteria.
market, psychographic and identity status of BBDO, and also overlap with the loyalty driver
of HIROSE. Brand trend measures the long-term performance of market share, expected
brand performance, the sensitivity of the brand planning, and competitive action. Those four
criteria are relative to brand leadership; market and marketing support those dimensions of
Interbrand; therefore, both brand trend and marketing support are regarded as comparative
weak and suggest to be ruled out of the assessment framework.
In the second analytical phase, we found that brand leadership, dominance of
relevance market, and loyalty drivers were the three dimensions that can be substituted
by brand status of BBDO. If we look carefully at the breakdown of BBDO in brand status,
it contains two indicators of brand strength and brand appeal; five different levels of the
dimension were constructed to measure the functional, market, psychographic, identify, and
legendary status of the brand. However, brand leadership of Interbrand mainly measures
market share, visibility, market position, and competitors outline which are very similar to
the market status of BBDO. Then the dimension dominance of relevance market from BBDO
mainly investigates the size of the brand revenue relative to the market leader which can
also be included in brand strength and brand appeal of the market status of the brand status.
Finally, the loyalty driver of HIROSE is also highly related to brand trust and brand loyalty
of the identity status of the brand status.
According to data analysis, the original 15 dimensions are extracted into the final 9
dimensions using appropriate research methods to establish an assessment framework which
can integrate the complete concept from various models of brand equity. From the analysis
above we can conclude that the brand status of BBDO has relatively strong substitution
capability since this dimension is more completely compared to other dimensions among
different models since it can both cover the brand leadership of Interbrand and dominance of
relevance market of BBDO. Therefore, it can be seen that, with the exception of dimensions
between the different models, there are possibilities for mutual substitution and that the
proposed model can efficiently extract the key criteria required.
5. Conclusions
In this paper, we used social network analysis and the maximum mean de-entropy algorithm
to extract key criteria from numerous existing criteria. We also applied the proposed model
to evaluating brand equity. The effect and feasibility of the hybrid model were demonstrated.
However, we were able to have discussions with only a few experts and applied the model
only to evaluating brand equity. In a follow-up study, we recommend that a further criteria
extraction projects be executed using the proposed hybrid model to demonstrate increased
feasibility of our model and allow the provision of more reliable conclusions.
References
1 B. S. Blanchard and W. J. Fabrycky, Systems Engineering and Analysis, Prentice Hall, New Jersey, NJ,
USA, 4th edition, 2005.
2 Wikipedia, Social Network Analysis, http://en.wikipedia.org/, 2012.
3 S. Wasserman and K. Faust, Social Network Analysis: Methods and Applications, Cambridge University
Press, New York, NY, USA, 1994.
4 M. Balconi, S. Breschi, and F. Lissoni, “Networks of inventors and the role of academia: an exploration
of Italian patent data,” Research Policy, vol. 33, no. 1, pp. 127–145, 2004.
5 M. Balconi and A. Laboranti, “University-industry interactions in applied research: the case of
microelectronics,” Research Policy, vol. 35, no. 10, pp. 1616–1630, 2006.
14 Mathematical Problems in Engineering
6 F. Murray, “Innovation as co-evolution of scientific and technological networks: exploring tissue
engineering,” Research Policy, vol. 31, no. 8-9, pp. 1389–1403, 2002.
7 C. S. Wagner and L. Leydesdorff, “Network structure, self-organization, and the growth of
international collaboration in science,” Research Policy, vol. 34, no. 10, pp. 1608–1618, 2005.
8 L. C. Yin, H. Kretschmer, R. A. Hanneman, and Z. Y. Liu, “Connection and stratification in research
collaboration: an analysis of the COLLNET network,” Information Processing and Management, vol. 42,
no. 6, pp. 1599–1613, 2006.
9 J. P. Scott, Social Network Analysis: A Handbook, Sage Publications, London, UK, 2nd edition, 2000.
10 D. Knoke and S. Yang, Social Network Analysis, Sage Publications, London, UK, 2nd edition, 2007.
11 L. C. Freeman, “Centrality in social networks conceptual clarification,” Social Networks, vol. 1, no. 1,
pp. 215–239, 1978.
12 P. Bonacich, “Factoring and weighting approaches to status scores and clique identification,” The
Journal of Mathematical Sociology, vol. 2, no. 1, pp. 113–120, 1972.
13 P. Bonacich, “Power and centrality: a family of measures,” American Journal of Sociology, vol. 92, no. 5,
pp. 1170–1182, 1987.
14 C. T. Butts and K. M. Carley, “Some simple algorithms for structural comparison,” Computational and
Mathematical Organization Theory, vol. 11, no. 4, pp. 291–305, 2005.
15 I. K. Aleksandr, Mathematical Foundations of Information Theory, Courier Dover, New York, NY, USA,
1957.
16 C. W. Li and G. H. Tzeng, “Identification of a threshold value for the DEMATEL method using
the maximum mean de-entropy algorithm to find critical services provided by a semiconductor
intellectual property mall,” Expert Systems with Applications, vol. 36, no. 6, pp. 9891–9898, 2009.
17 C. W. Li and G. H. Tzeng, “Identification of interrelationship of key customers’ needs based on
structural model for services/capabilities provided by a semiconductor-intellectual-property mall,”
Applied Mathematics and Computation, vol. 215, no. 6, pp. 2001–2010, 2009.
18 P. Bonacich and T. M. Liggett, “Asymptotics of a matrix valued Markov chain arising in sociology,”
Stochastic Processes and their Applications, vol. 104, no. 1, pp. 155–171, 2003.
19 N. Penrose, “Valuation of brand names and trade marks,” in Brand Valuation: Establishing a True and
Fair View, pp. 32–45, 1989.
20 R. Zimmermann, U. Klein-Bolting, B. Sander, and T. Murad-Aga, Brand Equity Review, BBDO Group,
Berlin, Germany, 2001.
21 R. Zimmermann, Brand Equity Excellence, BBDO Group, Berlin, Germany, 2002.
22 Department of Trade and Industry, “Report of brand valuation,” in Proceedings of the Seminar of
Corporate Legal System in the Ministry of Economy, Tokyo, Japan, 2002.
23 C. T. Butts, “Tools for social network analysis,” http://cran.r-project.org/, 2010.
24 C. T. Butts, “Software manual for the R sna package,” http://erzuli.ss.uci.edu/R.stuff/, 2010.
Hindawi Publishing Corporation
Mathematical Problems in Engineering
Volume 2012, Article ID 931092, 13 pages
doi:10.1155/2012/931092
Research Article
Solving Packing Problems by a Distributed Global
Optimization Algorithm
Copyright q 2012 Nian-Ze Hu et al. This is an open access article distributed under the Creative
Commons Attribution License, which permits unrestricted use, distribution, and reproduction in
any medium, provided the original work is properly cited.
Packing optimization problems aim to seek the best way of placing a given set of rectangular boxes
within a minimum volume rectangular box. Current packing optimization methods either find
it difficult to obtain an optimal solution or require too many extra 0-1 variables in the solution
process. This study develops a novel method to convert the nonlinear objective function in a
packing program into an increasing function with single variable and two fixed parameters. The
original packing program then becomes a linear program promising to obtain a global optimum.
Such a linear program is decomposed into several subproblems by specifying various parameter
values, which is solvable simultaneously by a distributed computation algorithm. A reference solu-
tion obtained by applying a genetic algorithm is used as an upper bound of the optimal solution,
used to reduce the entire search region.
1. Introduction
A packing optimization problem is to seek a minimal container, which can hold a given
number of smaller rectangular boxes. This problem is also referred to as a container load-
ing problem. Packing cartons into a container is concerning material handling in the manu-
facturing and distribution industries. For instance, workers in the harbor have to pack more
than one type of cartons into a container, and they often deal with this problem by the rule
of thumb but a systematic approach. Therefore, the utilization of the container is low , which
will cause additional costs.
2 Mathematical Problems in Engineering
Similar issues can be found in fields such as knapsack 1, 2, cutting stock 3, 4,
assortment problems 5, 6, rectangular packing 7, pallet loading 8, 9, and container
loading problems 10, 11. In addition, researchers have dealt with various related problems.
For instance, Dowsland 12 and Egeblad and Pisinger 1, He et al. 13, Wu et al. 14,
de Almeida and Figueiredo 15, Miyazawa and Wakabayashi 16, and Crainic et al. 17
proposed different heuristic methods for solving three-dimensional packing problems, Chen
et al. 10 formulated a mixed integer program for container loading problems, and Li
and Chang 6 developed a method for finding the approximate global optimum of the
assortment problem. However, Li and Chang’s method 6 requires using numerous 0-1
variables to linearize the polynomial objective function in their model, which would involve
heavy computation in solving packing problems. Moreover, Chen et al.’s approach 10 can
only find a local optimum of packing problems with nonlinear objective function. Recently,
many global optimization methods have been developed, where Floudas’ method 18 is one
of most promising methods for solving general optimization problems. Although Floudas’
method 18 can be applied to solve a packing problem to reach finite ε-convergence to
the global minimum, it requires successively decomposing the concave part of the original
problem into linear functions. Besides, it adopts lower boundary techniques, which is
time consuming. Li et al. 19 designed a distributed optimization method to improve the
computational efficiency for solving packing problems. Tsai and Li 20 also presented an
enhanced model with fewer binary variables and used piecewise linearization techniques
to transform the nonconvex packing problem into a mixed-integer linear problem, which is
solvable to find a global solution.
Due to the complexity and hardness of three-dimensional packing problems, most
results on this topic are based on heuristics 21. Furthermore, parallel computing 22 was
adopted to improve the efficiency of combinatorial computation. Parallel Genetic Algorithm
GA 23, parallel with heuristic 24, and parallel Tabu Search Algorithm TSA 25 were
proposed to solve container-packing problems under some conditions. These methods are
capable of obtaining solutions with good performance relative to test examples in the litera-
ture. However, the algorithms cannot guarantee to get a global optimum.
This paper proposes another method for finding the optimum of the packing problem.
The major advantage of this method is that it can reformulate the nonlinear objective function
of original packing problem as an increasing function with single variable and two given
parameters. In addition, distributed computation and genetic algorithm are adopted to
improve the efficiency and ensure the optimality. The proposed method then solves the
reformulated programs by specifying the parameters sequentially to reach the globally opti-
mal solution on a group of network-connected computers.
2. Problem Formulation
Given n rectangular boxes with fixed lengths, widths, and heights, a packing optimization
problem is to allocate these n boxes within a rectangular container having minimal volume.
Denote x, y, and z as the width, length, and height of the container; the packing optimization
problem discussed here is stated as follows:
minimize xyz
3 x ≤ x ≤ x, y ≤ y ≤ y, z ≤ z ≤ z
x, y, z, x, y, z are constants .
2.1
According to Chen et al. 10, the current packing model adopts the terminologies as
follows.
pi , qi , ri : Dimension of box i, pi is the length, qi is the width, and ri is the height,
and pi , qi , and ri are integral constants. i ∈ J, J {1, 2, 3, . . . , n} is the set of the given
boxes.
x, y, z: Variables indicating the length, width, and height of the container.
xi , yi , zi : Variables indicating the coordinates of the front-left-bottom corner of box
i.
lxi , lyi , lzi : Binary variables indicating whether the length of box i is parallel to the
X-axis, Y -axis, or Z-axis. The value of lxi is equal to 1 if the length of box i is parallel
to the X-axis; otherwise, it is equal to 0. It is clear that lxi lyi lzi 1.
wxi , wyi , wzi : Binary variables indicating whether the width of box i is parallel to
the X-axis, Y -axis, or Z-axis. The value of wxi is equal to 1 if the width of box i is
parallel to the X-axis; otherwise, it is equal to 0. It is clear that wxi wyi wzi 1.
hxi , hyi , hzi : Binary variables indicating whether the width of box i is parallel to
the X-, Y -, or Z-axis. The value of hxi is equal to 1 if the height of box i is parallel to
the X-axis; otherwise, it is equal to 0. It is clear that hxi hyi hzi 1.
For a pair of boxes i, k, where i < k, there is a set of 0-1 vector aik , bik , cik , dik , eik , fik defined
as
The front-left-bottom corner of the container is fixed at the origin. The interpretation
of these variables is illustrated in Figure 1. Figure 1 contains two boxes i and k, where box i
is located with its length along the X-axis and the width parallel to the Z-axis, and box k is
located with its length along the Z-axis and the width parallel to the X-axis. We then have lxi ,
wzi , hyi , lzk , wxk , and hyk equal to 1. In addition, since box i is located on the left-hand side
of and in front of box k, it is clear that aik dik 1 and bik cik eik fik 0.
According to Chen et al. 10 and Tsai and Li 20, the packing problem can be formu-
lated as follows.
4 Mathematical Problems in Engineering
Problem 1.
where
lxi , lyi , lzi , wxi , wyi , wzi , hxi , hyi , hzi , aik , bik , cik , dik , eik , fik are 0-1 variables, 2.19
M max x, y, z , xi , yi , zi ≥ 1, 1 ≤ x ≤ x ≤ x, 1 ≤ y ≤ y ≤ y, 1 ≤ z ≤ z ≤ z,
2.20
x, y, z, x, y, z are constants,
The objective of this model is to minimize the volume of the container. The constraints
2.3–2.9 are nonoverlapping conditions used to ensure that none of these n boxes overlaps
with each other. Constraints 2.10–2.12 ensure that all boxes are within the enveloping
container. Constraints 2.13–2.18 describe the allocation restrictions among logic variables.
For instance, 2.13 implies that the length of box i is parallel to one of the axes. 2.16 implies
that only one of the length, the width, and the height of box i is parallel to X-axis.
Mathematical Problems in Engineering 5
pi
ri
z
Z qi Box i
yi
(xi , yi , zi ) qk
rk
y
Box k
pk
X
(0, 0, 0) (xk , yk , zk )
x
3. Proposed Method
Consider the objective function Obj xyz in 2.2, where x ≥ y ≥ z, 1 ≤ x ≤ x ≤ x, 1 ≤ y ≤
y ≤ y, 1 ≤ z ≤ z ≤ z and x, y, z are positive variables. Denote r and s as two variables defined
as r x − y z, s x − z. Replace y by 2x − s − r and replace z by x − s. xyz then becomes
Obj as follows.
Obj x 2x2 − r 3sx rs s2 . 3.1
Proposition 3.1. Suppose r and s in 3.1 are fixed values, then Obj is an increasing function.
Proof. Since ∂Obj /∂x 6x2 − 2xr 3s rs s2 xy yz 2xz > 0, it is clear that Obj is an
increasing function.
Proof. Since dimensions of box i, pi , qi , ri , are integral constants for i 1, 2, . . . , n and all
of n boxes are nonoverlapping, therefore, x∗ , y∗ , and z∗ that indicate the optimal solution of
the container must be integral.
6 Mathematical Problems in Engineering
Proposition 3.3. 3 ni1 pi qi ri ≤ x y z/3, where pi is the length, qi is the width, and ri is the
height of the given box i. [19, 20]
n √ n
Proof. Since i1 pi qi ri ≤ xyz and 3 xyz ≤ x y z/3, then we can have 3
i1 pi qi ri ≤
x y z/3.
According to the above propositions and given values of r and s denoted as r and s,
consider the following program.
Problem 2.
Minimize Obj x 2x3 − r 3sx2 rs s2 x
n
x y z
subject to
3
pi qi ri ≤ ,
i1
3 3.2
Proposition 3.4. If xΔ , yΔ , zΔ is the solution of Problem 2 found by a genetic algorithm and
x∗ , y∗ , z∗ is the globally optimal solution of Problem 2, then x∗ ≤ xΔ .
Proof. Since Obj is an increasing function with single variable x following Proposition 3.1
and 2x∗ 3 − r 3sx∗ 2 r s s2 x∗ ≤ 2xΔ 3 − r 3sxΔ 2 r s s2 xΔ . Hence, x∗ ≤ xΔ .
Adding the constraint x ≤ xΔ to Problem 2 for reducing the search region of the opti-
mal solution, we can have the following two programs.
Problem 3.
Minimize Obj x 2x3 − r 3sx2 rs s2 x
subject to x ≤ xΔ ,
3.3
r x−y z, s x − z, 2.3 ∼ 2.21, 3.2,
Problem 4.
Minimize x
3.4
subject to all the constraints in Problem 3.
Proposition 3.5. Let (x∗ , y∗ , z∗ ) be the global optimum of Problem 3, then x∗ , y∗ , z∗ is also the
global optimum of Problem 4.
Mathematical Problems in Engineering 7
Proof. Since Obj x is an increasing function with single variable x following
Proposition 3.1, Problems 2 and 3 have the same global optimum x∗ , y∗ , z∗ .
4. Distributed Algorithm
The solution procedure for solving Problem 1 to obtain a global optimum is presented in the
following with a flow chart shown in Figure 2.
Step 1. Find an initial solution by GA. From Proposition 3.4, the obtained solution is xΔ ,
yΔ , zΔ and the constraint x ≤ xΔ can be utilized to reduce the searching space of the global
solution.
Step 2. Denote sm and s1 as the upper and lower bounds of s, respectively. Find the
bounds of s x − z by solving the following linear programs:
sm Max x − z | subject to 2.3 ∼ 2.21, 3.2 ,
4.1
s1 Min x − z | subject to 2.3 ∼ 2.21, 3.2 .
Step 3. Denote rn and r1 as the upper and lower bounds of r, respectively. Find the
bounds of r x − y z by solving the following linear programs:
rn Max x − y z | subject to 2.3 ∼ 2.21, 3.2, x − z si ,
4.2
r1 Min x − y z | subject to 2.3 ∼ 2.21, 3.2, x − z si ,
Step 4. Decompose main problem and perform distributed packing algorithm. According
to verity of r and s, the main problem can be decomposed into several subproblems. The
transformed subproblem of each iterative process is listed as follows.
8 Mathematical Problems in Engineering
Start
i=i−1
No
∗
xij < x
j =j +1
Yes
No
j=n
Yes
No
i =1
Yes
Stop
Problem Pij .
Minimize Objij x
subject to 2.3 − 2.21, 3.2, 3.3, x − z si, x − y zr j ,
Mathematical Problems in Engineering 9
Obj x 2x2 − r j 3si x r j si si2 ,
1 ≤ i ≤ m, 1 ≤ j ≤ n.
4.3
Step 7. The whole process is finished and the host computer obtains the optimal solution
x∗ , y∗ , z∗ with the objective value x∗ y∗ z∗ .
5. Numerical Examples
To validate the proposed method, several examples with different number of boxes are solved
by LINGO 11.0 26 with the distributed algorithm. Two of the test problems denoted as
Problems 1 and 2 are taken from Chen et al. 10. The other examples are arbitrarily generated.
Solving these problems by the proposed method, the obtained globally optimal solutions are
listed in Tables 1 and 3. Comparison results between GA and the proposed method are shown
in Table 2, and the associated graphs are presented in Figures 3 and 4.
Packing problems often arise in logistic application. The following example
Problem 5 demonstrates how to apply the proposed algorithm in transportation problem
and compare the result with traditional genetic algorithm.
Problem 5. Several kinds of goods are packed into a container so as to deliver to 6 different
stores on a trip. The dimensions of width and height of the container are 5 and 4. All goods
are packed in cubic boxes, which have three different sizes. In order to take less time during
unloading, boxes sent to the same store must be packed together. Different groups of boxes
cannot overlap each other. Moreover, the packing order to each group must be ordered of the
arriving time to each store. The boxes required to be sent to each store are listed in Table 4.
The objective is to determine the smallest dimensions of the container.
Solution 1. The arrangement of boxes can be treated as level assortment. The boxes packed
in the same level will be delivered to the same store. After performing the proposed method,
list of the optimal solutions are shown in Table 5, and illustrated graph is presented in
Figure 5.
10 Mathematical Problems in Engineering
Table 2: Solution comparison of the proposed algorithm and genetic algorithm GA.
GA Proposed method
Problem number
x, y, z Objective value x, y, z Objective value
1 4 boxes 30, 30, 6 5400 28, 26, 6 4368
2 6 boxes 33, 26, 7 6006 35, 28, 6 5880
3 7 boxes 25, 25, 10 6250 31, 16, 12 5952
Y
Z
6
Box 4
Box 1 Box 3
(8, 10, 2)
(16, 11, 0)
26
(0, 0, 6)
(8, 0, 5)
Box 2
(0, 0, 0) (8, 0, 0) X
28
S6
S5
S4
S3
S2
S1
6. Conclusions
This paper proposes a new method to solve a packing optimization problem. The proposed
method reformulates the nonlinear objective function of the original packing problem into a
linear function with two given parameters. The proposed method then solves the reformu-
lated linear 0-1 programs by specifying the parameters sequentially to reach the globally opti-
mal solution. Furthermore, this study adopts a distributed genetic algorithm and distributed
12 Mathematical Problems in Engineering
Store Goods
S1 A, A, A, B, B, B, B, C
S2 A, A, B, B, B, B, C, C
S3 A, A, A, B, B, B, B, C
S4 A, A, B, B, B, C, C, C
S5 A, A, A, A, B, B, B, C
S6 A, A, A, A, A, B, C, C
A: 1-inch cubic box; B: 2-inch cubic box; C: 3-inch cubic box.
Store S1 S2 S3 S4 S5 S6
x1 , y1 , z1 A 1, 2, 0 A 0, 1, 3 A 4, 0, 2 A 0, 0, 3 A 4, 0, 0 A 0, 0, 0
x2 , y2 , z2 A 3, 3, 3 A 5, 0, 0 A 0, 1, 0 A 0, 2, 3 A 4, 1, 3 A 5, 2, 0
x3 , y3 , z3 A 0, 4, 0 B 3, 0, 0 A 1, 2, 0 B 3, 0, 1 A 0, 2, 0 A 0, 3, 0
x4 , y4 , z4 B 4, 0, 0 B 1, 0, 2 B 3, 0, 0 B 7, 0, 1 A 4, 4, 1 A 2, 4, 0
x5 , y5 , z5 B 0, 0, 1 B 3, 0, 2 B 0, 3, 0 B 1, 0, 2 B 2, 0, 0 A 5, 1, 0
x6 , y6 , z6 B 2, 0, 0 B 1, 0, 0 B 1, 0, 2 C 0, 2, 0 B 2, 0, 2 B 3, 0, 2
x7 , y7 , z7 C 0, 2, 1 C 0, 2, 0 B 0, 3, 2 C 6, 2, 0 B 0, 0, 2 C 0, 1, 1
x8 , y8 , z8 C 3, 2, 0 C 3, 2, 0 C 2, 2, 1 C 3, 2, 0 C 1, 2, 0 C 3, 2, 1
Dimension of si 6×5×4 6×5×4 5×5×4 9×5×4 5×5×4 6×5×4
Volume of si 120 120 100 180 100 120
The global solution is 37, 5, 4, and the minimal volume of the container is 740.
Acknowledgments
The authors would like to thank the anonymous referees for contributing their valuable com-
ments regarding this paper and thus significantly improving its quality. The paper is partly
supported by Taiwan NSC Grant NSC 99-2410-H-027-008-MY3.
References
1 J. Egeblad and D. Pisinger, “Heuristic approaches for the two- and three-dimensional knapsack pack-
ing problem,” Computers and Operations Research, vol. 36, no. 4, pp. 1026–1049, 2009.
2 D. Fayard and V. Zissimopoulos, “An approximation algorithm for solving unconstrained two-
dimensional knapsack problems,” European Journal of Operational Research, vol. 84, no. 3, pp. 618–632,
1995.
3 M. Hifi and R. Ouafi, “Best-first search and dynamic programming methods for cutting problems: the
cases of one or more stock plates,” Computers and Industrial Engineering, vol. 32, no. 1, pp. 187–205,
1997.
4 J. F. Tsai, P. L. Hsieh, and Y. H. Huang, “An optimization algorithm for cutting stock problems in the
TFT-LCD industry,” Computers and Industrial Engineering, vol. 57, no. 3, pp. 913–919, 2009.
5 J. E. Beasley, “An algorithm for the two-dimensional assortment problem,” European Journal of Opera-
tional Research, vol. 19, no. 2, pp. 253–261, 1985.
Mathematical Problems in Engineering 13
6 H. L. Li and C. T. Chang, “An approximately global optimization method for assortment problems,”
European Journal of Operational Research, vol. 105, no. 3, pp. 604–612, 1998.
7 J. F. Tsai, P. C. Wang, and M. H. Lin, “An efficient deterministic optimization approach for rectangular
packing problems,” Optimization. In press.
8 F. H. F. Liu and C. J. Hsiao, “A three-dimensional pallet loading method for single-size boxes,” Journal
of the Operational Research Society, vol. 48, no. 7, pp. 726–735, 1997.
9 J. Terno, G. Scheithauer, U. Sommerweiß, and J. Riehme, “An efficient approach for the multi-pallet
loading problem,” European Journal of Operational Research, vol. 123, no. 2, pp. 372–381, 2000.
10 C. S. Chen, S. M. Lee, and Q. S. Shen, “An analytical model for the container loading problem,” Euro-
pean Journal of Operational Research, vol. 80, no. 1, pp. 68–76, 1995.
11 G. Scheithauer, “LP-based bounds for the container and multi-container loading problem,”
International Transactions in Operational Research, vol. 6, pp. 199–213, 1999.
12 W. B. Dowsland, “Three-dimensional packing-solution approaches and heuristic development,” Inter-
national Journal of Production Research, vol. 29, no. 8, pp. 1673–1685, 1991.
13 Y. He, Y. Wu, and R. de Souza, “A global search framework for practical three-dimensional packing
with variable carton orientations,” Computers and Operations Research, vol. 39, no. 10, pp. 2395–2414,
2012.
14 Y. Wu, W. Li, M. Goh, and R. de Souza, “Three-dimensional bin packing problem with variable bin
height,” European Journal of Operational Research, vol. 202, no. 2, pp. 347–355, 2010.
15 A. de Almeida and M. B. Figueiredo, “A particular approach for the Three-dimensional Packing Pro-
blem with additional constraints,” Computers and Operations Research, vol. 37, no. 11, pp. 1968–1976,
2010.
16 F. K. Miyazawa and Y. Wakabayashi, “Three-dimensional packings with rotations,” Computers and
Operations Research, vol. 36, no. 10, pp. 2801–2815, 2009.
17 T. G. Crainic, G. Perboli, and R. Tadei, “TS2 PACK: a two-level tabu search for the three-dimensional
bin packing problem,” European Journal of Operational Research, vol. 195, no. 3, pp. 744–760, 2009.
18 C. A. Floudas, Deterministic Global Optimization: Theory, Computational Methods and Applications,
Kluwer Academic Publishers, Dodrecht, The Netherlands, 2000.
19 H. L. Li, J. F. Tsai, and N. Z. Hu, “A distributed global optimization method for packing problems,”
Journal of the Operational Research Society, vol. 54, no. 4, pp. 419–425, 2003.
20 J. F. Tsai and H. L. Li, “A global optimization method for packing problems,” Engineering Optimization,
vol. 38, no. 6, pp. 687–700, 2006.
21 G. Wäscher, H. Haußner, and H. Schumann, “An improved typology of cutting and packing pro-
blems,” European Journal of Operational Research, vol. 183, no. 3, pp. 1109–1130, 2007.
22 B. Bourbeau, T. Gabriel Crainic, and B. Gendron, “Branch-and-bound parallelization strategies ap-
plied to a depot location and container fleet management problem,” Parallel Computing, vol. 26, no. 1,
pp. 27–46, 2000.
23 H. Gehring and A. Bortfeldt, “A parallel genetic algorithm for solving the container loading problem,”
International Transactions in Operational Research, vol. 9, pp. 497–511, 2002.
24 J. Btazewicz and R. Walkowiak, “A new parallel approach for multi-dimensional packing problems,”
Lecture Notes in Computer Science, vol. 2328, pp. 585–591, 2006.
25 A. Bortfeldt, H. Gehring, and D. Mack, “A parallel tabu search algorithm for solving the container
loading problem,” Parallel Computing, vol. 29, no. 5, pp. 641–662, 2003.
26 LINDO System, LINGO Release 11.0, LINDO System, Chicago, Ill, USA, 2008.
Hindawi Publishing Corporation
Mathematical Problems in Engineering
Volume 2012, Article ID 151590, 20 pages
doi:10.1155/2012/151590
Research Article
Hybrid Optimization Approach for the Design of
Mechanisms Using a New Error Estimator
Copyright q 2012 A. Sedano et al. This is an open access article distributed under the Creative
Commons Attribution License, which permits unrestricted use, distribution, and reproduction in
any medium, provided the original work is properly cited.
A hybrid optimization approach for the design of linkages is presented. The method is applied to
the dimensional synthesis of mechanism and combines the merits of both stochastic and determin-
istic optimization. The stochastic optimization approach is based on a real-valued evolutionary
algorithm EA and is used for extensive exploration of the design variable space when searching
for the best linkage. The deterministic approach uses a local optimization technique to improve the
efficiency by reducing the high CPU time that EA techniques require in this kind of applications. To
that end, the deterministic approach is implemented in the evolutionary algorithm in two stages.
The first stage is the fitness evaluation where the deterministic approach is used to obtain an
effective new error estimator. In the second stage the deterministic approach refines the solution
provided by the evolutionary part of the algorithm. The new error estimator enables the evaluation
of the different individuals in each generation, avoiding the removal of well-adapted linkages that
other methods would not detect. The efficiency, robustness, and accuracy of the proposed method
are tested for the design of a mechanism in two examples.
1. Introduction
The mechanical design of modern machines is often very complex and needs very
sophisticated tools to meet technological requirements. The design of linkages is no
exception, and modern applications in this field have increasingly demanding requirements.
The design of linkages consists in obtaining the best mechanism to fulfil a specific motion
characteristic demanded by a specific task. In many engineering design fields there are
three common requirements known as function generation, path generation, or rigid-body
guidance 1, 2. Dimensional synthesis deals with the determination of the kinematic
parameters of the mechanism necessary to satisfy the required motion characteristics.
2 Mathematical Problems in Engineering
Different techniques have been used for the synthesis of mechanisms including graphical and
analytical techniques 1–3. Graphical and analytical methods developed in the literature are
relatively restricted because they find the exact solution for a reduced number of prescribed
poses and variables. However, during the last decades numerical methods have enabled an
increase in the complexity of the problems by using numerical optimization techniques 4–8.
Despite the work done in dimensional synthesis over recent decades, the design of
mechanisms is still a task where the intuition and experience of the engineers play an
important role. One of the main reasons for this is the large number of variables involved in
a strongly nonlinear problem. Under these circumstances the design variable space contains
too many local minima and only some of them can be identified as local solutions. These local
solutions provide an error below a limit established by the designer and can be considered
acceptable. However, only the global minimum leads to the solution that provides the
greatest accuracy and this should be the main objective in the design of mechanisms.
The application of local optimization techniques to the synthesis of mechanisms took
place mainly during the 80s and 90s. Although other techniques have become more important
in recent years, they remain important so far. Some local search methods have been described
in references 4–6. The main disadvantage of these methods is their dependence on the
initial point, or initial guess, although they also require a differentiable objective function.
Several research works have been done to achieve exact differentiation, which improve the
accuracy and efficiency of these methods. For example, in 5 exact gradient is determined to
optimize a six-bar and eight-bar mechanism. In 6 a general synthesis procedure is obtained
by using exact differentiation and it is applied to different kinds of problems. However, the
dependence on the initial point cannot be avoided and therein lies the weakest point of local
search methods.
Global search methods avoid the dependence on the initial point, but there is a sharp
increase in the computational time necessary to achieve convergence. Genetic algorithms
GAs 7, 8, evolutionary algorithms EAs 9, and Particle Swarm PS are some of the
most frequently used optimization techniques in the literature. All these techniques mimic
the behaviour of processes found in nature and are based on biological processes.
Genetic and evolutionary algorithms apply the principles of evolution found in nature
to the problem of finding an optimal solution. Holland 10 was the first to introduce the GA
and DeJong 11 verified the usage. In GA the genes are usually encoded using a binary
language whereas in EA the decision variables and objective function are used directly. As
coding is not necessary, EAs are less complex and easier to implement for solving complicated
optimization problems. Cabrera et al. 8 used GAs applied to a four-bar linkage in a path
generation problem. Some years later Cabrera et al. 9 used EAs to solve more complex
problems in the design of mechanisms. In this case a multiobjective problem is formulated
including mechanical advantage in the objective function as a design requirement. In 12 a
genetic algorithm is used for the Pareto optimum synthesis of a four-bar linkage considering
the minimization of two objective functions simultaneously.
Hybrid algorithms with application to the synthesis of linkages have been studied
in recent years. Lin 13 developed an evolutionary algorithm by combining differential
evolution and the real-valued genetic algorithm. Khorshidi 14 developed a hybrid approach
where a local search is employed to accelerate the convergence of the algorithm. However,
these methods are limited to the four-bar mechanism and their application is restricted to
path generation problems.
The objective function is based on the synthesis error estimation. The most widely
used error estimator in the literature is Mean Square Distance MSD. The MSD is used to
Mathematical Problems in Engineering 3
measure the difference between the desired parameters and the generated ones. However,
this formulation has proven to be ineffective when seeking the optimal solution 15. In
many cases the MSD misleads the direction of the design and good linkages generated by
the algorithm can be underestimated. Therefore, error estimation is of the utmost importance
for deterministic and stochastic optimization. In EA the error estimator must be evaluated
for each individual in each generation and for this reason the lack of accuracy could lead
to poor efficiency in the optimization process. To avoid these problems Ullah and Kota 15
proposed the use of Fourier Descriptors that evaluate only the shape differences between the
generated and desired paths. However, the proposed formulation is limited to closed paths
in path generation problems. An energy-based error function is used in 16 where the finite
element method is used to assess the synthesis error. This formulation reduces the drawbacks
of MSD, but problems with the relative distance between desired and generated parameters
remain.
The aim of this work is to propose a new hybrid algorithm that combines an
evolutionary technique with a local search optimization approach. Some of the fundamentals
in mechanism synthesis studied in this paper have been extensively discussed in the
literature. However, the originality of this work lies in two aspects: the first one is the
introduction of a new error estimator which accurately compares the function generated by
the candidate mechanism with desired function. The second one is a novel approach based
on the combination of deterministic and stochastic optimization techniques in the so-called
hybrid methods. The flowchart for the optimization process is presented in the paper together
with the results and conclusions.
minimize Fqw, w
gqw, w ≤ 0,
where the objective function Fqw, w formulates the technological requirements of the
mechanism to be designed. The equality constraints Φqw, w formulate the kinematic
restrictions during the motion, and the inequality constraints gqw, w establish the
limitations in the geometrical dimensions. Vector qw is the vector of dependent coordinates
and w is the n-dimensional vector of design variables.
To illustrate the formulation the scheme of a four-bar mechanism in a path generation
synthesis problem is shown in Figure 1. The proposed method can be applied effortlessly to
any type of planar mechanism; however, the example in Figure 1 enables the formulation to
be easily understood. The equality constraints are formulated as follows:
⎧ ⎫
⎪
⎪ L1 cos θ1 L2 cosθ20 θ2i L3 cos θ3 L3 cos θ3 ⎪ ⎪
⎪
⎨ L sin θ L sinθ ⎪
1 1 2 20 θ2i L3 sin θ3 L3 sin θ3 ⎬
Φqw, wi 0, 2.2
⎪
⎪xg − x0 − L1 cos θ1 − L2 cosθ20 θ2i − L5 cosθ5 α⎪ ⎪
⎪
⎩ ⎪
⎭
yg − y0 − L1 sin θ1 − L2 sinθ20 θ2i − L5 sinθ5 α
4 Mathematical Problems in Engineering
Desired path
x3 y3
L5 L3
α θ4
θ5
θ3 L4
x2 y2
θ1
L2 x0 y0
θ2i
L1
x1 y1 θ20
where the vector of design variables contains the geometrical dimensions of the link. That is
wT x0 y0 θ1 θ20 L1 L2 L3 L4 L5 α , 2.3
qT θ3 θ4 xg yg . 2.4
where subscript d stands for desired points, g stands for the generated ones and p is the
number of precision points. Most of the works in dimensional synthesis propose the Mean
Square Distance to assess the error between the two curves. That is
p T 1
1
F g −d
i i
g −d
i i
g − dT g − d . 2.6
2 i1 2
The deterministic approach used in this paper is based on a local search procedure which
uses first-order differentiation to obtain the search direction. In the synthesis problem the
generated precision points depend on the vector of design variables and 2.6 should be
rewritten as follows:
1
Fqw, w g{qw, w} − dT g{qw, w} − d 2.7
2
which is the objective function that must be minimized subject to the equality and inequality
constraints to obtain the optimal dimensions of the mechanism. Differentiating 2.7 with
respect to the design variables and equating it to zero provides
With the aim of greater clarity hereafter the dependence on the variables is omitted. The term
between brackets in 2.8 can be expanded using Taylor series expansion as
∇F ≈ JT g − JT d JT JΔw 0. 2.10
wj 1 wj − α JT J JT g − JT d . 2.11
In this formula the stepsize α has been included in order to control the distance along the
search direction. In 6 the determination of the stepsize and the exact Jacobian is described.
Differentiation of equality constraints given in 2.1 using the chain rule yields
∂Φ ∂q ∂Φ
0. 2.12
∂q ∂w ∂w
6 Mathematical Problems in Engineering
g4
g3
g2
dp−1
dp d4 g1
d1 d3 gp
d2
gp−1
∂g ∂Φ −1 ∂Φ
J− . 2.13
∂q ∂q ∂w
All terms in 2.13 can be easily obtained from the objective function and constraints, and they
enable the exact Jacobian to be determined for use in the deterministic optimization method.
If there are inequality constraints in the optimization problem, they can be converted
to equality constraints through the addition of so-called slack variables. That is
In this way, each inequality constraint adds a new variable that must be included in the
formulation.
y
g4
g3
βg g2
x
βd
dp−1
g1
dp gp
d4
d1 d3
d2
gp−1
Thus, one can say that the error between two curves is minimum if they are compared
by the translation, rotation, and scaling. Therefore, 2.7 could underestimate the error unless
some transformations are introduced. The first transformation consists of the translation of
the generated curve towards the desired one. To do that, the geometric centroids of both
curves are determined by using the precision points as follows:
p
1
dc di ,
p i1
3.1
p
1
gc gi ,
p i1
where dc and gc are the coordinates of the geometric centroids for desired and generated
curves, respectively. The new coordinates of the precision points for the two paths are
obtained by translating the geometric centroids to the origin of the reference frame. That
is
di0 di − dc ,
3.2
gi0 gi − gc .
Figure 3 shows the translation of both curves. Thus, the error estimation can be reformulated
in the following way:
p T
1
E0 gi0 − di0 gi0 − di0 . 3.3
2 i1
Obviously, 3.3 reduces the error and is more accurate than 2.7. However, it should be
pointed out that it still depends on the order chosen for numbering the precision points. In
8 Mathematical Problems in Engineering
other words, 3.3 provides a comparison of the precision points with the same superscript,
which depends on the arbitrary choice previously made by the designer. Therefore, it could
be possible to reduce the error when the order of numbering is changed. Therefore, removing
the effect of the numbering requires the formulation of p error estimators. For the case of two
closed curves, as shown in Figure 3, the following matrix can be written:
⎡ p ⎤
g10 − d10 g2 − d10 ··· gp−1 − d10 g0 − d10
⎢ g20 − d20 g30 − d20 ··· ⎥ p
g0 − d20 g10 − d20
⎢ ⎥
⎢ .. .. .. ⎥ .. ..
⎢ ⎥ . 3.4
⎢ . . . ⎥ . .
⎢ p−1 p−1 p p−1 p−3 p−1 p−2 p−1 ⎥
⎣g0 − d0 g0 − d0 · · · g0 − d0 g0 − d0 ⎦
p p p p−2 p p−1 p
g0 − d0 g10 − d0 · · · g0 − d0 g0 − d0 p×p
Each column of 3.4 gives the terms of the error estimator for each possible combination.
Thus, each error estimator can be formulated as the summatory function defined by
p−j 1 2 p 2
i j−1 i j−p−1
Fj g0 − di0 g0 − di0 ; j 1, 2, . . . , p, 3.5
i1 ip−j 2
where subscript j stands for the estimator index. Therefore, Fj is a single-valued function
providing the estimation of the error. A vector can be formulated with all the values given by
3.5
FT F1 F2 · · · Fp . 3.6
Only one of the terms in this vector provides the minimum error and will be selected to form
the objective function. That is
Fm minF. 3.7
The matrix given by 3.4 and the summatory given by 3.5 are only valid for the comparison
of closed-closed curves. However, it is possible to have two other situations: open-open or
open-closed paths. The former case is shown in Figure 4, while the latter is shown in Figure 5.
In both cases the number of precision points may be different for the desired and generated
curves. Thus, the precision points are redefined as follows:
where c is the number of precision points for the generated curve. Similarly to the closed-
closed case, the centroid of the precision points is determined and the curves are translated
Mathematical Problems in Engineering 9
g1
g 2
g3
d1
d2
g4
d3
p
d4 dp−1 d
gc−1
gc
g4
g3
d1 g2
d2
d3
g1
dp
d4 dp−1
gc
c−1
g
Figure 5: Desired open path d and generated closed path g.
to the origin of the reference frame. The possible combinations that allow the estimation of
the error are given by the following matrix:
⎡ c−pc−p 1 ⎤
g10 − d10 g20 − d10 · · · g0 − d10 g0 − d10
⎢ c−p 1 c−p 2 ⎥
⎢ g20 − d20 g30 − d20 · · · g0 − d20 g0 − d20 ⎥
⎢ ⎥
⎢ .. .. .. .. .. ⎥
E⎢ . . . . . ⎥ . 3.9
⎢ ⎥
⎢ p−1 p−1 p p−1 p−1 p−1 ⎥
⎣g0 − d0 g0 − d0 · · · g0 − d0 g0 − d0 ⎦
c−2 c−1
p p p 1 p p p
g0 − d0 g0 − d0 · · · gc−1
0 − d0 gc0 − d0 p×c−p 1
p i j−1
2
Fj g0 − di0 ; j 1, 2, . . . , c − p 1. 3.10
i1
Equation 3.10 provides the different error estimators and the minimum value given by this
formula is selected as the objective function.
10 Mathematical Problems in Engineering
When the desired path is an open curve and the generated path is a closed curve see
Figure 5, the aforementioned process can be used. However, the error estimator should be
adapted to this situation. That is
⎡ ⎤
g10 − d10 g20 − d10 ··· gc−1
0 − d0
1
gc0 − d10
⎢ g20 − d20 g30 − d20 ⎥ ··· gc0 − d20 g10 − d20
⎢ ⎥
⎢ ⎥
⎢ .. .. ⎥ .. .. ..
E⎢ . . ⎥ . . . . 3.11
⎢ p−1 ⎥
⎢g − dp−1 gp − dp−1 · · · gp−3 − dp−1 gp−2 − dp−1 ⎥
⎣ 0 0 0 0 0 0 0 0 ⎦
p p p 1 p p−2 p p−1 p
g0 − d0 g0 − d0 · · · g0 − d0 g0 − d0 p×c
p i j−1
2
Fj g0 − di0 , 1≤j ≤c−p 1,
i1
3.12
c−j 1 2 p 2
i j−1 i j−c−1
Fj g0 − di0 g0 − di0 , c−p 1 ≤ j ≤ c.
i1 ic−j 2
Equations 3.5, 3.10, and 3.12 provide a better comparison because they remove the effect
of the translation and avoid the influence of the numbering. However, the error estimation
can be enhanced by rotation and scaling. Indeed, if the generated curve is rotated and scaled
with respect to the desired one, the difference between the two curves could be reduced. To
do this, two new parameters must be introduced. The first one is a reference angle which
provides the orientation of each curve. In Figure 2 the orientation angles are given by βd and
βg . In practical design of mechanisms the modification of βg implies the rotation of the whole
linkage in the plane, which is allowed for most of the cases. The second parameter is the
scaling factor s. This parameter allows the generated curve to be expanded or contracted to
reduce the difference with respect to the desired path. For the case of closed-closed curves
the introduction of the rotation and scaling factor in the formulation modifies equations as
follows:
p−j 1
i j−1 2
p i j−p−1
2
Fm βg , s sAg0 − di0 sAg0 − di0 ; j 1, 2, . . . , p, 3.13
i1 ip−j 2
where
cos βg − sin βg
A βg A , 3.14
sin βg cos βg
is the rotation matrix and provides the rotation of the generated precision points.
The error estimator given by 3.13 is now the objective function in a local optimization
subproblem with two variables, βg and s. This optimization subproblem attempts to find the
best orientation and size of the generated curve and also the linkage in order to reduce the
Mathematical Problems in Engineering 11
error with respect to the desired path. The objective functions for the open-open curves may
be readily derived as
p i j−1
2
Fm βg , s sAg0 − di0 ; j 1, 2, . . . , c − p 1, 3.15
i1
p i j−1
2
Fm βg , s sAg0 − di0 , 1≤j ≤c−p 1,
i1
3.16
c−j 1 i j−1
2 p i j−c−1
2
Fm βg , s sAg0 − di0 sAg0 − di0 , c−p 1 ≤ j ≤ c.
i1 ic−j 2
These expressions might suggest that the problem must be solved as an optimization with
two variables. However, the authors’ experience shows that better results are obtained
when the problem is solved independently for each variable. In other words, the results
obtained are very accurate when the rotation optimization problem is solved before the
scaling problem.
In summary, the aforementioned transformations are the core of the comparison
between the desired curve and the candidate, avoiding the influence of location, orientation,
and size all at once. This provides an important contribution that improves the efficiency in
the exploration of the search space when using evolutionary algorithms.
chromosomes are encoded using real-valued genes instead of a binary code because several
works 9, 13 have demonstrated the advantages of this procedure in the design of linkages.
Thus, each gene gives the real value of a design variable in the mechanism to be synthesized
and all genes are grouped in a chromosome which in classical optimization is known as the
vector of design variables. That is
where m represents the dimensionality of wr,g , g is the generation subscript, and rmax is
the number of individuals in each generation. The dimension of w is given by the type of
mechanism to be synthesized and the kind of coordinates used in their definition. In this work
natural coordinates are used for this purpose, as well as in the definition of the generated and
desired paths. The starting and successive populations are randomly generated:
where gmax is the number of generations. In this work rmax does not change during the
optimization process so the population neither increases nor decreases. After a generation has
been created, the fitness of each individual is evaluated in order to sort them for the selection.
The evaluation of the fitness depends on the type of curves involved in the problem, selecting
3.13, 3.15 or 3.16 according to the case. The algorithm uses an elitism strategy in order to
preserve the best individuals for the next generation. To obtain the number of best individuals
an elitism factor, ef, is used as follows:
where nE is the number of individuals whose genetic information is preserved for the
following generation. After that, the tournament selection starts and the parents are chosen
for reproduction. The first step in reproduction is to establish the number of offspring
generated by the crossover, whose valued is given by the following formula:
where nC is the number of offspring generated by the crossover operator and rf is the
reproduction factor. Mutation is another operator used to change the genes randomly during
the reproduction. The number of offspring affected by mutation is given by
Thus, the number of parents is twice the number of offspring selected for crossover plus the
number of individuals selected for mutation. To decide whether or not it should become a
member for reproduction, the roulette wheel method 8 is used for the selection of parents
from the complete population. The number of slots in the roulette is equal to the number of
individuals and the size of the slots is equal to their expectation. Once the parents are selected,
crossover is used to increase the diversity of the individuals in the complete population.
Mathematical Problems in Engineering 13
Crossover generates the offspring by taking genetic information from the two parents. The
chromosomes of the descendents are obtained using the arithmetic mean of the same genes
taken from each parent using a random coefficient with normal distribution. The mutation
operator is controlled by two coefficients. The first one is the scale, sM, which controls the
range of the variation allowed in the genes. The second one is the shrink coefficient, hM.
This coefficient allows mutations with a wider interval of variation in the first generations
but gradually reduces this interval in the following generations. In this way the algorithm
provides exploration and exploitation of the global optimum and maintains a suitable balance
during the optimization process. The authors have verified that the control of the shrink
coefficient is fundamental to obtain the optimal solution when the range of the design
variables is very different.
After the reproduction has finished for one generation, the new generation is evaluated
by using the fitness value for every individual and the same process is repeated until the
convergence of the evolutionary algorithm is achieved. Different convergence criteria may
be used to stop the algorithm. The first one is based on the accuracy obtained for the best
individual in the last generation, but a limit in the number of generations is also established
to stop the process. Once the convergence is achieved, the fitness of the last generation is
evaluated and a family of best individuals is obtained. The family of best individuals is
selected from those linkages whose fitness value is below a threshold. This family of linkages
is used as the initial guess for the deterministic approach to form the hybridization process
which is described in the following subsection.
5. Numerical Examples
In this section two examples are presented in order to demonstrate the capacity of the
hybrid algorithm. In the first example a four-bar mechanism is selected to be synthesized to
14 Mathematical Problems in Engineering
Setting EA parameters
Initial population, P0
Family 1 of solutions
Yes - Solution 1
Convergence?
- Solution 2
No - Etc.
EA operators: selection,
Gradient search
crossover, mutation, etc.
Family 2 of solutions
New generation - Solution 1
- Solution 2
- Etc.
Best solution
generate a right angle path. The example does not correspond to any actual implementation
in engineering design, but this type of path is a challenging objective and demonstrates
the accuracy, robustness, and efficiency of the proposed approach. The second example is
a practical application in the design of an actual machine. The results in these examples are
divided into two stages. The first one is the result obtained by the evolutionary algorithm
and the second one is the result obtained by the complete hybrid algorithm which includes
the local optimization approach in the dimensional synthesis.
Table 1: Desired path and the path generated at the convergence with the proposed algorithm.
Paths 1 2 3 4 5 6 7 8 9 10 11
x 0 0 0 0 0 0 3.00 6.00 9.00 12.00 15.00
Desired mm d
yd 15.00 12.00 9.00 6.00 3.00 0 0 0 0 0 0
xg 0.21 0.06 0.08 −0.24 −0.29 0.82 2.96 5.70 8.77 11.92 14.97
EA mm
yg 15.33 12.20 8.75 5.27 2.76 1.14 0.03 −0.59 −0.68 −0.17 0.94
xs 0.22 −0.16 −0.18 0.04 0.11 0.00 3.00 5.99 9.00 12.01 14.93
Hybrid mm
ys 14.95 12.01 8.99 6.00 2.98 0.24 0.05 −0.17 −0.26 −0.11 0.32
order to compare the result with the desired path, the third and fourth rows of Table 1 show
the coordinates of the generated points. The solution for the hybrid algorithm is shown in
Figure 7b where the path followed by the coupler point fits very well with the desired one.
In the last two rows of Table 1 the coordinates of the generated path are shown and the last
row of Table 3 shows the values for the design variables.
The error at convergence is 0.2025 mm2 and the time necessary to achieve the
convergence was 212 seconds, which is a very reasonable computational cost in this kind
of problem.
Since it is stochastic, the results differ each time the algorithm runs. In order to evaluate
the robustness, the algorithm was run 30 times and the sample mean error obtained at
convergence was 0.309 mm2 with a sample standard deviation of 0.211 mm2 . The sample
mean CPU time to achieve the convergence was 215.05 seconds with a standard deviation
of 19.031 seconds.
25 25
20 20
15 15
10 10
5 5
0 0
−5 −5
−10 −10
−15 −15
−20 −20
−25 −25
−10 0 10 20 30 −10 0 10 20 30
a b
Figure 7: a Solution with the evolutionary algorithm and b with the hybrid algorithm.
x0 x0 θ1 θ20 L1 L2 L3 L4 L5 α
Design variables
mm mm rad rad mm mm mm mm mm rad
EA solution −10.47 8.34 0.31 1.68 24.30 14.81 24.48 31.70 4.02 −0.68
Hybrid solution −8.01 4.91 0.28 5.50 28.16 16.94 23.05 32.68 7.13 −1.63
6. Concluding Remarks
In this paper a hybrid optimization approach has been presented with application to the
optimal dimensional synthesis of planar mechanisms. The objective function is selected
using a new error estimator defined by means of the precision points. This error estimator
enables the evaluation of the fitness of the function without influence of translation, rotation,
and scaling effects. The error estimation is done using a local optimization procedure
Mathematical Problems in Engineering 17
Injection mechanism
Electric motor
Injection system
a b
Figure 8: a Injection moulding machine and b detail of the injection system.
Table 4: Desired path and the path generated at convergence with the proposed algorithm.
providing a very efficient hybrid algorithm. The hybrid algorithm combines the advantages
of both stochastic and deterministic approaches to improve the robustness and accuracy.
Two examples have been presented in the paper to demonstrate the capacity of the method.
The examples show that the proposed method not only achieves the convergence but also
demonstrates how the accuracy is improved by the combination of the two procedures.
18 Mathematical Problems in Engineering
1000
900
Desired
EA solution
Hybrid solution
θ3
θ4
L3
L2
L4
θ2 α
L5
L1
L6
θ1
θ6
θ10
x0 y0
x5 y5
Design x0 y0 θ10 θ4 L1 L2 L3 L4 L5 L6 α x5
variables mm mm rad rad mm mm mm mm mm mm rad mm
EA solution 810.92 311.51 2.745 −1.23 517.03 992.82 991.66 1227.4 1161.7 722.41 −0.286 100
Hybrid
537.39 393.31 2.785 −1.4 416.52 603.75 554.85 724.41 611.08 372.27 0.2437 100
solution
To do this, the examples depict the solution for the case of the evolutionary algorithm
working alone, and then the solution improved by the hybrid algorithm. This shows
how the evolutionary algorithm provides an approximation to the solution and then the
local optimization improves the accuracy. In both examples the solution provides good
designs and the generated curves fit very well with the desired ones. In summary, the
hybrid algorithm is a valuable tool for the design of mechanisms when highly demanding
requirements are imposed. Thus, the conclusion we draw is that the appropriate combination
of stochastic and deterministic algorithms has an enormous potential in the more effective
solution of optimization problems in the design of mechanisms. This work will be further
developed for the solution of other mechanism design problems by adapting the algorithm.
Furthermore, another future task in this field aims to improve the efficiency of the hybrid
optimizer by using the most recent developments in metaheuristic approaches such as
Particle Swarm Optimization and Differential Evolution.
Acknowledgment
This paper has been developed in the framework of the Project DPI2010-18316 funded by the
Spanish Ministry of Economy and Competitiveness.
References
1 F. Freudenstein, “An analytical approach to the design of four-link mechanisms,” Transactions of the
ASME, vol. 76, pp. 483–492, 1954.
2 G. N. Sandor, A general complex number method for plane kinematics synthesis with applications [Ph.D.
thesis], Columbia University, New York, NY, USA, 1959.
3 A. G. Erdman, “Three and four precision point kinematic synthesis of planar linkages,” Mechanism
and Machine Theory, vol. 16, no. 3, pp. 227–245, 1981.
4 S. Krishnamurty and D. A. Turcic, “Optimal synthesis of mechanisms using nonlinear goal
programming techniques,” Mechanism and Machine Theory, vol. 27, no. 5, pp. 599–612, 1992.
5 J. Mariappan and S. Krishnamurty, “A generalized exact gradient method for mechanism synthesis,”
Mechanism and Machine Theory, vol. 31, no. 4, pp. 413–421, 1996.
6 R. Sancibrian, P. Garcı́a, F. Viadero, and A. Fernández, “A general procedure based on exact gradient
determination in dimensional synthesis of planar mechanisms,” Mechanism and Machine Theory, vol.
41, no. 2, pp. 212–229, 2006.
7 A. Kunjur and S. Krishnamurty, “Genetic algorithms in mechanical synthesis,” Journal of Applied
Mechanism and Robotics, vol. 4, no. 2, pp. 18–24, 1997.
8 J. A. Cabrera, A. Simon, and M. Prado, “Optimal synthesis of mechanisms with genetic algorithms,”
Mechanism and Machine Theory, vol. 37, no. 10, pp. 1165–1177, 2002.
9 J. A. Cabrera, F. Nadal, J. P. Muñoz, and A. Simon, “Multiobjective constrained optimal synthesis of
planar mechanisms using a new evolutionary algorithm,” Mechanism and Machine Theory, vol. 42, no.
7, pp. 791–806, 2007.
10 J. H. Holland, Adaptation in Natural and Artificial Systems, The University of Michigan Press, Ann
Arbor, Mich, USA, 1975, An introductory analysis with applications to biology, control, and artificial
20 Mathematical Problems in Engineering
intelligence.
11 K. A. DeJong, An analysis of the behaviour of a class of genetic adaptive system [Ph.D. thesis], University of
Michigan, Ann Arbor, Mich, USA, 1975.
12 N. Nariman-Zadeh, M. Felezi, A. Jamali, and M. Ganji, “Pareto optimal synthesis of four-bar
mechanisms for path generation,” Mechanism and Machine Theory, vol. 44, no. 1, pp. 180–191, 2009.
13 W. Y. Lin, “A GA-DE hybrid evolutionary algorithm for path synthesis of four-bar linkage,”
Mechanism and Machine Theory, vol. 45, no. 8, pp. 1096–1107, 2010.
14 M. Khorshidi, M. Soheilypour, M. Peyro, A. Atai, and M. S. Panahi, “Optimal design of four-bar
mechanisms using a hybrid multi-objective GA with adaptive local search,” Mechanism and Machine
Theory, vol. 46, no. 10, pp. 1453–1465, 2011.
15 I. Ullah and S. Kota, “Optimal synthesis of mechanisms for path generation using fourier descriptors
and global search methods,” Journal of Mechanical Design, vol. 119, no. 4, pp. 504–510, 1997.
16 I. Fernández-Bustos, J. Aguirrebeitia, R. Avilés, and C. Angulo, “Kinematical synthesis of 1-dof
mechanisms using finite elements and genetic algorithms,” Finite Elements in Analysis and Design,
vol. 41, no. 15, pp. 1441–1463, 2005.
Hindawi Publishing Corporation
Mathematical Problems in Engineering
Volume 2012, Article ID 780637, 14 pages
doi:10.1155/2012/780637
Research Article
Goal-Programming-Driven Genetic
Algorithm Model for Wireless Access Point
Deployment Optimization
Copyright q 2012 C.-S. Wang and C.-T. Chang. This is an open access article distributed under
the Creative Commons Attribution License, which permits unrestricted use, distribution, and
reproduction in any medium, provided the original work is properly cited.
Appropriate wireless access point deployment APD is essential for ensuring seamless user
communication. Optimal APD enables good telecommunication quality, balanced capacity
loading, and optimal deployment costs. APD is a typical NP-complex problem because improving
wireless networking infrastructure has multiple objectives MOs. This paper proposes a method
that integrates a goal-programming-driven model PM and a genetic algorithm GA to resolve
the MO-APD problem. The PM identifies the target deployment subject of four constraints: budget,
coverage, capacity, and interference. The PM also calculates dynamic capacity requirements to
replicate real wireless communication. Three experiments validate the feasibility of the PM. The
results demonstrate the utility and stability of the proposed method. Decision makers can easily
refer to the PM-identified target deployment before allocating APs.
1. Introduction
Appropriate wireless access point deployment APD is essential for ensuring seamless user
communication. Optimal APD enables good telecommunication quality, balanced capacity
loading, and optimal deployment costs. APD is a typical NP-complex problem 1 because
it involves multiple decision objectives, such as budget 2–4, coverage 2, 5–8, interference
3, 4, 7, and dynamic capacity 1, 4, 6–9. Furthermore, these objectives usually contradict
each other 7. For example, the number of APs is usually positively related to the wireless
signal coverage rate and telecommunication reliability 1. However, more APs increase
deployment costs. These conflicting criteria should be considered simultaneously when
solving APD problems 9, 10.
2 Mathematical Problems in Engineering
In the last decade, many studies have attempted to solve APD optimally by
considering multiple objectives MOs. There are four main objectives: budget, coverage
rate, capacity, and interference. Studies have attempted to identify maximal coverage. For
example, Huang et al. developed a growth-planning algorithm to establish the maximal
coverage range 11. Zhao et al. used a point-to-point signal strength strategy to implement
indoor AP location optimization for maximal coverage 12. For the capacity objective, the
capacity requirements of wireless networks compared to wired networks are particularly
difficult to evaluate because users are dynamic and can move from place to place. This
makes APD a dynamic and complex problem. The dynamic capacity requirement must
be addressed to resolve APD 13 because users can access particular APs to balance
loads 9, 14. Finally, for the interference objective, too many APs of the same type and
placed too close together may cause AP malfunction because of frequency interference. To
avoid communication interference, some studies 6, 15 have suggested that APs should
be arranged on different communication channels, but this leads to other communication
channel assignment problems.
This paper applies a goal-programming-driven model PM to the MO-APD problem.
It uses goal programming GP to infer and model the PM and a genetic algorithm GA to
search for near optimal solutions. These methods are easily applied to MO-APD problems to
reflect real situations. The remainder of this paper is organized as follows: Section 2 defines
the problem; Section 3 details the PM; Section 4 presents a discussion on the PM solution
process using a GA; Section 5 provides the results of numerical experiments which are given
in this section; lastly, Section 6 offers a conclusion and suggestions for future research.
Variable Definition
Bij The base station for the potential AP deployment area, where i is a row and j is a column.i
1, 2, . . . , n; j 1, 2, . . . , m.
xq Represents the type q AP allocated to Bij , where q ∈ 1, 2, 3 in Table 1. If xq 0, no AP has
been allocated to Bij .
cxq Represents the AP cost for each AP type, xq ∈ 1, 2, 3.
reqij The networking capacity requirement between Bi and Bj .
CSTT Cxq Bi,j , ∀Xq /
0, i 1, 2, . . . , n, j 1, 2, . . . , m, 2.1
i j
bgt 2.2
else θT .
CSTT
Figure 1 shows that to enable seamless user communication, two APs are allocated, but
two capacity requirements req2,1 and req2,3 have no signal coverage. The coverage function
CVGT evaluates the signal coverage area of T . Equation 2.3 evaluates the coverage
fulfillment rate:
CVGT
ΦT . 2.3
target area
Figure 2 shows a dynamic capacity scenario. The target area allocates two Type 1 APs AP1
and AP2 , and two APs simultaneously cover the capacity requirements req22 . For time
slot 1 T1 , the capacity requirements of req12 , req13 , and req22 access AP1 , and the capacity
requirements of req31 and req32 access AP2 . Therefore, for time slot T1 , AP1 and AP2 must
provide 55 and 25 mbit/s capacity, respectively. In time slot T2 , req22 shifts connection from
AP1 to AP2 for balance loading. Therefore, for time slot T2 , AP1 and AP2 must provide 35
and 45 mbit/s, respectively. Actual capacity requirements are difficult to evaluate accurately.
4 Mathematical Problems in Engineering
Target Area
req 1, m
req req
1, 1 1, 2
req 3, 3
req n, 1
req n, 2
Target Area
T1
req 2,2 = 20 mbps
req 2,2 = 20 mbps
T2
Figure 2: The dynamic capacity requirements of two APs. The blue area represents time slot 1 T1 and the
pink area represents time slot 2 T2 .
DCPij T 2.4
else ΨT .
reqij
Mathematical Problems in Engineering 5
Target Area
B1,1 B1,m
B1,2
Type 1 AP
Bn,1 B3,3
Type 1 AP Bn,2
IFTT
ωT 1 − . 2.5
target area
subject to
Equations 3.2–3.4 are the coverage, capacity, and interference constraints. Equation
3.1 is the objective function, which minimizes the total cost subject to multiple decision
constraints 3.2–3.4.
The second approach is performance oriented, and it maximizes the performance of
target deployment subjects to real constraints e.g., budget. This study reformulated the MO-
APD using the performance-oriented method, shown in Proposal 2 P2.
P2
subject to 3.1–3.4.
Equation 3.5 is the objective function in P2, which maximizes the coverage, capacity
fulfillment, and interference fulfillment rates of T subject to budgetary 3.1 and other
decision constraints 3.2–3.4. GP aids MO decision-making problem modeling. It was
first introduced by Charnes and Cooper 16 and further developed by Tamiz et al. 17,
Romero 18, and Chang 19. Various types of GP approaches exist, such as lexicographic
GP, weighted GP, MINMAX Chebyshev GP, and multichoice GP 19. To enable decision
makers to easily set the constraint weighting according to their preferences, this study
used a weighted GP approach to translate P2 into the PM. wcvg , wcp , and wIFT are the
important weights between 0 and 1 for the GAL-CVG, GAL-CP, and GAL-IFT constraints,
respectively.
PM
Min Wbgt bgt Wcvg cvg− Wcp cp− WIFT IFT− , 3.6
subject to
Fitness Initial
evaluation generation
GA manipulation
Next
Select generation
Target Yes No
Terminated? Crossover
deployment
Mutation
next generation. After meeting a number of iterations or predefined criteria, a near optimal
solution is found.
⎡ ⎤
B1,1 B1,2 · · · B1,m
⎢ B2,1 B2,2 · · · B2,m ⎥
⎢ ⎥
T ⎢ . .. .. .. ⎥. 4.1
⎣ .. . . . ⎦
Bn,1 Bn,2 · · · Bn,m
Fi Wbgt bgt Wcvg cvg− Wcp cp− WIFT IFT− . 4.2
4.3. GA Manipulations
1 Selection: roulette wheel selection ensures that highly fit chromosomes produce
more offspring. This method selects a candidate network according to its survival
probability, which is equal to its fitness relative to the whole population, as shown
in 4.3:
Fi
. 4.3
Fi
2 Crossover: the crossover method randomly selects two chromosomes from the
mating pool for mating. Crossover site C is randomly selected in the interval
8 Mathematical Problems in Engineering
4 8 3 10 5 7 10 6 8 10
7 9 5 1 3 1 6 2 2 9
0 2 5 6 5 6 4 5 3 8
6 1 8 7 5 4 9 6 2 6
0 4 7 7 8 8 10 8 3 3
10 6 8 1 4 4 6 8 9 5
4 1 9 2 8 0 8 2 7 4
3 9 2 5 8 7 1 4 8 9
8 2 7 4 4 5 8 8 10 7
6 1 9 0 6 1 5 4 6 6
1, n × m. Two new chromosomes, called offspring, are then obtained by swapping
all characters between position C and n × m.
3 Mutation: the combined reproduction and crossover methods occasionally lose
potentially useful chromosome information. Mutation is introduced to overcome
this. It is implemented by randomly complementing a bit 0 to 1 and vice versa.
This ensures that good chromosomes are not permanently lost.
Table 3: Decision variables and import weights for four subtests in Experiment 1.
Model I for subtests E1.a and E1.b Model II for subtests E1.c and E1.d
Min 0.25a 0.4b bgt 0.25a 0.2b cvg−
Min 0.25bgt 0.250cvg− 0.25cp− 0.25IFT−
b
0.25 0.2 cp
a −
0.25 0.2b IFT−
a
Subject to Subject to
CSTT − bgt bgt− 30000 CSTT − bgt bgt− 30000
ΦT − cvg cvg− 0.85 ΦT − cvg cvg− 0.85
ΨT − cp cp− 0.8 ΨT − cp cp− 0.8
ωT − IFT IFT− 0.8 ωT − IFT IFT− 0.8c 0.9d
bgt , bgt , cvg , cvg− , cp , cp− , IFT , IFT− ≥ 0
−
bgt , bgt− , cvg , cvg− , cp , cp− , IFT , IFT− ≥ 0
Note that ais for E1a, bfor E1b, cfor E1c, and dfor E1d.
as shown in Figure 6. The results also show that the decision maker must either increase the
budget or adjust the other decision objectives. For example, E1.b shows that the coverage and
capacity fulfillment rates can only reach 0.7 at the current budget. As the budget increases
from 30000 in E1.b to 35000 in E1.c and E1.d, the number of APs deployed increases to
22 on average. The coverage and capacity rates increase from 0.7 in E1.b to 0.82 in E1.c and
E1.d. E1.d deployed more APs 23 at a lower cost than E1.a 22 APs. Figure 6 shows that
the APs in E1.c and E1.d are spread evenly in the target area to avoid interference. Figure 7
shows the convergence trends for all subtests. T emerges after 100–150 iterations.
T3 T1
T2 T3
T1
T1
T1 T1 T2
T1 T1
T3
T1 T1 T1 T1 T1
T2 T2 T1
T3 T3
T1 T2 T2
T2
T1
T1
T1 T1 T1 T1 T1 T1 T2
T1 T1
E1.a E1.b
T2 T1 T1 T2
T3 T1 T1 T2 T1 T1 T1
T1 T1 T2 T1
T2 T2
T1
T2
T1 T1
T1 T1
T1 T1 T1 T2 T1
T1
T2 T1 T1
T1 T3
T1 T1 T1 T2
T1 T1
T1 T3 T1 T1
E1.c E1.d
requirements are scattered in the corners of the target area. Table 6 shows that the capacity
requirements and all default decision variables are identical for both tests. The GA param-
eters—population size 600, terminated generation 500, crossover rate 0.4, and mutation
rate 0.1—were fixed to enable result comparison. To avoid random GA effects, all subtests
were run three times with the same parameters on the same machine. The result averages are
reported.
Table 7 shows the experiment results analysis. As expected, APD follows the capacity
requirements, as shown in Figures 8a and 8b. Figures 8a and 8b also show that APs are
more central in E2.a than in E2.b to fulfill the capacity requirements. Although the capacity
requirements are the same in both experiments, E2.a requires nine APs, which is more than
Mathematical Problems in Engineering 11
0.96
0.94
0.92
Fitness value
0.9
0.88
0.86
0.84
0 100 200 300 400 500
Generations
E1.a E1.c
E1.b E1.d
20
T1 T3
T2
25 25 8
T2
T1 T2
20 8
T2 T3
17
T1
T1
10 35 17 10
T1 T2
T1 T2
35
T2 T1
a b
Figure 8: The capacity requirement and results of Experiment 2 for a subtest E2.a and b subtest E2.b.
The numbers represent the capacity requirements.
E2.b seven APs. Therefore, capacity requirements are dynamic, and T requires more APs to
manage the capacity requirement increase in E2.a.
6. Conclusion
Optimal wireless LAN WLAN design is important to ensure seamless user communication.
Appropriately locating wireless APs for WLANs is important. Optimal APD enables high
telecommunication quality, balanced capacity loading, and optimal deployment costs. This
study proposes a GP-driven model integrated with a GA to solve MO-APD subject to four
constraints: budget, capacity, interference, and coverage. The experiment results show that
Mathematical Problems in Engineering 13
the PM resolves many APD problems and achieves dynamic capacity replication. Results
confirm the ability of the PM to solve large-scale APD problems. Future research should focus
on other applications and further verification of PM.
Appendix
Figure 9 shows the capacity requirement for the E3.a subtest.
References
1 N. Weicker, G. Szabo, K. Weicker, and P. Widmayer, “Evolutionary multiobjective optimization for
base station transmitter placement with frequency assignment,” IEEE Transactions on Evolutionary
Computation, vol. 7, no. 2, pp. 189–203, 2003.
2 A. Mc Gibney, M. Klepal, and D. Pesch, “Agent-based optimization for large scale WLAN design,”
IEEE Transactions on Evolutionary Computation, vol. 15, no. 4, pp. 470–486, 2011.
3 L. Liao, W. Chen, C. Zhang, L. Zhang, D. Xuan, and W. Jia, “Two birds with one stone: wireless access
point deployment for both coverage and localization,” IEEE Transactions on Vehicular Technology, vol.
60, no. 5, pp. 2239–2252, 2011.
4 K. P. Scheibe and C. T. Ragsdale, “A model for the capacitated, hop-constrained, per-packet wireless
mesh network design problem,” European Journal of Operational Research, vol. 197, no. 2, pp. 773–784,
2009.
5 K. Collins, S. Mangold, and G. M. Muntean, “Supporting mobile devices with wireless LAN/MAN
in large controlled environments,” IEEE Communications Magazine, vol. 48, no. 12, pp. 36–43, 2010.
6 A. Hills, “Large-scale wireless LAN design,” IEEE Communications Magazine, vol. 39, no. 11, pp. 98–
104, 2001.
14 Mathematical Problems in Engineering
7 K. Jaffrès-Runser, J. M. Gorce, and S. Ubéda, “Mono- and multiobjective formulations for the indoor
wireless LAN planning problem,” Computers and Operations Research, vol. 35, no. 12, pp. 3885–3901,
2008.
8 R. Whitaker and S. Hurely, “Evolution of planning for wireless communication systems,” in Proceed-
ings of the 36th Hawaii International Conference on System Sciences, p. 10, January 2003.
9 J. H. Lee, B. J. Han, H. J. Lim, Y. D. Kim, N. Saxena, and T. M. Chung, “Optimizing access point allo-
cation using genetic algorithmic approach for smart home environments,” Computer Journal, vol. 52,
no. 8, pp. 938–949, 2009.
10 I. E. Liao and K. F. Kao, “Enhancing the accuracy of WLAN-based location determination systems
using predicted orientation information,” Information Sciences, vol. 178, no. 4, pp. 1049–1068, 2008.
11 X. Huang, U. Behr, and W. Wiesbecd, “Automatic cell planning for a low-cost and spectrum efficient
wireless network,” in Proceedings of the IEEE Global Telecommunications Conference (GLOBECOM ’00),
vol. 1, pp. 282–276, December 2000.
12 Y. Zhao, H. Zhou, and M. Li, “Indoor access points location optimization using Differential
Evolution,” in Proceedings of the International Conference on Computer Science and Software Engineering
(CSSE ’08), pp. 382–385, December 2008.
13 J. H. Lee, B. J. Han, H. K. Bang, and T. M. Chung, “An optimal access points allocation scheme based
on genetic algorithm,” in Proceedings of the International Conference on Future Generation Communication
and Networking (FGCN ’07), pp. 55–59, December 2007.
14 F. Guo and T. C. Chiueh, “Scalable and robust WLAN connectivity using access point array,” in
Proceedings of the International Conference on Dependable Systems and Networks, pp. 288–297, July 2005.
15 X. Ling and K. L. Yeung, “Joint access point placement and channel assignment for 802.11 wireless
LANs,” IEEE Transactions on Wireless Communications, vol. 5, no. 10, pp. 2705–2711, 2006.
16 A. Charnes and W. W. Cooper, Management Model and Industrial Application of Linear Programming, vol.
1, Wiley, New York, NY, USA, 1961.
17 M. Tamiz, D. Jones, and C. Romero, “Goal programming for decision making: an overview of the
current state-of-the-art,” European Journal of Operational Research, vol. 111, no. 3, pp. 569–581, 1998.
18 C. Romero, “Extended lexicographic goal programming: a unifying approach,” Omega, vol. 29, no. 1,
pp. 63–71, 2001.
19 C. T. Chang, “Revised multi-choice goal programming,” Applied Mathematical Modelling, vol. 32, no.
12, pp. 2587–2595, 2008.
20 D. E. Goldberg, Genetic Algorithms in Search, Optimization and Machine Learning, Addison-Wesley, 1989.
Hindawi Publishing Corporation
Mathematical Problems in Engineering
Volume 2012, Article ID 874761, 20 pages
doi:10.1155/2012/874761
Research Article
Multithreshold Segmentation Based on Artificial
Immune Systems
Copyright q 2012 Erik Cuevas et al. This is an open access article distributed under the Creative
Commons Attribution License, which permits unrestricted use, distribution, and reproduction in
any medium, provided the original work is properly cited.
Bio-inspired computing has lately demonstrated its usefulness with remarkable contributions to
shape detection, optimization, and classification in pattern recognition. Similarly, multithreshold
selection has become a critical step for image analysis and computer vision sparking considerable
efforts to design an optimal multi-threshold estimator. This paper presents an algorithm for
multi-threshold segmentation which is based on the artificial immune systemsAIS technique,
also known as theclonal selection algorithm CSA. It follows the clonal selection principle
CSP from the human immune system which basically generates a response according to the
relationship between antigens Ag, that is, patterns to be recognized and antibodies Ab, that
is, possible solutions. In our approach, the 1D histogram of one image is approximated through a
Gaussian mixture model whose parameters are calculated through CSA. Each Gaussian function
represents a pixel class and therefore a thresholding point. Unlike the expectation-maximization
EM algorithm, the CSA-based method shows a fast convergence and a low sensitivity to
initial conditions. Remarkably, it also improves complex time-consuming computations commonly
required by gradient-based methods. Experimental evidence demonstrates a successful automatic
multi-threshold selection based on CSA, comparing its performance to the aforementioned well-
known algorithms.
1. Introduction
Several image-processing applications aim to detect and classify relevant features which may
be later analyzed to perform several high-level tasks. In particular, image segmentation seeks
to group pixels within meaningful regions. Commonly, gray levels belonging to the object are
2 Mathematical Problems in Engineering
substantially different from those featuring the background. Thresholding is thus a simple
but effective tool to isolate objects of interest; its applications include several classics such as
document image analysis, whose goal is to extract printed characters 1, 2, logos, graphical
content, or musical scores; also it is used for map processing which aims to locate lines,
legends, and characters 3. Moreover, it is employed for scene processing, seeking for object
detection, marking 4, and for quality inspection of materials 5, 6.
Thresholding selection techniques can be classified into two categories: bilevel and
multilevel. In the former, one limit value is chosen to segment an image into two classes: one
representing the object and the other one segmenting the background. When distinct objects
are depicted within a given scene, multiple threshold values have to be selected for proper
segmentation, which is commonly called multilevel thresholding.
A variety of thresholding approaches have been proposed for image segmentation,
including conventional methods 7–10 and intelligent techniques 11, 12. Extending the
segmentation algorithms to a multilevel approach may cause some inconveniences: i they
may have no systematic or analytic solution when the number of classes to be detected
increases, and ii they may also show a slow convergence and/or high computational cost
11.
In this work, the segmentation algorithm is based on a parametric model holding
a probability density function of gray levels which groups a mixture of several Gaussian
density functions Gaussian mixture. Mixtures represent a flexible method of statistical
modelling as they are employed in a wide variety of contexts 13. Gaussian mixture has
received considerable attention in the development of segmentation algorithms despite its
performance is influenced by the shape of the image histogram and the accuracy of the
estimated model parameters 14. The associated parameters can be calculated considering
the expectation-maximization EM algorithm 15, 16 or gradient-based methods such as
Levenberg-Marquardt, LM 17. However, EM algorithms are very sensitive to the choice of
the initial values 18; meanwhile, gradient-based methods are computationally expensive
and may easily get stuck within local minima 14. Therefore, some researchers have
attempted to develop methods based on modern global optimization algorithms such as
the learning automata LA 19 and differential evolution algorithm 20. In this paper,
an alternative approach using a bio-inspired optimization algorithm for determining the
parameters of a Gaussian mixture is presented.
On the other hand, biological inspired methods can successfully be transferred into
novel computational paradigms as shown by the successful development of artificial neural
networks, evolutionary algorithms, swarming algorithms, and so on. The human immune
system HIS is a highly evolved, parallel, and distributed adaptive system 21 that
exhibits remarkable abilities that can be imported into important aspects in the field of
computation. This emerging field is known as artificial immune system AIS 22 which is a
computational system fully inspired by the immunology theory and its functions, including
principles and models. AISs have recently reached considerable research interest from
different communities 23, focusing on several aspects of optimization, pattern recognition,
abnormality detection, data analysis, and machine learning. Artificial immune optimization
has been successfully applied to tackle numerous challenging optimization problems with
remarkable performance in comparison to other classical techniques 24.
Clonal selection algorithm CSA 25 is one of the most widely employed AIS
approaches. The CSA is a relatively novel evolutionary optimization algorithm which
has been built on the basis of the clonal selection principle CSP 26 of HIS. The CSP
explains the immune response when an antigenic pattern is recognized by a given antibody.
Mathematical Problems in Engineering 3
In the clonal selection algorithm, the antigen Ag represents the problem to be optimized
and its constraints, while the antibodies Abs are the candidate solutions of the problem.
The antibody-antigen affinity indicates as well the matching between the solution and
the problem. The algorithm performs the selection of antibodies based on affinity either
by matching against an antigen pattern or by evaluating the pattern via an objective
function. In mathematical grounds, CSA has the ability of getting out of local minima while
simultaneously operating over a pool of points within the search space. It does not use the
derivatives or any of its related information as it employs probabilistic transition rules instead
of deterministic ones. Despite its simple and straightforward implementation, it has been
extensively employed in the literature for solving several kinds of challenging engineering
problems 27–29.
In this paper, the segmentation process is considered as an optimization problem
approximating the 1D histogram of a given image by means of a Gaussian mixture model.
The operation parameters are calculated through the CSA. Each Gaussian contained within
the histogram represents a pixel class and therefore belongs to the thresholding points. In
order to compare the segmentation results with other optimization methods, the number
of elements in the Gaussian mixture classes is considered already known or given by
the user. The experimental results, presented in this work, demonstrate that CSA exhibits
fast convergence, relative low computational cost, and no sensitivity to initial conditions by
keeping an acceptable segmentation of the image, that is, a better mixture approximation in
comparison to the EM- or gradient-based algorithms.
This paper organizes as follows. Section 2 presents the method following the Gaussian
approximation of the histogram. Section 3 provides information about the CSA while
Section 4 demonstrates the automatic threshold determination. Section 5 discusses some
implementation details. Experimental results for the proposed approach are presented in
Section 6, followed by the discussion summarized in Section 7.
2. Gaussian Approximation
Let consider an image holding L gray levels 0, . . . , L − 1 whose distribution is displayed
within a histogram hg. In order to simplify the description, the histogram is normalized
just as a probability distribution function, yielding
ng
h g , h g > 0,
N
2.1
L−1 L−1
N ng , h g 1,
g0 g0
where ng denotes the number of pixels with gray level g and N being the total number of
pixels in the image.
The histogram function can thus be contained into a mix of Gaussian probability
functions of the form
2
K K
Pi − x − μi
px Pi · pi x exp , 2.2
i1 i1 2πσi 2σi2
with Pi being the probability of class i, pi x being the probability distribution function of
gray-level random variable x in class i, μi and σi being the mean and standard deviation of
4 Mathematical Problems in Engineering
the ith probability distribution function, and K being the number of classes within the image.
In addition, the constraint K i1 Pi 1 must be satisfied.
The mean square error is used to estimate the 3K parameters Pi , μi , and σi ; i 1, . . . , K.
For instance, the mean square error between the Gaussian mixture pxi and the experimental
histogram function hxi is now defined as follows:
1 n 2
K
J p xj − h xj ω· Pi −1 . 2.3
n j1 i1
Assuming an n-point histogram as in 13 and ω being the penalty associated with the
constrain K i1 Pi 1. In general, the parameter estimation that minimizes the square error
produced by the Gaussian mixture is not a simple problem. A straightforward method is to
consider the partial derivatives of the error function to zero by obtaining a set of simultaneous
transcendental equations 13. However, an analytical solution is not always available
considering the nonlinear nature of the equation which in turn yields the use of iterative
approaches such as gradient-based or maximum likelihood estimation. Unfortunately, such
methods may also get easily stuck within local minima.
In the case of other algorithms such as the EM algorithm and the gradient-based
methods, the new parameter point lies within a neighbourhood distance of the previous
parameter point. However, this is not the case for the CSA which is based on stochastic
principles. New operating points are thus determined by a parameter probability function
that may yield points lying far away from previous operating points, providing the algorithm
with a higher ability to locate and pursue a global minimum.
This paper aims to compare its segmentation results to other optimization methods
that have been applied to similar segmentation tasks. Therefore, the number of elements in
the Gaussian mixture classes is considered as already known or provided by the user. The
number of classes, in most cases, represents the number of objects which are contained in the
image. However, if the selected number is lower than the object number, it can be assumed
that results actually favour the classification of bigger objects yet bearing the expense of
ignoring small subjects.
i maintenance of memory cells which are functionally disconnected from repertoire,
ii selection and cloning of most stimulated Abs,
iii suppression of nonstimulated cells,
iv affinity maturation and reselection of clones showing the highest affinities,
v mutation rate proportional to Abs affinities.
From immunology concepts, an antigen is any substance that forces the immune
system to produce antibodies against it. Regarding the CSA systems, the antigen concept
Mathematical Problems in Engineering 5
refers to the pending optimization problem which focuses on circle detection. In CSA, B cells,
T cells, and antigen-specific lymphocytes are generally called antibodies. An antibody is a
representation of a candidate solution for an antigen, for example, the prototype circle in
this work. A selective mechanism guarantees that those antibodies solutions that better
recognize the antigen and therefore may elicit the response are to be selected holding long
life spans. Therefore, such cells are to be named memory cells M.
3.1. Definitions
In order to describe the CSA, the notation includes boldfaced capital letters indicating
matrices and boldfaced small letters indicating vectors. Some relevant concepts are also
revisited below:
i antigen: the problem to be optimized and its constraints circle detection,
iii affinity: the objective function measurement for an antibody circle matching,
where the positive integer m is the size of antibody population D {d1 , d2 , . . . , dm }which is
an m-dimensional group of antibody d, being a spot within the antibody space I.
TCP A
TM TSC
Ak −→ Yk −→ Zk ∪ Ak −→ Ak 1. 3.2
Yk TPC Ak TPC a1 k, TPC a1 k, . . . , TPC an k , 3.3
vector. Function roundx gets x to the least integer bigger than x. There are various methods
for calculating qi . In this work, it is calculated as follows:
Fai k
qi k round Nc · n i 1, 2, . . . , n, 3.4
j1 F aj k
where Nc is called the clonal size. The value of qi k is proportional to the value of Fai k.
After clonal proliferation, the population becomes
where
Yi k yij k yi1 k, yi2 k, . . . , yiqi k ,
3.6
yij k a1 k, j 1, 2, . . . , qi , i 1, 2, . . . , n.
A
3.2.2. Affinity Maturation Operator (TM )
The affinity maturation operation is performed by hypermutation. Random changes are
introduced into the antibodies just like it happens in the immune system. Such changes may
A
lead to increase in the affinity. The hypermutation is performed by the operator TM which is
applied to the population Yk as it is obtained by clonal proliferation Zk TM Yk.
C
α e−ρ·Fab , 3.7
α beingthe mutation rate, F being the objective function value of the antibody ab as it is
normalized between 0,1, and ρ being a fixed step. In 31, it is demonstrated the importance
of including the factor ρ into 3.7 to improve the algorithm performance. The way ρ modifies
the shape of the mutation rate is shown by Figure 1.
The number of mutations held by a clone with objective function value F is equal to
L·α, considering L as the length of the antibody—22 bits are used in this paper. For the binary
encoding, mutation operation can be done as follows: each gene within an antibody may be
replaced by its opposite number i.e., 0-1 or 1-0.
Following the affinity maturation operation, the population becomes
A
where TM is the operator as it is defined by 3.7 and applied onto the antibody yij .
Mathematical Problems in Engineering 7
1
0.9
0.8
0.7
0.6
0.5
α 0.4 ρ=3
0.3 ρ=5
0.2 ρ = 10
ρ = 20
0.1
0
0 0.2 0.4 0.6 0.8 1
F
bi k if Fai k < Fbi k
TSC Zi k ∪ ai k 3.9
ai k if Fai k ≥ Fbi k,
where i 1, 2, . . . , n.
Each step of the CSA may be defined as follows.
1 Initialize randomly a population Pinit, a set h Pr n of candidate solutions of
subsets of memory cells M which is added to the remaining population Pr , with
the total population being PT Pr M, with M holding n memory cells.
2 Select the n best individuals of the population PT to build Ak, according to the
affinity measure objective function.
3 Reproduce TPC population Ak proportionally to their affinity with the antigen
and generate a temporary population of clones Yk. The clone number is an
increasing function of the affinity with the antigen 3.1.
4 Mutate TM A
the population Yk of clones according to the affinity of the antibody
to the antigen 3.4. A maturated antibody population Zk is thus generated.
5 Reselect TSC the best individuals from Zk and Ak to compose a new memory
set M Ak 1.
6 Add random Pr novel antibodies diversity introduction to the new memory cells
M to build PT .
7 Stop if any criteria are reached; otherwise return to 2.2.
Figure 2 shows the full draw of the CSA. The clone number in Step 3 is defined
according to 3.1. Although a unique mutation operator is used in Step 5, the mutated values
8 Mathematical Problems in Engineering
6 1
Pr Pinit
M
A(k + 1) 5 Select n
TSC
A(k)
3
TPC
Y (k)
4
A
TM
Z(k)
of individuals are inversely proportional to their fitness by means of 3.7; that is, the more
Ab shows a better fitness, the less it may change.
The similarity property 32 within the Abs can also affect the convergence speed
of the CSA. The idea of the antibody addition based on the immune network theory is
introduced for providing diversity to the newly generated Abs in M, which may be similar
to those already in the old memory M. Holding such a diverse Ab pool, the CSA can avoid
being trapped into local minima 30, contrasting to well-known genetic algorithms GAs
which usually tend to bias the whole population of chromosomes towards only the best
candidate solution. Therefore, it can effectively handle challenging multimodal optimization
tasks 33–36.
The management of population includes a simple and direct searching algorithm for
globally optimal multimodal functions. This is also another clear difference in comparison to
other evolutionary algorithms, like GA, because it does not require crossover but only cloning
and hypermutation of individuals in order to use affinity as selection mechanism. The CSA
is adopted in this work to find the parameters P , σ, and μ of Gaussian functions and their
corresponding threshold values for the image.
considering
Th
E1 Th ph 1 xdx,
−∞
∞ 4.2
E2 Th ph xdx.
Th
E1 Th is the probability of mistakenly classifying the pixels in the h 1th class belonging
to the hth class, while E2 Th is the probability of erroneously classifying the pixels in the hth
class belonging to the h 1th class. Pj s are the a priori probabilities within the combined
probability density function, and Th is the threshold value between the hth and the h 1th
classes. One Th value is chosen such as the error ETh is minimized. By differentiating ETh
with respect to Th and equating the result to zero, it is possible to use the following equation
to define the optimum threshold value Th :
considering
A σh2 − σh2 1 ,
B 2 · μh σh2 1 − μh 1 σh2 ,
4.4
2 2 σh 1 Ph
C σh μh 1 − σh 1 μh 2 · σh σh 1 2 · ln .
σh Ph 1
Although the above quadratic equation has two possible solutions, only one of them
is feasible; that is, a positive value falling within the interval.
5. Implementation Details
In this work, either an antibody or an antigen will be represented in binary form by a bit
chain of the form
c c 1 , c 2 , . . . , cL , 5.1
c ∈ SL . 5.2
Once the above steps are completed, the process is started again, until one individual
showing the highest affinity is found, that is, finding the minimum stated in 2.3. At this
work, the algorithm considers 3 Gaussians to represent the same number of classes. However,
it can be easily expanded to more classes. Each single Gaussian has the variables Pi , μi , σi
with i 1, 2, 3 representing the Hamming shape-space by means of an 22 bits word over
the following ranges:
Pi : 1, maxh
μi : min g , max g 5.3
∗
σi : 1, max g 0.5 ,
with g representing the grey level and h is the histogram of the grey level image. Hence, the
first step is to generate the initial individual population of the antibody population by means
of
AB 2 · r, N, Sp − 1; 5.4
with Sp representing the bit string assigned to each of the initial individuals N. Later, in order
to perform the mapping from binary string to real value, we use
21
c L , . . . , c2 , c 1 2 ci · 2
i
r. 5.5
i0 10
rmax
r r · , 5.6
222 − 1
1800
1400
1600
1200
Frequency
1000
800
600
400
200
0
0 50 100 150 200 250
Gray level
a b
Figure 3: a Original “The Cameraman” image and b its correspondent histogram.
6. Experimental Results
6.1. Presentation of Results
In this section, two experiments are tested to evaluate the performance of the proposed
algorithm. The first one considers the well-known image of the “The Cameraman” shown
in Figure 3a as its corresponding histogram is presented by Figure 3b. The goal is to
segment the image with 3 different pixel classes. During learning, the CSA algorithm adjusts 9
parameters in this test. In this experiment, a population of 100 N individuals is considered.
Each candidate holds 9 dimensions, yielding
IN σ1N , σ2N , σ3N , P1N , P2N , P3N , μN N N
1 , μ2 , μ3 , 6.1
with N representing the individual’s number, in this case, 100. The parameters P, σ, μ are
randomly initialized.
The experiments suggest that, after 130 iterations, the CSA algorithm has converged
to the global minimum. Figure 4a shows the obtained Gaussian functions pixel classes,
while Figure 4b shows the combined graph. The layout in Figure 4b suggests an easy
combination of the Gaussian functions to approximate the shape of the graph shown in
Figure 3b, representing the histogram of the original image. Figure 5 shows the segmented
image whose threshold values are calculated according to 3.4 and 3.5.
In order to test the performance, the algorithm gets to analyze the image shown in
Figure 6 whose histogram exhibits a remarkable problem a set of peaks regarding class
approximation. Such image, due to its complexity, is considered as a benchmark image for
other algorithms, including classical approaches as in 7, 10 or intelligent algorithms as in
11, 12. For this particular image, after 190 generations, the algorithm was capable to achieve
the minimum approximation value J see 2.3, considering three different classes. Figure 7
shows the approximation quality; meanwhile, Figure 8 presents the obtained segmented
image.
The third experiment considers a new image known as “The scene” shown by
Figure 9a. The histogram is presented in Figure 9b. Now, the algorithm considers 4 pixel
classes. The optimization is performed by the CSA algorithm resulting in the Gaussian
12 Mathematical Problems in Engineering
1800 1800
1600 1600
1400 1400
1200
Frequency
1200
Frequency
1000 1000
800
800
600
600
400
400 200
200 0
0 0 50 100 150 200 250
0 50 100 150 200 250 Gray level
Gray level
Pixel class 3
Pixel class 2
Pixel class 1
a b
Figure 4: Applying the CSA algorithm for 3 classes and its results: a Gaussian functions for each class,
b mixed Gaussian functions approaching the original histogram.
Figure 5: The image after the segmentation is applied, considering three classes.
4000
3500
3000
Frequency
2500
2000
1500
1000
500
0
0 50 100 150 200 250
Gray level
a b
3000 3000
2500 2500
2000 2000
Frequency
Frequency
1500 1500
1000 1000
500 500
0 0
0 50 100 150 200 250 0 50 100 150 200 250
Gray level Gray level
Pixel class 3
Pixel class 2
Pixel class 1
a b
Figure 7: CSA segmentation for 3 classes over the benchmark image: a Gaussian functions for each class.
b Mixed Gaussian functions approaching the look of the original histogram.
Figure 8: The benchmark image after segmentation considering all three classes.
functions shown by Figure 10a. Figure 10b presents the combined graph from the
addition of such Gaussian functions.
After the optimization by the CSA algorithm, the added layout including all 4
Gaussian functions is obtained as shown by Figure 11a. It is also evident that the resulting
function approaches the original histogram as Figure 11b shows the resulting image after
applying the segmentation algorithm.
4000
3500
3000
2500
Frequency
2000
1500
1000
500
0
0 50 100 150 200 250
Gray level
a b
Figure 9: Third experiment data: a the original image “The scene” and b its corresponding histogram.
3000 3500
2500 3000
2500
2000
Frequency
Frequency
2000
1500
1500
1000 1000
500 500
0 0
0 50 100 150 200 250 0 50 100 150 200 250
Gray level Gray level
Figure 10: Applying the CSA algorithm for 4 classes: a Gaussian functions at each class: b adding all
four Gaussian functions, it approaches the original histogram.
computing Gaussian mixture parameters. The comparison focuses on the following issues:
sensitivity to initial conditions, singularities, convergence, and computational costs.
4000
3500
3000
Frequency
2500
2000
1500
1000
500
0
0 50 100 150 200 250
Gray level
Gaussian approximation
Original histogram
a b
Figure 11: Segmentation of “The Scene” considering four classes for the CSA algorithm. a Comparison
between the original histogram and the Gaussian approach. b The image after the segmentation process.
×10−3
12
10
Frequency
8
6
4
2
Figure 12: a Original image used for the comparison on initial conditions and b its corresponding
histogram.
analysis on Table 1 shows the acute sensitivity of the EM algorithm to initial conditions.
By such sensitivity, it is observed in Figure 13 where a clear pixel misclassification appears
in some sections of the image. In case of the LM method, although it does not present
a considerable difference in comparison to optimal values, its deviation shows that it is
prone to get trapped into a local minimum. On the other hand, the CSA algorithm exhibits
the best performance as its parameter values fall the nearest to the optimal approximation
performance.
EM LM CSA
Figure 13: Segmented images after applying the EM, the LM, and the CSA algorithm with different initial
conditions.
Table 1: Comparison between the EM, the LM, and the CSA algorithm, considering two different initial
conditions.
segmentation literature. All the experiments consider four classes. Table 2 shows the
averaged measurements as they are obtained from 20 experiments. It is evident that the
EM is the slowest to converge iterations, and the LM shows the highest computational
cost time elapsed because it requires complex Hessian approximations. On the other hand,
the CSA shows an acceptable tradeoff between its convergence time and its computational
Mathematical Problems in Engineering 17
Original images
EM-segmented images
LM-segmented images
CSA-segmented images
Figure 14: Original benchmark images a–c and segmented images obtained by the EM, the LM, and
the CSA algorithms.
18 Mathematical Problems in Engineering
Table 2: Iterations and time requirements of the EM, the LM, and the CSA algorithm as they are applied to
segment benchmark images see Figure 14.
Iterations
a b c
Time elapsed
1855 1833 1870
EM
2.72 s 2.70 s 2.73 s
985 988 958
LM
4.03 s 4.04 s 4.98 s
201 188 282
CSA
0.21 s 0.18 s 0.25 s
cost. Finally, Figure 14 below shows the segmented images as they are generated by each
algorithm. By analyzing the images a–c in Figure 14, it is clear that the CSA approach
presents better results when the segmented images are compared with the original ones. In
case of the EM and LM algorithms, several stains are identified in regions where the intensity
level must be considered homogenous.
7. Conclusions
In this paper, an automatic image multi-threshold approach based on the clonal selection
algorithm CSA is proposed. The segmentation process is considered to be similar to an
optimization problem. The algorithm approximates the 1-D histogram of a given image using
a Gaussian mixture model whose parameters are calculated through the CSA. Each Gaussian
function approximating the histogram represents a pixel class and therefore one threshold
point.
Experimental evidence shows that the CSA has a compromise between its convergence
time and its computational cost when it is compared to the expectation-maximization EM
method and the Levenberg-Marquardt LM algorithm. Additionally, the CSA also exhibits
a better performance under certain circumstances initial conditions on which it is well
reported in the literature 14, 18 that the EM has underperformed. Finally, the results
have shown that the stochastic search accomplished by the CSA method shows a consistent
performance with no regard of the initial value and still showing a greater chance to reach
the global minimum.
Although Table 2 indicates that the CSA method can yield better results in comparison
to the EM and gradient-based methods, notice that the aim of our paper is not intended
to beat all segmentation algorithms which have been proposed earlier but to show that the
artificial immune systems can effectively serve as an attractive alternative to evolutionary
algorithms which have been employed before to successfully segment images.
References
1 T. Abak, U. Baris, and B. Sankur, “The performance evaluation of thresholding algorithms for optical
character recognition,” in Proceedings of the 4th International Conference on Document Analysis and
Recognition, pp. 697–700, August 1997.
2 M. Kamel and A. Zhao, “Extraction of binary character/graphics images from grayscale document
images,” Graphical Models and Image Processing, vol. 55, no. 3, pp. 203–217, 1993.
Mathematical Problems in Engineering 19
3 O. D. Trier and A. K. Jain, “Goal-directed evaluation of binarization methods,” IEEE Transactions on
Pattern Analysis and Machine Intelligence, vol. 17, no. 12, pp. 1191–1201, 1995.
4 B. Bhanu, “Automatic target recognition: state of the art survey,” IEEE Transactions on Aerospace and
Electronic Systems, vol. 22, no. 4, pp. 364–379, 1986.
5 M. Sezgin and B. Sankur, “Selection of thresholding methods for non-destructive testing applica-
tions,” in Proceedings of the IEEE International Conference on Image Processing (ICIP ’01), pp. 764–767,
October 2001.
6 M. Sezgin and R. Taşaltı́n, “A new dichotomization technique to multilevel thresholding devoted to
inspection applications,” Pattern Recognition Letters, vol. 21, no. 2, pp. 151–161, 2000.
7 R. Guo and S. M. Pandit, “Automatic threshold selection based on histogram modes and a
discriminant criterion,” Machine Vision and Applications, vol. 10, no. 5-6, pp. 331–338, 1998.
8 N. R. Pal and S. K. Pal, “A review on image segmentation techniques,” Pattern Recognition, vol. 26, no.
9, pp. 1277–1294, 1993.
9 P. K. Sahoo, S. Soltani, A. K. C. Wong, and Y. C. Chen, “A survey of thresholding techniques,”
Computer Vision, Graphics and Image Processing, vol. 41, no. 2, pp. 233–260, 1998.
10 W. Snyder, G. Bilbro, A. Logenthiran, and S. Rajala, “Optimal thresholding-a new approach,” Pattern
Recognition Letters, vol. 11, no. 12, pp. 803–810, 1990.
11 S. Chen and M. Wang, “Seeking multi-thresholds directly from support vectors for image
segmentation,” Neurocomputing, vol. 67, no. 4, pp. 335–344, 2005.
12 L. Chih-Chin, “A novel image segmentation approach based on particle swarm optimization,” IEICE
Transactions on Fundamentals of Electronics, Communications and Computer Sciences, vol. 89, no. 1, pp.
324–327, 2006.
13 R. C. Gonzalez and R. E. Woods, Digital Image Processing, Addison Wesley, Boston, Mass, USA, 1992.
14 L. Gupta and T. Sortrakul, “A gaussian-mixture-based image segmentation algorithm,” Pattern
Recognition, vol. 31, no. 3, pp. 315–325, 1998.
15 A. P. Dempster, A. P. Laird, and D. B. Rubin, “Maximum likelihood from incomplete data via the EM
algorithm,” Journal of the Royal Statistical Society B, vol. 39, no. 1, pp. 1–38, 1997.
16 Z. Zhang, C. Chen, J. Sun, and L. Chan, “EM algorithms for gaussian mixtures with split-and-merge
operation,” Pattern Recognition, vol. 36, no. 9, pp. 1973–1983, 2003.
17 H. Park, S. Amari, and K. Fukumizu, “Adaptive natural gradient learning algorithms for various
stochastic models,” Neural Networks, vol. 13, no. 7, pp. 755–764, 2000.
18 H. Park and T. Ozeki, “Singularity and slow convergence of the em algorithm for gaussian mixtures,”
Neural Processing Letters, vol. 29, no. 1, pp. 45–59, 2009.
19 E. Cuevas, D. Zaldivar, and M. Perez-Cisneros, “Seeking multi-thresholds for image segmentation
with learning automata,” Machine Vision and Applications, vol. 22, no. 5, pp. 805–818, 2011.
20 E. Cuevas, D. Zaldivar, and M. Pérez-Cisneros, “A novel multi-threshold segmentation approach
based on differential evolution optimization,” Expert Systems with Applications, vol. 37, no. 7, pp. 5265–
5271, 2010.
21 G. A. Goldsby, T. J. Kindt, J. Kuby, and B. A. Osborne, Immunology, W. H. Freeman, New York, NY,
USA, 5th edition, 2003.
22 L. N. de Castro and J. Timmis, Artificial Immune Systems: A New Computational Intelligence Approach,
Springer, London, UK, 2002.
23 D. Dasgupta, “Advances in artificial immune systems,” IEEE Computational Intelligence Magazine, vol.
1, no. 4, pp. 40–49, 2006.
24 X. Wang, X. Z. Gao, and S. J. Ovaska, “Artificial immune optimization methods and applications—a
survey,” in Proceedings of the IEEE International Conference on Systems, Man and Cybernetics, (SMC ’04),
pp. 3415–3420, The Hague, The Netherlands, October 2004.
25 L. N. de Castro and F. J. von Zuben, “Learning and optimization using the clonal selection principle,”
IEEE Transactions on Evolutionary Computation, vol. 6, no. 3, pp. 239–251, 2002.
26 G. L. Ada and G. Nossal, “The clonal-selection theory,” Scientific American, vol. 257, no. 2, pp. 50–57,
1987.
27 C. A. C. Coello and N. C. Cortés, “Solving multiobjective optimization problems using an artificial
immune system,” Genetic Programming and Evolvable Machines, vol. 6, no. 2, pp. 163–190, 2005.
28 F. Campelo, F. G. Guimarães, H. Igarashi, and J. A. Ramı́rez, “A clonal selection algorithm for
optimization in electromagnetics,” IEEE Transactions on Magnetics, vol. 41, no. 5, pp. 1736–1739, 2005.
20 Mathematical Problems in Engineering
29 W. Dong, G. Shi, and L. Zhang, “Immune memory clonal selection algorithms for designing stack
filters,” Neurocomputing, vol. 70, no. 4–6, pp. 777–784, 2007.
30 M. Gong, L. Jiao, L. Zhang, and H. Du, “Immune secondary response and clonal selection inspired
optimizers,” Progress in Natural Science, vol. 19, no. 2, pp. 237–253, 2009.
31 V. Cutello, G. Narzisi, G. Nicosia, and M. Pavone, “Clonal selection algorithms: a comparative case
study using effective mutation potentials,” in Proceedings of the 4th International Conference on Artificial
Immune Systems, (ICARIS ’05), C. Jacob et al., Ed., pp. 13–28, August 2005.
32 M. Gong, L. Jiao, and X. Zhang, “A population-based artificial immune system for numerical
optimization,” Neurocomputing, vol. 72, no. 1–3, pp. 149–161, 2008.
33 J. Yoo and P. Hajela, “Immune network simulations in multicriterion design,” Structural and
Multidisciplinary Optimization, vol. 18, no. 2, pp. 85–94, 1999.
34 X. Wang, X. Z. Gao, and S. J. Ovaska, “A hybrid optimization algorithm in power filter design,” in
Proceedings of the 31st Annual Conference of IEEE Industrial Electronics Society (IECON ’05), pp. 1335–
1340, Raleigh, NC, USA, November 2005.
35 X. Xu and J. Zhang, “An improved immune evolutionary algorithm for multimodal function
optimization,” in Proceedings of the 3rd International Conference on Natural Computation (ICNC ’07), pp.
641–646, Haikou, China, August 2007.
36 T. Tang and J. Qiu, “An improved multimodal artificial immune algorithm and its convergence
analysis,” in Proceedings of the 6th World Congress on Intelligent Control and Automation (WCICA ’06),
pp. 3335–3339, Dalian, China, June 2006.
Hindawi Publishing Corporation
Mathematical Problems in Engineering
Volume 2012, Article ID 193246, 14 pages
doi:10.1155/2012/193246
Research Article
Quality Improvement and Robust Design Methods
to a Pharmaceutical Research and Development
Copyright q 2012 B. R. Cho and S. Shin. This is an open access article distributed under the
Creative Commons Attribution License, which permits unrestricted use, distribution, and
reproduction in any medium, provided the original work is properly cited.
Researchers often identify robust design, based on the concept of building quality into products
or processes, as one of the most important systems engineering design concepts for quality
improvement and process optimization. Traditional robust design principles have often been
applied to situations in which the quality characteristics of interest are typically time-insensitive.
In pharmaceutical manufacturing processes, time-oriented quality characteristics, such as the
degradation of a drug, are often of interest. As a result, current robust design models for
quality improvement which have been studied in the literature may not be effective in finding
robust design solutions. In this paper, we show how the robust design concepts can be applied
to the pharmaceutical production research and development by proposing experimental and
optimization models which should be able to handle the time-oriented characteristics. This is
perhaps the first attempt in the robust design field. An example is given, and comparative studies
are discussed for model verification.
1. Introduction
Continuous quality improvement has become widely recognized by many industries as a crit-
ical concept in maintaining a competitive advantage in the marketplace. It is also recognized
that quality improvement activities are efficient and cost-effective when implemented during
the design stage. Based on this awareness, Taguchi 1 introduced a systematic method for
applying experimental design, which has become known as robust design which is often
referred to as robust parameter design. The primary goal of this method is to determine the
best design factor settings by minimizing performance variability and product bias, that is,
the deviation from the target value of a product. Because of the practicability in reducing
2 Mathematical Problems in Engineering
the inherent uncertainty associated with system performance, the widespread application
of robust design techniques has resulted in significant improvements in product quality,
manufacturability, and reliability at low cost. Although the main robust design principles
have been implemented in a number of different industrial settings, our literature study
indicates that robust design has been rarely addressed in the pharmaceutical design process.
In the pharmaceutical industry, the development of a new drug is a lengthy process
involving laboratory experiments. When a new drug is discovered, it is important to design
an appropriate pharmaceutical dosage or formulation for the drug so that it can be delivered
efficiently to the site of action in the body for the optimal therapeutic effect on the intended
patient population. The Food and Drug Administration FDA requires that an appropriate
assay methodology for the active ingredients of the designed formulation be developed and
validated before it can be applied to animal or human subjects. Given this fact, one of the
main challenges faced by many researchers during the past decades is the optimal design
of pharmaceutical formulations to identify better approaches to various unmet clinical
needs. Consequently, the pharmaceutical industry’s large investment in the research and
development R&D of new drugs provides a great opportunity for research in the areas of
experimentation and design of pharmaceutical formulations. By definition, pharmaceutical
formulation studies are mixture problems. These types of problems take into account the
proportions within the mixture, not the amount of the ingredient; thus, the ingredients in such
formulations are inherently dependent upon one another, and consequently experimental
design methodologies commonly used in many manufacturing settings may not be effective.
Instead, for mixture problems, a special kind of experimental design, referred to as a mixture
experiment, is needed. In mixture experiments, typical factors in question are the ingredients
of a mixture, and the quality characteristic of interest is often based on the proportionality of
each of those ingredients. Hence, the quality of the pharmaceutical product is influenced by
such designs when they are applied in the early stages of drug development.
In this paper, we propose a new robust design model in the context of pharmaceutical
production R&D. The main contribution of this paper is twofold. First, traditional experi-
mental design methods have often applied to situations in which the quality characteristics
of interest are typically time-insensitive. In pharmaceutical manufacturing processes, time-
oriented quality characteristics, such as the degradation of a drug, are often of interest,
and these time-oriented data often follow a Weibull distribution. Since it may take a long
time to observe the degradation of a drug product, the concept of censored samples can be
integrated in designing optimal pharmaceutical formulations. In this paper, we develop a
censored sample-based experimental design model for optimal pharmaceutical formulations
by integrating the main robust design principles. Second, we then show how the response
surface methodology, which is a well-established statistical tool, can be integrated with the
proposed censored sample-based robust deign model. Finally, we show how the maximum
likelihood method is implemented in estimating mean and variance of censored Weibull data.
A numerical example is given, and comparison studies for the two estimation methods are
discussed for model verification. This paper is organized as follows. In the next section, we
present a literature review on mixture design and robust design. In Section 3, we describe
our proposed censored robust design model for the optimal design of pharmaceutical
formulations in detail. The maximum likelihood method is then studied, and optimization
models are proposed. In Section 4, we demonstrate our proposed methods using a numerical
example and compare the results under the two different optimization models. In the last
section, we conclude the paper with a discussion of our findings.
Mathematical Problems in Engineering 3
2. Literature Study
In this section, the literature of robust design and mixture designs is discussed.
that estimated empirical models are often subject to random error. In order to obtain a more
precise robust design solution in the presence of the error, Xu and Albin 34 developed a
model which can be resistant to the error by considering all points in the confidence intervals
associated with the estimated model. When multiple quality characteristics are considered,
those characteristics are often correlated. Govindaluri and Cho 23, investigated the effect of
correlations of quality characteristics on robust design solutions, while Egorov et al. 35 and
Kovach et al. 32 studied optimal robust design solutions using the indirect optimization
algorithm and physical programming, respectively. Finally, Shin and Cho 25 studied trade-
off studies on minimizing variance and achieving the predetermined target value.
Scheffe 36 first introduced his theory on the prediction of the responses of mixture
experiments based on their proportions. The theory defines xi as the proportion of ingredient
i in the mixture. Furthermore, the proportionality idea of this theory provides the experiment
with a property in which the proportions of the k ingredients within the mixture must equal
100 percent, as illustrated by the equation ki1 xi x1 x2 · · · xk 1, where xi ≥ 0 for all
i 1, 2, . . . , k. Scheffe 36 employed a simplex lattice design to represent the design points of
the feasible experimental region of the ingredients. The simplex lattice design is defined by
the notation {k, m}, where m 1 defines the number of equally spaced proportion values
from 0 to 1 for each experiment and those proportions are determined by the equation
xi 0, 1/m, 2/m, . . . , 1. All possible combinations of the proportions are used to determine
the design points within the simplex lattice design. In general, the number of design points
in a {k, m} simplex lattice design is defined n k m − 1!/m!k − 1! Scheffe 37 also
modified this simplex lattice design to introduce the simplex centroid design for experiments
that include the overall centroid of the region at the coordinate1/k, 1/k, . . . , 1/k.
Augmented designs of both the simplex lattice and simplex centroid designs exist.
Cornell 38 analyzed both an augmented simplex lattice design and an augmented simplex
centroid design with ten design points each. Applications of mixture experiments revealed
other design possibilities. The most natural obstacle is the limitation on the proportion of a
certain ingredient within a mixture. The limitation could be found in the form of lower, upper,
and both lower and upper bounds or constraints. This led researchers to develop other ways
to obtain design points that are within the feasible region given the constraints. An example
of such models is the extreme vertices design for mixture experiments. First introduced
by Mclean and Anderson 39, extreme vertices designs for mixture problems consider the
extreme points of the irregular polyhedron, formed by constraints in experimental runs, in
addition to the centroids of each facet. The major disadvantage with this design is the possible
large number of design points that can be obtained with the given constraints, specifically as
the number of ingredients increases and the feasible region becomes more complex. Snee and
Marquart 40 presented an algorithm to determine the appropriate subset of design points
when the vertices of the feasible region are too many to handle. They compare the efficiency
of their approach to G- and D-optimal designs, both of which are common techniques used
for determining the appropriate points at which to take observations. Bayesian D-optimal
designs shown by DuMouchel and Jones 41 are a modification of D-optimal designs,
which reduces the dependency of the design on the assumed model. Using such models
as a leverage point, Andere-Rendon et al. 42 investigated the Bayesian D-optimal design
specifically for mixture experiments which include both potential and primary model terms
Mathematical Problems in Engineering 5
in order to form the Bayesian slack variable model. The results favored the Bayesian D-
optimal design with smaller bias errors and better-fitted models. Along the same lines, Goos
and Donev 43 extended the work of Donev 44 with the implementation of D-optimal
designs for blocked mixture experiments. Unlike other research that used orthogonally
blocked experiments see 45, 46, they employed mixture designs that are not orthogonally
blocked and used an algorithm that provided a simpler methodology to construct blocked
mixture designs.
The simplified polynomials, also referred to as canonical polynomials, are widely used
throughout the literature and are embedded in software packages for mixture experiments.
However, these designs have been scrutinized, especially because of their lack of squared
terms. Piepel et al. 47 proposed a partial quadratic mixture model that includes the linear
Scheffé terms but augments them with the appropriate squared or quadratic cross product
terms. Extending from alternative models proposed by Snee and Rayner 48, Cornell
and Gorman 49 explained how highly constrained regions, such as those in mixture
experiments having components with considerably smaller ranges than others, result in
skewed responses, thus creating fitted models that have inherent collinearity. Both models
attempt to modify the scale on the feasible region that results in the experiment’s constraints
in an effort to eliminate the collinearity between components. For other collinearity research
and discussions, refer to the publications of Sengupta et al. 50 and Prescott et al. 51. Other
research publications have employed robust design methodologies for mixture experiments.
Steiner and Hamada 52 modeled a mixture experiment to include the coefficient and
terms that account for the interactions of mixture and controllable process variables and the
interactions of the mixture and noise variables. This model utilizes the Taguchi loss function
1, 2 to reduce the noise variables. The practicality of the design may be questionable as it
does not account for constraints within the system, which opens the opportunity for further
research. A different approach to dealing with noise factors can be studied in Goldfarb et al.
53. Continuing in similar research, Goldfarb et al. 54 introduced a three-dimensional
variance dispersion graph with the purpose of comparing competing mixture experimental
designs based on their prediction variance properties. Goldfarb et al. 55 proposed an
interesting addition to this research by implementing genetic algorithms within an objective
function to minimize the maximum scaled prediction variance in the mixture design region.
This investigation showed that with a few runs of the genetic algorithm, the scaled prediction
variance can be significantly reduced, allowing the experimenter to control noise variables
inherent in the experiment.
3.1. Notations
Notations associated with parameters and variables used in paper are defined as follows:
x x1 , x2 , . . . , xk vector of k control factors,
y vector of output observations,
yi i output observations,
T censored observations,
6 Mathematical Problems in Engineering
Ti i censored observations,
θ vector of parameter θ,
n
m
Ly, T; θ f yi ; θ 1 − F Tj ; θ , 3.1
i1 j1
Mathematical Problems in Engineering 7
n
m
ly, T; θ ln f yi ; θ ln 1 − F Tj ; θ . 3.2
i1 j1
n
ly, T; θ ln f yi ; θ m ln1 − FT ; θ. 3.3
i1
The values of the components of θ that maximize the loglikelihood function will constitute
the maximum likelihood estimates. These are computed as the solutions to the system of
equations:
∂
ly, T; θ 0, 3.4
∂θk
where the θk ’s are the components of the vector of parameters θ. If the underlying
distribution follows a Weibull distribution with parameters α and β, then θ α, β, and
for the n actual observations y1 , y2 , . . . , yn , and m censored observations T1 , T2 , . . . , Tm , the
loglikelihood function is
n
n
β
m
β
l y, T; μ; σ n ln α ln β β − 1 ln yi − α yi − α Tj . 3.5
i1 i1 j1
n
n
β
l y, T; μ; σ n ln α ln β β − 1 ln yi − α yi − αmT. 3.6
i1 i1
The maximum likelihood estimates of α and β are the solutions to the system of equations:
n
n β
− yi − mT β 0,
α i1
3.7
n n
n β
ln yi − α yi ln yi − αmT β ln T 0.
β i1 i1
Equation 3.7 is obtained by taking the derivatives of the loglikelihood function with respect
to α and β. Similarly, a system of equations can be derived using 3.6 for varying censoring
8 Mathematical Problems in Engineering
times. The solutions to 3.7, namely, α are the maximum likelihood estimates of α and
and β,
β, and we use them in estimating the mean and standard deviation as follows:
1
α
μ 1/β Γ 1 ,
β
3.8
2
α α 2/β Γ 1 −μ
2 .
β
k 3.9
s.t. xk 1.
i1
It is noted that the sum of pharmaceutical component proportion is one. By considering the
usual approximation of the Taguchi’s loss function 1, we can also find the solution to the
following optimization model Model 2:
1 σ 2 x
min 1 3 2
x∈Ω μ x
2 μ x
3.10
k
s.t. xk 1.
i1
By inspection, we notice that this function decreases as μx increases and as σx decreases.
Also, a part of the feasibility requirements for these proposed objective functions is that the
mean response is nonzero, that is, μx / 0, which is the case for censored samples where
the objective is to get μx as large as possible. We will demonstrate that both proposed
optimization models yield optimal solutions in the next section.
x1 x2 x3 Observations
1 0 0 30, 30, 27.3130, 30 , 25.3841, 18.4713, 26.3209, 25.1345, 32.6499, 30
0 1 0 30, 20.5247, 9.9338, 22.1908, 24.0196, 16.9328, 18.3113, 21.5325, 21.4749, 17.7806
0 0 1 11.6089, 24.6178, 18.2190, 21.9588, 27.6887, 21.6017, 26.7790, 13.8795, 24.6055, 30
1/2 1/2 0 24.8139, 27.3573, 30 , 23.8765, 30 , 15.8716, 30, 30, 25.4216, 30
1/2 0 1/2 29.3682, 30, 34.0853, 28.2447, 21.2239, 19.7356, 30, 23.1654, 20.4046
0 1/2 1/2 23.8148, 14.5317, 25.6167, 23.4480, 30, 18.7073, 30, 41.6825, 24.0814, 14.2608
1/6 1/6 2/3 30, 27.6718, 30, 27.1105, 18.4573, 28.0881, 30, 28.7328, 21.7883, 23.6375
1/6 2/3 1/6 14.8984, 19.5179, 27.8163, 24.6057, 30, 26.3950, 28.0956, 18.4866, 18.7682, 27.7717
2/3 1/6 1/6 23.0132, 25.5782, 30, 30, 28.1479, 30, 30, 30, 19.7557, 30
2/3 1/3 1/3 30, 28.9475, 19.3873, 30, 27.4342, 28.5594, 30, 30, 17.7312, 21.8007
10 samples are subject to each design point, and the experiment is run for 30 hours. The
manufacturer believes that the degradation times follow a Weibull distribution. Notice that
this is a censoring problem because the experiment is terminated after 30 hours, and for the
rest of the samples denoted by 30 , we only know that their degradation times exceed 30
hours but we do not know their actual times. Using the functions in the methods of Sections
3.2 and 3.3, estimates of the means and standard deviations are obtained for each design
point. Table 2 shows the experimental design together with the observed responses that have
failed. Note that the number of observations per design point is not the same throughout the
experiment.
In Table 3, we show the estimates of the mean and standard deviation obtained by
our proposed algorithm and by the method of ML estimation. Using the MLE algorithm, the
response surfaces for the mean and standard deviation are, respectively, found as
x 27.222x1
μ 18.601x2 2.190x3 − 0.318x1 x2 5.680x1x 3 14.908x2 x3 ,
4.1
σ x 4.884x1 5.088x2 5.300x3 − 0.568x1 x2 − 1.917x1x 3 7.957x2 x3 .
10 Mathematical Problems in Engineering
30
25
20
Standard deviation
8
Mean
15
7
10
6
5
5 1
0
1 4 0.8
1 0.6
0.5 0.8 0.4
x1 0 0.6
0.4 0.2 x1
0.4 0.2
0
1 0.8 0.6 x2 0.2 0 0
x2
30 30
25 25
20 20
Mean
Mean
15 15
10 10
5 1 5 1
0 0.5 0 0.5
0 0.2 x3 0 0.2 x3
0.4 0.6 0 0.4 0.6 0
0.8 1 0.8 1
x1 x1
7.5
Standard deviation
7
6.5
30
6
25
5.5
20
Mean
5
15
0 4.5
10 0
5
0.5
0 x3 x2 0.5
1 0.8 0.8 1
0.6 0.4 1 1 0.4 0.6
0.2 0 0 0.2 x3
x2
Using 4.1, Figure 1 illustrates surface plots of the process mean and standard deviation
associated with input control factors x1 , x2 , x3 . In order to represent tradeoffs between the
mean and standard deviation, a criterion space is demonstrated in Figure 2.
Table 4 shows the results obtained using the two different optimization models we
studied earlier. For this particular example, Model 1 is more effective in increasing the mean
value, while Model 2 turns out be to more effective in decreasing standard deviation. The
basic goal of both optimization models is to obtain the optimal factor level settings which
maximize the mean and minimize the standard deviation. We recommend practitioners use
Mathematical Problems in Engineering 11
Standard deviation
6.5
5.5
4.5
0 5 10 15 20 25 30
Mean
both optimization models and pick one of them, depending on their priority of mean and
standard deviation.
5. Conclusions
Robust design has been demonstrated to be a successful process improvement methodology
for diverse system engineering problems. However, its ability to provide sound solutions
for time-oriented data is currently limited since the traditional robust design method is
mainly applicable to time-insensitive data, and an appropriate robust design method for
time-oriented data has not been reported. This paper has proposed new robust design
experimental and optimization models for time-oriented data by developing a set of methods
in three separate phases, namely, modeling, estimation, and optimization phases. More
specifically, this paper has proposed a censored sample-based robust design model by
developing a censored experimental design framework, a censored maximum likelihood
estimation method in the context of robust design, and optimization models for the censored
robust design. For verification, a proposed optimization model has been compared with
the traditional Taguchi optimization model. The proposed experimental methodology is
particularly useful for experiments with time-oriented pharmaceutical characteristics, such
as the degradation of a drug, with an unequal number of observations at each design point.
With suitable modifications, our proposed methodology could be extended to the case in
which multiple quality characteristics are often correlated under study. Another area for
12 Mathematical Problems in Engineering
future study includes the development of optimal designs, known as computerized designs,
for the case in which physical experimental constraints are imposed.
Acknowledgment
This work was supported by the 2011 Inje University research grant.
References
1 G. Taguchi, Introduction to Quality Engineering, Asian Productivity Organization, UNIPUB/Kraus
International, White Plains, NY, USA, 1986.
2 G. Taguchi, Systems of Experimental Design: Engineering Methods to Optimize Quality and Minimize Cost,
UNIPBUK/Kraus International, White Plains, NY, USA, 1987.
3 D. M. Steinberg and D. Bursztyn, “Noise factors, dispersion effects, and robust design,” Statistica
Sinica, vol. 8, no. 1, pp. 67–85, 1998.
4 R. V. León, A. C. Shoemaker, and R. N. Kacker, “Performance measures independent of adjustment.
An explanation and extension of Taguchi’s signal-to-noise ratios,” Technometrics, vol. 29, no. 3, pp.
253–285, 1987.
5 G. E. P. Box, “Signal-to-noise ratios, performance criteria, and transformations,” Technometrics, vol. 30,
no. 1, pp. 1–17, 1988.
6 G. E. P. Box, S. Bisgaard, and C. Fung, “An explanation and critique of Taguchi’s contributions to
quality engineering,” Quality and Reliability Engineering International, vol. 4, no. 2, pp. 123–131, 1988.
7 V. N. Nair, “Taguchi’s parameter design: a panel discussion,” Technometrics, vol. 34, no. 2, pp. 127–161,
1992.
8 K. L. Tsui, “An overview of Taguchi method and newly developed statistical methods for robust
design,” IIE Transactions, vol. 24, pp. 44–57, 1992.
9 G. G. Vining and R. H. Myers, “Combining Taguchi and response surface philosophies: a dual res-
ponse approach,” Journal of Quality Technology, vol. 22, pp. 38–45, 1990.
10 E. Del Castillo and D. C. Montgomery, “A nonlinear programming solution to the dual response
problem,” Journal of Quality Technology, vol. 25, no. 3, pp. 199–204, 1993.
11 K. A. F. Copeland and P. R. Nelson, “Dual response optimization via direct function minimization,”
Journal of Quality Technology, vol. 28, no. 3, pp. 331–336, 1996.
12 R. Khattree, “Robust parameter design: a response surface approach,” Journal of Quality Technology,
vol. 28, no. 2, pp. 187–198, 1996.
13 K. J. Kim and D. K. J. Lin, “Dual response surface optimization: a fuzzy modeling approach,” Journal
of Quality Technology, vol. 30, no. 1, pp. 1–10, 1998.
14 D. K. J. Lin and W. Tu, “Dual response surface optimization,” Journal of Quality Technology, vol. 27, no.
1, pp. 34–39, 1995.
15 J. S. R. Jayaram and Y. Ibrahim, “Multiple response robust design and yield maximization,” Interna-
tional Journal of Quality and Reliability Management, vol. 6, pp. 826–837, 1999.
16 B. R. Cho, Y. J. Kim, D. L. Kimbler, and M. D. Phillips, “An integrated joint optimization procedure for
robust and tolerance design,” International Journal of Production Research, vol. 38, no. 10, pp. 2309–2325,
2000.
17 Y. J. Kim and B. R. Cho, “Economic integration of design optimization,” Quality Engineering, vol. 12,
no. 4, pp. 561–567, 2000.
18 Y. J. Kim and B. R. Cho, “Determining the optimum process mean for a skewed process,” International
Journal of Industrial Engineering, vol. 10, no. 4, pp. 555–561, 2003.
19 R.-X. Yue, “Model-robust designs in multiresponse situations,” Statistics & Probability Letters, vol. 58,
no. 4, pp. 369–379, 2002.
20 C. Park and B. R. Cho, “Development of robust design under contaminated and non-normal data,”
Quality Engineering, vol. 15, no. 3, pp. 463–469, 2003.
21 G. Miro-Quesada and E. Del Castillo, “A dual response approach to the multivariate robust parameter
design problem,” Technometrics, vol. 46, no. 2, pp. 176–187, 2004.
22 B. R. Cho and C. Park, “Robust design modeling and optimization with unbalanced data,” Computers
and Industrial Engineering, vol. 48, no. 2, pp. 173–180, 2005.
Mathematical Problems in Engineering 13
23 S. M. Govindaluri and B. R. Cho, “Integration of customer and designer preferences in robust design,”
International Journal of Six Sigma and Competitive Advantage, vol. 1, no. 3, pp. 276–294, 2005.
24 S. Shin and B. R. Cho, “Bias-specified robust design optimization and its analytical solutions,”
Computers and Industrial Engineering, vol. 48, no. 1, pp. 129–140, 2005.
25 S. Shin and B. R. Cho, “Studies on a biobjective robust design optimization problem,” IIE Transactions,
vol. 41, no. 11, pp. 957–968, 2009.
26 S. B. Lee, C. Park, and B. R. Cho, “Development of a highly efficient and resistant robust design,” Inter-
national Journal of Production Research, vol. 45, no. 1, pp. 157–167, 2007.
27 W. R. Myers, W. A. Brenneman, and R. H. Myers, “A dual-response approach to robust parameter
design for a generalized linear model,” Journal of Quality Technology, vol. 37, no. 2, pp. 130–138, 2005.
28 T. J. Robinson, S. S. Wulff, D. C. Montgomery, and A. I. Khuri, “Robust parameter design using
generalized linear mixed models,” Journal of Quality Technology, vol. 38, no. 1, pp. 65–75, 2006.
29 J. Kovach and B. R. Cho, “A D-optimal design approach to robust design under constraints: a new
Design for Six Sigma tool,” International Journal of Six Sigma and Competitive Advantage, vol. 2, no. 4,
pp. 389–403, 2006.
30 J. Kovach and B. R. Cho, “Development of a D-optimal robust design model for restricted experi-
ments,” International Journal of Industrial Engineering, vol. 14, pp. 117–128, 2007.
31 J. Kovach and B. R. Cho, “Solving multiresponse optimization problems using quality function-based
robust design,” Quality Engineering, vol. 20, no. 3, pp. 346–360, 2008.
32 J. Kovach, B. R. Cho, and J. Antony, “Development of an experiment-based robust design paradigm
for multiple quality characteristics using physical programming,” International Journal of Advanced
Manufacturing Technology, vol. 35, no. 11-12, pp. 1100–1112, 2008.
33 H. Ginsburg and I. Ben-Gal, “Designing experiments for robust-optimization problems: the Vs-opti-
mality criterion,” IIE Transactions, vol. 38, no. 6, pp. 445–461, 2006.
34 D. Xu and S. L. Albin, “Robust optimization of experimentally derived objective functions,” IIE Tran-
sactions, vol. 35, no. 9, pp. 793–802, 2003.
35 I. N. Egorov, G. V. Kretinin, I. A. Leshchenko, and S. V. Kuptzov, “Multi-objective approach for robust
design optimization problems,” Inverse Problems in Science and Engineering, vol. 15, no. 1, pp. 47–59,
2007.
36 H. Scheffé, “Experiments with mixtures,” Journal of the Royal Statistical Society. Series B, vol. 20, pp.
344–360, 1958.
37 H. Scheffé, “The simplex-centroid design for experiments with mixtures,” Journal of the Royal Statistical
Society. Series B, vol. 25, pp. 235–263, 1963.
38 J. A. Cornell, Experiments with Mixtures: Designs, Models, and the Analysis of Mixture Data, John Wiley
& Sons Inc., New York, NY, USA, 1981.
39 R. A. McLean and V. L. Anderson, “Linear estimates of parameters in the extreme value distribution,”
Technometrics, vol. 8, pp. 454–477, 1966.
40 D. D. Snee and D. W. Marquardt, “Extreme vertices designs for linear mixture models,” Technometrics,
vol. 16, no. 3, pp. 399–408, 1974.
41 W. DuMouchel and B. Jones, “A simple bayesian modification of d-optimal designs to reduce
dependence on an assumed model,” Technometrics, vol. 36, pp. 37–47, 1994.
42 J. Andere-Rendon, D. C. Montgomery, and D. A. Rollier, “Design of mixture experiments using
bayesian D-optimality,” Journal of Quality Technology, vol. 29, no. 4, pp. 451–463, 1997.
43 P. Goos and A. N. Donev, “The D-optimal design of blocked experiments with mixture components,”
Journal of Quality Technology, vol. 38, no. 4, pp. 319–332, 2006.
44 A. N. Donev, “Design of experiments with both mixture and qualitative factors,” Journal of the Royal
Statistical Society. Series B, vol. 51, no. 2, pp. 297–302, 1989.
45 A. K. Nigam, “Correction to: “Block designs for mixture experiments”,” The Annals of Statistics,vol. 4,
no. 6, pp. 1294–1295, 1976.
46 P. W. M. John, “Experiments with mixtures involving process variables,” Tech. Rep. 8, Center for Sta-
tistical Sciences, University of Texas, Austin, Tex, USA, 1984.
47 G. F. Piepel, J. M. Szychowski, and J. L. Loeppky, “Augmenting Scheffé linear mixture models with
squared and/or crossproduct terms,” Journal of Quality Technology, vol. 34, no. 3, pp. 297–314, 2002.
48 R. D. Snee and A. A. Rayner, “Assessing the accuracy of mixture model regression calculations,”
Journal of Quality Technology, vol. 14, no. 2, pp. 67–79, 1982.
14 Mathematical Problems in Engineering
49 J. A. Cornell and J. W. Gorman, “Fractional design plans for process variables in mixture experi-
ments,” Journal of Quality Technology, vol. 16, no. 1, pp. 20–38, 1984.
50 T. K. Sengupta, R. K. Nandy, S. Mukhopadhyay et al., “Mixture models based on homogeneous
polynomials,” Journal of Statistical Planning and Inference, vol. 71, no. 1-2, pp. 303–311, 1998.
51 P. Prescott, N. R. Draper, A. M. Dean, and S. M. Lewis, “Mixture experiments: ILL-conditioning and
quadratic model specification,” Technometrics, vol. 44, no. 3, pp. 260–268, 2002.
52 S. H. Steiner and M. Hamada, “Making mixtures robust to noise and mixing measurement errors,”
Journal of Quality Technology, vol. 29, no. 4, pp. 441–450, 1997.
53 H. B. Goldfarb, C. M. Borror, and D. C. Montgomery, “Mixture-process variable experiments with
noise variables,” Journal of Quality Technology, vol. 35, no. 4, pp. 393–405, 2003.
54 H. B. Goldfarb, C. M. Borror, D. C. Montgomery, and C. M. Anderson-Cook, “Three-dimensional vari-
ance dispersion graphs for mixture-process experiments,” Journal of Quality Technology, vol. 36, no. 1,
pp. 109–124, 2004.
55 H. B. Goldfarb, C. M. Borror, D. C. Montgomery, and C. M. Anderson-Cook, “Using genetic algo-
rithms to generate mixture-process experimental designs involving control and noise variables,” Jour-
nal of Quality Technology, vol. 37, no. 1, pp. 60–74, 2005.
Hindawi Publishing Corporation
Mathematical Problems in Engineering
Volume 2012, Article ID 569429, 13 pages
doi:10.1155/2012/569429
Research Article
A Label Correcting Algorithm for Partial
Disassembly Sequences in the Production Planning
for End-of-Life Products
Copyright q 2012 Pei-Fang Jennifer Tsai. This is an open access article distributed under the
Creative Commons Attribution License, which permits unrestricted use, distribution, and
reproduction in any medium, provided the original work is properly cited.
Remanufacturing of used products has become a strategic issue for cost-sensitive businesses.
Due to the nature of uncertain supply of end-of-life EoL products, the reverse logistic
can only be sustainable with a dynamic production planning for disassembly process. This
research investigates the sequencing of disassembly operations as a single-period partial
disassembly optimization SPPDO problem to minimize total disassembly cost. AND/OR graph
representation is used to include all disassembly sequences of a returned product. A label
correcting algorithm is proposed to find an optimal partial disassembly plan if a specific reusable
subpart is retrieved from the original return. Then, a heuristic procedure that utilizes this
polynomial-time algorithm is presented to solve the SPPDO problem. Numerical examples are
used to demonstrate the effectiveness of this solution procedure.
1. Introduction
Product recovery, or remanufacturing, has been considered as one of the most profitable
options in dealing with the end-of-life EoL products. The benefit of product recovery is
especially more attractive when a facility is capable of performing both manufacturing and
remanufacturing processes, and the coordination of these two processes can be included in
the production planning and scheduling. Griese et al. 1 discussed the economic challenges
for reuse and the main technical obstacles in three product categories: medical equipment,
automotive electronics, and computers. They argued that, for personal computers, reuse and
repair appeared to have more potential than pure recycling materials. Similar benefit was
confirmed by Grenchus et al. 2 from the practice at the IBM Endicott asset recovery center.
They found that, with little disassembly effort, functional parts that were recoverable had
more resale value than plain material recovery.
2 Mathematical Problems in Engineering
Commercial
Design
use
Raw Customer
material Build Distribute
use
Reuse Resell
Service
Reuse
Recycling Remanufacturing
Collectors
Disassembling
Incineration
Disposal
Landfill
Figure 1: Integrated supply chain for original equipment manufacturers. Adopted from 4.
For any original equipment manufacturer OEM with the capacity in performing
assembly and disassembly operations, the retrieved components from returned products can
be integrated with their forward production. As shown in Figure 1, a complete product life
cycle is defined by forward logistics solid line and reverse logistics dotted line. Informa-
tion bold line from disassembly operation can further assist in design and manufacturing
new products. Retrieving spare parts from returned products has been one of the most
prominent strategies. Fleischmann 3 observed that, when the returned used equipments
were integrated into the spare part planning, a push policy in which returned equipments
were tested and dismantled as soon as available can achieve higher service level in IBM’s
spare part management.
It remains as one of the biggest challenges to develop the techniques in production
planning system for product recovery to be sustainable 5. For regular assembly, demand
can be determined in advance and hence the required resources can be planed and scheduled
along the time horizon. However, for disassembly process planning, variation in quantity
and quality of returned products is so huge that it fails to fit into any available planning
scheme. Even when the demand for remanufactured products is known with a set of available
returned products, it is still challenging to decide how these products should be dismantled
to minimize the disassembly effort for those refurbished products. Kasmara et al. 6 used an
integer programming model that included sales and returns in each period with the objective
function as maximization of profits. Clegg et al. 7 presented a linear programming model
of production planning for both new and remanufactured products. Some studies focused on
the effect of average flowtime for both assembly and disassembly operations under different
scenarios in planning mixes 8–13.
The ability to salvage the value of these returned products relies both on the disassem-
bly capacity and the ability to find the most cost-effective disassembly sequences to retrieve
valuable parts. This research is motivated to find disassembly sequences with minimum
operation costs in the production planning for EoL products. A single period planning is
considered due to the inherent fluctuation in the demand and supply of EoL products in
different periods. Moreover, a partial disassembly policy is considered for better profit in
product recovery.
Mathematical Problems in Engineering 3
V5 : extension card
V10 : power supply 0
V4 : extension card E2, 3 : insert E0, 10
V3 : extension card d e f h E2, 4 : insert
10 E0, 5
E2, 5 : insert
5 E0, 11 E0, 3
10 E0, 3 : screw 5
4
c E0, 4 : screw 11
V2 : system board 3 m E0, 4 E2, 5
b V1 : drive case E0, 5 : screw E0, 1
E0, 2
11 E0, 2 : screws E2, 3
a 6 E1, 6 3
g
z E0, 11 : screws E1, 7 1
V11 : drive bay 2 1
y E1, 7 : screws
E1, 9 : screws 7 2 4
n
k 6 7 i E1, 6 : screws E1, 8 E E2, 4
1, 9
0 E1, 8 : screws 8
V0 : main case l 8 9 j x 9
E0, 10 : screws
E0, 1 : screws
V7 : floppy drive
V6 : hard drive V8 : sucket V8 : sucket
(a) (b)
Figure 2: a The structure and b component-fastener graph for a partially disassembled PC 24.
ABCDEF E
F D
1
1 2
ABCDE ABCDF
2 B A
3 4 3 C
A = 10
ABCD (a)
B = 11 12 4 11
C = 12
D = 13 1 2 3 4 5 6 7 8 9 10 11 12 13
5 1 −1 −1 . . . . . . . . . . .
E = 14 13
F = 15 2 1 . −1 . . . . . . . . −1 .
3 . 1 . −1 . . . . . . −1 . .
4 . . 1 1 −1 . . . . . . . −1
ABF BCD
5 . . . . . −1 . . . . 1 . .
5 6
6 . . . . 1 . −1 . . . . 1 .
7 . . . . . 1 . −1 . . . . 1
6 7
8 . . . . . . . . −1 . . 1. .
9 . . . . . . 1 . . −1 1 . 1
10 . . . . 1 . . 1 1 . . . .
AB AE CD 11 . . . . . . 1 1 . . . . .
12 . . . . . . . . . 1 . . .
7 8 9
13 . . . . . . . . . 1 . . .
8 9 10
14 . 1 1 . . . . . 1 . . . .
15 1 . . 1 . 1 . . . . . . .
(b) (c)
Figure 3: a The structure, b AND/OR disassembly tree, and c transition matrix for a ballpoint 21.
between subassemblies and parts where separation operations are possible. A component-
fastener graph can be used to represent the assembly relationship 23, 24. As shown in
Figure 2, if two components are attached or joined by fasteners, then these two components
are connected by an undirected edge in the component-fastener graph.
A disassembly AND/OR graph is another useful representation for all possible
disassembly sequences. An AND/OR graph is a directed hypergraph where subassemblies
are represented as nodes. If more than one subassembly/part can be separated from a
parent assembly in one disassembly operation, a hyperarc AND arc is used to indicate
this operation and connects the parent node to all child nodes. Otherwise, a directed arcs
OR arcs are used. Figure 3a illustrates the drawing of a ballpoint product with associated
AND/OR graph in Figure 3b. The assembly {ABCDE} yields two subassemblies {BCD}
and {AE} through the disassembly operation {12}, and this operation is represented as an
AND arc “∪” arc. Moreover, this hypergraph is compact which requires a reduced number
of nodes/arcs to enumerate all partial sequences of disassembly operations 25.
This AND/OR graph can also be represented completely via a transition matrix T .
Suppose I be the set of subassemblies/subparts and J be the set of operations, the element Tij
has a value of 1 if a subassembly i ∈ I is released by some operation j ∈ J or −1 if subassembly
Mathematical Problems in Engineering 5
SPPDO
Minimize cj · yj , 3.1
j∈J
Subject to Tij · yj ≥ bi , ∀i ∈ I0 ∪ If , 3.2
j∈J
Tij · yj ≥ 0, ∀i ∈ I \ I0 ∪ If , 3.3
j∈J
This SPPDO model is a generalized minimum cost flow problem where the arc in a graph
include both hyperarcs AND arcs and regular directed arcs OR arcs. The constraint set
3.2 is to ensure that the demands of reusable subparts are fulfilled or the required quantities
for original returns are still sufficient. The constraint sets 3.3 and 3.4 are nonnegative
constraints on the resulting quantity of intermediate subassemblies/parts and the number
of operations needed. It is worth noting that this AND/OR graph is acyclic, and the sum of
degrees from all nodes might not be zero. Further note that, since the transition matrix T does
6 Mathematical Problems in Engineering
not have the property of total unimodularity, this SPPDO formulation can only be solved as
a pure integer programming IP problem.
In the literature, a searching algorithm for an AND/OR graph with different inter-
pretation is available, but it is not applicable to solve the SPPDO problem directly. AND/OR
graph is often used to represent a problem-solving process which transforms the original
problem into several subproblems 28, 29. Each node represents a distinct problem. The node
represents the original problem is referred to as the starting node or root node. A terminal
node, or leaf node, in this graph represents a problem whose solution is either known to exist
or not to exist. A directed arc is linked from a node problem to its associated successive
nodes subproblems. For an OR arc, the problem is solved when the immediately successive
subproblem is solved. If the problem is linked by an AND arc, the problem is only solved
when all the successive subproblems are solved. Hence, the solution for the original problem
is to search for a tree that connects the root node with terminal nodes only 28.
Zhang and Kuo 23 had extended this searching algorithm in finding a solution
tree from an AND/OR graph to obtain the optimal assembly sequence toward the final
product. Even with the assumption of reversible operation, the sequence obtained from
the solution tree might not be directly used for generating disassembly sequence if partial
disassembly is allowed. The main reason is that the disassembly level needs to be identified
if full disassembly is not desirable. The selection in a proper set of terminal nodes can
be combinatorial 25. Considering the example as in Figures 4a and 4b, the searching
algorithm can be used to find the best trees for the assembly operations from leaf nodes to
the root node as in Figure 5. But to retrieve a stick from the returned problem, the optimal
sequences for partial disassembly operations {1}, {8} or {4}, {7} can only be obtained if
and only if the set of terminal nodes selected are {7, 9, 10} or {6, 10, 12}, respectively.
Initialization
For each source node k in k ∈ Is , set the minimum cost dk 0, the predecessor set pk φ, and
update the set of labelled node L L ∪ {k}. For node k in I \ Is , set the minimum cost dk ∞,
the predecessor set pk φ. Select the first labelled node k in L.
Step 1. Determine the set Sk for unlabelled nodes that are immediately connected from node
k; that is, Sk {n | n ∈ tj, j ∈ Fk } \ L. If Sk is not empty, then go to Step 2. Otherwise, go to
Step 3.
2 3
[5,{1}] [0, φ]
5 6 8 9
7 10
8
5 4
6 [6,{3}] 7
[7,{1}] 11
[7,{1,2}] [2,{3}]
13 14
15
[7,{1,3}]
9 10
12
Handle
Cap Stick Receptacle
[5,{1}]
[5,{1}] [2,{3}] [6,{7}]
b
Figure 4: a The structure and b the AND/OR disassembly tree for a simple product 25.
1 1
3
2
8
7
7
6
14
13
9 10 12 9
11 10 11 12
a b
Figure 5: Alternated trees of disassembly sequence to remove the stick for the simple product in a 25.
2 Determine the set Ui for unlabelled nodes that are immediately connected toward
node i; that is, Ui {n|n ∈ fj, j ∈ Ri } \ L. If Ui is empty, make node i as labelled
and update the set of labelled nodes L L ∪ {i}.
Step 3. Mark the current node k as solved. Move to the next unsolved node in L as new node
k, go to Step 1. If all nodes are solved, stop.
8 Mathematical Problems in Engineering
If only original returns can be dismantle for reusable subparts, we have Is I0 . For
any solved node i, the di from the label Li di , pi represents the minimum disassembly cost
to retrieve one subpart i directly from a starting assembly. Moreover, the complexity of this
procedure is polynomial bound by the number of nodes; that is, O|I|2 30.
Lemma 3.1. If only one target subpart has positive demand, the label correcting algorithm solves the
SPPDO problem.
Proof. In the SPPDO formulation, it is assumed that target subparts are retrieved from
returned products directly. If only one target subpart i has positive demand, it is equivalent
to find an optimal sequence of disassembly operations with the minimum total cost to reach
the destination node i from the source node. This minimum cost is also defined as di in the
label Li di , pi for node i, i ∈ If .
Lemma 3.2. If more than one target subpart has positive demands, then the corresponding di from
the label correcting algorithm for associated nodes forms an upper bound for the optimal solution in
the SPPDO problem. That is, suppose yj∗ , j ∈ J, be the optimal solution for the SPPDO problem, then
one has j∈J cj · yj∗ ≤ i∈If bi · di .
Proof. Suppose two target subparts, say l and m, l / m, have positive demands. Let Pl
{jl1 , jl2 , . . . , jlp } and Pm {jm1 , jm2 , . . . , jmp } be the optimal disassembly sequences for node
l and node m, respectively. Without loss of generality, we assume that bl ≥ bm > 0. Suppose
there exists some node k ∈ tj , j ∈ Pl and also k ∈ fj , j ∈ Pm . Then the shortest disassembly
path from the source node to node k is overlapped in the optimal disassembly sequences from
the source node to node l and node m. Hence, an upper bound for the optimal disassembly
cost is j∈J cj · yj∗ ≤ bl · dl bm · dm − bm · dk .
Moreover, it can be concluded that j∈J cj · yj∗ i∈If bi · di if and only if each original
return can be retrieved for no more than one target subpart only.
Initialization
Apply the label correcting algorithm with Is Io to obtain the initial labels Li di , pi for all
subparts i, i ∈ I. Set the variable yj 0, j ∈ J. Let xi be the quantity of subassembly i available
for further dismantle, set xi 0, i ∈ I, and xi |bi |, i ∈ I0 . Note that i∈I0 |bi | ≥ i∈If bi .
Phase 1 Path construction. Select a target node i which has the maximum total potential cost
in unfulfilled demand; that is, i argk∈If max{bk − xk · dk | bk > xk }. Break ties arbitrarily.
Obtain the minimum cost disassembly sequences, Pi {ji1 , ji2 , . . . , jip }, which forms a
directed path from source node s, s fji1 toward the demand node i; i ∈ tjip . Break ties
arbitrarily. Find the maximum flow for this path, Δ min{xs , bi − xi }.
Mathematical Problems in Engineering 9
For each arc j, j ∈ Pi , start from the first arc, perform the following updates sequen-
tially.
yj yj Δ,
xk xk − Δ, if k ∈ f j , 3.5
xk xk Δ, if k ∈ t j .
Phase 2 Termination test. Update the total disassembly cost z z Δ · di and check for any
unfulfilled demand; that is, xk < bk , k ∈ If . If all demands are fulfilled, then stop. The current
solution of yj , j ∈ J, is feasible for the SPPDO problem.
Otherwise, update the set of source nodes to include intermediate nodes with positive
quantity; that is, Is I0 ∪ {k}, if xk > 0 for k ∈ I \ {I0 ∪ If }. Update labels Li di , pi for all
subparts i, i ∈ I using the label correcting algorithm in Section 3.3 with the new set of source
nodes Is . Go to Phase 1.
Lemma 3.3. The objective value of solution obtained from this heuristic procedure, referred to as zH ,
is at least as good as that obtained from the label correcting algorithm, referred to as zLC .
Proof. In Phase 2, the labels for subparts are altered only when there exists some intermediate
subassembly with positive quantity to be further dismantled for those target subparts
with lower costs. Otherwise, the initial labels remain unchanged. So, the contribution of
disassembly cost for a target subpart in zH is no higher than that in zLC .
Furthermore, let z∗ j∈J cj · yj be the optimal solution for the SPPDO problem and
zLC i∈If bi · di . We have z∗ ≤ zH ≤ zLC from Lemmas 3.2 and 3.3.
4. Numerical Examples
In this section, the simple product in Figure 4a is used to demonstrate how the heuristic
procedure in Section 3.4 works to generate partial disassembly sequences for the SPPDO
problem. There are totally twelve different subassemblies, I {1, . . . , 12}. The original
return is represented by node {1}, and reusable subparts are nodes {9}, {10}, {11}, {12} for
cap, stick, receptacle, and handle, respectively, that is, I0 {1} and If {9, 10, 11, 12}. A total
of fifteen disassembly operations can be used to dismantle this product, J {1, . . . , 15} with
the associated costs C {5, 7, 7, 5, 6, 7, 2, 2, 7, 6, 2, 1, 4, 4, 1}.
The construction of labels using the label correcting algorithm is shown in Figure 6.
The minimum disassembly cost to retrieve node {10} stick is 7 with two alternative optimal
disassembly sequences: P10 {4, 7} or P10 {1, 8}. This optimal solution is consistent with
observation in Section 3.2.
Next, we demonstrate the use of heuristic algorithm with the consideration of the
following demands for target subparts: three caps node{9} and one stick node{11}; that
is, b9 3. b11 1. It implies that the supply of original returns should be at least four; that
is, b1 −4. In the initialization step, all variables x, y are set to zero except x1 4 and the
initial labels are obtained from Figure 6. In Phase 1, since node {9} cap has a higher total
unfulfilled cost than node {11} stick, node {9} is selected along with the associated directed
10 Mathematical Problems in Engineering
Initialization Iteration 1:
1 1
Cap Stick Receptacle Handle Cap Stick Receptacle Handle
C2 = 7 C3 = 7
2 3 2 3
[∞, φ] [∞, φ] [5,{1}], labeled [5,{1}], labeled
8 8
5 4 5 4
6 [∞, φ] 7 6 [∞, φ] 7
[∞, φ] [7,{1}]
[∞, φ] [∞, φ] [7,{1}] [7,{1}]
[∞, φ] [7,{1}]
9 12 9 12
10 11 Handle
Cap
10 11 Handle
Cap
Stick Receptacle
[∞, φ] Stick Receptacle
[5,{1}]
[∞, φ] [5,{1}] [∞, φ] [∞, φ]
[∞, φ] [∞, φ]
Iteration 2: Iteration 3:
1 1
Cap Stick Receptacle Handle Cap Stick Receptacle Handle
C2 = 7 C3 = 7 C2 = 7 C3 = 7
2 3 2 3
[5,{1}], labeled, solved [5,{1}], labeled, solved [5,{1}], labeled, solved
[5,{1}], labeled C8 = 2
C5 = 6 5
6 7 C7 = 2 C5 = 6 5 6 7 8 9 C =6
10 10
8 C9 = 7 8
5 C6 = 7 4 5 C6 = 7 4
[11,{2}] 7 [11,{2,3}], labeled
6 6 7
[7,{1}], [7,{1}],
labeled [7,{1,2}], labeled [7,{1}] labeled [7,{1,2}], labeled [7,{1,3}], labeled
[7,{1}] [7,{1}],
labeled
9 12 9 12
Cap
10 11 Handle 10 11 Handle
Cap
[5,{1}]
Stick Receptacle Stick Receptacle
C3 = 7
C2 = 7
2 3
9 12
Cap 10 11 Handle
Figure 6: Brief illustration of stepwise label correcting algorithm for the product structure as in Figure 5a.
path P9 {1}. Calculate the flow Δ min{x1 , b9 − x9 } min{4, 3 − 0} 3 on this path.
Since P9 has only one arc j 1, update the solution y1 3, x1 4 − 3 1, x3 3, and x9 3.
Continue to Phase 2 to check for the termination criteria. Update the current total
disassembly cost z z Δ · d9 0 3 ∗ 5 15. There still exists unfulfilled demand since
x11 0 < b11 . So, first update the set of source nodes Is {1, 3} to include the intermediate
node {3} as x3 3 > 0. Then labels for subparts are updated using the label correcting
Mathematical Problems in Engineering 11
C1 = 5 4 C4 = 5
1 2 3
C2 = 7 C3 = 7
2 3
[5,{1}] [0, φ]
C7 = 2 C10 = 6
5 6 C8 = 2 8 9
7 10
C5 = 6 C6 = 7 C9 = 7 8
5
4
[6,{3}]
[7,{1}] 6 7
12 14
C14 = 4
13
C12 = 1
C13 = 4 15
C15 = 1
9 12
10
11 Handle
Cap
[5,{1}]
[5,{1}] Stick Receptacle
[2,{3}] [6,{7}]
Figure 7: Updated labels with an additional source node {3} as in the shaded area.
algorithm. It is worth noting that, as shown in Figure 7, not every label is changed, and the
effected nodes with associated arcs are shown in bold in the shaded area.
In the second iteration, select node {11} to fulfil the remaining demand b11 1 with
the disassembly path P11 {8, 14}. Calculate the maximum flow Δ min{x3 , b11 − x11 }
min{3, 1−0} 1 and update the solution starting from the first arc j 8: y8 1, x3 3−1 2,
x7 1, and x10 1. Then update for the second arc j 14: y14 1, x7 1 − 1 0, x11 1,
and x12 1. This procedure terminates since all demands are fulfilled with the total cost
z 15 1 ∗ 6 21. The resulting quantities of subparts are x1 1, x3 2, x9 3, x11 1, and
x12 1, and the required disassembly operations are y1 3, y8 1 and y14 1 for three caps
and one stick.
5. Conclusions
In this paper, we investigate a single period partial disassembly optimization SPPDO
problem to generate an optimal disassembly sequence in product recovery of the end-of-life
EoL products. An AND/OR graph representation and associated transition matrix are used
in the mathematical formulation of the SPPDO problem to minimize the total disassembly
cost. Since the transition matrix does not have the property of total unimodularity, this
SPPDO model can only be solved as a pure integer programming IP problem, which is
NPcomplete.
A label correcting algorithm is proposed to find an optimal disassembly sequence
when the reusable subpart is retrieved directly from original return. To solve the SPPDO
12 Mathematical Problems in Engineering
problem in general case, this paper presents a heuristic procedure that utilizes this poly-
nomial-time algorithm to find a solution. This heuristic procedure can quickly provide a good
disassembly plan for problems with more complicated disassembly structures in a real-world
setting within a reasonable computation effort. It can be further integrated in the production
planning for end-of-life EoL products to improve the profitability of product recovery.
Acknowledgment
This research was supported by the National Science Council of Taiwan, China, under Grant
no. NSC 98-2218-E-027-019.
References
1 H. Griese, H. Poetter, K. Schischke, O. Ness, and H. Reichl, “Reuse and lifetime extension strategies in
the context of technology innovations, global markets, and environmental legislation,” in Proceedings
of the IEEE International Symposium on Electronics and the Environment, pp. 173–178, Washington, DC,
USA, May 2004.
2 E. J. Grenchus, R. A. Keene, and C. R. Nobs, “A decade of processing end of life IT equipment—lessons
learned at the IBM endicott asset recovery center,” in Proceedings of the IEEE International Symposium
on Electronics and the Environment, pp. 195–198, Washington, DC, USA, May 2004.
3 M. Fleischmann, Quantitative Models for Reverse Logistics, Springer, New York, NY, USA, 2001.
4 E. Grenchus, S. Johnson, and D. McDonnell, “Improving environmental performance through reverse
logistics at IBM,” in Proceedings of the IEEE International Symposium on Electronics and the Environment,
pp. 236–240, May 2001.
5 V. D. R. Guide Jr., “Production planning and control for remanufacturing: industry practice and
research needs,” Journal of Operations Management, vol. 18, no. 4, pp. 467–483, 2000.
6 A. Kasmara, M. Muraki, S. Matsuoka, A. Sukoyo, and K. Suryadi, “Production planning in reman-
ufacturing/ manufacturing production system,” in Proceedings of the 2nd International Symposium on
Environmentally Conscious Design and Inverse Manufacturing, pp. 708–713, 2001.
7 A. J. Clegg, D. J. Williams, and R. Uzsoy, “Production planning for companies with remanufacturing
capability,” in Proceedings of the IEEE International Symposium on Electronics and the Environment (ISEE
’95), pp. 186–191, May 1995.
8 V. D. R. Guide Jr., “Scheduling using drum-buffer-rope in a remanufacturing environment,” Interna-
tional Journal of Production Research, vol. 34, no. 4, pp. 1081–1091, 1996.
9 V. D. R. Guide Jr. and M. S. Spencer, “Rough-cut capacity planning for remanufacturing firms,”
Production Planning and Control, vol. 8, no. 3, pp. 237–244, 1997.
10 V. Daniel, V. D. R. Guide Jr., and R. Srivastava, “An evaluation of order release strategies in a reman-
ufacturing environment,” Computers and Operations Research, vol. 24, no. 1, pp. 37–47, 1997.
11 V. D. R. Guide Jr., R. Srivastava, and M. E. Kraus, “Product structure complexity and scheduling of
operations in recoverable manufacturing,” International Journal of Production Research, vol. 35, no. 11,
pp. 3179–3199, 1997.
12 V. D. R. Guide Jr., M. E. Kraus, and R. Srivastava, “Scheduling policies for remanufacturing,”
International Journal of Production Economics, vol. 48, no. 2, pp. 187–204, 1997.
13 M. E. Ketzenberg, G. C. Souza, and V. D. R. Guide Jr., “Mixed assembly and disassembly operations
for remanufacturing,” Production and Operations Management, vol. 12, no. 3, pp. 320–335, 2003.
14 P. Dewhurst, “Product design for manufacture: design for disassembly,” Industrial Engineering, vol.
25, no. 9, pp. 26–28, 1993.
15 K. N. Taleb and S. M. Gupta, “Disassembly of multiple product structures,” Computers and Industrial
Engineering, vol. 32, no. 4, pp. 949–961, 1997.
16 I. M. Langella, “Heuristics for demand-driven disassembly planning,” Computers and Operations Re-
search, vol. 34, no. 2, pp. 552–577, 2007.
17 A. Gungor and S. M. Gupta, “An evaluation methodology for disassembly processes,” Computers and
Industrial Engineering, vol. 33, no. 1-2, pp. 329–332, 1997.
18 D. Navin-Chandra, “The recovery problem in product design,” Journal of Engineering Design, vol. 5,
no. 1, pp. 65–86, 1994.
Mathematical Problems in Engineering 13
Research Article
Improved Degree Search Algorithms in
Unstructured P2P Networks
Copyright q 2012 Guole Liu et al. This is an open access article distributed under the Creative
Commons Attribution License, which permits unrestricted use, distribution, and reproduction in
any medium, provided the original work is properly cited.
Searching and retrieving the demanded correct information is one important problem in networks;
especially, designing an efficient search algorithm is a key challenge in unstructured peer-to-peer
P2P networks. Breadth-first search BFS and depth-first search DFS are the current two typical
search methods. BFS-based algorithms show the perfect performance in the aspect of search
success rate of network resources, while bringing the huge search messages. On the contrary, DFS-
based algorithms reduce the search message quantity and also cause the dropping of search success
ratio. To address the problem that only one of performances is excellent, we propose two memory
function degree search algorithms: memory function maximum degree algorithm MD and mem-
ory function preference degree algorithm PD. We study their performance including the search
success rate and the search message quantity in different networks, which are scale-free networks,
random graph networks, and small-world networks. Simulations show that the two performances
are both excellent at the same time, and the performances are improved at least 10 times.
1. Introduction
Searching and retrieving the demanded correct information is becoming more and more
important with the emergence of the huge amounts of information and the growth in the
size of computer networks 1. Especially, in unstructured P2P networks, the node’s joining
and failure are both random and dynamic 2, and in this case, it is unfeasible and unpractical
that each node of the network has known and stored the global information about the whole
2 Mathematical Problems in Engineering
network topology and the location of queried resources. Thus, designing efficient search algo-
rithms according to the local network information is critical to the performance of unstruc-
tured P2P networks.
Considerable amount of work has been done in this field, so far, a number of search
algorithms have been proposed, including BFS algorithm 1, modified BFS algorithm 1,
3, 4, local search algorithm 5, 6, rumor broadcasting algorithm 7–10, the betweenness
11, 12, shortest path algorithm 13, iterative deepening algorithm 1, 14, 15, update propa-
gation algorithm 16, and random walks search 17–30. These search algorithms can be
classified into two categories: BFS-based method and DFS-based method. Although these
algorithms achieve relatively satisfying effects, these two types of search algorithms tend to
be inefficient, either generating too much load on the networks 1, 2 or not meeting the search
success rate of network resources. On the one hand, BFS-based algorithm shows the perfect
performance in the aspect of search success rate of network resources, but at the same time,
it brings the huge search messages. The number of search messages will grow exponentially
with the hop counts in the search process 17, 18. On the other hand, DFS-based algorithm
generates the search message quantity far smaller compared with BFS-based algorithm, and
the search message quantity will grow linearly with the hop counts 20. The main drawback
is to drop the search success ratio of network resources.
The references 3, 4, 7–10, 17–30 adopt the BFS-based method or DFS-based method
to achieve their goal, respectively, but do not overcome the problem that only one of perfor-
mances of the network loads, and the search success rate is excellent. To address this problem,
in this paper, according to the degree of nodes 2, we propose memory function maximum
degree algorithm MD and memory function preference degree algorithm PD. These two
algorithms can combine the advantages of the BFS-based algorithm and DFS-based algo-
rithm, which can be efficiently used to search random graph 31, 32 networks and power-law
networks such as scale-free networks and small-world networks 33, 34. We have studied
their performances in the search success rate of network resources and the search message
quantity. Simulations illustrate their validity and feasibility. The results show that MD
algorithm is better than PD algorithm in the search success rate. The search success rate of MD
algorithm is average 14 times better than the standard random walks algorithm; the search
message quantity is the same order of magnitude with it. Compared with modified BFS
algorithm, the search success rate of MD algorithm is higher than it, and the search message
quantity averagely reduces by over 18 times. Although PD algorithm can reduce the huge
search message quantity, the search success rate of it is inferior to the modified BFS algorithm.
information. The second is that one node has to remember the return node’s ID according
to its memory information. This requires the nodes of networks to save their neighbors’ ID
and, at the same time, save their related degree information. It is the advantage of these
two algorithms from the point of view reducing the unnecessary search messages, and it is
the shortcoming from the point of view occupying the storage space. Compared with the MB
algorithm and the random walks algorithms, they double the storage space. In the context, the
degree of a node is the number of connections or edges the node has to other nodes, including
the traversed nodes.
2 5 8 2 5 8
1 1 2
1 3 6 9 11 1 3 6 9 11
4 7 10 4 7 10
(a) (b)
2 5 8 2 5 8
1 2 3 4 1 2 3
1 3 6 9 11 1 3 6 9 11
3 3
4 7 10 4 7 10
4
(d) (c)
of node 3 is the largest, so the search message is sent to node 3. Figure 2b is the second
step; node 3 sends the search message to node 6. Figure 2c is the third step; the neighbor’s
nodes of node 6 have the same degree, and it sends two search messages to the neighbors.
Figure 2d is the forth step; the algorithm returns failed. If the resource is located in the
path composed by the highest-degree nodes, such as node 3, 6, 7, 9, 10, and 11, the maximum
degree algorithm can easily find these nodes. Figures 2e and 2f are the difference with MD
algorithm. Obviously, the MD algorithm easily finds the resource node. The maximum degree
algorithm searches the resource in failure, while MD algorithm searches the same resource in
success.
On the other hand, MD algorithm can search more resources according to the memory
of the ID and degree information. Figure 3 shows how to search the resource node with mem-
ory function using MD algorithm, assuming that the age value is large enough. In Figure 3a,
the source node 1 traverses all its neighbors and labels the neighbors according to BFS
method; then according to DFS method it broadcasts the search message to node 3 whose
degree is the highest among the neighbors. Figures 3b–3d repeat the search process. In
Figure 3e, node 7 and node 11 are the terminal nodes; the search messages return node 9 and
node 10 according to the memory information which are the ID information of their precur-
sors. Node 9 and node 10 remember the returned node’s ID, which the search messages do not
broadcast along the direction of these nodes. In Figure 3f, the neighbors of node 9 and node
10 are all traversed, so the search messages continue to return. In Figure 3g, the search mes-
sage returns to node 3. In Figure 3h, the search message will broadcast along the node with
the second highest degree, because it does not find the resource along the direction of nodes
with the highest degree. In Figure 3i, the search message finds the resource node, then node
8 responds to the required messages.
In summary, MD algorithm can search both categories resource nodes: highest degree
nodes and nonhighest degree nodes. It labels the nodes with different flags and uses the
nodes’ ID to reduce messages and, at the same time, to improve the search success ratio.
Mathematical Problems in Engineering 5
2 5 8 2 5 8
1 1 2
1 3 6 9 11 1 3 6 9 11
4 7 10 4 7 10
(a) (b)
2 5 8 2 5 8
1 2 3 4 1 2 3
1 3 6 9 11 1 3 6 9 11
3 3
4 7 10 4 7 10
4
(d) (c)
2 5 8 2 5 8
2
1 1
1 3 6 9 11 1 3 6 9 11
4 7 10 4 7 10
(e) (f)
Figure 2: Search process of the maximum degree algorithm and the difference with the MD algorithm.
n
d di , i 1, 2, . . . , n,
i 1
2.1
di
Πi , i −→ Ai .
d
In the worst conditions, the PD algorithm degenerates into the standard DFS algorithm
for either same degree case or different-degree case. The search process of this algorithm is
6 Mathematical Problems in Engineering
2 5 8 2 5 8 2 5 8
1 1 2 1 2 3
1 3 6 9 11 1 3 6 9 11 1 3 6 9 11
3
4 7 10 4 7 10 4 7 10
2 5 8 2 5 8 2 5 8
1 2 1 2 3 1 2 3 4
1 3 6 9 11 1 3 6 9 11 1 3 6 9 11
6 5 5
3
9
6
4 10 7 10 4 7 10 7 11 4 7 10
5 5 4
6
6
6
4 7 10 4 7 10 4 7 10
5 5 5
Figure 3: Process of the MD algorithm searching the resource node with memory function.
2 5 8 2 5 8
p2
p1 p1
1 3 6 9 11 1 3 6 9 11
p′
4 7 10 4 7 10
(a) (b)
p3 p3
2 5 8 2 5 8
p2 p2
p1 p1
1 3 6 9 11 1 3 6 9 11
4 7 10 4 7 10
(d) (c)
random, so the success rates are stochastic. The search process of this algorithm is shown
in Figure 4. Unlike the MD algorithm, it randomly generates a preference probability p
according to the neighbor nodes’ degree. Figures 4a–4d are the specific process in the ideal
conditions. In Figure 4b, if the node 3 chooses the neighbor node 6 according to the stochas-
tic preference, this algorithm will search resources in failure. It is shown as the dotted line box
in Figure 4b.
Mathematical Problems in Engineering 7
T
age
φ d·q , 3.1
age 1
where T is the threshold, d is the average degree of a node, q in the chosen ratio and is general
not more than 0.5. Here it is 0.25 in our simulations.
When considering the random walks case, the requesting node sends out one search
message to a randomly chosen neighbor, that is, standard random walks algorithm. This
search message is seen as a walker. Then the walker directly keeps in touch with the source
node in the process of walking and asks whether to proceed to the next step. If the requestor is
agreed to continue walking the termination conditions have not been satisfied, it randomly
chooses a neighbor to forward the walker. Otherwise, the algorithm terminates the walking
process. The search message quantity of this algorithm is related to the age value; thus, it
reduces the network loads and achieves a message reduction by over an order of magnitude
compared to the standard BFS algorithm it is also called flooding search algorithm in some
literatures 1, 2. In order to improve the search success rate, the requesting node sends
out k search messages to its k neighbors, that is, k random walks algorithm, assuming that
8 Mathematical Problems in Engineering
the number of search messages for each hop keeps fixed as k, that is, the number of walkers.
Therefore, the total number of search messages for random walks algorithm is
T
φ k · age, k 1, 2, . . . , N, 3.2
age 1
where T is the threshold of age. When the walker k meets k 1, it is standard random walks
algorithmR1, and when k meets k > 1, it is k random walks algorithm RK. k is general
not more than 16 22, and here it is 4 in our simulations.
In degree search algorithms, the query messages spread to these neighbor’s nodes with
the characteristics the preference degree or the maximum degree every step in the search
process. In the context, degree search algorithms include the MD and PD. The characteristics
include the maximum degree and preference degree. So the total number of search messages
for degree search algorithms is
T
φ m · age, 3.3
age 1
where m is the number of nodes with the characteristics. For instance, in Figures 1a and
1b, m value is 1 because the number of nodes with the maximum degree is 1. In Figure 1c,
m value is 2 because node 9 and node 10 have the same maximum degree.
Figure 6 shows the search message quantity φ generated by the various algorithms
in different topology networks, where the average degree D of the networks is chosen from
2 to 10. Figure 6a is the scale-free BA network, Figure 6b is the ER random graph network,
and Figure 6c is the WS small-world network. The number of nodes of the networks is 5,000.
Simulations show that in the three cases, the search message quantity is increasing
with the growth of the average degree of the networks. The search message quantity of MB
algorithm confirms the exponential growth, and random walk algorithms reduce the message
quantity into linear growth. The search message quantity of MD algorithm and PD algorithm
is slightly less than the standard random walks algorithm, but, in general, they are the same
order of magnitude according to the simulations. Due to memory function of MD algorithm
and PD algorithm proposed in this paper, they reduce unnecessary search messages in the
search process. Compared with k random walks algorithm, the search message quantity of
MD algorithm and PD algorithm is less. These two algorithms can decrease the message
quantity about 18 times than MB algorithm.
C·r
Ψ 3.4
N
Mathematical Problems in Engineering 9
×106 ×106
4.5 4.5
4 4
3 3
φ
φ
2 2
1 1
0 0
2 4 6 8 10 2 4 6 8 10
D D
a b
6
×10
4.5
4
3
φ
0
2 4 6 8 10
D
MD RK
PD MB
R1
c
Figure 6: Message quantity φ generated by the various algorithms in different topology networks.
where r is the replication ratio, C is the number of nodes covered by the algorithm, and N is
the total number of nodes of the network. This formula shows that the search success rate
highly depends on the coverage of the search algorithms. The age value determines the cover-
age of the search algorithms. MB algorithm, random walks algorithm, and PD algorithm have
random factors. Thus, their search success rates vary greatly depending on network topology
and the random choices which have been made.
According to the results of our simulations, the age value of the search algorithms is
not too large except the standard random walks algorithm. The maximum age value is chosen
100 in our simulations. In random walks algorithms, the walker k is chosen 4.
Figure 7 shows the search success ratio Ψ of the various algorithms in different
topology networks, where the average degree D of the networks is chosen from 2 to 10.
Figure 7a is the scale-free BA network, Figure 7b is the ER random graph network, and
Figure 7c is the WS small-world network. Simulations show that in the three cases, the
search success rates of the various search algorithms are increasing with the growth of the
average degree of the networks. The search success rate of MD algorithm is the highest; the
search success rate of RK algorithm is slightly higher than that of MB algorithm. The search
success rate of R1 algorithm is the least. The search success rate of PD algorithm is better than
the R1 algorithm and is inferior to the MB algorithm. The messages of MD algorithm can
10 Mathematical Problems in Engineering
0.8 0.8
0.6 0.6
0.4 0.4
Ψ
Ψ
0.2 0.2
0 0
2 4 6 8 10 2 4 6 8 10
D D
a b
0.8
0.6
0.4
Ψ
0.2
0
2 4 6 8 10
D
MD RK
PD MB
R1
c
Figure 7: Search success ratio Ψ of the various algorithms in different topology networks.
return the precursor, so it has more search scope. In RK algorithm, there are k random
walkers to search the resources at the same time; the success rate is higher than that of the R1
algorithm. The MB algorithm, a random algorithm, can find the resources quickly if these
resources locate in the chosen search paths. And it will generate the massive redundancy
messages if the chosen search paths do not exit the resources. In our simulation, we calculate
the mean value; the chosen proportion of MB algorithm is 0.25 and small, so the success rate
is slightly inferior to that of MD and RK algorithm.
Figure 8 shows the search success ratio Ψ of the various algorithms in scale-free BA
networks, where the age value is changed from 1 to 100. Figure 8a is the MD algorithm,
Figure 8b is the PD algorithm, Figure 8c is the R1 algorithm, Figure 8d is the RK algo-
rithm, and Figure 8e is the MB algorithm. We can see that all the search success rates are
increasing with the growth of the average degree. With the constraint of search ages, only the
search success rate of MB algorithm is not affected. In particular, in the same scale of network
and the same average degree of network, the search success rate of MD algorithm is the
highest among the five algorithms, that of R1 algorithm is the least, and the search success rate
of PD algorithm is between the R1 algorithm and the MB algorithm. To the search ages, the
age value of MB algorithm is the least when the search success rate reaches the maximum, and
the age value of R1 algorithm is the largest. The age values of MD algorithm and PD algorithm
are between the R1 algorithm and the MB algorithm.
Mathematical Problems in Engineering 11
1 0.5
0.8 0.4
0.6 0.3
Ψ
Ψ
0.4 0.2
0.2 0.1
0 0
1 10 20 30 40 50 60 70 80 90 100 1 10 20 30 40 50 60 70 80 90 100
Hops (age) Hops (age)
a b
0.3 0.8
0.6
0.2
0.4
Ψ
Ψ
0.1
0.2
0 0
1 10 20 30 40 50 60 70 80 90 100 1 10 20 30 40 50 60 70 80 90 100
Hops (age) Hops (age)
c d
1
0.8
0.6
Ψ
0.4
0.2
0
1 10 20 30 40 50 60 70 80 90 100
Hops (age)
D=2 D=8
D=4 D = 10
D=6
e
Figure 8: Search success ratio Ψ of the various algorithms in scale-free BA networks.
Figure 9 shows the search success ratio Ψ of the various algorithms in ER random
graph networks, where the age value is changed from 1 to 100. Figure 9a is the MD algo-
rithm, Figure 9b is the PD algorithm, Figure 9c is the R1 algorithm, Figure 9d is the RK
algorithm, and Figure 9e is the MB algorithm. The simulations show that the search success
rate is increasing with the growth of the average degree. In particular, in the same scale of
network and the same average degree of network, the search success rate of MD algorithm is
still the highest among the five algorithms, and that of R1 algorithm is the least. The search
success rate of PD algorithm is between the R1 algorithm and the MB algorithm. To the search
12 Mathematical Problems in Engineering
1 0.7
0.8
0.5
0.6
Ψ
Ψ
0.4 0.3
0.2
0.1
0 0
1 10 20 30 40 50 60 70 80 90 100 1 10 20 30 40 50 60 70 80 90 100
Hops (age) Hops (age)
a b
0.5 1
0.4 0.8
0.3 0.6
Ψ
Ψ
0.2 0.4
0.1 0.2
0 0
1 10 20 30 40 50 60 70 80 90 100 1 10 20 30 40 50 60 70 80 90 100
Hops (age) Hops (age)
c d
0.8
0.6
0.4
Ψ
0.2
0
1 10 20 30 40 50 60 70 80 90 100
Hops (age)
D=2 D=8
D=4 D = 10
D=6
e
Figure 9: Search success ratio Ψ of the various algorithms in ER random graph networks.
hops, the age value of MB algorithm is the least when the search success rate reaches the
maximum, and that of R1 algorithm is the largest. The age values of MD algorithm and PD
algorithm are between the R1 algorithm and the MB algorithm.
Figure 10 shows the search success ratio Ψ of the various algorithms in WS small-
world networks, where the age value is changed from 1 to 100. Figure 10a is the MD algo-
rithm, Figure 10b is the PD algorithm, Figure 10c is the R1 algorithm, Figure 10d is the
RK algorithm, and Figure 10e is the MB algorithm. We can see that the search success rate is
increasing with the growth of the average degree. In particular, in the same scale of network
Mathematical Problems in Engineering 13
0.8 0.5
0.4
0.6
0.3
0.4
Ψ
Ψ
0.2
0.2
0.1
0 0
1 10 20 30 40 50 60 70 80 90 100 1 10 20 30 40 50 60 70 80 90 100
Hops (age) Hops (age)
a b
0.2 0.8
0.15 0.6
0.1 0.4
Ψ
Ψ
0.05 0.2
0 0
1 10 20 30 40 50 60 70 80 90 100 1 10 20 30 40 50 60 70 80 90 100
Hops (age) Hops (age)
c d
0.8
0.6
0.4
Ψ
0.2
0
1 10 20 30 40 50 60 70 80 90 100
Hops (age)
D=2 D=8
D=4 D = 10
D=6
e
Figure 10: Search success ratio Ψ of the various algorithms in WS small-world networks.
and the same average degree of network, the search success rate of MD algorithm is the
highest among the five algorithms, that of R1 algorithm is the least, and the search success
rate of PD algorithm is between the R1 algorithm and the RK algorithm. To the search hops,
the age value of MB algorithm is the least when the search success rate reaches the maximum,
and that of R1 algorithm is the largest. The age values of MD algorithm and PD algorithm are
between the R1 algorithm and the MB algorithm.
In general, the search success rate of MD algorithm excels the MB algorithm and the
random walks algorithm in the same conditions. But considering the smaller value of age,
14 Mathematical Problems in Engineering
Table 1: The message quantity of the MD and MB algorithms in different topology networks.
0.3 0.15
0.25
0.2 0.1
0.15
Ψ
Ψ
0.1 0.05
0.05
0 0
1 2 4 6 8 10 1 2 4 6 8 10
Hops (age) Hops (age)
a b
Figure 11: Search success ratio Ψ of the MD and MB algorithms in different topology networks BA, ER,
and WS.
the success rate of MB is slightly higher than or equal to that of MD e.g., the success rate of
MB when the age value is 10 shown in Figures 8, 9, and 10. This is because the MD algorithm
does not have enough hops to return the nodes with second highest degree when it cannot
find the resource along the direction of nodes with highest degree. Thus, it returns search
failure. With the age increasing, the success rate of MD algorithm gradually transcends that
of MB algorithm. To PD algorithm, the search success rate is better than R1 algorithm but is
inferior to RK algorithm and MB algorithm.
It is obvious that the search success rate highly depends on the coverage of the search
algorithms defined by the age value. However, a large value of age will incur an essential
large response time for a requester to obtain search results. A smaller value of age is more
appropriate. Thus, we have given the MD and MB algorithm success rate and message quan-
tity in the case of small age value. The response time of these two algorithms is very short
and almost the same. The results are shown in Figure 11 and Table 1. From Figure 11, we
can see that the success rate of MD algorithm is slightly better than that of MB algorithm
when the average degree is small. Instead, the success rate of MB is higher when the average
degree is large. And at the same time, MB algorithm generates a mass of messages as shown
Mathematical Problems in Engineering 15
1500 1500
1000 1000
(ms)
(ms)
500 500
0 0
2 4 6 8 10 2 4 6 8 10
D D
a b
1500
1000
(ms)
500
0
2 4 6 8 10
D
MD RK
PD MB
R1
c
Figure 12: Average response time generated by the various algorithms in different topology networks.
in Table 1; it is about 15 times higher than that of MD algorithm. Therefore, the comprehensive
performances of MD algorithm are slightly higher than MB algorithm.
consuming. The average response time of MD and PD algorithm is about twice as much as
that of MB algorithm.
Although the average response time of MD algorithm is about twice higher than that
of MB as shown in Figure 12, the message quantity of MB algorithm is about 18 times higher
than that of MD as shown in Figure 6. When the age value is small as shown in Figure 11,
the response time of MD and MB algorithm is almost the same, and the search success rate of
these two algorithms is also close under the same values of D. However, the message quantity
of MB is about 15 times higher than that of MD. The algorithm does not exit whose all perfor-
mance indexes are perfect. Thus, in view of the comprehensive indexes, MD algorithm out-
performs MB algorithm.
4. Conclusion
This paper presents the design and evaluation of two memory function search algorithms
over the unstructured P2P networks, which, respectively, built on the top of scale-free BA net-
works, ER random graph networks, and WS small-world networks. The performance of these
two algorithms has been compared with the current algorithms used in existing unstructured
P2P networks. The search success rate of MD algorithm is averagely 14 times better than the
standard random walks algorithm, while the search message quantity is the same order of
magnitude with it. Compared with modified BFS algorithm, the search success rate of MD
algorithm is higher than that, while the search message quantity averagely reduces by over
18 times. Although PD algorithm can reduce the huge search message quantity, the search
success rate of it is inferior to the modified BFS algorithm.
Acknowledgments
This paper is supported by the National Natural Science Foundation of China Grant nos.
61170269, 61121061, the Foundation for the Author of National Excellent Doctoral Disser-
tation of PR China Grant no. 200951, the Asia Foresight Program under NSFC Grant no.
61161140320, the National Science Foundation of China Innovative Grant no. 70921061, the
CAS/SAFEA International Partnership Program for Creative Research Teams and the Pro-
gram for New Century Excellent Talents in University of the Ministry of Education of China
Grant no. NCET-10-0239.
References
1 E. Meshkova, J. Riihijarvi, M. Petrova, and P. Mahonen, “A survey on resource discovery mechanisms,
peer-to-peer and service discovery frameworks,” Computer Networks, vol. 52, no. 11, pp. 2097–2128,
2008.
2 A. Iamnitchi, M. Ripeanu, I. Foster et al., “Small-world file-sharing communities,” in Proceedings of the
IEEE INFOCOM, vol. 2, pp. 952–963, 2004.
3 V. Kalogeraki, D. Gunopulos, and D. Zeinalipour-Yazti, “A local search mechanism for peer-to-peer
networks,” in Proceedings of the 11th International Conference on Information and Knowledge Management
(CIKM ’02), pp. 300–307, 2002.
4 B. F. Cooper and H. Garcia-Molina, “Ad Hoc, self-supervising peer-to-peer search networks,” ACM
Transactions on Information Systems, vol. 23, no. 2, pp. 169–200, 2005.
5 J. Risson and T. Moors, “Survey of research towards robust peer-to-peer networks: search methods,”
Computer Networks, vol. 50, no. 17, pp. 3485–3521, 2006.
Mathematical Problems in Engineering 17
6 J. Jeong and P. Berman, “Low-cost search in scale-free networks,” Physical Review E, vol. 75, no. 3,
Article ID 036104, 4 pages, 2007.
7 V. Isham, J. Kaczmarska, and M. Nekovee, “Spread of information and infection on finite random
networks,” Physical Review E, vol. 83, no. 4, Article ID 046128, 12 pages, 2011.
8 D. Trpevski, W. K. S. Tang, and L. Kocarev, “Model for rumor spreading over networks,” Physical
Review E, vol. 81, no. 5, Article ID 056102, 11 pages, 2010.
9 D. Shah and T. Zaman, “Rumors in a network: who’s the culprit?” Institute of Electrical and Electronics
Engineers. Transactions on Information Theory, vol. 57, no. 8, pp. 5163–5181, 2011.
10 S. Boyd, A. Ghosh, B. Prabhakar, and D. Shah, “Randomized gossip algorithms,” Institute of Electrical
and Electronics Engineers. Transactions on Information Theory, vol. 52, no. 6, pp. 2508–2530, 2006.
11 L. A. Adamic, R. M. Lukose, A. R. Puniyani, and B. A. Huberman, “Search in powerlaw networks,”
Physical Review E, vol. 64, no. 4, Article ID 046135, 8 pages, 2001.
12 A. Tizghadam and A. Leon-Garcia, “On traffic-aware betweenness and network criticality,” in
Proceedings of the INFOCOM IEEE Conference on Computer Communications Workshops, pp. 1–6, 2010.
13 S. Misra and B. J. Oommen, “An efficient dynamic algorithm for maintaining all-pairs shortest paths
in stochastic networks,” IEEE Transactions on Computers, vol. 55, no. 6, pp. 686–702, 2006.
14 A. Reinefeld and T. Marsland, “Enhanced iterative-deepening search,” IEEE Transactions on Pattern
Analysis and Machine Intelligence, vol. 16, no. 7, pp. 701–710, 1994.
15 Y. Yao and J.-C. Yao, “On modified iterative method for nonexpansive mappings and monotone
mappings,” Applied Mathematics and Computation, vol. 186, no. 2, pp. 1551–1558, 2007.
16 Z. Wang, S. K. Das, M. Kumar, and H. Shen, “An efficient update propagation algorithm for P2P
systems,” Computer Communications, vol. 30, no. 5, pp. 1106–1115, 2007.
17 Q. Lv, P. Cao, E. Cohen, K. Li, and S. Shenker, “Search and replication in unstructured peer-to-peer
networks,” in Proceedings of the 16th international conference on Supercomputing (ICS ’02), pp. 84–95,
2002.
18 C. Gkantsidis, M. Mihail, and A. Saberi, “Random walks in peer-to-peer networks,” in Proceedings of
the IEEE INFOCOM, vol. 1, pp. 1–12, 2004.
19 A. Kumar, J. Xu, and E. W. Zegura, “Effcient and scalable query routing for unstructured peer-to-peer
networks,” in Proceedings of the IEEE INFOCOM, vol. 2, pp. 1162–1173, 2005.
20 C. Gkantsidis, M. Mihail, A. Saberi et al., “Random walks in peer-to-peer networks: algorithms and
evaluation,” Performance Evaluation, vol. 63, no. 3, pp. 241–263, 2006.
21 M. Zhong, K. Shen, and J. Seiferas, “The convergence-guaranteed random walk and its applications in
peer-to-peer networks,” Institute of Electrical and Electronics Engineers. Transactions on Computers, vol.
57, no. 5, pp. 619–633, 2008.
22 T. Lin, P. Lin, H. Wang, and C. Chen, “Dynamic search algorithm in unstructured peer-to-peer
networks,” IEEE Transactions on Parallel and Distributed Systems, vol. 20, no. 5, pp. 654–666, 2009.
23 X. Ming, Z. Shuigeng, G. Jihong, and H. Xiaohua, “A path-traceable query routing mechanism for
search in unstructured peer-to-peer networks,” Journal of Network and Computer Applications, vol. 33,
no. 2, pp. 115–127, 2010.
24 K. Oikonomou, D. Kogias, I. Stavrakakis et al., “Probabilistic flooding for efficient information
dissemination in random graph topologies,” Computer Networks, vol. 54, no. 10, pp. 1615–1629, 2010.
25 R. Beraldi, L. Querzoni, and R. Baldoni, “Low hitting time random walks in wireless networks,”
Wireless Communications and Mobile Computing, vol. 9, no. 5, pp. 719–732, 2009.
26 F. Guofu, Z. Jincheng, L. Sanglu, C. Guihai, and C. Daoxu, “RER: a replica efficiency based replication
strategy,” Chinese Journal of Electronics, vol. 17, no. 4, pp. 589–594, 2008.
27 G. Huiwei and I. H. S. Horace, “A study of parallel data mining in a peer-to-peer network,” Concurrent
Engineering-Research and Applications, vol. 15, no. 3, pp. 281–289, 2007.
28 N. Bisnik and A. A. Alhussein, “Optimizing random walk search algorithms in P2P networks,”
Computer Networks, vol. 51, no. 6, pp. 1499–1514, 2007.
29 Y. Liu, J. Han, J. Wang et al., “Rumor riding: anonymizing unstructured peer-to-peer systems,” IEEE
Transactions on Parallel and Distributed Systems, vol. 22, no. 3, pp. 464–475, 2011.
30 C.-K. Chau and P. Basu, “Analysis of latency of stateless opportunistic forwarding in intermittently
connected networks,” IEEE/ACM Transactions on Networking, vol. 19, no. 4, pp. 1111–1124, 2011.
31 M. E. J. Newman, “Scientific collaboration networks. I. Network construction and fundamental
results,” Physical Review E, vol. 64, no. 1, Article ID 016131, 8 pages, 2001.
32 A. Fronczak, P. Fronczak, and J. A. Holyst, “Average path length in random networks,” Physical
Review E, vol. 70, no. 5, Article ID 056110, 7 pages, 2004.
18 Mathematical Problems in Engineering
33 L. A. Adamic and B. A. Huberman, “Growth dynamics of the world wide web,” Nature, vol. 401, p.
131, 1999.
34 M. Ostilli, A. L. Ferreira, and J. F. F. Mendes, “Critical behavior and correlations on scale-free small-
world networks: application to network design,” Physical Review E, vol. 83, no. 6, Article ID 061149,
21 pages, 2011.
Hindawi Publishing Corporation
Mathematical Problems in Engineering
Volume 2012, Article ID 276519, 12 pages
doi:10.1155/2012/276519
Research Article
The Number of Students Needed for
Undecided Programs at a College from
the Supply-Chain Viewpoint
Copyright q 2012 Jin-Ling Lin et al. This is an open access article distributed under the Creative
Commons Attribution License, which permits unrestricted use, distribution, and reproduction in
any medium, provided the original work is properly cited.
The objective of this study is to determine how to do forward-looking analysis to determine the
future need for various professionals, so that educational institutes can produce an appropriate
labor force for national development. The concept of component part commonality, which derives
from research on effective inventory management using material resource planning MRP, can
also be applied to human resource planning and educational resource management in higher
education systems. Therefore, this paper proposed a systematic method to analyze student
recruitment numbers for future needs, based on the concept of MRP. The research studied the
relationship between a curricular structure tree and the associated commonalities. It explored
the relationship between the commonality of students and the aggregated level of student
replenishment. Based on that, systematic guidelines for curriculum design can be established for
undeclared programs at collages. Two simple examples were used to illustrate the implementation
of MRP in analysis of the replenishment levels necessary safety stock levels in an education
system such as an engineering college.
1. Introduction
Technology can improve on conventional methods of allocating a nation’s resources, includ-
ing human resources. Meanwhile, international economic competition is increasing with the
widening trade. Government protected industries, especially in newly industrialized coun-
tries, such as Taiwan and its neighbors, are encountering stronger competition than ever
before. Therefore, these countries must join with foreign countries, even if involuntarily, to
2 Mathematical Problems in Engineering
open their domestic markets. To deal with this situation, these newly industrialized countries
need to coordinate and integrate their limited resources to obtain optimal utilization. In
other words, they should take steps to speed up technological research and development
to increase their ability to compete. Success in these steps depends strongly on whether there
are sufficient human resources in each professional area 1. Therefore, a proper human
resource policy is one of the determining factors for a national survival in light of this
keen economic competition. Since educating a professional usually takes several years, a
responsible government studies the future needs and develops policies, in advance, so that
the education system can develop enough of the needed professionals 2.
In Taiwan, colleges or universities are the major institutes for training high quality
workers. Qualified entrants have to fulfill strict curricular requirements before graduating.
The study procedure in schools is analogous to the production process in industrial
manufacturing. Thus, the concept of material requirements planning MRP can be applied to
an education system. The very nature of an MRP system revolves around commonality, which
means different final products still consist of some common components. An MRP system,
based on the degree of commonality, determines the minimum aggregate safety stock levels
needed to prevent out-of-stock situations and satisfy customer demands. If an undeclared
programs were developed in universities, the designing of an appropriate curriculum for
undeclared students needs to be prepared in advance. Curriculum decisions would be
dependent on the undeclared program’s student recruitment totals.
Nevertheless, a college student’s major is determined by his score on the National
Union College Entry Examination in Taiwan. A semigovernmental organization assigns
prospective students to a department at a college or university according to their score
on the examination and their choice list. Most students are assigned to a college by their
examination scores, in spite of their interests. Meanwhile, in the second year of high school
students are categorized into four groups according to their future majors. A type of
professional streaming is premature for pupils, who are just fifteen or sixteen year-old. Early
streaming reduces the breadth of a student’s knowledge so that student may lack the
capability to integrate related professional areas. Confucian concepts emphasize respect for
the individual’s will and educational interests. Therefore, students should be encouraged
not to make any premature career decisions before gathering enough information and
understanding their abilities and interests. Therefore, the Taiwanese Ministry of Education
MOE has issued a proposal suggesting that the college education system should postpone
professional streaming until the junior year of university. The government expects that the
higher education system should be able to educate a professional with strongly integrated
technological abilities by postponing professional curricular streaming 3. Recently some
Taiwanese universities, such as National Tsing Hua University and National Sun Yat-
sen University, have provided undeclared programs for incoming freshmen. The primary
objective of the program is to let each student develop an interest-oriented academic career
and to broaden the recruitment pool for prospective student.
The objective of this study is to determine how to do forward-looking analysis to deter-
mine the future need for various professionals, so that educational institutes can produce an
appropriate labor force for national development. According to the Ministry of Education
proposal, universities should recruit students without major streaming at the college entry
level. Recruited students can choose not to declare their major until they have finished a two-
year common curriculum. Students are then assigned to a department based on their interests
and capabilities as determined by their academic performance. This method is completely
different from the current one, where a student’s major is determined when they obtain
Mathematical Problems in Engineering 3
admission to a university. However, two problems arise from this proposal. First, a school’s
reputation would be the main consideration for a student instead of their major after the
two-year common curriculum. This situation will worsen the current problem encountered
by the Taiwanese education system, where most national universities are too crowded, while
many private universities cannot recruit enough students, causing poor allocation of limited
educational resources 4.
Second, since students in the same institute would have a common curriculum,
their professional training would have a high degree of commonality. Although student
choices among various professional majors would be more flexible, it can be predicted that
most students would gravitate to popular departments, even if they were not interested or
capable. This will lead to a problem where some departments will not be able to recruit as
many students as under the present curricular streaming system. Moreover, each university
will invest most of its budget in popular departments to attract more students without
considering the social costs. From a long-term point of view, the planning and training of the
professional labor force may be distorted, and some resources for certain professions could
be diluted 5; this will surely have educational and social costs.
The dropout rate, especially in the upper years, is high in private universities. The
phenomenon leaves the fewer college students, especially in the upper years, and it has been
a problem in Taiwan for a long time. The higher education system deals with this problem by
recruiting transfer students from the technical education system High schools in Taiwan are
categorized into two groups: the so-called ordinary schools, where students are educated to
enter colleges, or technical schools, which provide practical vocational training, so students
can be employed by industry upon graduation or go to technical colleges. However, due
to convention, most students from technical schools would rather compete with students
from ordinary high schools to enter ordinary universities, instead of technical colleges.
This phenomenon totally distorts the authority’s original plan for the higher education
system. This approach lacks an integrated view and forward-looking consideration of future
industrial needs and resource allocation. Since educational resources are being drained and
curricular streaming is going to be postponed, how to allocate the limited resources efficiently
is an important matter for Taiwanese educational authorities. Moreover, Taiwan’s MOE
subsidizes universities by the number of students recruited, but penalizes universities with
dropout or vacancy rates that are too high. Under such circumstance, accurate estimation
student number is crucial for curriculum design. This paper proposes a systematic method
for analyzing student recruitment numbers to meet future needs based on commonality of
academic content if streaming were postponed until the junior year.
2. Research Approach
Material requirement planning MRP 6 has been applied to inventory control management
manufacturing in recent years and can likewise be applied to determining educational needs.
This approach calculates the total supplies of product components at various manufacturing
stages based on the commonality of each component part among the manufactured products,
the product structures, and lead-time factors, such as ordering, manufacturing, and delivery.
In other words, it uses the relationship between commonality and material costs, production,
ordering, and delivery lead-time offsets, to consolidate and transfer the final products
demanded during different time periods into the needed subassemblies of components. The
objective of this approach is to combine independent orders for final products and maintain
the quantity of orders at the highest possible level, so a factory can maintain raw materials
4 Mathematical Problems in Engineering
and component stocks at lower levels since it is economical. This approach reduces total
costs by minimizing the cost per unit and setup costs. The commonality of component
parts among different products has a significant impact on inventory levels for components
and subassemblies. As commonality increases, the same component can be placed in many
different products. When demand for a final product changes, the inventory of the lowest
common component used must be properly adjusted to accommodate the change and avoid
waste. Hence, the determination of the inventory level must be based on a global view so that
limited resources can be utilized efficiently and flexibly to satisfy variations in the quantity
of final products ordered.
The concept of component part commonality derives from research on effective inven-
tory management using MRP. However, the concept can also be applied to human resource
planning and educational resource management in higher education systems. In MRP, the
master production schedule MPS must be established first according to predictions about
end-product needs. This ensures that varying demand over different time periods can be met
by production capacity. Similarly, MPS concepts can also be applied to long-term human
resource planning, where the future demand for various professionals dictates the skills
needed by the talent pool. The education and production processes are very similar as value-
added processes. The fulfillment of curriculum requirements by a student is analogous to
the production process of any industrial product. College freshmen, like raw material, are
prepared for various professions through different core courses taken during the learning
process. Alumni with different majors can be viewed as different end products. Hence, a
learning process which trains the freshmen for different professions is analogous to the
assembly process in a production line. Like MRP where the end product can be represented
by the bill of materials BOM, the training of a student can be represented by a similar
hierarchical structure.
During the learning process, college students have a high degree of commonality after
they have completed the basic courses. Once students begin to take the core courses in their
major; however, even though some core courses are common to different majors curricula,
the streaming process begins. During the streaming process, students’ core courses follow a
predetermined order. Those who fail a prerequisite course must retake the course before they
can proceed to advanced courses. Some students may be forced to change their major or drop
out if they continuously fail in core courses. This process is just like the production process,
where defective products must be reworked or repaired before they can be processed onto
the next stage of manufacturing. Those with serious defects which cannot be repaired have
to be scrapped. Hence, from the operational viewpoint, the learning process is very similar to
the production process.
In MRP, to satisfy the demand for the end product, it is necessary to have some safety
stock to compensate for unavoidable waste during production. The safety stock is a cushion
for production variations; it ensures that raw materials and components can be supplied in
time without shortage. Due to limited resources, the amount of safety stock must be planned
from the aggregate viewpoint to reduce unnecessary costs and to increase production
efficiency. Similarly, the safety stock concept can be applied to the education process; some
students drop out of the system during the process and cause a shortage that can usually
be replenished by enrolling more transfer students or freshmen. However, graduate quotas
are based on the individual program requirements; that is, each program decides how many
supplementary transfer students or freshmen it needs without considering the replenishment
of students from an aggregate viewpoint. This is deemed a waste of precious resources. The
problem of not satisfying the preset target for professionals can be resolved by means of
Mathematical Problems in Engineering 5
the commonality concept in MRP, which uses an aggregate student replenishment scheme to
solve the unavoidable problem of a shortage of students.
Since similarities exist between the education process and the production process, it is
possible to implement an MRP and commonality models in the education system. This paper
studied the relationship between a curricular structure tree and the associated commonalities.
Two simple examples were used to illustrate the implementation of MRP in analysis of
the replenishment levels necessary safety stock levels in an education system such as an
engineering college.
3. Model Analysis
This paper describes how to apply the MRP method to analysis of student replenishment
and demonstrates that the MRP concept can be suitably implemented in an education system
7. The compatibility of implementation between manufacturing systems and education
systems can be observed from the MRP structure trees for a product and curriculum. In a
manufacturing system, the MRP structure tree of a product represents the relationship among
final products, semiproducts, and components. This hierarchical structure is composed
of components at the lowest level, semiproducts and final products at the highest level.
Similarly, the common and advanced courses in a curriculum can be used to construct an
MRP tree based on a predetermined sequence, as shown in Figure 1.
In Figure 1, each node element denotes a course, which a student should take to
gain essential knowledge. From another viewpoint, each node also represents the learning
training status of a student. Therefore, for the purpose of illustration, in this paper, it was
assumed that a node, that is, a course offered, also represented a student whose learning
status has reached this course’s level. Therefore, the inventory control method in MRP can
also be applied to analyze student recruiting in an education system. The following example
shows how the concept of commonality between products can significantly affect the safety
levels of inventory required. MRP can be used to analyze how the commonality among
qualified students in different disciplinary curricula affects the number of supplementary
students.
The degree of commonality index, DCI, denoted a measurement of commonality, the degree
of difference among some products 8–10. It was expressed as C dj1 φj /d, where φj
was the number of parents of course j in the curricular structure tree; d was the number of
prerequisite courses, which had any ancestor; the index range was 1 ≤ C ≤ dj1 φj β. For
example, in Figure 2, the DCIs were evaluated.
In other words, the DCI is a ratio between the number of prerequisite courses at the
lower levels and the total number of their parents related upper year courses. It can be
used to represent the degree of commonality between students through different curricular
disciplines 11, 12. If two curricula in a college were independent, then C 1, as shown in
Figure 3.
Example 3.1. A simplified two-level curricular structure tree were used to represent the
case where a college recruits students without postponing streaming, Figure 4a. In this
6 Mathematical Problems in Engineering
Graduate of A
engineering
Computer Electrical
engineering B C
engineering
1 2 1 2 1 2
3 4 3 4 3 3 4 3
6 5 6 6 5 6 6 5 5 6 7 5 7 5 6 7
1+1+2+1 2+1+2+2+1 2+1+3+2+3
C= 4 = 1.25 C= 4 =2 C= 4 = 2.2
example, there are ten different curricula majors in a college. Students should take the
course represented by the upper node after they fulfill the lower node course’s requirements.
If the college postponed streaming, the freshman and sophomore students took the
common core courses regardless of their future majors. Thus, most of the prerequisite courses
at the lower level were aggregated, as shown in Figure 4b, for example. The prerequisites
11· · · 20 for these ten majors were merged into a common prerequisite 21. Using this
example, the main effect resulting from increased commonality can be studied.
The DCI can be applied to any node or level in the curricular tree. Therefore, the
students that reached a certain node, that is, who took the prerequisite courses antecedents
for that node, were tagged by the DCI. For the course or subject j, the level of student
replenishment Sj was equal to the value of the safety parameter k multiplied by the predicted
standard deviation σj . If the actual demand y exceeded the expected demand u and
was greater than the probability of student replenishment level, it can be expressed by
Tchebysheff’s inequality:
1
Prob y − u ≥ kσ j ≤ , 3.1
k2 1
Mathematical Problems in Engineering 7
1 2
1+1+1+1
3 4 C= 4 =1
5 6
1 2 3 4 5 10 1 2 3 4 5 ··· 10
···
11 12 13 14 15 20 21
a b
where parameter k represents the upper bounds of the safety parameter. In the worst case, the
probability of a student replenishment shortage at any level can be derived by substituting
Sj /σj for k:
σj2
Ψ Prob y − u ≥ Sj . 3.2
s2j σj2
If there were a demand for students called differentiation-type students who already
took the various prerequisite courses, 11· · · 20 in Figure 4, then these students could be
replaced by students called common-type students who have taken the common course,
for example, 21 in Figure 4. The variance of the independent variable C is
1
σj2 · σ2. 3.4
d c
Substituting 3.4 into 3.2, the supplementary level of the j-type student is
1−Ψ 1 3.5
Sj · σc .
Ψ d
1 d
Sc √ Sj . 3.7
C j1
Equation 3.7 expresses the relationship between the number of incoming common-
type students and the number of replaced differentiation-type students. Due to the degree
of commonality, the number of needed supplementary common-type students was less than
that of differentiation-type students. From this example, it can be noted that
a a simplified two-level structure tree was used. However, if the number of structure
levels were more than two, the demand would not have been fully independent
due to the correlation between the elements in the tree,
b if N common-type students were needed and each of them was equivalent to d
differentiation-type students, then the CDI was
N
d Nd
C d, 3.8
i
N N
Example 3.2. From the above example, it was noted that the demand for supplementary
students can be reduced if the degree of commonality increases. The other example, where
there are two simplified two-level curricula, can be used to show how the common-type
student demand and differentiation-type student demand interacted with each other.
Example 3.1 depicts a simple case where the differentiation-type students were
replaced by common-type students. However, further study is needed on the situation
where there are some partially common components. The following example assumed two
professions were needed, and that the distribution of demand probability for both professions
was uniform: 0, b1 for Profession1 and 0, b2 for Profession2. If b1 ≥ b2 and Z1 , Z2 was
a vector representing the demand for these professions, then the example describes the
resolution of an optimization problem where the total cost of student replenishment should
be minimized, subject to the service level. The aggregate service level, ASL, is the probability
that all the demands will be satisfied 13. ASL was assumed to be γ in this example.
Case 1. There was no commonality between the two curricula as shown in Figure 5.
1 2
3 4 5 6
Minimize T S3 S4 S5 S6 ,
3.10
S.T. ASL γ,
where Si represents the number of i-type students; T is the total number of students.
Since the number of supplementary students was equal to the difference between the
total number of students and the expected demand, that is, the safe supplementary level T −
b1 b2 , minimizing the number of supplementary students is equivalent to minimizing
the total number of students. Therefore, to satisfy all the constraints, the conditions Z1
minS3 , S4 and Z2 S5 , S6 must exist. Let S3 S4 and S5 S6 , then ASL Prob{Z1
Z3 } ∗ Prob{Z2 Z6 } S3 S6 /b1 b2 γ.
Meanwhile,
Minimize T S3 S4 S5 S6 2S3 S6
Table 1 lists the solutions for S3 and S6 in different ranges of ASL and T 8.
Case 2. There was some commonality between the two curricula as shown in Figure 6.
In this example, the common-type node 7 replaced the differentiation-type node 4 and
node 5 from the previous example. The inequality of S7 S3 , S7 S6 , and S7 S3 S6
must hold. Meanwhile, ASL Prob{Z1 S3 and Z2 S6 and Z1 Z2 S7 }. The shadow
region in Figure 7 represents the probability of the ASL.
From Figure 7, the ASL was evaluated as S7 − S3 S7 − S6 S3 S6 − S7 S7 − S3
S3 S6 − S7 S7 − S6 S3 S6 − S7 /2/b1 b2 . That is S7 − S3 S7 − S6 S3 S6 − S7 3S7 −
S3 − S6 /2/b1 b2 . The problem was reformulated as follows:
Minimize T S3 S6 S7
3.12
S.T. 2S7 − S3 S7 − S6 S3 S6 − S7 3S7 − S3 − S6 2rb1 b2 .
10 Mathematical Problems in Engineering
1 2
3 7 6
i The total number of students T ∗ decreases when there is commonality among the
curricula.
ii If the students who have taken course 7 replace the ones who should have taken
courses 4 and 5, the number of common-type students is less than the number of
differentiation-type students, that is, S7 < S4 S5 .
iii When commonality among the professional curricula increases, the number of
differentiation-type students who have taken courses 3 and 6 also increases.
Since the number of supplementary students is equal to the difference between the
total number of students and the expected demand, the number of supplementary students
is proportional to the total number. The above comments actually describe the relationship
between the number of supplementary students and the total number. When streaming is
postponed, common courses must replace some professional courses. It means that some
commonality exits in the curriculum. In order to satisfy industrial demand, the student
recruiting policy must be flexible. To maintain the necessary overstock level, for safety,
transfer students are the primary resource to replenish programs, especially in schools with
high quality controls, where some students may leave for failure to fulfill the curricular
requirements. If a school wants to go with the streaming postponement policy and replen-
ishment is needed, the number of transfer students who have been streamed should be
increased. In other words, if a college needs to recruit transfer students to maintain the safety
stock level, it should take transfer students in their junior year. In addition, it can establish
systematic guidelines to curriculum design for undeclared programs at colleges.
4. Conclusion
The direction of higher education is to allow more people have opportunities to attend
college. However, as the government budget for education is shrinking, the need to expand
private contributions has grown. Despite pressure from a limited budget, high quality
education still needs to be maintained while costs are minimized. From the economic point of
view, the optimal balance among cost per student, governmental support, and self-generated
Mathematical Problems in Engineering 11
Z2
S7
b2
S6
Z1
S3 b1 S7
Figure 7: The probability of the aggregate service level.
income would lead to effective management of resources. This is also a good index for
planning and controlling the number of students enrolling in a school with undeclared status.
In accordance with the current interest of undergraduates maintaining undeclared majors
until their upper years and the need to maintain a certain level of professionals in training,
many universities have proposed a variety of curricular programs to broaden the recruitment
pool of prospective students. The proposed approach is based on the commonality among
college level curricula. By decreasing the total number of students recruited from different
departments, this program can reduce costs. For schools with student recruiting problems,
their enrollment depends on transfer and extension students. The proposed program is
more cost-effective because recruitment takes place at the college level rather than at the
department level. At the same time, advanced students who have declared their majors can be
recruited. In other words, students can transfer during all four years of their undergraduate
program.
This paper also proposes a quantified model of commonality and recruitment planning
for appropriate curriculum design. Two simple cases are used to illustrate the method,
which can be divided into two phases—before undeclared students declare majors and after.
However, with the challenges higher education is facing with majors, general education,
and professional programs, our model still needs more research into quantity and quality.
The future direction of the proposed quantification model is two-fold. First, optimization
of resource allocation needs to be studied when the commonality changes depending on
different types of professional training. Secondly, in order to satisfy the needs of the system,
methods for controlling student distribution need to provide for transfer student quality, and
appropriate and a sufficient curriculum needs to be offered accordingly.
References
1 W. C. Lou, “The review of the development of Taiwanese college education in recent forty years,” The
Ideal College Education, pp. 21–36, 2004.
12 Mathematical Problems in Engineering
2 Ministry of Education, Taiwan, “Trend and demand in professional education for Taiwan in the 21st
century,” 2001.
3 A. Geddam, “Mechatronics for engineering education: undergraduate curriculum,” International Jour-
nal of Engineering Education, vol. 19, no. 4, pp. 575–580, 2003.
4 G. Psachoropoulos and M. Woodhall, Education for Development: An Analysis of Investment Choices,
Wu-Nan Publication, 2005, Chinese version interpreted by Bo-In Sheai.
5 G. Wolf, “Foundations of competency assessment & program review,” Evaluation in Education, 2000.
6 C. A. Ptak, C. Smith, and J. Orlicky, Material Requirements Planning, McGraw-Hill, New York, NY,
USA, 2011.
7 J. Jiao and M. M. Tseng, “Customizability analysis in design for mass customization,” CAD Computer
Aided Design, vol. 36, no. 8, pp. 745–757, 2004.
8 D. A. Collier, “Aggregated safety stock levels and component part commonality,” Management Science,
vol. 28, no. 11, pp. 1296–1303, 1982.
9 D. A. Collier, “The measurement and operating benefits of component part commonality,” Decision
Science, vol. 12, no. 1, pp. 85–96, 1991.
10 A. Khajavirad and J. J. Michalek, “An extension of the commonality index for product family
optimization,” in Proceedings of the 33rd Design Automation Conference, presented at—2007 ASME Inter-
national Design Engineering Technical Conferences and Computers and Information in Engineering
Conference (IDETC/CIE ’07), pp. 1001–1010, Las Vegas, Nev, USA, September 2007.
11 T. Blecker and N. Abdelkafi, “The development of a component commonality metric for mass cus-
tomization,” IEEE Transactions on Engineering Management, vol. 54, no. 1, pp. 70–85, 2007.
12 M. A. Wazed, S. Ahmed, and N. Yusoff, “Commonality and its measurement in manufacturing re-
sources planning,” Journal of Applied Sciences, vol. 9, no. 1, pp. 69–78, 2009.
13 K. R. Baker, M. J. Magazine, and H. L. W. Nuttle, “The effect of commonality on safety stock in a
simple inventory model,” Management Science, vol. 32, no. 8, 1986.
Hindawi Publishing Corporation
Mathematical Problems in Engineering
Volume 2012, Article ID 756023, 15 pages
doi:10.1155/2012/756023
Review Article
A Review of Deterministic Optimization Methods
in Engineering and Management
Copyright q 2012 Ming-Hua Lin et al. This is an open access article distributed under the Creative
Commons Attribution License, which permits unrestricted use, distribution, and reproduction in
any medium, provided the original work is properly cited.
1. Introduction
The field of optimization has grown rapidly during the past few decades. Many new
theoretical, algorithmic, and computational contributions of optimization have been pro-
posed to solve various problems in engineering and management. Recent developments of
optimization methods can be mainly divided into deterministic and heuristic approaches.
Deterministic approaches take advantage of the analytical properties of the problem to
generate a sequence of points that converge to a global optimal solution. Heuristic approaches
have been found to be more flexible and efficient than deterministic approaches; however,
the quality of the obtained solution cannot be guaranteed. Moreover, the probability
of finding the global solution decreases when the problem size increases. Deterministic
approaches e.g., linear programming, nonlinear programming, and mixed-integer nonlinear
programming, etc. can provide general tools for solving optimization problems to
obtain a global or an approximately global optimum. With the increasing reliance on
2 Mathematical Problems in Engineering
Optimization
problems
Convex Nonconvex
IP Nonconvex Convex
MILP relaxation relaxation
Nonconvex Convex
MINLP MINLP
researchers in the last few decades. In this paper, we investigate the advances in deterministic
global optimization of nonconvex nonlinear programming NLP problems and nonconvex
mixed-integer nonlinear programming MINLP problems. For NLP problems, we focus on
signomial programming problems that are an important class of NLP problems and have
played a crucial role in applications.
The rest of this paper is organized as follows. Section 2 discusses the deterministic
methods for signomial programming problems. Section 3 reviews the theoretical and
algorithmic developments of mixed-integer nonlinear programming problems. Conclusions
are made in Section 4.
2. Signomial Programming
Signomial programming SP is an optimization technique for solving a class of nonconvex
nonlinear programming problems. Although SP problems occur frequently in engineering
and management science, SP problems with nonconvex functions are still difficult to be
solved to obtain a global optimum. The term geometric program GP was introduced by
Duffin et al. 7 since the analysis of geometric programs relied heavily upon geometric-
arithmetic mean inequality. The early work by Duffin and Peterson 45 solved the
posynomial geometric program analytically via the dual problem. Then, Duffin 46
developed a numerical method for GPs based on solving a sequence of linear programs
47. For solving SP problems, Duffin and Peterson 48 reformulated an SP problem as
a geometric program with reversed constraints. The reversed constraints give rise to a
nonconvex feasible region that the local minima for SP problems are not guaranteed to be
global minima 49. The developed methods for SP can be divided into two approaches.
The first class of SP approaches includes various heuristic techniques such as genetic
algorithms, simulated annealing, tabu search, ant colony optimization, and particle swarm
optimization. Although the heuristic methods have the advantage of easy implementation
and offer a better potential for complex problems, the obtained solution is not guaranteed
to be a globally optimal solution. The second class of SP approaches is the deterministic
method. For example, Maranas and Floudas 50, Floudas and Pardalos 51, Maranas and
4 Mathematical Problems in Engineering
Floudas 22, and Floudas 19–21 developed global optimization algorithms for solving
SP problems based on the exponential variable transformation, the convex relaxation, and
the branch and bound type algorithm. These methods transform the original nonconvex
problem into a convex problem and then solve it to obtain the global optimum. The use of the
logarithmic/exponential transformation in global optimization algorithms on SP problems
restricts these exponential-based methods to handle the problems with strictly positive
variables. Although positive variables are employed frequently to represent engineering
and scientific systems, it is also common to introduce nonpositive variables in modeling
the management problems or the system behavior, such as investment decisions, stresses,
temperatures, electrical currents, velocities, and accelerations. For treating free variable
x, Pörn et al. 52 suggested a simple translation, x τ eX . However, inserting the
transformed result into the original signomial term will bring additional signomial terms and
therefore increase the computation burden. Tsai et al. 11 proposed an approach to treat zero
boundary signomial discrete programming problems and suggested some convexification
rules. Li and Tsai 53, Tsai and Lin 54–56, Tsai et al. 57, Tsai 58, and Li and Lu
59 applied convexification strategies and piecewise linearization techniques to solve SP
problems with free discrete/continuous variables. However, the optimal solution obtained is
an approximate solution by the piecewise linearization approach. Lin and Tsai 60 presented
a generalized method to solve signomial discrete programming problems with free variables
for finding exactly alternative optima. Tsai and Lin 61 also integrated the convexification
techniques and the bounding schemes to solve a posynomial geometric program with
separable functions for finding a global optimal solution efficiently.
Convexification strategies for signomial terms are important techniques in global
optimization for SP problems. With different convexification approaches, an SP problem can
be reformulated into another convex program solvable to obtain an approximately global
optimum. For solving SP problems, Pörn et al. 52 integrated the exponential transformation
and piecewise linear approximations for reformulating nonconvex signomial problems. The
results were extended by Björk 62, Björk et al. 63, and Pörn et al. 64, by including certain
power transformations for convexification of nonconvex signomial terms. They discussed
that the right choice of transformation for convexifying nonconvex signomial terms has
a clear influence on the efficiency of the optimization approach. The concept of power
convex functions is introduced to improve the solution efficiency for certain SP problems.
T. Westerlund and J. Westerlund 65 proposed the generalized geometric programming
extended cutting plane GGPECP algorithm for nonconvex optimization problems by using
the cutting plane and transformation techniques. The GGPECP algorithm was described in
more detail in Westerlund 66. Lundell and Westerlund 67 and Lundell et al. 68 combined
the GGPECP algorithm with an optimization framework for the transformations used to
convexify the signomial terms into a signomial global optimization algorithm. The signomial
global optimization algorithm was further extended by Lundell and Westerlund 69, 70. Lin
and Tsai 71 and Tsai and Lin 56 also presented similar reformulation and range reduction
techniques to enhance the computational efficiency for solving SP problems.
For solving an SP problem, the above-mentioned convexification techniques are
used to reformulate the original SP problem into a convex and underestimating problem
solvable by a standard mixed-integer nonlinear programming MINLP solver 72–81.
Different transformations for positive and negative signomial terms have been proposed
and discussed by Björk et al. 63, Westerlund 66, Lundell and Westerlund 67, Pörn et al.
64, Lundell et al. 68, and Lundell and Westerlund 69, 70. For a positive signomial term
αm αm1
cxα1 1 x2α2 · · · xm xm1 · · · xnαn c > 0, α1 , . . . , αm > 0 and αm1 , . . . , αn < 0, they suggested either the
Mathematical Problems in Engineering 5
exponential transformation ET or the power convex transformation PCT is applied based
on the characteristics of the problems. The ET strategy and the PCT strategy are described as
follows 63.
The ET strategy:
⎧
⎪
⎨xi e ,
Xi
i 1, . . . , m,
αm αm1 α1 X1 ···αm Xm
cx1α1 x2α2 · · · xm xm1 · · · xnαn ⇐⇒ e 2.1
⎪
⎩c |αm1 | |α |
.
xm1 · · · xn n
The PCT strategy: this technique aims at constructing 1-convex signomial terms. First,
transform all variables with positive exponents by an inverse transformation IT, x X −1 ,
n
except the one with the greatest exponent denoted as αmax . Let S be defined as S i 1 |αi | −
αmax . If αmax < S 1, then transform the variable with the exponent αmax to that with the
exponent S 1. If αmax > S 1, then change one of the ITs, with the exponent αj to xj Xj−τ ,
where τ > 1 so that αmax S 1 τ − 1αj .
Since some negative signomial terms may exist in SP problems, they suggested the
αm αm1
potential transformation PT for a negative signomial term cxα1 1 x2α2 · · · xm xm1 · · · xnαn c <
0, α1 , . . . , αm > 0 and αm1 , . . . , αn < 0 expressed as follows.
The PT strategy:
⎧
⎪
⎪
1/R
⎪xi Xi ,
⎪
i 1, . . . , m,
⎨
αm αm1 xi X −1/R , i m 1, . . . , n,
cx1α1 x2α2 · · · xm xm1 · · · xnαn ⇐⇒ i 2.2
⎪
⎪ n
⎪
⎪
α1 /R
· · · Xmαm /R |αm1 |/R |α |/R
· · · Xn n , |αi |.
⎩cX1 Xm1 R
i 1
studied trilinear monomials with positive or negative domains, derived explicit expressions
for the facets of the convex and concave envelopes, and showed that these outperform
the previously proposed relaxations based on arithmetic intervals or recursive arithmetic
intervals. Meyer and Floudas 90 presented explicit expressions for the facets of convex
and concave envelopes of trilinear monomials with mixed-sign domains. Tardella 91
studied the class of functions whose convex envelope on a polyhedron coincides with the
convex envelope based on the polyhedron vertices and proved important conditions for
a vertex polyhedral convex envelope. Meyer and Floudas 92 described the structure of
the polyhedral convex envelopes of edge-concave functions over polyhedral domains using
geometric arguments and proposed an algorithm for computing the facets of the convex
envelopes.
Caratzoulas and Floudas 93 proposed novel convex underestimators for trigonomet-
ric functions, which are trigonometric functions themselves. Akrotirianakis and Floudas 94,
95 introduced a new class of convex underestimators for twice continuously differentiable
nonlinear programs, studied their theoretical properties, and proved that the resulting convex
relaxation is improved compared to the αBB one. Meyer and Floudas 90 proposed two new
classes of convex underestimators for general C2 nonlinear programs, which combine the
αBB underestimators within a piecewise quadratic perturbation, derived properties for the
smoothness of the convex underestimators, and showed the improvements over the classical
αBB convex underestimators for box-constrained optimization problems.
Three popular convex underestimation methods, arithmetic intervals AIs 96, recur-
sive arithmetic intervals rAIs 50, 85, 96, and explicit facets EFs for convex envelopes of
trilinear monomials 89, 90, are effective to underestimate a trilinear term x1 x2 x3 for xi to
be bounded variables. However, these existing methods have difficulty to treat a posynomial
function. According to Ryoo and Sahinidis 85, for underestimating a multilinear function
nk n/2 n
x1 x2 · · · xn with n variables, the AI scheme needs to use n−1 k 2 Θk i 1 2i linear constraints
maximally. Θk denotes the number of linear functions that the AI scheme generates to lower
bound k-cross-product terms, k 2, 3, . . . , n − 1. Since the number of linear constraints of
convex envelopes for a multilinear function with n variables grows doubly exponentially
in n, AI bounding scheme may only treat n ≤ 3 cases. It is more difficult for AI to treat
a posynomial function for n > 3 cases. Moreover, applying rAI scheme to underestimate
a multilinear function x1 x2 · · · xn needs to use the maximum of exponentially many 2n−1
linear inequalities. Therefore, the rAI bounding scheme has difficulty to treat posynomial
functions as well as the AI scheme. EF 89, 90 provided the explicit facets of the convex
and concave envelopes of trilinear monomials and demonstrated that these result in tighter
bounds than the AI and rAI techniques. An important difference between EF and other
bounding schemes is that these explicit facets are linear cuts, which were proven to
define the convex envelope. Explicit facets EFs of the convex envelope are effective in
treating general trilinear monomials, but the derivation of explicit facets for the convex
envelope of general multilinear monomials and signomials is an open problem. Li et al.
97 and Lu et al. 98 proposed a novel method for the convex relaxation of posynomial
functions. The approach is different from the work of Maranas and Floudas 50, which
provided an alternative way of generating convex underestimators for generalized geometric
programming problems via the exponential transformation and linear underestimation of the
concave terms. Applications of this approach include the area of process synthesis and design
of separations, phase equilibrium, nonisothermal complex reactor networks, and molecular
conformation problems e.g., 99–101.
Mathematical Problems in Engineering 7
and constraints and tested the proposed method on several benchmark problems arising in
process synthesis and scheduling applications. Parthasarathy and El-Halwagi 109 studied
a systematic framework for the optimal design of condensation, which is an important
technology for volatile organic compounds, and formulated the problem as a nonconvex
MINLP model. They also proposed an iterative global optimization approach based on
physical insights and active constraint principles that allow for decomposition and efficient
solution and applied it to a case study for the manufacture of adhesive tapes. Adjiman et
al. 82, 83, 110, 111 proposed two global optimization approaches, SMIN-αBB and GMIN-
αBB, for nonconvex MINLP problems based on the concept of branch-and-bound. These
two approaches rely on optimization or interval-based variable-bound updates to enhance
efficiency. Although one possible approach to circumvent nonconvexities in nonlinear
optimization models is reformulation, for instance, using the exponential transformation
to treat the generalized geometric programming problems in which a signomial term
β
x1α x2 is transferred into an exponential term eα ln x1 β ln x2 , the exponential transformation
technique can only be applied to strictly positive variables and is thus unable to deal with
nonconvex problems with free variables. Tsai et al. 11 proposed an approach to treat zero
boundary optimization problems and suggested some convexification rules for the signomial
terms with only three nonnegative discrete/integer variables. Björk and Westerlund 112
studied the global optimization of heat exchanger network synthesis through the simplified
superstructure representation that allows only series and parallel schemes and applied
convexification approaches for signomials by piecewise linear approximations. They also
formulated convex MINLP lower bounding models using the Patterson formula for the
log mean temperature difference considering both isothermal and nonisothermal mixing.
Ostrovsky et al. 113 studied nonconvex MINLP models in which most variables are in
the nonconvex terms and the number of linear constraints is much larger than the number
of nonlinear constraints for solvent design and recovery problems. The work presents a
tailored branch-and-bound approach using linear underestimators for tree functions based
on a multilevel function representation and shows that there is a significant reduction in
the branching variable space. Tawarmalani and Sahinidis 114 developed a branch and
bound framework for the global optimization of MINLP problems. The framework involves
novel linear relaxation schemes, a Lagrangian/linear duality-based theory for domain
and range reduction, and branching strategies that guarantee finiteness of the solution
sequence for certain classes of problems. They also discuss implementation issues and present
computational results with a variety of benchmark problems. Kesavan et al. 115 presented
outer-approximation algorithms for finding an optimal solution of a separable nonconvex
MINLP program. Emet and Westerlund 116 conducted a computational comparison of
solving a cyclic chromatographic separation problem using MINLP methods and reported
that the extended cutting plane method compares favourably against traditional outer-
approximation and branch-and-bound methods. A review of the recent advances in MINLP
optimization of planning and design problems in the process industry was presented by
Kallrath 117. Tawarmalani and Sahinidis 118 introduced a polyhedral branch-and-cut
approach in global optimization. Their algorithm exploits convexity in order to generate the
polyhedral cuts and relaxations for multivariate nonconvex problems. Meyer and Floudas
119 studied superstructures of pooling networks, which are important to the petrochemical,
chemical, and wastewater treatment industries, and formulated this generalized pooling
problem as a nonconvex MINLP problem that involves many bilinear terms in the constraint
functions. They proposed a global optimization algorithm based on a novel piecewise linear
reformulation-linearization technique RLT formulation. Karuppiah and Grossmann 120
Mathematical Problems in Engineering 9
addressed the problem of optimal synthesis of an integrated water system, where water
using processes and water treatment operations are jointly considered. The designed MINLP
model was solved with a new deterministic spatial branch and contract algorithm, in
which piecewise under- and overestimators are used for constructing the relaxations at
each node. Bergamini et al. 121 formulated an MINLP model for the global optimization
of heat exchanger networks and presented a new solution methodology that is based on
outer-approximation and utilizes piecewise underestimation. Rigorous constraints obtained
from physical insights are also included in the formulation, and the authors reported
computationally efficient global solutions for problems with up to nine process streams.
Tsai and Lin 54, 56 proposed a method for solving a signomial MINLP problem with free
variables by the convexification strategies and piecewise linearization techniques. However,
the optimal solution obtained is an approximate solution by the piecewise linearization
approach. Karuppiah et al. 122 presented an outer-approximation algorithm to globally
solve a nonconvex MINLP formulation that corresponds to the continuous time scheduling of
refinery crude oil operations. The solution procedure relies on effective mixed-integer linear
relaxations that benefit from additional cuts derived after spatially decomposing the network.
Foteinou et al. 123 presented a mixed-integer optimization framework for the synthesis
and analysis of regulatory networks. Their approach integrates prior biological knowledge
regarding interactions between genes and corresponding transcription factors, in an effort to
minimize the complexity of the problem. Misener et al. 124 proposed an extended pooling
problem to maximize the profit of blending reformulated gasoline on a predetermined
network structure of feed stocks, intermediate storage tanks, and gasoline products subject to
applicable environmental standards. They formulated the problem as a nonconvex MINLP
model due to the presence of bilinear, polynomial, and fractional power terms. A mixed-
integer linear programming relaxation of the extended pooling problem is proposed for
several small- to large-scale test cases. Misener et al. 125 introduced a formulation for
the piecewise linear relaxation of bilinear functions with a logarithmic number of binary
variables and computationally compared their performance of this new formulation to the
best performing piecewise relaxations with a linear number of binary variables. They also
unified the new formulation into the computational tool APOGEE that globally optimizes
standard, generalized, and extended pooling problems. Westerlund et al. 126 considered
some special but fundamental issues related to convex relaxation techniques in nonconvex
MINLP optimization, especially for optimization problems including nonconvex inequality
constraints and their relaxations.
The alternative global optima of an MINLP problem can be found if more than one
solution satisfies the same optimal value of the objective function. In practice, alternative
optima are useful because they allow the decision maker to choose from many solutions
without experiencing any deterioration in the objective function. For the case involving only
0-1 variables, Balas and Jeroslow 127 introduced the well-known binary cut with only
one constraint and no additional variables. Duran and Grossmann 73 used this binary
cut in their OA algorithm to exclude binary combinations. Tawarmalani and Sahinidis 38
mentioned that BARON can identify the K best solutions for a mixed-integer nonlinear
program, where K is an option specified by the user. Tsai et al. 57 proposed a general integer
cut to identify all alternative optimal solutions of a general integer linear programming
problem. Lin and Tsai 60 proposed a generalized method to find multiple optimal
solutions of an MINLP problem with free variables by means of variable substitution
and convexification strategies. The problem is first converted into another convex MINLP
problem solvable to obtain an exactly global optimum. Then, a general cut is utilized to
10 Mathematical Problems in Engineering
exclude the previous solution and an algorithm is developed to locate all alternative optimal
solutions.
4. Conclusions
Given the rapid advances in computing technology over the past decades, large optimization
theories and algorithms have been proposed to solve various real-world engineering and
management problems. Therefore, to give a systematic overview of the extant literature
is a challenge and motivates this study, particularly for that the field of optimization
has grown and evolved rapidly. This work first reviewed methods for continuous
variable optimization and survey advances in signomial programming. Then, mixed-integer
nonlinear programming methods for optimization problems with discrete components were
introduced. Contributions related to theoretical and algorithmic developments, formulations,
and applications for these two classes of optimization problems were also discussed.
Although deterministic approaches take advantage of analytical properties of the
problem to generate a sequence of points that converge to a global solution, heuristic
approaches have been found to be more flexible and efficient than deterministic approaches.
For solving nonconvex or large-scale optimization problems, deterministic methods may
not be easy to derive a globally optimal solution within reasonable time due to the high
complexity of the problem. Heuristic approaches therefore are presented to reduce the
computational time of solving an optimization problem, but the obtained solution is not
guaranteed to be a feasible or globally optimal solution. These two types of optimization
methods have different pros and cons. Therefore, integrating deterministic and heuristic
approaches may be a good way of solving large-scale optimization problems for finding a
global optimum. It is hoped that this paper will stimulate further research on developing
more advanced deterministic and heuristic methods to enhance the computational efficiency
of finding a globally optimal solution for various real application problems.
Acknowledgment
The research is supported by Taiwan NSC Grants NSC 99-2410-H-158-010-MY2 and NSC 99-
2410-H-027-008-MY3.
References
1 W. Sharpe, “A linear programming approximation for the general portfolio analysis,” Journal of
Financial and Quantitative Analysis, vol. 6, pp. 1263–1275, 1971.
2 B. Stone, “A linear programming formulation of the general portfolio selection model,” Journal of
Financial and Quantitative Analysis, vol. 8, pp. 621–636, 1973.
3 M. R. Young, “A minimax portfolio selection rule with linear programming solution,” Management
Science, vol. 44, no. 5, pp. 673–683, 1998.
4 C. S. Chen, S. M. Lee, and Q. S. Shen, “An analytical model for the container loading problem,”
European Journal of Operational Research, vol. 80, no. 1, pp. 68–76, 1995.
5 L. Faina, “A global optimization algorithm for the three-dimensional packing problem,” European
Journal of Operational Research, vol. 126, no. 2, pp. 340–354, 2000.
6 H. L. Li, C. T. Chang, and J. F. Tsai, “Approximately global optimization for assortment problems
using piecewise linearization techniques,” European Journal of Operational Research, vol. 140, no. 3,
pp. 584–589, 2002.
7 R. J. Duffin, E. L. Peterson, and C. Zener, Geometric programming: Theory and application, John Wiley
& Sons Inc., New York, 1967.
Mathematical Problems in Engineering 11
8 J. S. Arora, Introduction to Optimum Design, McGraw-Hill, New York, NY, USA, 1989.
9 E. Sandgren, “Nonlinear integer and discrete programming in mechnical design optimization,”
Journal of Mechanisms, Transmissions, and Automation in Design, vol. 112, no. 2, pp. 223–229, 1990.
10 J. F. Fu, R. G. Fenton, and W. L. Cleghorn, “A mixed integer-discrete -continuous programming
method and its application to engineering design optimization,” Engineering Optimization, vol. 17,
pp. 263–280, 1991.
11 J. F. Tsai, H. L. Li, and N. Z. Hu, “Global optimization for signomial discrete programming problems
in engineering design,” Engineering Optimization, vol. 34, no. 6, pp. 613–622, 2002.
12 J. F. Tsai, “Global optimization for signomial discrete programming problems in engineering
design,” Engineering Optimization, vol. 42, no. 9, pp. 833–843, 2010.
13 D. Rotem, G. A. Schloss, and A. Segev, “Data allocation for multidisk databases,” IEEE Transactions
on Knowledge and Data Engineering, vol. 5, no. 5, pp. 882–887, 1993.
14 X. Lin and M. Orlowska, “An integer linear programming approach to data allocation with the
minimum total communication cost in distributed database systems,” Information Sciences, vol. 85,
no. 1–3, pp. 1–10, 1995.
15 R. Sarathy, B. Shetty, and A. Sen, “A constrained nonlinear 0-1 program for data allocation,” European
Journal of Operational Research, vol. 102, no. 3, pp. 626–647, 1997.
16 J. F. Tsai and H. L. Li, “Technical note—on optimization approach for multidisk vertical allocation
problems,” European Journal of Operational Research, vol. 165, no. 3, pp. 835–842, 2005.
17 M. H. Lin, “An optimal workload-based data allocation approach for multidisk databases,” Data and
Knowledge Engineering, vol. 68, no. 5, pp. 499–508, 2009.
18 H. S. Ryoo and N. V. Sahinidis, “Global optimization of nonconvex NLPs and MINLPs with
applications in process design,” Computers and Chemical Engineering, vol. 19, no. 5, pp. 551–566, 1995.
19 C. A. Floudas, “Global optimization in design and control of chemical process systems,” Journal of
Process Control, vol. 10, no. 2, pp. 125–134, 2000.
20 C. A. Floudas, “Recent advances in global optimization for process synthesis, design and control:
enclosure of all Solutions,” Computers and Chemical Engineering, vol. 23, no. 1, pp. S963–S973, 1999.
21 C. A. Floudas, Deterministic Global Optimization: Theory, Methods and Application, Kluwer Academic
Publishers, Boston, Mass, USA, 2000.
22 C. D. Maranas and C. A. Floudas, “Global optimization in generalized geometric programming,”
Computers and Chemical Engineering, vol. 21, no. 4, pp. 351–369, 1997.
23 J. G. Ecker, M. Kupferschmid, C. E. Lawrence, A. A. Reilly, and A. C. H. Scott, “An application of
nonlinear optimization in molecular biology,” European Journal of Operational Research, vol. 138, no.
2, pp. 452–458, 2002.
24 J. Mockus, W. Eddy, L. Mockus, and G. Reklaitis, Bayesian Heuristic Approach to Discrete and Global
Optimization, Kluwer Academic Publishers, London, UK, 1996.
25 I.E. Grossmann, Global Optimization in Engineering Design, Kluwer Academic Publishers, London,
UK, 1996.
26 I. M. Bomze, T. Csendes, R. Horst, and P. M. Pardalos, Developments in Global Optimization, Kluwer
Academic Publishers, London, UK, 1997.
27 A. Migdalas, P. M. Pardalos, and P. Varbrand, Multilevel Optimization: Algorithms and Applications,
Kluwer Academic Publishers, London, UK, 1997.
28 E. S. Mistakidis and G. E. Stavroulakis, Nonconvex Optimization in Mechanics: Algorithm, Heuristics
and Engineering Applications by the F.E.M, Kluwer Academic Publishers, London, UK, 1997.
29 R. De Leone, A. Murli, P. M. Pardalos, and G. Toraldo, High Performance Software for Nonlinear
Optimization, Kluwer Academic Publishers, London, UK, 1998.
30 E. M. T. Hendrix, Global optimization at work, Ph.D. thesis, Agriculture University Wageningen,
Wageningen, The Netherlands, 1998.
31 G. F. Corliss and R. B. Kearfott, “Rigorous global search: industrial applications,” in Developments in
Reliable Computing, T. Csendes, Ed., pp. 1–16, Kluwer Academic Publishers, London, UK, 1999.
32 C. A. Floudas and P. M. Pardalos, Optimization in Computational Chemistry and Molecular Biology-Local
and Global Approaches, Kluwer Academic Publishers, London, UK, 2000.
33 M. Laguna and J. L. González-Velarde, Computing Tools for Modeling Optimization and Simulation,
Kluwer Academic Publishers, London, UK, 2000.
34 P. M. Papalambros and D. J. Wilde, Principles of Optimal Design-modeling and Computation, Cambridge
University Press, Cambridge, UK, 2nd edition, 2000.
35 A. Atkinson, B. Bogacka, and A. A. Zhigljavsky, Optimum Design 2000, Kluwer Academic, London,
UK, 2001.
12 Mathematical Problems in Engineering
64 R. Pörn, K.-M. Björk, and T. Westerlund, “Global solution of optimization problems with signomial
parts,” Discrete Optimization, vol. 5, no. 1, pp. 108–120, 2008.
65 T. Westerlund and J. Westerlund, “GGPECP—An algorithm for solving non-convex MINLP
problems by cutting plane and transformation techniques,” Chemical Engineering Transactions, vol.
3, pp. 1045–1050, 2003.
66 T. Westerlund, “Some transformation techniques in global optimization,” in Global Optimization, vol.
84, pp. 45–74, Springer, New York, NY, USA, 2006.
67 A. Lundell and T. Westerlund, “Optimization of power transformations in global optimization,”
Chemical Engineering Transactions, vol. 11, pp. 95–100, 2007.
68 A. Lundell, J. Westerlund, and T. Westerlund, “Some transformation techniques with applications in
global optimization,” Journal of Global Optimization, vol. 43, no. 2-3, pp. 391–405, 2009.
69 A. Lundell and T. Westerlund, “Convex underestimation strategies for signomial functions,”
Optimization Methods & Software, vol. 24, no. 4-5, pp. 505–522, 2009.
70 A. Lundell and T. Westerlund, “On the relationship between power and exponential transformations
for positive signomial functions,” Chemical Engineering Transactions, vol. 17, pp. 1287–1292, 2009.
71 M.-H. Lin and J. F. Tsai, “Range reduction techniques for improving computational efficiency in
global optimization of signomial geometric programming problems,” European Journal of Operational
Research, vol. 216, no. 1, pp. 17–25, 2012.
72 B. Borchers and J. E. Mitchell, “An improved branch and bound algorithm for mixed integer
nonlinear programs,” Computers & Operations Research, vol. 21, no. 4, pp. 359–367, 1994.
73 M. A. Duran and I. E. Grossmann, “An outer-approximation algorithm for a class of mixed-integer
nonlinear programs,” Mathematical Programming, vol. 36, no. 3, pp. 307–339, 1986.
74 R. Fletcher and S. Leyffer, “Solving mixed integer nonlinear programs by outer approximation,”
Mathematical Programming, vol. 66, no. 3, Ser. A, pp. 327–349, 1994.
75 C. A. Floudas, Nonlinear and Mixed Integer Optimization:Fundamentals and Applications, Oxford
University Press, New York, NY, USA, 1995.
76 A. M. Geoffrion, “Generalized Benders decomposition,” Journal of Optimization Theory and
Applications, vol. 10, pp. 237–260, 1972.
77 S. Leyffer, “Integrating SQP and branch-and-bound for mixed integer nonlinear programming,”
Computational Optimization and Applications, vol. 18, no. 3, pp. 295–309, 2001.
78 S. Lee and I. E. Grossmann, “New algorithms for nonlinear generalized disjunctive programming,”
Computers and Chemical Engineering, vol. 24, no. 9-10, pp. 2125–2142, 2000.
79 I. Quesada and I. E. Grossmann, “An LP/NLP based branch and bound algorithm for convex MINLP
optimization problems,” Computers and Chemical Engineering, vol. 16, no. 10-11, pp. 937–947, 1992.
80 R. A. Stubbs and S. Mehrotra, “A branch-and-cut method for 0–1 mixed convex programming,”
Mathematical Programming A, vol. 86, no. 3, pp. 515–532, 1999.
81 T. Westerlund and F. Pettersson, “An extended cutting plane method for solving convex MINLP
problems,” Computers and Chemical Engineering, vol. 19, no. 1, supplement, pp. S131–S136, 1995.
82 C. S. Adjiman, S. Dallwig, C. A. Floudas, and A. Neumaier, “A global optimization method, αBB,
for general twice-differentiable constrained NLPs-I. Theoretical advances,” Computers and Chemical
Engineering, vol. 22, no. 9, pp. 1137–1158, 1998.
83 C. S. Adjiman, I. P. Androulakis, and C. A. Floudas, “A global optimization method, αBB, for general
twice-differentiabe constrained NLPs-II. Implementation and computational results,” Computers and
Chemical Engineering, vol. 22, no. 9, pp. 1159–1179, 1998.
84 M. Tawarmalani and N. V. Sahinidis, “Semidefinite relaxations of fractional programs via novel
convexification techniques,” Journal of Global Optimization, vol. 20, no. 2, pp. 137–158, 2001.
85 H. S. Ryoo and N. V. Sahinidis, “Analysis of bounds for multilinear functions,” Journal of Global
Optimization, vol. 19, no. 4, pp. 403–424, 2001.
86 M. Tawarmalani, S. Ahmed, and N. V. Sahinidis, “Product disaggregation in global optimization and
relaxations of rational programs,” Optimization and Engineering, vol. 3, no. 3, pp. 281–303, 2002.
87 M. Tawarmalani, S. Ahmed, and N. V. Sahinidis, “Global optimization of 0-1 hyperbolic programs,”
Journal of Global Optimization, vol. 24, no. 4, pp. 385–416, 2002.
88 L. Liberti and C. C. Pantelides, “Convex envelopes of monomials of odd degree,” Journal of Global
Optimization, vol. 25, no. 2, pp. 157–168, 2003.
89 C. A. Meyer and C. A. Floudas, “Trilinear Monomials with Positive or Negative Domains: Facets of
Convex and Concave Envelopes,” in Frontiers in global optimization, C. A. Floudas and P. M. Pardalos,
Eds., pp. 327–352, Kluwer Academic Publishers, Santorini, Greece, 2003.
14 Mathematical Problems in Engineering
90 C. A. Meyer and C. A. Floudas, “Trilinear monomials with mixed sign domains: facets of the convex
and concave envelopes,” Journal of Global Optimization, vol. 29, no. 2, pp. 125–155, 2004.
91 F. Tardella, “On the existence of polyhedral convex envelopes,” in Frontiers in Global Optimization, C.
A. Floudas and P. M. Pardalos, Eds., pp. 563–573, Kluwer Academic Publishers, Santorini, Greece,
2003.
92 C. A. Meyer and C. A. Floudas, “Convex envelopes for edge-concave functions,” Mathematical
Programming B, vol. 103, no. 2, pp. 207–224, 2005.
93 S. Caratzoulas and C. A. Floudas, “Trigonometric convex underestimator for the base functions in
Fourier space,” Journal of Optimization Theory and Applications, vol. 124, no. 2, pp. 339–362, 2005.
94 I. G. Akrotirianakis and C. A. Floudas, “Computational experience with a new class of convex
underestimators: box-constrained NLP problems,” Journal of Global Optimization, vol. 29, no. 3, pp.
249–264, 2004.
95 I. G. Akrotirianakis and C. A. Floudas, “A new class of improved convex underestimators for twice
continuously differentiable constrained NLPs,” Journal of Global Optimization, vol. 30, no. 4, pp. 367–
390, 2004.
96 A. S. E. Hamed, Calculation of bounds on variables and underestimating convex functions for nonconvex
functions, Ph.D. thesis, The George Washington University, 1991.
97 H. L. Li, J. F. Tsai, and C. A. Floudas, “Convex underestimation for posynomial functions of positive
variables,” Optimization Letters, vol. 2, no. 3, pp. 333–340, 2008.
98 H. C. Lu, H. L. Li, C. E. Gounaris, and C. A. Floudas, “Convex relaxation for solving posynomial
programs,” Journal of Global Optimization, vol. 46, no. 1, pp. 147–154, 2010.
99 A. Aggarwal and C. A. Floudas, “Synthesis of general distillation sequences-nonsharp separations,”
Computers and Chemical Engineering, vol. 14, no. 6, pp. 631–653, 1990.
100 A. R. Ciric and C. A. Floudas, “A retrofit approach for heat exchanger networks,” Computers and
Chemical Engineering, vol. 13, no. 6, pp. 703–715, 1989.
101 A. C. Kokossis and C. A. Floudas, “Optimization of complex reactor networks-II. Nonisothermal
operation,” Chemical Engineering Science, vol. 49, no. 7, pp. 1037–1051, 1994.
102 L. T. Biegler and I. E. Grossmann, “Retrospective on optimization,” Computers and Chemical
Engineering, vol. 28, no. 8, pp. 1169–1192, 2004.
103 I. E. Grossmann, “Review of nonlinear mixed-integer and disjunctive programming techniques,”
Optimization and Engineering, vol. 3, no. 3, pp. 227–252, 2002.
104 I. E. Grossmann and L. T. Biegler, “Part II. Future perspective on optimization,” Computers and
Chemical Engineering, vol. 28, no. 8, pp. 1193–1218, 2004.
105 H. D. Sherali, W. P. Adams, and P. J. Driscoll, “Exploiting special structures in constructing a
hierarchy of relaxations for mixed integer problems,” Operations Research, vol. 46, no. 3, pp. 396–405,
1998.
106 M. Sinha, L. E. K. Achenie, and G. M. Ostrovsky, “Environmentally benign solvent design by global
optimization,” Computers and Chemical Engineering, vol. 23, no. 10, pp. 1381–1394, 1999.
107 I. Harjunkoski, T. Westerlund, and R. Pörn, “Numerical and environmental considerations on
a complex industrial mixed integer non-linear programming MINLP problem,” Computers and
Chemical Engineering, vol. 23, no. 10, pp. 1545–1561, 1999.
108 R. Pörn and T. Westerlund, “A cutting plane method for minimizing pseudo-convex functions in the
mixed integer case,” Computers and Chemical Engineering, vol. 24, no. 12, pp. 2655–2665, 2000.
109 G. Parthasarathy and M. M. El-Halwagi, “Optimum mass integration strategies for condensation
and allocation of multicomponent VOCs,” Chemical Engineering Science, vol. 55, no. 5, pp. 881–895,
2000.
110 C. S. Adjiman, I. P. Androulakis, and C. A. Floudas, “Global optimization of MINLP problems in
process synthesis and design,” Computers and Chemical Engineering, vol. 21, no. 1, pp. S445–S450,
1997.
111 C. S. Adjiman, I. P. Androulakis, and C. A. Floudas, “Global optimization of mixed-integer nonlinear
problems,” AIChE Journal, vol. 46, no. 9, pp. 1769–1797, 2000.
112 K. M. Björk and T. Westerlund, “Global optimization of heat exchanger network synthesis problems
with and without the isothermal mixing assumption,” Computers and Chemical Engineering, vol. 26,
no. 11, pp. 1581–1593, 2002.
113 G. M. Ostrovsky, L. E. K. Achenie, and M. Sinha, “On the solution of mixed-integer nonlinear
programming models for computer aided molecular design,” Computers and Chemistry, vol. 26, no.
6, pp. 645–660, 2002.
Mathematical Problems in Engineering 15
Research Article
A Hybrid Genetic Algorithm for the Multiple
Crossdocks Problem
Copyright q 2012 Zhaowei Miao et al. This is an open access article distributed under the Creative
Commons Attribution License, which permits unrestricted use, distribution, and reproduction in
any medium, provided the original work is properly cited.
We study a multiple crossdocks problem with supplier and customer time windows, where any
violation of time windows will incur a penalty cost and the flows through the crossdock are
constrained by fixed transportation schedules and crossdock capacities. We prove this problem
to be NP-hard in the strong sense and therefore focus on developing efficient heuristics. Based
on the problem structure, we propose a hybrid genetic algorithm HGA integrating greedy
technique and variable neighborhood search method to solve the problem. Extensive experiments
under different scenarios were conducted, and results show that HGA outperforms CPLEX solver,
providing solutions in realistic timescales.
1. Introduction
As companies seek more profitable supply chains, there has been a desire to optimize
distribution networks to reduce logistics costs. This includes finding the best locations for
facilities, minimizing inventories, and minimizing transportation costs. A distinct recent
industry example is the successful implementation of crossdocking strategy at Wal-Mart,
whose crossdocks require coordinating 2000 dedicated trucks over a large network of
warehouses, crossdocks, and retail points 1. While there is a rich literature on conventional
facility location problems, crossdocking strategies—which minimize inventory by processing
goods quickly for reshipment—have recently attracted the attention of researchers see,
e.g., 2–4. In conventional transshipment-inventory models, a common assumption is that
demand usually stochastic that cannot be met from one supply point can be fulfilled
through some other point. The objective is then to evaluate a control policy for replenishment.
Work on this subject has been extensive and can be found in, for example, Krishnan and
Rao 5, Karmarkar and Patel 6, Karmarkar 7, Tagaras 8, Robinson 9, Rudi et al. 10,
2 Mathematical Problems in Engineering
Grahovac and Chakravarty 11, Herer and Tzur 12, Herer et al. 13, Axsäter 14, and
Axsäter 15. For the general n location transshipment model, heuristics were proposed by
Robinson 9 who developed a large-scale LP by discretizing demand. Although these studies
considered inventory and transshipment costs, they did not address time constraints that
occur during the transshipment process, for example, constraints imposed by transportation
schedules. There has been relatively limited research on distribution and system design,
which includes crossdocks. Some recent attempts can be found in Donaldson et al. 16,
Ratcliff et al. 17, Gumus and Bookbinder 4, Li et al. 3, Miao et al. 18, and Boysen et
al. 19. In particular, Gumus and Bookbinder 4 modeled location-distribution networks
that include crossdock facilities to determine the impact on the supply chain. Li et al. 3
developed a heuristic algorithm to find JIT schedules within a single crossdock. Miao et al.
18 and Boysen et al. 19 studied how to schedule the inbound and outbound trucks to
achieve high operational efficiency within a single crossdock.
Our work differs from the above research in that we study a kind of multiple
crossdocks problem where transportation is available at fixed schedules, and where both
shipping and delivery at supply and demand locations can be executed within specified
time windows with normal transportation cost, and any shipment that cannot be met will be
fulfilled through external channels, which causes penalty costs. Supplier time windows allow,
for example, flexibility in planning for the best shipping times to fit production and operating
schedules. Time windows at demand points satisfy customer requirements, for example,
when service deadlines must be met. Moreover, in the real-world applications, sometimes
the time windows are allowed to be violated. There are two primary reasons for this: one is
exogenous, for example, it might not be practical to satisfy all the time windows constraints
when the demands are too high during a certain period; the other is endogenous, for example,
some shipments are emergent and have to be executed outside the normal time windows.
Clearly, such abnormal arrangements usually incur additional costs we call penalty in this
paper. Furthermore, we consider inventory cost at the crossdocks, which includes storing
cost and handling cost and are one of the cost sources in any transshipment strategy.
To the best of our knowledge, there are some papers closet to our work, including Lim
et al. 20, Chen et al. 21, and Ma et al. 22. Lim et al. 20 studied the complexity of different
types of multiple crossdocks problems where transportation schedules such as flexible
schedules and fixed schedules, and time constraints at manufacturers and customers are
included. Our model extends one of the cases studied by them, which is called single shipping
and single delivery by fixed schedules. However, their problem requires all the demands
should be satisfied by the suppliers, which is very difficult to achieve even to find a feasible
solution. We relax this constraint and allow demands of some customers to be unfulfilled,
but penalty cost will be incurred. In addition, most importantly, in the earlier works, they did
not set up any mathematical formulation and provide any implementable algorithms that
were able to solve this practical problem. We model this problem as an integer programming
problem, prove it to be NP-hard in the strong sense, and then focus on developing efficient
heuristic algorithms. Chen et al. 21 extended another case studied by Lim et al. 20, which
is called single shipping and single delivery by flexible schedules, simplified the problem by
discretizing the time horizon, and designed metaheuristic algorithms to solve it. Despite the
constraint of single shipping and single delivery, Ma et al. 22 took into consideration setup
cost of each vehicle in their multiple crossdocks problem, which is also strong NP-hard and
solved by a two-stage heuristic algorithm developed by them.
While one of the main objectives in this paper is to develop an efficient heuristic
algorithm to solve the above multiple crossdocks problem, we find that numerous researches
Mathematical Problems in Engineering 3
satisfy time windows constraints and also helpful to potential consolidation. Third, shipping
schedules offered by transportation providers are fixed, that is, departure and arrival times
of any schedule are fixed. For example, the schedules in the railway network or in airline
operations are usually fixed. We assume that each schedule has a set of associated shipping
costs and route capacities. Fourth, the setup cost of each shipment sometimes is very high
in real world, so in order to reduce setup cost as much as possible, it requires each supplier
make only one batch shipment to some crossdock within its specified time window, and each
customer can receive goods only one time from some crossdock within its time window,
which is called single shipping and single delivery case 20, and for this case, consolidation
is quite important. Finally, the objective of our problem is to satisfy the demands of the
customers with minimum total costs including shipping cost, inventory cost and penalty
cost without violating the capacity constraints of crossdocks and routes through given fixed
schedules.
The underlying problem can be represented by a network. Let Σ : {1, . . . , n} be the set
of supply nodes suppliers where, for each i ∈ Σ, si units of goods are available, which can
be shipped released in the time window bir , eir , Δ : {1, . . . , m} the set of demand nodes
customers where each k ∈ Δ requires dk units of goods, which must be delivered accepted
within the time window bka , eka , and X : {1, . . . , l} the set of crossdocks, where each j ∈ X
has inventory capacity cj and inventory cost hj per unit per time. Take S1 to denote all fixed
scheduled routes between points in Σ and points in X, that is, routes serviced by transport
providers, each with a scheduled departure begin and arrival end time, capacity and unit
transportation cost. Similarly, let S2 denote the set of fixed schedules between the crossdocks
X and customers Δ.
Si,j : set of fixed transportation schedules between supplier i and crossdock j, and
|Si,j | γi,j , where | · | represents the cardinality of a set.
Sj,k : set of fixed transportation schedules between crossdock j and customer k, and
|Sj,k | γj,k
.
bi,j,q
r r
, ei,j,q : qth fixed transportation schedule in Si,j , where bi,j,q
r r
and ei,j,q are the
beginning time point and ending time point of this fixed schedule, respectively.
bj,k,q
a a
, ej,k,q : qth fixed transportation schedule in Sj,k , where bj,k,q
a a
and ej,k,q are the
beginning time point and ending time point of this fixed schedule, respectively.
ci,j,q : unit shipping cost from supplier i to crossdock j through qth fixed
transportation schedule in Si,j .
cj,k,q : unit shipping cost from crossdock j to customer k through qth fixed
transportation schedule in Sj,k .
Pi : unit penalty cost for supplier i if its cargo cannot be shipped out.
Pk : unit penalty cost for customer k if its demand cannot be met.
Tj : set of ending time points of all fixed transportation schedules in ∪ni 1 Si,j , that is,
Tj {ei,j,q
r
: 1 ≤ i ≤ n, 1 ≤ q ≤ γi,j }.
Mathematical Problems in Engineering 5
Tj : set of beginning time points of all fixed transportation schedules in ∪m S , that
k 1 j,k
is, Tj {bj,k,q
a
: 1 ≤ k ≤ m, 1 ≤ q ≤ γj,k }.
Tj : set of time points when the inventory level of crossdock j is likely to be changed,
that is, Tj Tj ∪ Tj . Let |Tj | τj and all the elements in Tj are sorted in an increasing
order, and let tj,g g 1, 2, . . . , τj correspond to these τj time points such that
tj,1 ≤ tj,2 ≤ · · · ≤ tj,τj . Using this notation, we can easily formulate the set of flow
conservation constraints later.
θi,j,q : a binary parameter, which is 0 if the beginning time point of qth fixed
transportation schedule in Si,j is within the time window of supplier i, that is,
r
bi,j,q ∈ bir , eir , and 1 otherwise.
θj,k,q : a binary parameter, which is 0 if the ending time point of qth fixed
transportation schedule in Sj,k is within the time window of customer k, that is,
a
ej,k,q ∈ bka , eka , and 1 otherwise.
xi,j,q : binary, which is 1 if to deliver cargos from supplier i is bound for crossdock j
through qth fixed transportation schedule in Si,j , and 0 otherwise.
xj,k,q : binary, which is 1 if to receive cargos from crossdock j to customer k is through
qth fixed transportation schedule in Sj,k , and 0 otherwise.
yj,tj,g : integer, which is inventory level in crossdock j at time tj,g , where tj,g ∈ Tj .
We are now ready to formulate the transshipment problem, which hereafter is called
problem P:
where
γ
n
l
γi,j
m
l
j,k
COSTTransportation ci,j,q si xi,j,q cj,k,q dk xj,k,q ,
i 1 j 1q 1 k 1j 1q 1
⎛ ⎞ ⎛ γ
⎞
n
l γi,j
m
l
j,k
l
τj
COSTInventory hj tj,g − tj,g−1 yj,tj,g−1 ,
j 1g 1
6 Mathematical Problems in Engineering
s.t.
xi,j,q ≤ 1 − θi,j,q 1 ≤ i ≤ n, 1 ≤ j ≤ l, 1 ≤ q ≤ γi,j , 2.2
xj,k,q ≤ 1 − θj,k,q 1 ≤ j ≤ l, 1 ≤ k ≤ m, 1 ≤ q ≤ γj,k , 2.3
l
γi,j
xi,j,q ≤ 1 1 ≤ i ≤ n, 2.4
j 1 q 1
γ
l
j,k
xj,k,q ≤1 1 ≤ k ≤ m 2.5
j 1q 1
si xi,j,q ≤ CAPi,j,q 1 ≤ i ≤ n, 1 ≤ j ≤ l, 1 ≤ q ≤ γi,j ,
2.6
dk xj,k,q ≤ CAPj,k,q
1 ≤ k ≤ m, 1 ≤ j ≤ l, 1 ≤ q ≤ γj,k ,
yj,tj,g ≤ cj 1 ≤ j ≤ l, 1 ≤ g ≤ τj ,
2.7
yj,tj,0 0 1 ≤ j ≤ l, tj,0 0 ,
n
m
yj,tj,g yj,tj,g−1 si xi,j,q − dk xj,k,q 1 ≤ j ≤ l, 1 ≤ g ≤ τj , 2.8
i 1 {q:ei,j,q
r
tj,g } k 1 q:ba t
j,k,q j,g
xi,j,q ∈ {0, 1} 1 ≤ i ≤ n, 1 ≤ j ≤ l, 1 ≤ q ≤ γi,j ,
xj,k,q ∈ {0, 1} 1 ≤ j ≤ l, 1 ≤ k ≤ m, 1 ≤ q ≤ γj,k , 2.9
yj,tj,g ∈N 1 ≤ j ≤ l, 1 ≤ g ≤ τj .
In the above formulation, the objective is to minimize total cost, including transporta-
tion cost, penalty cost, and inventory cost. Note that we impose the penalty cost on both
supplier and customer sides here because unfulfilled demands have different impact on each
side in general. Constraint 2.2 ensures that each available fixed transportation schedule is
within the time window of suppliers. Similarly, the available fixed transportation schedule of
customers is given by 2.3. Constraint 2.4 ensures that each delivery is fulfilled within each
supplier specified time window at most once and 2.5 forces each customer to receive cargos
within its time window for no more than one time, which is required by single shipping
and single delivery constraint. The capacity constraints of fixed schedules are given by 2.6.
The capacity constraint of every crossdock is restricted by 2.7 and we also set a zero initial
inventory for each crossdock. The changes of inventory level of each crossdock are recorded
in 2.8, which ensures cargo flow conservation.
We have the following proposition whose proof is given in the appendix.
Proposition 2.1. The multiple crossdocks problem P is NP-hard in the strong sense, even if supply
and demand time windows and crossdock and route capacities are relaxed.
From the above proposition, we know that to find minimum cost of this problem is
NP-hard in the strong sense. Hence, it is unlikely to find a polynomial or pseudopolynomial
time algorithm to solve the problem unless P NP. As a result, we focus on efficient heuristics
Mathematical Problems in Engineering 7
to solve problem P. In the next section, we describe a heuristic that exploits the problem
structure and solves the problem efficiently.
Step 2. Evaluate the fitness of each individual with respect to the objective function.
Step 3. Select a group of best individuals as the population pool, which guarantees that the
best genes can be preserved in offsprings.
Step 5. Apply mutation to diversify the pool by changing some genes in specified
chromosomes.
Step 6. Apply variable neighborhood search to each new generated offspring, and go to Step 2
until one of the termination conditions is satisfied.
Note that both crossover and mutation operators are only applied when assigning for
suppliers and any change in assignments for suppliers will trigger changes in assignments
for customers. Furthermore, in HGA, variable neighborhood search is applied to improve
the solution. This evaluation-selection-reproduction-local search cycle is repeated until one
8 Mathematical Problems in Engineering
ν1 χ1 χ2 χ3 χ4 χ5 χ6 χ7 χ8
ν2 ψ1 ψ2 ψ3 ψ4 ψ5 ψ6 ψ7 ψ8
of the termination conditions is satisfied, namely, either the maximum number of iterations
is reached or the best solution cannot be improved within a certain number of iterations.
γ
n
l
γi,j
m
l
j,k
ci,j,q − Pi si xi,j,q cj,k,q − Pk dk xj,k,q
C, 3.1
i 1 j 1q 1 k 1j 1q 1
n m
where constant C i 1 Pi si k 1 Pk dk and each coefficient is negative because we assume
that the unit penalty cost is higher than unit transportation cost. Let Ci,j,q represent the cost
supplier i can save if he ships cargos to crossdock j by bi,j,q r r
, ei,j,q and Cj,k,q represent the
cost customer k can save if he receives cargos from crossdock j by bj,k,q , ej,k,q , where Ci,j,q
a a
Pi − ci,j,q and Cj,k,q Pk − cj,k,q
, respectively.
From 3.1, we can see that, for a supplier, the most important point is saving cost Ci,j,q ,
which is the primary factor that determines which crossdock to ship to and which route to
be chosen between these two locations. This decision subsequently affects the holding cost in
the corresponding crossdock it ships to. To reflect this fact, we set probabilities for supplier-
crossdock assignments so that a higher cost saving of an assignment would result in a higher
probability for that assignment to be chosen. Formally, the probability that supplier i ships
cargos to crossdock j by schedule bi,j,q r r
, ei,j,q is calculated as follows:
Ci,j,q
Probi,j,q 1 ≤ i ≤ n, 1 ≤ j ≤ l, 1 ≤ q ≤ γi,j . 3.2
max Ci,j ,q
j ,q
For customers, we know which crossdocks have cargo after the first stage, and then we
assign one of these crossdocks to each customer and choose a route to deliver. The assignment
strategy of crossdocks and schedules for customer is similar to that of the suppliers, and we
also can calculate the probability similar to 3.2. Only one difference is that the customers
just can be assigned to those crossdocks that have been already assigned to suppliers, instead
of all the crossdocks. After that, we need adjust the solution according to time match criterion
to guarantee feasibility. During adjustment, we needs to eliminate those infeasible issues such
as time conflicts, overflow of capacity, and nonconservation of cargo flows. By adjustment,
infeasible solutions will scarcely be generated. However, for some infeasible solutions that
are too difficult to repair, we just need to unfulfill those customers who incur infeasibility to
get a feasible solution.
3.1.3. Crossover
In order to preserve efficient genes in a chromosome, the two-point crossover operator is
applied to generate offspring, which is widely adopted in GA see, e.g., 3, 18. The crossover
operator is illustrated by Figure 2. First, two individuals, we call parent 1 and parent 2, are
selected randomly from the population pool, and then two points are randomly selected
between genes representing assignment for suppliers, and because crossover operators
10 Mathematical Problems in Engineering
ν1 χ1 χ2 χ3 χ4 χ5 χ6 χ7 χ8
Supplier segment of parent 1
ν2 ψ1 ψ2 ψ3 ψ4 ψ5 ψ6 ψ7 ψ8
applied among customers will generate lots of infeasible solutions, assignments for customers
are determined by assignments for suppliers. The symbols outside the two crossover points
are directly inherited from parent 1 to offspring 1, and the other genes of offsprings 1
are transferred from the symbols of parent 2 in corresponding positions. After crossover,
crossdocks and schedules are reassigned for customers using the aforementioned greedy
technique. After changing the roles of parents, the same procedure is applied to generate
offspring 2.
3.1.4. Mutation
It is obvious that the initial population generated by two-stage greedy method has poor
ability to carry the genetic diversity because the cost saving priority and time match priority
in the greedy technique reduce the chance for crossdocks and schedules with relatively
low cost saving to be chosen. As a result, the greedy technique causes population pool
homogeneity. In order to overcome this limitation, we use the mutation operator with a
given individual mutation probability Pim to mutate every individual and apply the greedy
technique to mutation of some gene representing the assignment of crossdock to some
supplier with gene mutation probability Pgm calculated by 3.2 in the selected individual.
After that, we also need to adjust the new solution to be feasible by the strategy which
is similar to that of initial solutions. Different from prior crossover operator slightly, the
mutation operator may deteriorate the current solution in terms of fitness. However, its goal
is not only to preserve the best genes but also to attain inferior genes with some probability
to diversify the pool.
2 can ship out cargos earlier than customer 1 when the penalty cost of customer 2 is relatively
high. So the strategy aims to give a priority to customer 2.
4. Computational Experiments
We generate a great variety of problem instances and apply HGA to solve them. For
comparison purposes, we also use ILOG CPLEX 11.0 solver to solve the instances, which
is widely adopted by many papers see, e.g., 3, 21, 22. Both HGA and the CPLEX solver
were run on a personal computer with an Intel 2.4 GHz Pentium 4 CPU and 1G memory.
The test data generation, parameter settings of HGA, and detailed computational results are
reported in the following content.
in such a way that they can represent realistic situations and can cover different scenarios,
which is suggested by Chen et al. 21, Li et al. 3, and Ma et al. 22, and the parameters
in HGA are based on numerous computational experiments, and they are effective to attain
desirable results.
The test data generation procedure requires three basic parameters: the number of
suppliers n, the number of customers m, and the number of crossdocks l. The time horizon is
fixed at 48 hours 2 days in the test sets; note that this is usually the longest-time shipments
by railways between two cities. The n start points of supplier i time window bir 1 ≤ i ≤
n were then randomly generated from a uniform distribution U0, 12. The end points
of supplier i time window eir were also randomly generated from a uniform distribution
U12, 36. For customers, their time windows are generated as bka , eka , where bka ∼ U12, 24
and eka ∼ U24, 48. The number of fixed transportation schedules between two points is
randomly generated in the interval 6, 8. Meanwhile, the beginning time of the first fixed
transportation schedule from supplier to crossdock is generated according to penalty cost and
transportation cost, so the fixed schedule is as begin, end, where begin ∼ Ubri × Pi /Pi −
ci,j,1 , 12 and end ∼ U13, 24, which means that a higher penalty cost provides a supplier
with a higher motivation to ship out cargos. Other schedules are generated as begin, end,
where begin ∼ Uthefirst schedule begin time, 12 and end ∼ U13, 24. Similarly, the
arrival time of the last delivery schedule for customers is generated according to penalty cost
and delivery cost, so the time window of the last delivery schedule is as begin, end, where
begin ∼ U12, 35 and end ∼ U36, eka × cj,k,γ Pk /Pk . For others, the time window is as
j,k
begin, end, where begin ∼ U12, 35 and end: U36, the last schedule arrival time. Next,
because pickups usually follow deliveries within short times, we take the inventory cost at
crossdocks to be small relative to transportation costs. This reflects the fact that handling
costs are usually smaller than transportation costs. Based on this, the transportation cost
per unit cargo of each fixed scheduled route is uniformly generated in the interval 10, 30
and inventory handling cost per unit per hour is uniformly generated in the interval 1, 3,
which on average is 1/10 of transportation cost. The penalty cost is set to be relatively higher
compared to the transportation cost, which can enforce suppliers and customers to deliver
cargoes on time, so the penalty cost per unit cargo is uniformly generated in the interval
30, 90. Lastly, the amount of supplied cargo si demanded cargo dk is uniformly generated
in the interval 100, 500. The capacity of each crossdock is set to αΣ1≤i≤n si , where α is
randomly generated from a uniform distribution U0.5, 0.8. Also the capacity of each route is
set to βΣ1≤i≤n si /n, where β is randomly generated from a uniform distribution U2, 5. The
following values of parameters are used: # max iter 104 , #terminate iter 100, #pop 200,
and #crossover 80. The mutation probability Pim is taken to be 0.02, which is proved to be
effective in experiments.
Table 1: Result of CPLEX and HGA on random instances with small scale.
Problem size 10×4×10 12×4×12 14×4×14 16×4×16 18×4×18 10×6×10 12×6×12 14×6×14
LBs 104818 105058 139556 167451 144959 114548 104734 145931
CPLEX Objective 111094 107805 142217 174498 152658 123788 108595 155670
Times >3600 >3600 3301.1 >3600 >3600 >3600 >3600 >3600
Gap 5.52% 2.21% 1.85% 4.14% 4.02% 6.24% 3.36% 5.71%
Objective 110834 107808 142124 173815 150214 122921 108113 154486
HGA Times 472.93 781.9 530.79 922.69 562.30 726.26 608.8 1084.69
Gap 5.30% 2.21% 1.80% 3.79% 2.81% 5.59% 2.97% 5.14%
Table 2: Result of CPLEX and HGA on random instances with medium scale.
Problem size 16 × 10 ×
20×4×20 22×4×22 16×6×16 18×6×18 20×6×20 16×8×16 18×8×18
16
LBs 153165 197595 143037 161171 185072 131554 149797 132087
CPLEX Objective 158295 202441 154712 169924 195823 142704 161564 150179
Times >5000 >5000 >5000 >5000 >5000 >5000 >5000 >5000
Gap 2.64% 2.37% 7.48% 4.95% 5.17% 7.74% 6.90% 12.06%
Objective 157928 202398 153922 167337 194017 138126 159152 147352
HGA Times 1080.17 1741.77 1139.11 677.49 1326.22 675.84 1224.98 609.23
Gap 2.48% 2.34% 6.97% 3.69% 4.40% 4.40% 5.71% 10.22%
time, the gaps between value attained by both CPLEX and HGA, and the low bound attained
by CPLEX when terminated.
1 Small-size instances: the results are shown in Table 1. In this category, eight small
scale instance groups are generated with the size n and m ranging from 10 to 14,
and l ranging from 4 to 6. We use these instances to compare the performance of
the CPLEX solver and HGA. We find that, only in one group, CPLEX solver reaches
the LB within time limit set as 3600 s; for other cases, CPLEX fails to get the better
solutions within 3600 s comparing HGA, which gets better solutions more quickly,
and of which the average gaps are apparently smaller than CPLEX solver.
2 Medium-size instances: the results are reported in Table 2. In this category, eight
instance groups with the size n and m ranging from 16 to 22 and l ranging from 4 to
10 are tested. Time limit is set to more than 5000 s. Also CPLEX fails to get the better
solutions within the time limit for all the instance groups, while HGA performs
well in no more than 1200 s.
3 Large-size instances: the results can be found in Table 3. In this category, large-scale
instance groups are generated and categorized into 8 groups with the size n and m
ranging from 20 to 24 and l ranging from 8 to 12. The CPLEX solver is unable to
obtain the better solutions within the time limit, which is set to 7200 s; only in one
group CPLEX gets a better solution than HGA. However, HGA can attain much
better solutions in no more than 1800 s in the other seven cases.
All the three categories of 24 instance groups show that HGA performs fairly well and
is preferable over the commercial CPLEX solver.
The main feature of our proposed HGA is to integrate variable neighborhood search
VNS into a general GA framework so that it has the ability to get better solutions, especially
Mathematical Problems in Engineering 15
Table 3: Result of CPLEX and HGA on random instances with large scale.
Problem size 18 × 10 × 20 × 10 × 22 × 10 × 22 × 12 × 24 × 12 ×
20×8×20 22×8×22 24×8×24
18 20 22 22 24
LBs 179309 189188 196688 1495661 162114 170686 208110 192499
CPLEX Objective 195510 207942 217234 160386 186614 194127 232170 227976
Times >7200 >7200 >7200 >7200 >7200 >7200 >7200 >7200
Gap 7.68% 8.64% 9.20% 6.38% 13.12% 11.80% 10.36% 14.95%
Objective 194431 204953 216949 160445 183441 189251 227295 221557
HGA Times 1159.11 1718.59 1696.08 1314.28 1298.37 1316.45 1419.08 1545.45
Gap 7.30% 7.34% 9.07% 6.43% 11.71% 9.77% 8.44% 12.96%
for large-scale problem instances. Hence, by comparing the results of HGA and GA without
VNS, we can identify how much the solutions can be improved for large-size problem
instances. The numerical results are reported in Table 4, where each problem size has 5
instances, and Gap is defined by
The results show that our proposed HGA provides better solutions without sacrificing
much computational efforts compared with the GA. Specifically, the results show that HGA
outperforms GA for all the cases in terms of solution quality. Although the speed of HGA
is slower than GA, it is reasonable because HGA requires more time to search for a better
solution by applying VNS. This gives us a clearer idea of the performance of the proposed
HGA for the large-sized problems. That is, in general, HGA can provide high-quality
solutions in realistic timescales for large-size problems.
Note that our problem has many characteristics e.g., fixed transportation schedules,
inventory capacity, etc. different from the vehicle routing problem VRP, although the VRP
also considers how to find an optimal transportation scheme to satisfy customer demands.
The heuristic algorithms that can be very effective for the VRP cannot be applied to our
problem because these two types of problems have different structures and constraints.
5. Conclusions
In this paper, we consider multiple crossdocks problem through fixed transportation
schedules with time windows, capacity, and penalty. The objective is to minimize the total
costs including shipment costs, penalty cost, and inventory cost. Since we prove that the
16 Mathematical Problems in Engineering
Appendix
Proof of Proposition 2.1
We provide a reduction of the strongly NP-complete 3-partition problem: given positive
integers, w, D, and Γ {1, 2, . . . , 3w} with positive integer values γi where, for each
i ∈ Γ, i∈Γ γi wD and D/4 < γi < D/2 for i ∈ Γ, can Γ be partitioned into w disjoint sets
Γ1 , Γ2 , . . . ,Γw such that |Γk | 3 and i∈Γk γi D for k 1, . . . ,w? From an arbitrary instance
of 3-partition, we consider a polynomial reduction to an instance of our multiple crossdocks
problem and ask if there exists a feasible solution whose objective value is no greater than
2wD. For w suppliers given in Σ let Σ {1, 2, . . . , w} and 3w customers in Δ let Δ Γ,
let si be the supply and si D for i ∈ Σ with unit penalty cost 3, while for each k ∈ Δ, let
dk γk be the demand also with unit penalty cost 3. Exactly one crossdock, χ, say, with
inventory holding cost 1 per unit product per time, exists linking suppliers with customers.
For each supplier i i ∈ Σ, there is only one fixed transportation schedule i, i 1 with
unit transportation cost 1. On the other hand, for each customer k k ∈ Δ, there is w fixed
transportation schedule {q 1, q 2 : q 1, . . . , w} connected with crossdock χ also with
unit shipping cost 1.
We now show that a feasible schedule exists whose objective value is no greater than
2wD if and only if the 3-partition has a feasible solution. On the one hand, if 3-partition has a
feasible solution Γ1 ,. . . , Γw , note that we needs pay attention to the single shipping and single
delivery condition, and hence we should ship all goods provided by supplier i i ∈ Σ to χ
Mathematical Problems in Engineering 17
through fixed schedule i, i 1, respectively, and transship all of them to customer j j ∈ Γi
through i 1, i 2, which satisfies the demand γj for customer j j ∈ Γi exactly. It is
easy to verify that such a schedule is feasible and total cost is 2wD. On the other hand, if
a feasible schedule exists with objective no greater than 2wD, then it is optimal since it is
easy to prove that 2wD is the lower bound of our instance, whose reason is because the total
transportation cost is 2wD at least, and if any cargo is delayed in crossdock or any demand
is unfulfilled, then the total cost is definitely greater than 2wD. Hence, this optimal solution
must satisfy the following two conditions: 1 there is no inventory in crossdock at any time;
2 no penalty cost is incurred. We can then construct a partition by setting Δi to be the subset
of k ∈ Δ whose demand is satisfied by supplier i for 1 ≤ i ≤ w. Because of conditions 1
and 2, the demand of customer k k ∈ Δ is γk which should be satisfied immediately
by fixed schedule i 1, i 2. Moreover, because of the single shipping and single delivery
condition, we have k∈Δi dk D. Since D/4 < dk < D/2 for k ∈ Δ, we have |Δi | 3. Hence,
Δ1 , . . . , Δw is a feasible partition for the instance of 3-partition and this completes the proof.
Acknowledgments
The authors wish to acknowledge one anonymous referee whose comments have helped
to greatly improve this paper. This research was supported in part by NSFC 70802052,
70701039, 71072090, MOE NCET-10-0712, NCET-10-0847, the Fundamental Research
Funds for the Central Universities 2010221025, 10wkpy20, the Academic Outstanding
Youthful Research Talent Plan of Fujian Province JA10001S, and Soft Science Projects of
Fujian Province 2011R0081.
References
1 D. Simchi-Levi, P. Kaminsky, and E. Simchi-Levi, Designing and Managing the Supply Chain, McGraw-
Hill—Irwin, Boston, Mass, USA, 2nd edition, 2003.
2 T. Brockmann, “21 Warehousing trends in the 21st century,” IIE Solutions, vol. 31, pp. 36–40, 1999.
3 Y. Li, A. Lim, and B. Rodrigues, “Crossdocking—JIT scheduling with time windows,” Journal of the
Operational Research Society, vol. 55, no. 12, pp. 1342–1351, 2004.
4 M. Gumus and J. H. Bookbinder, “Cross-docking and its implication in location-distribution
systems,” Journal of Business Logistics, vol. 25, no. 2, pp. 199–228, 2004.
5 K. Krishnan and V. Rao, “Inventory control in N warehouses,” Journal of Industrial Engineering, vol.
16, pp. 212–215, 1965.
6 U. S. Kamarkar and N. R. Patel, “The one-period n-location distribution problem,” Naval Research
Logistics, vol. 24, no. 4, pp. 559–575, 1977.
7 U. S. Karmarkar, “The multilocation multiperiod inventory problem: bounds and approximations,”
Management Science, vol. 33, no. 1, pp. 86–94, 1987.
8 G. Tagaras, “Effects of pooling on the optimization and service levels of two-location inventory
systems,” IIE Transactions, vol. 21, no. 3, pp. 250–257, 1989.
9 L. W. Robinson, “Optimal and approximate policies in multiperiod, multilocation inventory models
with transshipments,” Operations Research, vol. 38, no. 2, pp. 278–295, 1990.
10 N. Rudi, S. Kapur, and D. F. Pyke, “A two-location inventory model with transshipment and local
decision making,” Management Science, vol. 47, no. 12, pp. 1668–1680, 2001.
11 J. Grahovac and A. Chakravarty, “Sharing and lateral transshipment of inventory in a supply chain
with expensive low-demand items,” Management Science, vol. 47, no. 4, pp. 579–594, 2001.
12 Y. T. Herer and M. Tzur, “The dynamic transshipment problem,” Naval Research Logistics, vol. 48, no.
5, pp. 386–408, 2001.
18 Mathematical Problems in Engineering
13 Y. T. Herer, M. Tzur, and E. Yücesan, “Transshipments: an emerging inventory recourse to achieve
supply chain leagility,” International Journal of Production Economics, vol. 80, no. 3, pp. 201–212, 2002.
14 S. Axsäter, “A new decision rule for lateral transshipments in inventory systems,” Management Science,
vol. 49, no. 9, pp. 1168–1179, 2003.
15 S. Axsäter, “Evaluation of unidirectional lateral transshipments and substitutions in inventory
systems,” European Journal of Operational Research, vol. 149, no. 2, pp. 438–447, 2003.
16 H. Donaldson, E. L. Johnson, H. D. Ratliff, and M. Zhang, “Schedule-driven crossdocking networks,”
Research Report, Georgia Institute of Technology, 1999.
17 D. H. Ratcliff, J. van de Vate, and M. Zhang, “Network design for load-driven cross-docking systems,”
Research Report, Georgia Institute of Technology, 1999.
18 Z. Miao, A. Lim, and H. Ma, “Truck dock assignment problem with operational time constraint within
crossdocks,” European Journal of Operational Research, vol. 192, no. 1, pp. 105–115, 2009.
19 N. Boysen, M. Fliedner, and A. Scholl, “Scheduling inbound and outbound trucks at cross docking
terminals,” OR Spectrum, vol. 32, no. 1, pp. 135–161, 2009.
20 A. Lim, Z. Miao, B. Rodrigues, and Z. Xu, “Transshipment through crossdocks with inventory and
time windows,” Naval Research Logistics, vol. 52, no. 8, pp. 724–733, 2005.
21 P. Chen, Y. Guo, A. Lim, and B. Rodrigues, “Multiple crossdocks with inventory and time windows,”
Computers and Operations Research, vol. 33, no. 1, pp. 43–63, 2006.
22 H. Ma, Z. Miao, A. Lim, and B. Rodrigues, “Crossdocking distribution networks with setup cost and
time window constraint,” Omega-International Journal of Management Science, vol. 39, no. 1, pp. 64–72,
2011.
23 L. Hai and A. Lim, “A metaheuristic for the pickup and delivery problem with time windows,”
International Journal on Artificial Intelligent Tools, vol. 12, no. 2, pp. 173–186, 2003.
24 M. Gendreau, F. Guertin, J. Y. Potvin, and R. Séguin, “Neighborhood search heuristics for a dynamic
vehicle dispatching problem with pick-ups and deliveries,” Transportation Research Part C, vol. 14, no.
3, pp. 157–174, 2006.
Hindawi Publishing Corporation
Mathematical Problems in Engineering
Volume 2012, Article ID 463976, 40 pages
doi:10.1155/2012/463976
Research Article
A Nonlinear Multiobjective Bilevel Model for
Minimum Cost Network Flow Problem in a
Large-Scale Construction Project
Copyright q 2012 Jiuping Xu et al. This is an open access article distributed under the Creative
Commons Attribution License, which permits unrestricted use, distribution, and reproduction in
any medium, provided the original work is properly cited.
The aim of this study is to deal with a minimum cost network flow problem MCNFP in a large-
scale construction project using a nonlinear multiobjective bilevel model with birandom variables.
The main target of the upper level is to minimize both direct and transportation time costs. The
target of the lower level is to minimize transportation costs. After an analysis of the birandom
variables, an expectation multiobjective bilevel programming model with chance constraints is
formulated to incorporate decision makers’ preferences. To solve the identified special conditions,
an equivalent crisp model is proposed with an additional multiobjective bilevel particle swarm
optimization MOBLPSO developed to solve the model. The Shuibuya Hydropower Project is
used as a real-world example to verify the proposed approach. Results and analysis are presented
to highlight the performances of the MOBLPSO, which is very effective and efficient compared to
a genetic algorithm and a simulated annealing algorithm.
1. Introduction
Network flow optimization is a large part of combinatorial optimization. The minimum cost
network flow problem MCNFP is made up of a wide category of problems 1, 2. MCNFP
plays a very important role in many real-world applications such as communications 3, 4,
informatics 5, and transportation 6. Other well-known problems like the shortest path
problem and the assignment problem are considered to be special MCNFP cases 7.
In recent decades, the MCNFP has been well researched with many models and algo-
rithms being developed, for example, 8–13. These studies, however, have not often taken
carrier type selection and transportation time into account when looking at the transportation
network. Yet, both cost and time control are important in construction projects, especially in
2 Mathematical Problems in Engineering
large-scale construction projects where transportation costs are largely based on the rates
charged by carriers, which have a significant influence on transportation time. In real condi-
tions, there is increasing pressure to shorter transportation time to reduce or eliminate extra
project expenses, with the early arrival of materials shortening the completion time of the
construction project and improving construction efficiency. In these cases, it is necessary to
include both the carrier type selection or transportation time in the transportation network
analysis. In this paper, a multiobjective bilevel MCNFP is studied. On the upper level, the con-
struction contractor determines the material flow of each transportation network path with
the criteria being the minimization of both direct costs and total transportation time costs. On
the lower level, the transportation manager controls each carrier’s flow so that total transpor-
tation costs are minimized.
The research presented previously has a common foundation in that they were all
based on a deterministic transportation network. However, transportation systems are often
complex, so decision makers inevitably encounter uncertain parameters when making a
decision. Within the last two decades the use of multimodels with uncertain parameters in the
study of network flow problems has been increasingly exploited. For example, Watling 14
studied a user equilibrium traffic network assignment problem with stochastic travel times
and a late arrival penalty. Chen and Zhou 15 developed an α-reliable mean-excess traffic
equilibrium model with stochastic travel times. Lin 16 constructed a revised stochastic flow
network to model a realistic computer network in which each arc has a lead time and a
stochastic capacity. Sumalee et al. 17 dealt with a reliable network design problem which
was looked at uncertain demand and total travel time reliability. In actual analyses, ran-
domness is considered one important source of uncertainty. Yet with MCNFP randomness is
seen to be increasingly complex because of the often incomplete or uncertain information.
To date, there has been little research which considers multilevel twofold uncertainty coeffi-
cients for MCNFP. Therefore this research concentrates on the problem under a birandom
environment with the logic behind this choice of birandom variables illustrated in Section 2.
The MCNFP proposed in this paper is a multiobjective bilevel programming problem,
first introduced by Geoffrion and Hogan 18, and consequently developed by researchers
such as Tarvainen and Haimes 19, Osman et al. 20, Zhang et al. 21, and Calvete and
Galéb 22. Multiobjective bilevel programming has been greatly improved in both the
theoretical and practical areas. While these studies have significantly contributed to a variety
of applications, to the best of our knowledge, there is still no known research considering the
modeling for the MCNFP. With bilevel programming problems being intrinsically difficult,
it is not surprising that most exact algorithms to date have focused on the simplest cases of
bilevel programs, that is, problems with relatively easy to determine properties such as linear,
quadratic, or convex objective and/or constraint functions 23. Since the proposed bilevel
MCNFP model is nonlinear, nonconvex, and nondifferentiable, it follows that the search for
exact algorithms which are formally efficient is all but futile and it is necessary instead to
search for effective heuristic algorithms to solve the MCNFP. Determining the global optimal
solution is of great importance in MCNFP. Specifically, this paper deals with the multiple
objectives by employing the concept of nondominated solutions instead of applying
weighted sum scalarization. In this study, an effort is made to develop a multiobjective bilevel
particle swarm optimization MOBLPSO to solve a real world MCNFP in the Shuibuya
Hydropower Project.
The remainder of this paper is structured as follows. In Section 2, an introduction to
the bilevel MCNFP is presented along with the motivation for employing birandom variables
in the problem. An expectation multiobjective bilevel programming model with chance
Mathematical Problems in Engineering 3
constraints under a birandom environment is established in Section 3, for which the equi-
valent crisp model is derived in Section 4. In addition, an MOBLPSO is illustrated in Section 5.
In Section 6, an application to earth-rock work transportation in a large construction project
is given in order to show the validity and efficiency of the proposed models and algorithms.
Concluding remarks and further discussion are in Section 7.
2. Problem Statement
The MCNFP discussed considers both the construction contractor and the transportation
manager as the two participants. In a large-scale construction project, material often has
supply origin and receipt destination nodes, with the construction contractor generally
assigning a specialized Transportation Company. The bilevel model considers the construc-
tion contractor and the transportation manager in the specialized Transportation Company
concurrently, gives priority to contractor benefit, and considers the influence of the contrac-
tor’s decision making on the flow distribution of the transportation manager’s carriers. As
both cost and time control are important in construction projects, their effectiveness needs to
be considered. The construction contractor assigns the flow of material to each transportation
path to minimize direct costs and transportation time costs, while the transportation manager
aims to minimize transportation costs by making decisions about flow of material by each
carrier through the transportation path based on the construction contractor’s decision
making, which in turn influences contractor’s decision-making through adjustments to the
flow of material by each carrier along the transportation path.
Therefore, the MCNFP in this paper can be abstracted as a bilevel programming prob-
lem. To model the problem conveniently, the involved transportation network is considered
a bipartite network represented by a graph with sets of nodes and arcs. In the network,
a node represents the facilities in the network, for instance, a station or a yard, and an arc re-
presents a line between two adjacent facilities. The model structure of the MCNFP is in
Figure 1.
The birandom environment has been successfully studied and applied in many areas, such as
flow shop scheduling problem 24, portfolio selection 25, and vendor selection 26. These
studies show the necessity of considering birandom environment in practical problems. There
is a strong motivation for considering birandom environment for the MCNFP.
In real conditions, the transportation plan is usually made before the occurrence
of any transportation activity; thus the determined values of some parameters cannot be
obtained in advance; so there is a strong need to consider uncertainty in transportation
4 Mathematical Problems in Engineering
Transportation
manager
1 1 1 1 1 1
Transportation network
Birandom variable: transportation time, transportation cost
Minimize the
transportation cost
Transportation environment
The speed of the carrier is influenced by the Mean transportation time μ is determined
vehicle condition by the speed of the carrier
k
Specify a realistic distribution for μ ti ∼ N μ, σ 2 , μ ∼ N μ, σ0 2
3. Modelling
In this section, the relevant assumptions and notations are first outlined, some basic know-
ledge about the birandom variable is introduced, and then the nonlinear multiobjective bi-
level MCNFP model under a birandom environment is formulated.
3.2. Notations
The following mathematical notations are used to describe the MCNFP.
Certain Parameters
ri : maximal passing capacity of arc i;
wk : weight for carrier k in the transportation network;
vk : volume capacity of carrier k;
Tj : transportation time constraint represents the time of transportation path j in Tj
units in which the material demand between all OD pairs has to be transported;
j
γi : a binary variable equal to 1 if and only if arc i is a segment of transportation path j
for carrier k;
Qod : transportation demand of the material from origin node o to destination node d;
cj : direct cost of unit volume material using transportation j;
: ceiling operator rounding upward to integer.
Uncertain Parameters
k
ei : unit transportation cost of material flow on arc ai for carrier k;
tk : free transportation time of material flow on arc i for carrier k;
i0
Decision Variables
xj : volume of material flow on transportation path j, which is the decision variable of
the upper level;
ykj : volume of material flow transported by carrier k through path j, which is the
decision variable of the lower level.
Mathematical Problems in Engineering 7
Definition 3.1 see 32. A birandom variable ξ is a mapping from a probability space
Ω, A, Pr to a collection S of random variables such that for any Borel subset B of the real
line the induced function Pr{ξω ∈ B} is a measurable function with respect to ξ.
For each given Borel subset B of the real line , the function Pr{ξω ∈ B} is a random
variable defined on the probability space Ω, A, Pr.
Example 3.2. Let Ω {ω1 , ω2 , . . . , ωm }, and Pr{ω1 } Pr{ω2 } ··· Pr{ωm−1 } Pr{ωm } 1.
Assume that ξ is a function on Ω, A, Pr as follows:
⎧
⎪
⎪
⎪ ξ1 , if ω ω1 ,
⎪
⎪
⎪
⎪
⎪
⎪ ξ2 , if ω ω2 ,
⎪
⎨
..
ξω . 3.1
⎪
⎪
⎪
⎪
⎪
⎪ξm−1 , if ω ωm−1 ,
⎪
⎪
⎪
⎪
⎩ξ , if ω ωm ,
m
where ξ1 is a uniformly distributed random variable on 0, 1, ξ2 , . . . , ξm−1 are normally
distributed random variables with a mean 1 and a standard variance 0.5, and ξm is a standard
normally distributed random variable with a mean 0 and a standard variance 1, that is,
ξ1 ∼ U0, 1, ξ2 ∼ N1, 0.5, . . . , ξm−1 ∼ N1, 0.5, and ξm ∼ N0, 1. From Definition 3.1, ξ
is clearly a birandom variable as shown in Figure 3.
Example 3.3. Assume that a and b are two random variables defined on Ω , A , Pr , and for
any ω∗ ∈ Ω , bω∗ ≥ aω∗ holds; then random variable ξ ∼ Naω∗ , bω∗ is a birandom
variable.
Example 3.4. Let ξ be a random variable defined on the probability space Ω, A, Pr satisfying
is also a normally distributed random variable on Ω , A , Pr with the
μ, σ 2 , where μ
ξ ∈ N
mean μ and variance σ ∗2 . Then ξ is a birandom variable.
Definition 3.5. A birandom variable ξ is said to be normal, if for each ω, ξω is a random
variable with the following probability density function:
2
1 x − μω
φx √ exp − , 3.2
σω 2π 2σω2
where the number of random variable of μω and σω is not less than one. The normal
birandom variable is denoted by Nμω and σω.
8 Mathematical Problems in Engineering
Ω
∈
ω
p x p x
ξ m−1 ωm−1 ξ m ωm
p ω x
ξ
ωm
1 p x
ξ 1 ω1
ωm−1
p x
ξ 2 ω2
··
··
··
ω3
−4 ω2
−3
−2 ω1
−1
0
x∈
R 1
2
3
4
the expected value of the total transportation cost is the objective of the lower level. Denote
the expected total transportation cost of material as Cykj ; then the objective function of the
lower level model can be formulated as
⎡ ⎤
j k ykj
min C ykj E⎣ γi ei ⎦. 3.3
k∈Ψ j∈Ω i∈Φ
vk
⎧
⎨1, if i ∈ Aj , j ∈ Ω,
j
γi 3.4
⎩0, otherwise.
where α and β are user-defined parameters and, in this problem, are set to 0.15 and 2.0,
respectively.
k
Technically, it is not possible to strictly ensure that the random event i∈Φ γi ti does
j
k
not exceed T because of the birandom variable t . In practical problem, the decision makers
j i0
often provide an appropriate budget Tj in advance, to ensure that the restriction is, to a certain
k
extent, satisfied, that is to maximize the probability of the random event Pr{ i∈Φ γi ti0 ω1
j
j β
α j∈Ω k∈Ψ γi ykj /vk /ri ≤ Tj } under a given confidence level, which can be written as
follows:
⎧ ⎧ ⎡ ⎛ ⎤ ⎫ ⎫
⎪ ⎪ j ⎞β ⎪ ⎪
⎨ ⎨ γ y /v ⎬ ⎬
k
⎢ ⎠ ⎥
γi ti0 ω⎣1
kj k
α⎝
j j∈Ω k∈Ψ i
Pr ω | Pr ⎦ ≤ T j ≥ θ ≥ δ, j ∈ Ω, k ∈ Ψ.
⎪
⎩ ⎪
⎩i∈Φ ri ⎪
⎭ ⎪
⎭
3.6
Here the decision makers’ aspiration level is indicated as θ, so we use a “Pr” to ensure that
the constraint holds at the predetermined confidence level. Additionally, based on probability
theory, a further “Pr” is needed to describe the random elements presented in Section 2,
10 Mathematical Problems in Engineering
which guarantee the establishment of a certain confidence level δ, resembling the P-model
probability maximization model presented in 34.
The transportation flow may exceed some arcs’ capacity because of uncertainties such
as the condition of the construction project road. Such conditions may require the manager
to select another path. Thus, the total amount of capacity on arc i cannot exceed the maximal
capacity of the arc i, which produces the following constraint:
j ykj
γi ≤ ri , i ∈ Φ. 3.7
j∈Ω k∈Ψ
vk
Actually, it is difficult to ensure that each service carrier is fully loaded; so the sum
of all flows transported by all kinds of carriers through each path cannot be less than the
material flow assigned to it; thus the following constraint is obtained:
ykj ≥ xj , j ∈ Ω. 3.8
k∈Ψ
The path flow in path j used by carrier k should not be negative, such that
ykj ≥ 0, j ∈ Ω, k ∈ Ψ. 3.9
k
is a birandom variable, i∈Φ γi ti can be regarded as a special birandom variable. Similarly,
j
the expected value of the total transportation time cost is one of the objectives on the
k
upper level. Different carriers are given different weights, and ti tk 1
i0
j β
α j∈Ω k∈Ψ γi ykj /vk /ri , so the second objective function of the upper level can be
described as follows:
⎡ ⎡ ⎛ ⎞ ⎤⎤
j β
⎢ j ⎢ k∈Ψ γi ykj /vk
⎠ ⎥⎥
k
γi ti0 ⎣1 α⎝
j∈Ω
T xj , ykj E⎣ wk ⎦⎦. 3.11
k∈Ψ j∈Ω i∈Φ
ri
The following constraint ensures that the sum of the weights is equal to 1:
wk 1. 3.13
k∈Ψ
In order to describe the nonnegative variables, the constraints in 3.14 are presented:
xj ≥ 0, j ∈ Ω. 3.14
min C xj , ykj cj xj ,
o,d∈E j∈Pod
⎡ ⎡ ⎛ ⎤⎤
j ⎞β
⎢ j ⎢ γ y /v
⎠ ⎥ ⎥
k
γi ti0 ⎣1
kj k
α⎝
j∈Ω k∈Ψ i
min T xj , ykj E⎣ wk ⎦⎦,
k∈Ψ j∈Ω i∈Φ
ri
12 Mathematical Problems in Engineering
⎧
⎪
⎪ xj Qod , o, d ∈ E,
⎪
⎪
⎪
⎪ j∈Pod
⎪
⎪
⎪
⎪ wk 1,
⎪
⎪
⎪
⎪
⎪
⎪
k∈Ψ
⎪
⎪ xj ≥ 0, j ∈ Ω,
⎪
⎪ ⎡ ⎤
⎪
⎪
⎪
⎪ ykj
⎪ j k
⎪
⎪
⎪ min C ykj E⎣ γi ei ⎦,
⎪
⎪ vk
⎪
⎪ ⎧ ⎧ ⎧k∈Ψ j∈Ω i∈Φ ⎡ ⎤ ⎫ ⎫
⎪
⎪ ⎛ ⎞β
⎪
⎪ ⎪ ⎪ ⎪ j ⎪ ⎪
⎪ ⎪
⎪ ⎨ ⎨ γ y /v ⎬ ⎬
⎪ ⎪ k
⎢ ⎠ ⎥
γi ti0 ω⎣1 α⎝
kj k
⎪
⎨ ⎪
⎪ Pr ω | Pr
j j∈Ω k∈Ψ i
⎦ ≤ Tj ≥ θ ≥ δ,
⎪
⎪ ⎪ ⎪ ri ⎪ ⎪
s.t. ⎪
⎪ ⎩ ⎩ i∈Φ ⎭ ⎭
⎪ ⎪
⎪
⎪
⎪ ⎪
⎪
⎪
⎪ ⎪
⎪ j ∈ Ω, k ∈ Ψ,
⎪
⎪ ⎪
⎪
⎪
⎪ ⎪
⎪
⎪
⎪ ⎪
⎪ j ykj
⎪
⎪ ⎨ γi ≤ ri , i ∈ Φ,
⎪
⎪
⎪
⎪ s.t. vk
⎪
⎪ ⎪
⎪ ⎧
j∈Ω k∈Ψ
⎪
⎪ ⎪
⎪ ⎨1, if i ∈ Aj , j ∈ Ω,
⎪
⎪ ⎪
⎪
⎪
⎪ ⎪
⎪ γ
j
⎪
⎪ ⎪ i ⎩
⎪
⎪
⎪ ⎪
⎪ 0, otherwise,
⎪
⎪ ⎪
⎪
⎪
⎪ ⎪
⎪
⎪
⎪ ⎪
⎪ ykj ≥ xj , j ∈ Ω,
⎪
⎪ ⎪
⎪
⎪
⎪ ⎪
⎪ k∈Ψ
⎩ ⎩ y ≥ 0, j ∈ Ω, k ∈ Ψ.
kj
3.15
Definition 4.1 see 35. Let ξ be a random variable on the probability space Ω, A, Pr. Then
the expected value of ξ is defined by
∞ 0
Eξ Pr{ξ ≥ r}dr − Pr{ξ ≤ r}dr. 4.1
0 −∞
Note that the terms expected value, expectation, and mean value can be used inter-
changeably.
Definition 4.2 see 36. Let ξ be a birandom variable on the probability space Ω, A, Pr; then
the expected value of birandom variable ξ can be defined as follows:
∞ 0
Eξ Pr{ω ∈ Ω | Eξω ≥ t}dt − Pr{ω ∈ Ω | Eξω ≤ t}dt, 4.2
0 −∞
Lemma 4.3 see 32. Let ξ be a birandom variable on the probability space Ω, A, Pr. If the
expected value Eξω of the random variable ξω is finite for each ω, then Eξω is a random
variable on Ω, A, Pr.
Remark 4.4. It should be noted that the expected value operator E, which appears on both
sides of the previous definition of Eξ, is overloaded. In fact, symbol E represents different
meanings. That is to say, overloading allows us to use the same symbol E for different
expected value operators, because we can deduce the meaning from the type of argument.
k
Lemma 4.5. Let ti0 ω be a normal birandom variable, subject to a normal distribution
μtki0 ω, σ1t
N 2
k ω where μ
tki0 ∼ Nμtki0 , σ2t
2
k . Then the objective function in 3.15
i0 i0
⎡ ⎡ ⎛ ⎤⎤
j ⎞β
⎢ j ⎢ γ y /v
⎠ ⎥ ⎥
k
γi ti0 ⎣1
kj k
α⎝
j∈Ω k∈Ψ i
T xj , ykj E⎣ wk ⎦⎦ 4.3
k∈Ψ j∈Ω i∈Φ
ri
⎡ ⎛ ⎞ ⎤
j β
⎢ k∈Ψ γi ykj /vk
T xj , ykj wk
j
γi μtki0 ⎣1 α⎝
j∈Ω
⎠ ⎥ ⎦. 4.4
k∈Ψ j∈Ω i∈Φ
ri
k
Proof. From the assumption, for any ω ∈ Ω, ti0 ω ∼ N
μtki0 ω, σ1t
2
k ω is an independent
i0
tki0 ∼ Nμtki0 , σ2t
random variable, where μ 2
k . From Definition 4.2,
i0
k ∞ ! k " 0 ! k "
E ti0
Pr ω ∈ Ω | E ti0 ω ≥ t dt − Pr ω ∈ Ω | E ti0 ω ≤ t dt, 4.5
0 −∞
k
since t1i0 ω ∼ N
μtki0 ω, σt2k ω, and by Definition 4.1, function 4.5 is transformed as
i0
follows:
k ∞ # $ 0 # $
E ti0 Pr μ tki0 ω ≥ t dt − Pr μ tki0 ω ≤ t dt μtki0 , 4.6
0 −∞
and then
14 Mathematical Problems in Engineering
⎡ ⎡ ⎛ ⎤⎤
j ⎞β
⎢ j ⎢ γ y /v
⎠ ⎥ ⎥
k
γi ti0 ⎣1
kj k
α⎝
j∈Ω k∈Ψ i
T xj , ykj E⎣ wk ⎦⎦
k∈Ψ j∈Ω i∈Φ
ri
⎡ ⎛ ⎤
k j ⎞β
⎢ γ y /v
γi E ti0 ⎣1 ⎠ ⎥
kj k
α⎝
j j∈Ω k∈Ψ i
T xj , ykj wk ⎦
k∈Ψ j∈Ω i∈Φ
ri
⎡ ⎛ ⎤
j ⎞β
⎢ γ y /v
⎠ ⎥
kj k
α⎝
j j∈Ω k∈Ψ i
wk γi μtki0 ⎣1 ⎦.
k∈Ψ j∈Ω i∈Φ
ri
4.7
j ykj
C ykj γi μeik . 4.8
k∈Ψ j∈Ω i∈Φ
vk
k
Lemma 4.6. Assume that ti0 ω ∼ N μtki0 ω, σ1t
2
k ω is a normal birandom variable, where μ tki0 is
i0
j
a normal distributed random variable characterized by μ tki0 ∼ Nμtki0 , σ2t
2
k ; then tki0 ω1
i∈Φ γi μ
i0
j β
α j∈Ω k∈Ψ γi ykj /vk /ri − Tj ∈ Nμr , σr2 is also a random variable, where
⎡ ⎛ ⎤
j ⎞β
⎢ γ y /v
⎠ ⎥
kj k
α⎝
j j∈Ω k∈Ψ i
μr γi μtki0 ω⎣1 ⎦ − Tj ,
i∈Φ
ri
%
& 4.9
⎡ ⎛ ⎤
& j ⎞β 2
& ( j )2 k∈Ψ γi ykj /vk
& 2 ⎢ ⎠ ⎥
α⎝
j∈Ω
σr ' γi σ2t k ⎣1 ⎦ .
i∈Φ
i0 ri
Then the following constraint for the first constraint is derived in 3.15:
⎧ ⎧ ⎡ ⎛ ⎞ ⎤ ⎫ ⎫
⎪ ⎪ j β ⎪ ⎪
⎨ ⎨ k∈Ψ γi ykj /vk
⎬ ⎬
k
⎢ ⎠ ⎥
γi ti0 ω⎣1 α⎝
j j∈Ω
Pr ω | Pr ⎦ ≤ Tj ≥θ ≥ δ, 4.10
⎪
⎩ ⎪
⎩ i∈Φ ri ⎪
⎭ ⎪
⎭
4.11
k
Proof. From the assumption, it is known that for any ω ∈ Ω, ti0 ω ∼ N
μtki0 ω, σ1t
2
k ω is an
i0
independent random variable, so it follows that
⎡ ⎛ ⎤
j ⎞β
γ y /v ( )
k
⎢ ⎠ ⎥
γi ti0 ω⎣1
kj k
α⎝
j j∈Ω k∈Ψ i
⎦ − Tj ∼ N μ r ω, σr2 ω 4.12
i∈Φ
ri
⎡ ⎛ ⎤
j ⎞β
⎢ γ y /v
⎠ ⎥
kj k
α⎝
j j∈Ω k∈Ψ i
r ω
μ tki0 ω⎣1
γi μ ⎦ − Tj ,
i∈Φ
ri
⎡ ⎤ 4.13
⎛ j ⎞β 2
( j )2 γ y /v
⎢ kj k
⎠ ⎥
α⎝
j∈Ω k∈Ψ i
σr2 ω γi σ1t2
k ω⎣1 ⎦ .
i∈Φ
i0 ri
Then
⎧ ⎡ ⎛ ⎤ ⎫
⎪ j ⎞β ⎪
⎨ γ y /v ⎬
k
⎢ ⎠ ⎥
γi ti0 ω⎣1
kj k
α⎝
j j∈Ω k∈Ψ i
Pr ⎦ ≤ Tj ≥ θ
⎪
⎩ i∈Φ ri ⎪
⎭
⎧* ( )β
+ ⎫
j
k
⎪ j ⎪
⎪
⎨ i∈Φ γi ti0 ω 1 α j∈Ω k∈Ψ γi ykj /vk /ri − Tj − μ
r ω ⎪
μr ω⎬
−
⇐⇒ Pr ≤ ≥θ
⎪
⎪ σr ω ⎪
σr ω ⎪
⎩ ⎭
* +
−μr ω
⇐⇒ Φ r ω ≤ −Φ−1 θ1r1
≥ θ ⇐⇒ μ σr ω
σr ω
⎡ ⎛ ⎤
j ⎞β
⎢ γ y /v
⎠ ⎥
kj k
tki0 ω⎣1 α⎝
j j∈Ω k∈Ψ i
⇐⇒ γi μ ⎦ − Tj
i∈Φ
ri
%
& ⎡ ⎛ ⎤
& j ⎞β 2
& ( j )2 k∈Ψ γi ykj /vk
& ⎢ ⎠ ⎥
α⎝
−1 j∈Ω
≤ −Φ θ' γi σ1t2
k ω⎣1 ⎦ .
i∈Φ
i0 ri
4.14
16 Mathematical Problems in Engineering
j j β
tki0 ∼ Nμtki0 , σ2t
Since μ 2
k , then tki0 ω1
i∈Φ γi μ α j∈Ω k∈Ψ γi ykj /vk /ri − Tj ∈
i0
Nμr , σr2 , where
⎡ ⎛ ⎤
j ⎞β
⎢ γ y /v
⎠ ⎥
kj k
α⎝
j j∈Ω k∈Ψ i
μr γi μtki0 ω⎣1 ⎦ − Tj ,
i∈Φ
ri
%
& 4.15
⎡ ⎛ ⎤
& j ⎞β 2
& ( j )2 k∈Ψ γi ykj /vk
& 2 ⎢ ⎠ ⎥
α⎝
j∈Ω
σr ' γi σ2t k ⎣1 ⎦ ,
i∈Φ
i0 ri
⎧ ⎧ ⎡ ⎛ ⎤ ⎫ ⎫
⎪ ⎪ j ⎞β ⎪ ⎪
⎨ ⎨ γ y /v ⎬ ⎬
k
⎢ ⎠ ⎥
γi ti0 ω⎣1
kj k
α⎝
j j∈Ω k∈Ψ i
Pr ω | Pr ⎦ ≤ T j ≥ θ ≥δ
⎪
⎩ ⎪
⎩ i∈Φ ri ⎪
⎭ ⎪
⎭
⎧ ⎡ ⎛ ⎞ ⎤ ⎫
⎪ j β ⎪
⎨ k∈Ψ γi ykj /vk
⎬
⎢ ⎠ ⎥
α⎝
j j∈Ω
⇐⇒ Pr ω | tki0 ω⎣1
γi μ ⎦ − Tj ≤ M ≥δ
⎪
⎩ i∈Φ
ri ⎪
⎭
⎧ ( ⎫
⎪ j j )β ⎪
⎪
⎨ tki0 ω 1
i∈Φ γi μ α k∈Ψ γi ykj /vk /ri − T j − μr ⎪
j∈Ω
M − μr ⎬
⇐⇒ Pr ω | ≤ ≤δ
⎪
⎪ σr σr ⎪ ⎪
⎩ ⎭
%
& ⎡ ⎛ ⎤
& j ⎞β 2
& ( j )2 k∈Ψ γi ykj /vk
& ⎢ ⎠ ⎥
α⎝
−1 −1 j∈Ω
⇐⇒ M ≥ μr σr Φ δ ⇐⇒ Φ θ' γi σ1t2
k ω⎣1 ⎦
i∈Φ i0 ri
μr σr Φ−1 δ ≤ 0,
4.16
,
j 2 2 j β 2
where M −Φ−1 θ i∈Φ γi σ1tk ω 1 α j∈Ω k∈Ψ γi ykj /vk /ri .
i0
Based on Lemmas 4.3 and 4.5, it is determined that the expectation multiobjective
bilevel programming model with chance constraints 3.15 is equivalent to the following crisp
nonlinear multiobjective bilevel programming problem:
Mathematical Problems in Engineering 17
min C xj , ykj cj xj ,
o,d∈E j∈Pod
⎡ ⎛ ⎤
j ⎞β
⎢ γ y /v
⎠ ⎥
kj k
α⎝
j j∈Ω k∈Ψ i
min T xj , ykj wk γi μtki0 ⎣1 ⎦,
k∈Ψ j∈Ω i∈Φ
ri
⎧
⎪
⎪ xj Qod , o, d ∈ E,
⎪
⎪
⎪
⎪ j∈Pod
⎪
⎪
⎪
⎪ wk 1,
⎪
⎪
⎪
⎪ k∈Ψ
⎪
⎪
⎪
⎪ xj ≥ 0, j ∈ Ω,
⎪
⎪
⎪
⎪
⎪
⎪ ykj
⎪
⎪ min C ykj
j
γi μeik ,
⎪
⎪ vk
⎪
⎪ ⎧
⎪
⎪ %
&
k∈Ψ j∈Ω i∈Φ
⎡ ⎤
⎪
⎪ ⎪ ⎛ ⎞β 2
⎪
⎪ ⎪
⎪ & j
⎪
⎪ ⎪
⎪ & ( ) γ y /v
⎪ ⎪ & 2 ⎢ kj k
⎠ ⎥
α⎝
j∈Ω k∈Ψ i
⎨ ⎪ −1 j
⎪Φ θ' k ω⎣1
2
⎪ γi σ1t ⎦
⎪
⎪ ri
⎪
i0
s.t. ⎪ i∈Aj
⎪
⎪ ⎪
⎪
⎪
⎪ ⎪
⎪
⎪
⎪ ⎪
⎪ μr σr Φ−1 δ ≤ 0, j ∈ Ω, k ∈ Ψ,
⎪
⎪ ⎪
⎪
⎪
⎪ ⎪
⎪
⎪
⎪ ⎨ j ykj
⎪
⎪ γi ≤ ri , i ∈ Φ,
⎪
⎪ s.t vk
⎪
⎪ ⎪
⎪ j∈Ω k∈Ψ ⎧
⎪
⎪ ⎪
⎪
⎪
⎪ ⎪
⎪ ⎨1, if i ∈ Aj , i ∈ Ω,
⎪
⎪ ⎪
⎪ j
⎪
⎪ ⎪
⎪ γ
⎪
⎪ ⎪
⎪
i ⎩0, otherwise,
⎪
⎪ ⎪
⎪
⎪
⎪ ⎪
⎪
⎪
⎪ ⎪
⎪ ykj ≥ xj , j ∈ Ω,
⎪
⎪ ⎪
⎪
⎪
⎪ ⎪k∈Ψ
⎪
⎪
⎩ ⎪
⎩y ≥ 0, j ∈ Ω, k ∈ Ψ,
kj
4.17
j j β
where μr i∈Φ γi μtki0 ω1 α j∈Ω k∈Ψ γi ykj /vk /ri − Tj , σr
,
j 2 j β 2
i∈Φ γi σ2t
2
k 1 α j∈Ω k∈Ψ γi ykj /vk /ri and θ and δ are predicted confidence
i0
5. Solution Approach
To obtain an analytical optimal solution for a bilevel programming problem BLPP is
difficult, yet there is theoretical evidence supporting these observations since BLPP is in fact
NP-hard even in its linear form 37. Moreover, this problem is nonlinear and nondifferen-
tiable, and the MCNFP in a large-scale construction project has various nodes and links.
On the other hand, the nondifferentiable piecewise objective functions and constraints
18 Mathematical Problems in Engineering
presented in the MCNFP bring computational difficulties. Thus, the possibility of finding
a solution to the complexity is increased, and it is difficult to solve with any exact algorithm.
Therefore, here, an MOBLPSO is proposed to solve the MCNFP in a large-scale construction
project. Because particle swarm optimization PSO is computationally tractable compared
with other heuristic algorithms, it is easy to implement, and does not require any gradient
information for an objective function but the value.
PSO is a population-based self-adaptive search stochastic optimization technique pro-
posed by Kennedy and Eberhart 38, which was inspired by the social behaviour of animals
such as fish schooling and birds flocking. Similar to other population-based algorithms, such
as evolutionary algorithms, PSO can solve a variety of difficult optimization problems but
has shown a faster convergence rate than other evolutionary algorithms on some problems
39. PSO is influenced little by the continuity of the objective function; it just uses the
primary math operators and receives good results in static, noisy, and continuously changing
environments 40. Another advantage of PSO is that it has very few parameters to adjust,
which makes it particularly easy to implement. Since PSO can be implemented easily and
effectively, it has been applied in solving real-world optimization problems in recent years,
such as 41–45. In PSO, the following formulas 38 are applied to update the position and
velocity of each particle:
- . - .
vdl τ1 1 wτvdl τ1 cp r1 pdl,best τ − pdl τ cg r2 gdl,best τ − pdl τ ,
5.1
pdl τ1 1 pdl τ vdl τ 1,
where vdl is the velocity of lth particle at the dth dimension in the τth iteration, wτ is the
inertia weight, pdl τ is the position of lth particle at the dth dimension, r1 and r2 are random
numbers in the range 0, 1, cp and cg are personal and global best position acceleration
constants, respectively, while, pdl,best τ is personal best position of the lth particle at the dth
dimension, and gdl,best τ is the global best position at the dth dimension.
As mentioned before, it is very difficult to solve the bilevel model, especially when
the model is nonlinear. The contribution of this paper is that a universal effective algorithm
for solving the bilevel model is put forward, which is based on the hierarchical iteration. The
key idea of the algorithm is that the optimum of the bilevel model can be approached step by
step through repeatedly iterative calculations between the upper and lower models. The main
body of the proposed approach is a type of PSO—multiobjective PSO MOPSO with Pareto-
Archived Evolution Strategy PAES—which is designed only to cope with the upper level
programming problem based on the follower’s optimal response. Another type of improved
PSO—PSO with passive congregation PSOPC—is embedded to deal with the lower level
programming problem and obtain the optimal response of the follower for each given
decision variable of the upper level programming. The follower’s optimal reaction is then
returned to upper level programming problem as the implementation base for the MOPSO.
The algorithm is called the MOBLPSO, the notations of which for the MCNFP are listed as
follows:
τ1 , τ2 : iteration index of the upper and lower level, τ1 1, 2, . . . , T1 and τ2 1, 2,
. . . , T2 ;
l1 , l2 : particle index of the upper and lower level, l1 1, 2, . . . , L1 and l2 1, 2, . . . , L2 ;
j: index of transportation path, j 1, 2, . . . , J;
Mathematical Problems in Engineering 19
k: index of carrier, k 1, 2, . . . , K;
r1 , r2 , r3 , r4 , r5 : uniform distributed random number within 0,1;
w1 τ1 : inertia weight in the τ1 th iteration of the upper level;
w2 τ2 : inertia weight in the τ2 th iteration of the lower level;
Rlkj2 τ2 : particle selected randomly from the swarm at the kjth dimension in the
τ2 th iteration;
c1p , c1g : personal and global best position acceleration constant of the upper level;
c2p , c2g : personal and global best position acceleration constant of the lower level;
cpc : passive congregation coefficient of the lower level;
p1 max , p1 min : maximum and minimum position value of the upper level;
P l1 τ1 : vector position of the l1 th particle in the τ1 th iteration, P l1 τ1 p1l1 τ1 ,
p2l1 τ1 , . . . , pJl1 τ1 ;
P l2 τ2 : vector position of the l2 th particle in the τ2 th iteration, P l2 τ2 p1l2 τ2 ,
p2l2 τ2 , . . . , pKJ
l1
τ2 ;
V l1 τ1 : vector velocity of the l1 th particle in the τ1 th iteration, V l1 τ1 v1l1 τ1 ,
v2l1 τ1 , . . . , vJl1 τ1 ;
V l2 τ2 : vector velocity of the l2 th particle in the τ2 th iteration, V l2 τ2 v1l2 τ2 ,
v2l2 τ2 , . . . , vKJ
l2
τ2 ;
P l1 ,best τ1 : vector personal best position of the l1 th particle in the τ1 th iteration,
P l1 ,best τ1 p1l1 ,best τ1 , p2l1 ,best τ1 , . . . , pJl1 ,best τ1 ;
P l2 ,best τ1 : vector personal best position of the l2 th particle in the τ2 th iteration,
P l2 ,best τ2 p1l2 ,best τ2 , p2l2 ,best τ2 , . . . , pJl2 ,best τ2 ;
Gbest τ2 : vector global best position in the τ2 th iteration, Gbest τ2 g1best τ2 ,
g2best τ2 , . . . , gKJ
best
τ2 ;
ARC: the positions of the particles that represent nondominated vectors in the
repository;
FP l1 τ1 : fitness value of P l1 τ1 ; and
FP l2 τ2 : fitness value of P l2 τ2 .
Algorithm 1
Procedure Test
if the archive is not full;
add pjl1 τ1 1 to the archive;
if pjl1 τ1 1 is in a less crowded region of the archive than pjl1 ,best τ1 ;
accept pjl1 ,best τ1 1 pjl1 τ1 1;
else maintain pjl1 ,best τ1 1 pjl1 ,best τ1 ;
else
If pjl1 τ1 1 is in a less crowded region of the archive than pjl1 ,best τ1 for some
member pjl1 ,best τ1 on the archive;
add pjl1 τ1 1 to the archive, and remove a member of the archive from the most
crowded region;
if pjl1 τ1 1 is in a less crowded region of the archive than pjl1 ,best τ1 ;
accept pjl1 ,best τ1 1 pjl1 τ1 1;
else maintain pjl1 ,best τ1 1 pjl1 ,best τ1 ;
else
if pjl1 τ1 1 is in a less crowded region of the archive than pjl1 ,best τ1 ;
accept pjl1 ,best τ1 1 pjl1 τ1 1;
else maintain pjl1 ,best τ1 1 pjl1 ,best τ1 .
Algorithm 2
The details for the PAES procedure, test procedure, and selection procedure are stated
hereinafter Algorithms 1 and 2, in which P l1 ,best τ1 is initially set equal to the initial position
of particle l1 .
The repository that stores the positions of the particles that represent nondominated
vectors is denoted by ARC; then the velocity of each l1 th particle of the upper level is updated
using the following equations:
- . - .
vjl1 τ1 1 w1 τ1 vjl1 τ1 c1p r1 pjl1 ,best τ1 − pjl1 τ1 c1g r2 ARCh τ1 − pjl1 τ1 , 5.2
where ARCh τ1 is a solution randomly selected from the repository in iteration τ1 , which
can improve significantly the ability of global convergence by avoiding being trapped in
Mathematical Problems in Engineering 21
a stagnant state in finite iterations 49, 50. The index h is selected in the following way:
the hypercubes containing more than one particle are assigned a fitness equal to the result
of dividing any number into the number of particles they contain. This aims to decrease the
fitness of those hypercubes that contain more particles which is seen as a form of fitness
sharing 51. Then, a roulette-wheel selection is applied using these fitness values to select
the hypercube from which the corresponding particle is taken. Once the hypercube has been
selected, a particle is selected randomly within the hypercube.
- .
l2 l2 l2 ,best l2
vkj τ2 1 w2 τ2 vkj τ2 c2p r3 pkj τ2 − pkj τ2
- . ( ) 5.3
l2
best
c2g r4 gkj τ2 − pkj τ2 cpc r5 Rlkj2 τ2 − pkj
l2
τ2 .
Parameter Selection
In order to guarantee the convergence of MOBLPSO, the parameters are selected on the basis
of empirical results that are carried out to observe the behaviour of the algorithm in different
parameter settings. By comparing several sets of parameters, including population size, iter-
ation number, acceleration coefficients, and inertia weight, the empirical results have shown
that the constant acceleration coefficients with c1p c1g 1.5 for the upper level, c2p 1,
c2g 2, and cpc 1.5 i.e., passive congregation coefficient 52 for the lower level, and
the adaptive inertia weights 53 provide good convergent behaviour in this study, which
is in accordance with the results provided by Eberhart and Shi 54. The adaptive inertia
22 Mathematical Problems in Engineering
weights for the upper level i.e., e 1 and lower level i.e., e 2 are set to be varying with
iteration as follows:
τe − Te
we τe we Te − we 1 − we Te , 5.4
1 − Te
where the iteration numbers are T1 500 i.e., for the upper level and T2 100 i.e., for the
lower level, and we 1 0.9 and we Te 0.1 for e 1, 2. Since the probability of becoming
trapped in the stagnant state can be reduced dramatically by using a large number of particles
49, the population sizes are set to be 300 for the upper level and 100 for the lower level.
Initialize
In the upper level, set iteration τ1 1. For l1 1, 2, . . . , L1 , generate the position of the particle
l1 with an integer random position note that every particle in the upper level consists of j
dimensions in this study. In the lower level, set iteration τ2 1. For l2 1, 2, . . . , L2 , generate
the position of the particle l2 with an integer random position note that every particle in the
lower level consists of k × j dimensions in this study.
Fitness Value
The fitness value used to evaluate the particle is the value of objective function in each level.
There are two objectives in the upper level; particle P l1 τ1 ’s fitness value FP l1 τ1 is a 1 × 2
matrix, namely,
⎡ ⎡ ⎛ ⎞ ⎤⎤
j β
( ) k∈Ψ γi ykj /vk
⎢ ⎢ ⎠ ⎥⎥
α⎝
j j∈Ω
F P l1 τ1 ⎣ cj xj , wk γi μtki0 ⎣1 ⎦⎦.
o,d∈E j∈Pod k∈Ψ j∈Ω i∈Φ
ri
5.5
( )
j ykj
F P l2 τ2 γi μeik . 5.6
k∈Ψ j∈Ω i∈Φ
vk
Mathematical Problems in Engineering 23
Particle p1l1 (τ1 ) p2l1 (τ1 ) ··· pjl1 (τ1 ) ··· pJl1 (τ1 )
xjl1 = pjl1
l2 l2 l2 l2
p11 (τ2 ) p21 (τ2 ) pk1 (τ2 ) pK1 (τ2 )
l2
p12 (τ2 ) l2
p22 (τ2 ) l2 l2
pK2 (τ2 )
pk2 (τ2 )
. . ··· . ··· .
Particle . . . .
. . . .
. . . .
. . . .
. . . .
l2 l2
l2
p1J (τ2 ) l2
p2J (τ2 ) pkJ (τ2 ) pKJ (τ2 )
l2 l2
ykj = pkj
Figure 4: Decoding method and mapping between PSO particles and solutions of two levels.
Figure 5 shows the schematic procedure for the MOBLPSO to generate solutions
for the proposed multiobjective bilevel model. Such a repeated interaction and cooperation
between two types of PSO reflects and simulates the decision process of multiobjective bilevel
programming and is able to solve multiobjective bilevel programming sequentially.
6. Practical Application
6.1. Project Description
In this section, an earth-rock work transportation project in a large-scale water conservancy
and hydropower construction project is taken as an example for our optimization method.
The Shuibuya Hydropower Project was conducted in Badong County, which is located in the
middle reaches of Qingjiang River in Sichuan province, China. The project is the first cascaded
project on the Qingjiang main stream and the third important project following Geheyan and
Gaobazhou in China. Once completed, it will provide a major power source to meet the peak
load demand in the Central China Power Grid. The installed capacity and annual output of
Shuibuya Power Plant are 1,600 MW and 3.92 GWh, respectively. The project has a powerful
regulating ability with a normal pool level of 400 m and reservoir capacity of 4.58 × 109 m3 .
The project consists of a concrete-faced rock fill dam CFRD, underground power house, a
chute spillway on the left bank, and the sluice tunnel on the right bank. The dam is 233 m
high and is the tallest of its kind in the world at present with a total volume of 15.64 × 106 m3 .
24 Mathematical Problems in Engineering
Start
Initialize the decision variable of upper level and the Set initial iteration
τ2 = 1
parameters of multiobjective aPSO
Yes
No In the feasible
Evaluate the initial particles to get the personal best solution?
position P l1 , best = P l1 (1)(l1 = 1, 2, ..., L1 )
Yes
x
Evaluate the initial particles to get the personal best
Solve the lower-level programming by PSOPC for position
each given initial result
(x, y)
Caculate each fitness value of the lower level particles
Caculate each fitness value by multiobejective method
Go to next iteration
Go to next iteration τ2 = τ2 + 1 Update the inertia weight ω2 (τ 2 )
τ1 = τ1 + 1 Update the inertia weight ω1 (τ1 )
No
Update the best position by using Pareto-Archived τ 2 = T2 ?
Evolution Strategy (PAES)
Yes
Yes
Stop
Output the optimum results
O
Borrow areas
1 ··· j ··· J
D O D
Dumpling sites Excavation projects Filling projects
1 ··· w ··· W 1 ··· i ··· I 1 ··· m ··· M
O+D
Stockpile areas
1 ··· n ··· N
The transportation network in the project includes an internal road network and an
external road network. In this case, only the internal road network is considered. The internal
road network is composed of 20 trunk roads located on the left and right banks, with 11 on the
left and 9 on the right, with a cross-river bridge connecting the left and right banks.
Therefore, 21 links are considered in this paper. In order to apply the proposed methods more
conveniently, adjacent roads of the same type have been combined and road shapes have been
ignored. An abstracted transportation network is illustrated in Figure 8. For each link in the
k
transportation network, there are a free flow birandom travel time t and birandom trans-
i0
k
portation cost ei . The corresponding data of which are stated in Table 3. In order to collect
the data of transportation time and costs, investigations and surveys were made to obtain
historical data from the financial department and experienced engineers of construction
team in Hubei Qingjiang Shuibuya Project Construction Company. Since the transportation
time and costs for each arc of the path change over time, the data are classified into
different parts based on different periods. Both transportation time and costs are assumed
to approximately follow normal distributions for each period, and the two parameters
expected value and variance for the normal distributions are estimated by using maximum
likelihood estimation, which justified by a chi-square goodness-of-fit test. By comparing
the normal distributions for the same transportation time and costs in different periods,
it can be found that the expected values of the aforementioned normal distributions also
approximately follow a similar random distribution pattern, which has also been justified
using a chi-square goodness-of-fit test. It should be noted that since the variance fluctuations
are quite insignificant, the median values of the variances in different periods are selected as
the variance for the previous normal distribution. The predicted confidence levels given by
the decision maker are, respectively, θ 0.9 and δ 0.85.
26 Mathematical Problems in Engineering
Orphan bank 1
(Liushutang orphan bank)
Orphan bank 2
(Heimagou orphan bank)
Borrow area 2 Stockpile area 3
(Qiaogou borrow area) (Xiaqiaogou stockpile area)
Figure 7: Layout of borrow areas, dumpling sites and stockpile areas of Shuibuya Hydropower Project.
#3 #9
#1 #5 #7 #15 #12
#2 #4 #10
Node
Arc
Excavation project 3,4 → filling project 3,4 2300 8 ⇔ #11 → #13 → #16 1.90 1.45
9 ⇔ #11 → #10 → #13 → #16 2.70 2.30
Mathematical Problems in Engineering
Excavation project 3,4 → stockpile area 3 1380 15 ⇔ #11 → #10 → #14 → #15 2.10 2.33
k k
Table 3: Birandom free flow travel time ti0 and birandom transportation cost ei .
Arc capacity ri Transportation cost eik CNY/unit Free transportation time tki0 h
Arc i Corresponding
nodes n/h Carrier 1 Carrier 2 Carrier 3 Carrier 1 Carrier 2 Carrier 3
Nμ, 0.64, with Nμ, 1.00, with Nμ, 0.01, with Nμ, 0.25, with Nμ, 0.81,with Nμ, 0.16, with
1 #1, #3 105
μ ∼ N 5.2, 0.10 μ ∼ N 6.0, 0.09 μ ∼ N 6.8, 1.00 μ ∼ N 0.2, 0.01 μ ∼ N 0.24, 0.09 μ ∼ N0.28, 0.04
Nμ, 1.00, with Nμ, 0.49, with Nμ, 0.81, with Nμ, 0.36, with Nμ, 0.25, with Nμ, 0.09, with
2 #1, #5 110
μ ∼ N3.2, 0.64 μ ∼ N 3.5, 0.16 μ ∼ N 3.8, 0.25 μ ∼ N 0.15, 0.04 μ ∼ N 0.16, 0.04 μ ∼ N 0.18, 0.01
Nμ, 4.00, with Nμ, 0.81, with Nμ, 0.64, with Nμ, 0.25, with Nμ, 0.36, with Nμ, 0.16, with
3 #1, #6 135
μ ∼ N 7.2, 1.00 μ ∼ N 7.5, 0.64 μ ∼ N7.8, 0.49 μ ∼ N 0.30, 0.16 μ ∼ N 0.35, 0.09 μ ∼ N 0.38, 0.01
Nμ, 1.00, with Nμ, 4.00, with Nμ, 0.01, with Nμ, 0.81, with Nμ, 0.36, with Nμ, 0.64, with
4 #2, #4 115
μ ∼ N 3.7, 0.49 μ ∼ N 4.0, 0.81 μ ∼ N 4.3, 0.01 μ ∼ N 0.14, 0.16 μ ∼ N 0.18, 0.09 μ ∼ N 0.19, 0.01
Nμ, 0.36, with Nμ, 0.81, with Nμ, 0.64, with Nμ, 0.01, with Nμ, 0.25, with Nμ, 0.09, with
5 #3, #5 108
μ ∼ N 2.2, 0.01 μ ∼ N 2.4, 0.09 μ ∼ N 2.6, 0.16 μ ∼ N 0.14, 0.01 μ ∼ N 0.18, 0.01 μ ∼ N 0.22, 0.01
Nμ, 0.81, with Nμ, 1.00, with Nμ, 0.49, with Nμ, 0.64, with Nμ, 0.09, with Nμ, 0.25, with
6 #4, #6 112
μ ∼ N 3.2, 0.16 μ ∼ N 4.0, 0.09 μ ∼ N 4.8, 0.09 μ ∼ N 0.12, 0.16 μ ∼ N 0.15, 0.01 μ ∼ N 0.18, 0.01
Nμ, 1.00, with Nμ, 0.64, with Nμ, 0.36, with Nμ, 0.01, with Nμ, 0.04, with Nμ, 0.01, with
7 #5, #7 180
μ ∼ N5.0, 0.01 μ ∼ N 5.2, 0.25 μ ∼ N 5.4, 0.16 μ ∼ N 0.25, 0.01 μ ∼ N 0.28, 0.01 μ ∼ N 0.32, 0.01
Nμ, 0.49, with Nμ, 0.36, with Nμ, 0.49, with Nμ, 0.09, with Nμ, 0.04, with Nμ, 0.09, with
8 #6, #7 145
μ ∼ N 4.5, 0.01 μ ∼ N 4.2, 0.16 μ ∼ N 3.4, 0.16 μ ∼ N 0.16, 0.01 μ ∼ N 0.19, 0.01 μ ∼ N 0.23, 0.01
Nμ, 0.16, with Nμ, 0.81, with Nμ, 1.00, with Nμ, 0.04, with Nμ, 0.01, with Nμ, 0.09, with
9 #6, #8 160
μ ∼ N 3.6, 0.25 μ ∼ N 4.0, 1.00 μ ∼ N 4.4, 0.09 μ ∼ N 0.13, 0.01 μ ∼ N 0.15, 0.04 μ ∼ N 0.17, 0.01
Nμ, 1.00, with Nμ, 1.00, with Nμ, 0.64, with Nμ, 0.01, with Nμ, 0.04, with Nμ, 0.09, with
10 #7, #8 165
μ ∼ N 4.0, 0.81 μ ∼ N 4.4, 0.49 μ ∼ N 4.8, 0.25 μ ∼ N 0.16, 0.01 μ ∼ N 0.19, 0.01 μ ∼ N 0.23, 0.04
Nμ, 4.00, with Nμ, 0.16, with Nμ, 0.25, with Nμ, 0.01, with Nμ, 0.04, with Nμ, 0.09, with
11 #7, #9 100
μ ∼ N 7.2, 0.09 μ ∼ N 7.6, 0.81 μ ∼ N 8.0, 0.16 μ ∼ N 0.36, 0.01 μ ∼ N 0.39, 0.01 μ ∼ N 0.42, 0.01
Nμ, 0.81, with Nμ, 0.64, with Nμ, 1.00, with Nμ, 0.01, with Nμ, 0.01, with Nμ, 0.01, with
12 #7, #15 130
μ ∼ N 8.0, 0.81 μ ∼ N 8.5, 0.50 μ ∼ N 8.7, 0.01 μ ∼ N 0.33, 0.01 μ ∼ N 0.36, 0.01 μ ∼ N 0.38, 0.01
Nμ, 4.00, with Nμ, 0.64, with Nμ, 0.25, with Nμ, 0.01, with Nμ, 0.01, with Nμ, 0.01, with
13 #10, #11 225
μ ∼ N 5.0, 0.81 μ ∼ N 6.0, 0.16 μ ∼ N 6.5, 0.09 μ ∼ N 0.30, 0.01 μ ∼ N 0.34, 0.01 μ ∼ N 0.36, 0.01
Nμ, 0.81, with Nμ, 1.00, with Nμ, 0.04, with Nμ, 0.01, with Nμ, 0.04, with Nμ, 0.04, with
14 #10, #13 155
μ ∼ N 5.5, 0.36 μ ∼ N 5.6, 0.49 μ ∼ N 5.8, 0.16 μ ∼ N 0.22, 0.09 μ ∼ N 0.26, 0.16 μ ∼ N 0.30, 0.09
Nμ, 0.64, with Nμ, 1.00, with Nμ, 0.09, with Nμ, 0.04, with Nμ, 0.01, with Nμ, 0.01, with
15 #10, #14 150
μ ∼ N 4.5, 0.49 μ ∼ N 4.6, 0.25 μ ∼ N 4.8, 0.09 μ ∼ N 0.22, 0.01 μ ∼ N 0.25, 0.04 μ ∼ N 0.27, 0.01
Nμ, 0.64, with Nμ, 0.09, with Nμ, 0.16, with Nμ, 0.04, with Nμ, 0.09, with Nμ, 0.01, with
16 #11, #13 170
Mathematical Problems in Engineering
μ ∼ N 3.6, 0.16 μ ∼ N 4.0, 0.01 μ ∼ N 4.4, 0.25 μ ∼ N 0.16, 0.01 μ ∼ N 0.18, 0.04 μ ∼ N 0.20, 0.01
Table 3: Continued.
Arc capacity ri Transportation cost eik CNY/unit Free transportation time tki0 h
Arc i Corresponding
Mathematical Problems in Engineering
Transportation path j
Material volume xj 950.000 543.763 1672.488 256.317 177.432 1700.000 766.646 1411.958 57.966 63.430 1279.675
Material volume y1j 426.432 309.723 576.206 71.020 93.874 505.161 371.885 449.943 36.224 46.715 520.477
Material volume y2j 413.345 107.514 587.051 92.521 44.360 620.743 267.585 521.378 42.172 48.536 364.225
Material volume y3j 369.912 159.467 509.230 52.741 127.025 574.096 285.792 440.637 34.154 48.542 394.974
Transportation path j 12 13 14 15 16 17 18 19 20 211 22
Material volume xj 370.325 161.357 1318.643 1380.000 1700.000 346.384 903.616 1150.000 1300.000 747.349 512.651
Material volume y1j 144.947 43.373 487.119 398.457 650.263 158.874 427.931 381.462 382.515 363.827 63.158
Material volume y2j 195.901 42.686 421.532 476.339 598.634 111.618 200.388 430.262 446.545 264.172 80.990
Material volume y3j 155.662 46.672 409.992 505.204 451.103 171.539 325.558 338.276 470.940 373.284 484.426
31
32
12.82 12.82
12.81 12.81
12.8 12.8
3.2 3.4 3.6 3.8 4 3.2 3.4 3.6 3.8 4
The total direct cost ×104 The total direct cost ×104
(Iteration 200) (Iteration 300)
12.83 12.83
12.82 12.82
12.81 12.81
12.8 12.8
3.2 3.4 3.6 3.8 4 3.2 3.4 3.6 3.8 4
The total direct cost ×104 The total direct cost ×104
(Iteration 400) (Iteration 500)
The transportation time
12.815 12.815
12.81 12.81
12.805 12.805
12.8 12.8
3.2 3.4 3.6 3.8 4 3.2 3.4 3.6 3.8 4
The total direct cost ×104 The total direct cost ×104
Figure 9: Pareto optimal solutions of upper level programming for Shuibuya Hydropower Project.
between the best and the worst and between the best and the average solutions for MCNFP-
rm are wider than the gaps of their counterparts for the birandom type. This shows that
randomness creates a much larger solution space when uncertainty is introduced. Fortu-
nately, the widened solution space with a further stochastic nature in the MCNFP-birm
provides better solutions and they were successfully located by MOBLPSO, evidenced by the
narrower gaps between the best and the worst and between the best and the average solutions
found for the MCNFP-birm. This suggests that MOBLPSO is an effective and relatively
efficient approach for solving the MCNFP under a birandom environment.
Since the PSOPC for the lower level is nested in the MOPSO for the upper level, and the
MOPSO is the main body of the proposed MOBLPSO, the evaluation of the MOPSO was
mainly paid attention to. In the MOPSO, a multiobjective method is introduced to derive
the Pareto optimal solution set for the upper level programming. This provides effective and
nondominated alternate schemes for the construction contractor. Compared to the weight-
sum method dealing with multiobjectives in 21, the solutions here are confirmed to be more
practical.
34
Iteration The average distance metric The distribution metric The extent metric
1 0.0781 0.1474 48.6227
100 0.0738 0.3460 58.1423
200 0.0878 0.4031 61.1722
300 0.0526 0.6967 69.2970
400 0.0554 0.8082 64.7773
500 0.0435 0.8938 68.8648
Table 8: Computation time and memory used by MOBLPSO, MOBLGA and MOBLSA.
Type Combination of weights Minimal weight sum value of the two objectives Average computation time min
ω1 ω2 MOBLPSO MOBLGA MOBLSA MOBLPSO MOBLGA MOBLSA
1 ω1 0.1 ω2 0.9 1152.027 1160.989 1152.531 20.856 25.899 34.458
2 ω1 0.2 ω2 0.8 1024.272 1032.419 1024.273 20.926 26.032 33.921
3 ω1 0.3 ω2 0.7 896.238 903.848 897.214 21.052 25.679 34.068
4 ω1 0.4 ω2 0.6 769.472 775.277 769.555 21.158 25.329 35.568
5 ω1 0.5 ω2 0.5 641.786 646.706 641.897 20.963 26.169 35.012
6 ω1 0.6 ω2 0.4 514.148 518.136 514.216 20.678 26.969 35.142
7 ω1 0.7 ω2 0.3 386.486 388.529 386.521 21.203 26.135 34.791
8 ω1 0. 8 ω2 0.2 258.792 260.836 258.792 20.734 25.475 35.291
9 ω1 0.9 ω2 0.1 131.081 132.143 131.081 21.134 26.189 34.815
Mathematical Problems in Engineering
Mathematical Problems in Engineering 37
for the different combination of weights i.e., ω1 and ω2 represent the weights of the two
objectives, resp.. It is demonstrated that the MOBLPSO for the MCNFP can perform opti-
mizing better than the MOBLGA, since MOBLGA may lead to a local search and need more
computation time. On the other hand, the MOBLSA could get similar results to those from
MOBLPSO, but computation was much slower than the MOBLPSO.
Acknowledgments
This paper was supported by the Key Program of NSFC Grant no. 70831005 and “985”
Program of Sichuan University “Innovative Research Base for Economic Development and
Management.” The authors would like to give their great appreciates to the editors and
38 Mathematical Problems in Engineering
anonymous referees for their helpful and constructive comments and suggestions, which
have helped to improve this paper.
References
1 R. K. Ahuja, T. L. Magnanti, J. B. Orlin, and M. R. Reddy, “Applications of network optimization,” in
Network Models, vol. 7, pp. 1–83, North-Holland, Amsterdam, The Netherlands, 1995.
2 F. Glover, D. Klingman, and N. Phillips, Network Models in Optimization and Their Applications in
Practice, Wiley, New York, NY, USA, 1st edition, 1992.
3 E. L. Hahne and R. G. Gallager, “Round robin scheduling for fair flow control in data communication
networks,” NASA STI/Recon Technical Report N, vol. 86, pp. 30–47, 1986.
4 M. Pióro and D. Medhi, Routing, Flow, and Capacity Design in Communication and Computer Networks,
Morgan Kaufmann, 2004.
5 A. S. Avestimehr, S. N. Diggavi, and D. N. C. Tse, “Wireless network information flow: a deterministic
approach,” IEEE Transactions on Information Theory, vol. 57, no. 4, pp. 1872–1905, 2011.
6 W. S. Lee, W. I. Lim, and P. H. Koo, “Transporter scheduling based on a network flow model under a
dynamic block transportation environment,” in Proceedings of the International Conference on Computers
& Industrial Engineering, pp. 311–316, 2009.
7 K. Paparrizos, N. Samaras, and A. Sifaleras, “An exterior simplex type algorithm for the minimum
cost network flow problem,” Computers & Operations Research, vol. 36, no. 4, pp. 1176–1190, 2009.
8 C. Dang, Y. Sun, Y. Wang, and Y. Yang, “A deterministic annealing algorithm for the minimum
concave cost network flow problem,” Neural Networks, vol. 24, no. 7, pp. 699–708, 2011.
9 D. B. M. M. Fontes, E. Hadjiconstantinou, and N. Christofides, “A dynamic programming approach
for solving single-source uncapacitated concave minimum cost network flow problems,” European
Journal of Operational Research, vol. 174, no. 2, pp. 1205–1219, 2006.
10 D. Goldfarb and Z. Jin, “A new scaling algorithm for the minimum cost network flow problem,”
Operations Research Letters, vol. 25, no. 5, pp. 205–211, 1999.
11 A. V. Goldberg, “An efficient implementation of a scaling minimum-cost flow algorithm,” Journal of
Algorithms, vol. 22, no. 1, pp. 1–29, 1997.
12 A. Sedeño-Noda and C. González-Martı́n, “An algorithm for the biobjective integer minimum cost
flow problem,” Computers & Operations Research, vol. 28, no. 2, pp. 139–156, 2001.
13 X. Zhu, Q. Yuan, A. Garcia-Diaz, and L. Dong, “Minimal-cost network flow problems with variable
lower bounds on arc flows,” Computers & Operations Research, vol. 38, no. 8, pp. 1210–1218, 2011.
14 D. Watling, “User equilibrium traffic network assignment with stochastic travel times and late arrival
penalty,” European Journal of Operational Research, vol. 175, no. 3, pp. 1539–1556, 2006.
15 A. Chen and Z. Zhou, “The α-reliable mean-excess traffic equilibrium model with stochastic travel
times,” Transportation Research B, vol. 44, no. 4, pp. 493–513, 2010.
16 Y. K. Lin, “Reliability evaluation of a revised stochastic flow network with uncertain minimum time,”
Physica A, vol. 389, no. 6, pp. 1253–1258, 2010.
17 A. Sumalee, D. P. Watling, and S. Nakayama, “Reliable network design problem: case with uncertain
demand and total travel time reliability,” Transportation Research Record, vol. 1964, no. 1, pp. 81–90,
2006.
18 A. M. Geoffrion and W. W. Hogan, “Coordination of two-level organizations with multiple
objectives,” in Techniques of Optimization, pp. 455–466, Academic Press, New York, NY, USA, 1972.
19 K. Tarvainen and Y. Y. Haimes, “Coordination of hierarchical multiobjective systems: theory and
methodology,” IEEE Transactions on Systems, Man, and Cybernetics, vol. 12, no. 6, pp. 751–764, 1982.
20 M. S. Osman, M. A. Abo-Sinna, A. H. Amer, and O. E. Emam, “A multi-level non-linear multi-
objective decision-making under fuzziness,” Applied Mathematics and Computation, vol. 153, no. 1, pp.
239–252, 2004.
21 G. Zhang, J. Lu, and T. Dillon, “Decentralized multi-objective bilevel decision making with fuzzy
demands,” Knowledge-Based Systems, vol. 20, no. 5, pp. 495–507, 2007.
22 H. I. Calvete and C. Galéb, “Linear bilevel programs with multiple objectives at the upper level,”
Journal of Computational and Applied Mathematics, vol. 234, no. 4, pp. 950–959, 2010.
23 B. Colson, P. Marcotte, and G. Savard, “Bilevel programming: a survey,” 4OR: A Quarterly Journal of
the Belgian, French and Italian Operations Research Societies, vol. 3, no. 2, pp. 87–107, 2005.
Mathematical Problems in Engineering 39
24 J. Xu and X. Zhou, “A class of multi-objective expected value decision-making model with birandom
coefficients and its application to flow shop scheduling problem,” Information Sciences, vol. 179, no.
17, pp. 2997–3017, 2009.
25 L. Yan, “One type of optimal portfolio selection in birandom environments,” Modern Applied Science,
vol. 3, no. 6, pp. 126–131, 2009.
26 J. Xu and C. Ding, “A class of chance constrained multiobjective linear programming with birandom
coefficients and its application to vendors selection,” International Journal of Production Economics, vol.
131, no. 2, pp. 709–720, 2011.
27 S. Gao and I. Chabini, “Optimal routing policy problems in stochastic time-dependent networks,”
Transportation Research B, vol. 40, no. 2, pp. 93–122, 2006.
28 N. E. El Faouzi and M. Maurin, “Reliability Metrics for path travel time under log-normal
distribution,” in Proceedings of the 3rd International Symposium on Transportation Network Reliability,
Delft, The Netherlands, July 2007.
29 H. Rakha, I. El-Shawarby, and M. Arafeh, “Trip travel-time reliability: issues and proposed solutions,”
Journal of Intelligent Transportation Systems, vol. 14, no. 4, pp. 232–250, 2010.
30 E. Mazloumi, G. Currie, and G. Rose, “Using GPS data to gain insight into public transport travel time
variability,” Journal of Transportation Engineering, vol. 136, no. 7, Article ID 006007QTE, pp. 623–631,
2010.
31 L. R. Ford Jr. and D. R. Fulkerson, Flows in Networks, Princeton University Press, Princeton, NJ, USA,
1962.
32 J. Peng and B. Liu, “Birandom variables and birandom programming,” Computers and Industrial
Engineering, vol. 53, no. 3, pp. 433–453, 2007.
33 Y. Sheffi, Urban Transportation Network, Prentice-Hall, Englewood Cliffs, NJ, USA, 1984.
34 A. Charnes and W. W. Cooper, “Deterministic equivalents for optimizing and satisficing under chance
constraints,” Operations Research, vol. 11, pp. 18–39, 1963.
35 K. Krickeberg, Probability Theory, Addison-Wesley, Reading, UK, 1965.
36 J. Xu and L. Yao, Random-Like Multiple Objective Decision Making, Sringer, 2009.
37 P. Hansen, B. Jaumard, and G. Savard, “New branch-and-bound rules for linear bilevel program-
ming,” Society for Industrial and Applied Mathematics, vol. 13, no. 5, pp. 1194–1217, 1992.
38 J. Kennedy and R. Eberhart, “Particle swarm optimization,” in Proceedings of the IEEE Conference on
Neural Networks, Piscataway, NJ, USA, 1995.
39 J. Kennedy and R. Eberhart, Swarm Intelligence, Morgan Kaufmann Publishers, 2001.
40 M. P. Song and G. C. Gu, “Research on particle swarm optimization: a review,” in Proceedings of
International Conference on Machine Learning and Cybernetics, pp. 2236–2241, chn, August 2004.
41 D. Y. Sha and C. Y. Hsu, “A hybrid particle swarm optimization for job shop scheduling problem,”
Computers & Industrial Engineering, vol. 51, no. 4, pp. 791–808, 2006.
42 S. H. Ling, H. H. C. Iu, K. Y. Chan, H. K. Lam, B. C. W. Yeung, and F. H. Leung, “Hybrid particle swarm
optimization with wavelet mutation and its industrial applications,” IEEE Transactions on Systems,
Man, and Cybernetics B, vol. 38, no. 3, pp. 743–763, 2008.
43 J. Xu, F. Yan, and S. Li, “Vehicle routing optimization with soft time windows in a fuzzy random
environment,” Transportation Research E, vol. 47, no. 6, pp. 1075–1091, 2011.
44 R.-M. Chen and C.-M. Wang, “Project scheduling heuristics-based standard PSO for task-resource
assignment in heterogeneous grid,” Abstract and Applied Analysis, vol. 2011, Article ID 589862, 20
pages, 2011.
45 J. Xu and Z. Zeng, “A discrete time optimal control model with uncertainty for dynamic machine
allocation problem and its application to manufacturing and construction industries,” Applied
Mathematical Modelling, vol. 36, no. 8, pp. 3513–3544, 2012.
46 C. C. Coello and M. S. Lechuga, “MOPSO: a proposal for multiple objective particle swarm
optimization,” in Proceedings of the Congress on Evolutionary Computation, 2002.
47 C. A. Coello Coello, G. T. Pulido, and M. S. Lechuga, “Handling multiple objectives with particle
swarm optimization,” IEEE Transactions on Evolutionary Computation, vol. 8, no. 3, pp. 256–279, 2004.
48 J. D. Knowles and D. W. Corne, “Approximating the nondominated front using the Pareto Archived
Evolution Strategy,” Evolutionary Computation, vol. 8, no. 2, pp. 149–172, 2000.
49 F. van den Bergh and A. P. Engelbrecht, “A convergence proof for the particle swarm optimiser,”
Fundamenta Informaticae, vol. 105, no. 4, pp. 341–374, 2010.
50 Z. Zhigang and G. Xinyi, “Particle swarm optimization based algorithm for bilevel programming
problems,” in Proceedings of the 6th ISDA International Conference on Intelligent Systems Design and
Applications, pp. 951–956, October 2006.
40 Mathematical Problems in Engineering
51 K. Deb and D. E. Goldberg, “An investigation of niche and species formation in genetic function
optimization,” in Proceedings of the 3rd International Conference of Genetic Algorithms, pp. 42–50, 1989.
52 S. He, Q. H. Wu, J. Y. Wen, J. R. Saunders, and R. C. Paton, “A particle swarm optimizer with passive
congregation,” BioSystems, vol. 78, no. 1–3, pp. 135–147, 2004.
53 Y. Shi and R. C. Eberhart, “Particle swarm optimization,” in Proceedings of the IEEE International
Conference on Neural Networks, 1998.
54 R. C. Eberhart and Y. Shi, “Comparing inertia weights and constriction factors in particle swarm
optimization,” in Proceedings of the IEEE Congress on Evolutionary Computation (CEC ’00), pp. 84–88,
San Diego, Calif, USA, July 2000.
55 E. Zitzler, K. Deb, and L. Thiele, “Comparison of multiobjective evolutionary algorithms: empirical
results,” Evolutionary Computation, vol. 8, no. 2, pp. 173–195, 2000.
56 Y. Yin, “Multiobjective bilevel optimization for transportation planning and management problems,”
Journal of Advanced Transportation, vol. 36, no. 1, pp. 93–105, 2002.
57 H. Liu, L. Wang, and H. Song, “Bi-level model of railway transportation existing network and
simulated annealing algorithm,” in Proceedings of the 2nd International Conference on Transportation
Engineering (ICTE ’09), vol. 4, pp. 3314–3319, July 2009.
Hindawi Publishing Corporation
Mathematical Problems in Engineering
Volume 2012, Article ID 293137, 14 pages
doi:10.1155/2012/293137
Research Article
A Selection Approach for Optimized
Problem-Solving Process by Grey Relational Utility
Model and Multicriteria Decision Analysis
Copyright q 2012 C.-K. Ke and M.-Y. Wu. This is an open access article distributed under the
Creative Commons Attribution License, which permits unrestricted use, distribution, and
reproduction in any medium, provided the original work is properly cited.
In business enterprises, especially the manufacturing industry, various problem situations may
occur during the production process. A situation denotes an evaluation point to determine the
status of a production process. A problem may occur if there is a discrepancy between the actual
situation and the desired one. Thus, a problem-solving process is often initiated to achieve the
desired situation. In the process, how to determine an action need to be taken to resolve the
situation becomes an important issue. Therefore, this work uses a selection approach for optimized
problem-solving process to assist workers in taking a reasonable action. A grey relational utility
model and a multicriteria decision analysis are used to determine the optimal selection order of
candidate actions. The selection order is presented to the worker as an adaptive recommended
solution. The worker chooses a reasonable problem-solving action based on the selection order.
This work uses a high-tech company’s knowledge base log as the analysis data. Experimental
results demonstrate that the proposed selection approach is effective.
1. Introduction
Problem solving is an important process that enables corporations to create competitive
advantages. In manufacturing industries, various problem situations may occur during the
production process 1, 2. A situation denotes an evaluation point to determine the status
of a production process. A problem may occur if there is a discrepancy between the actual
situation and the desired one 3. Thus, a problem-solving process is often initiated to achieve
the desired situation. In the process, workers determine what action needs to be taken to
resolve the situation. For a given problem, a situation may occur with various features
2 Mathematical Problems in Engineering
according to the context at that time. Due to the uncertain characteristics of situations, several
causes and possible actions may exist for a specific situation. Workers may observe a problem
situation, collect relevant information from the enterprise knowledge repository, explore
possible causes, and identify operational conditions in order to decide appropriate action
3, 4.
Quality of Service QoS is an important consideration in evaluating a problem-
solving solution. Worker feedback of an evaluating process can be represented as a utility
model reflecting the satisfaction a worker observes from taking an action. The worker
provides such a utility model 5 before committing to take an action. Grey relational
analysis 6, 7 can quantify all influences of various factors and their relation to consolidate
the utility model. Therefore, the worker’s grey relational utility model can be applied to
the monitoring information in order to evaluate the action’s QoS. The worker will get
the expected value of the issue of interest from taking an action. Based on various issues
of interest, how to select the reasonable action from a large number of candidate actions
requires a multicriteria decision analysis. A multicriteria decision analysis 8 is concerned
with structuring and solving decision and planning problems involving multiple criteria.
The purpose is to support decision makers facing such problems. Typically, there does not
exist a unique optimal action for such problems, and it is necessary to use decision maker’s
preferences to differentiate between actions. Therefore, a multicriteria decision analysis is
required to discover the selection order of the various actions for a specific situation. The
discovered selection order helps worker to solve the situation.
This work uses a selection approach to candidate actions to assist the worker in
acquiring a reasonable problem-solving action. Action formalization, a grey relational utility
model, and a multicriteria decision analysis are used to obtain an optimal selection order
for candidate actions. Then, the selected action for a specific situation is taken through a
problem-solving process. The result is considered a reasonable problem-solving solution for
the worker. This work explores a high-tech company’s knowledge base log as the analysis
data. The prototype system and use cases are proposed in the previous research 4. In this
work, we have an experiment to demonstrate that the proposed approach is effective. The
contribution of this research is in demonstrating a method which is easy to implement in a
problem solving knowledge recommendation system for selecting a reasonable solution.
The remainder of this paper is organized as follows. Section 2 reviews related works
on problem solving, grey relational analysis, multicriteria decision analysis, knowledge
management, and retrieval. Section 3 introduces a selection approach for optimized problem-
solving process by a grey relational utility model and a multicriteria decision analysis.
Section 4 describes an experimental data of a high-tech company’s knowledge base. An
experiment of the knowledge base log and discussions are showed in Section 5. Finally,
Section 6 presents our conclusions.
2. Related Works
The related literature covers problem solving, grey relational analysis and utility model,
multicriteria decision analysis, knowledge management, and retrieval.
system overload, and low machine utilization. A situation denotes an evaluation point to
determine the status i.e., desirable or undesirable of a production process. A problem may
occur if there is a discrepancy between the actual situation and the desired one 3. For
example, when the current production output is below the desired level, the production line
may have some problems. Thus, a problem solving process is often initiated to achieve the
desired situation. Problem solving is the thought process that resolves various difficulties
and obstacles spread in the gap between the current problem and its desired solution
4.
Various approaches have been proposed to support problem solving. Allen et al.
enforced a problem-solving process based on the knowledge gained from solving previous
similar problems 9. Chang et al. implemented a self-improvement helpdesk service system
1, and Park et al. developed a decision support system for problem solving in a complex
production process 2. More recently, Yang et al. proposed integrating the case-based
reasoning CBR approach with ART-Kohonen neural networks ART-KNNs to enhance
fault diagnosis in electric motors 10. Moreover, Guardati introduced RBCShell as a tool
for constructing knowledge-based systems, whereby previously solved problems are stored
in the case memory to support problem solving in new cases 11.
In a complex production process, problem solving is usually knowledge intensive.
Past experience or knowledge, routine problem-solving procedures, and previous decisions
can be used to enhance problem solving. The types of knowledge are investigated to use for
problem solving and suggest the circulation of knowledge to avoid knowledge inertia 12.
In the problem-solving process, workers take several problem-solving steps to determine
what action needs to be taken to resolve the situation 3. Such action involves both
human wisdom and enterprise knowledge. Workers may observe a problem situation, collect
relevant information from the enterprise knowledge repository, explore possible causes, and
identify operational conditions in order to decide appropriate action 4.
The grey relational analysis GRA is an important method in the grey system theory 7. The
GRA has been widely used in a number of areas, such as manufacturing 13, transportation
14, and the building trade 6. In the grey system theory, the GRA is essentially believed
to have captured the similarity measurements or relations in a system. Generally, the
procedure of grey relation analysis includes grey relation generation and grey relational
grade calculation steps. The grey relation generation step removes anomalies associated with
different measurement units and scales by the normalization of raw data. The grey relational
grade calculation step uses the grey relational coefficient to describe the trend relationship
between an objective series and a reference series at a given point in a system 15.
Quality of Service QoS is an important consideration in evaluating a problem-
solving action. Worker feedback of an evaluating process can be represented as a utility model
reflecting the satisfaction a worker observes from taking an action. The worker provides such
a utility model 5 before committing to take an action. GRA can quantify all influences of
various factors and their relation to consolidate the utility model. Therefore, the worker’s
grey relational utility model can be applied to the monitoring information in order to evaluate
the action’s QoS. The worker will get the expected value of the issue of interest from taking
an action.
4 Mathematical Problems in Engineering
Let A {a1 , a2 , . . . , an } denote the set of actions, and let a ∈ A. Let AP denote the set
of actions providers, let b ∈ AP , and let function S : AP → P A denote the actions provided
by an action provider, where P represents the power set operator. Let SW denote the set of
worker of the system, and let w ∈ SW. Each action has associated issues of interest, denoted
by set I, whose workers are interested in monitoring, and i ∈ I. Function IS represents the
set of issues of interest for an action: IS : A → P I. Function Ow : A × AP × I → R
denotes the expectation of the worker w for the actions he takes, where R denotes the real
w,b
numbers. Notation va,i represents the expectation of worker w on issue i of action a supplied
by provider b.
In a problem-solving environment, a potential issue of interest could be the QoS. Based
on the expectations, a worker can develop a utility model which reflects the satisfaction he
perceives from taking an action.
Larger-the-better is as follows:
w,b w,b
Ua,i v − mini Ua,i v
U∗ w,b
a,i v w,b w,b
. 3.1
maxi Ua,i v − mini Ua,i v
Smaller-the-better is as follows:
w,b w,b
maxi Ua,i v − Ua,i v
U∗ w,b
a,i v w,b w,b
. 3.2
maxi Ua,i v − mini Ua,i v
Then the grey relation equation 25, 32, 33 is used to calculate the grey relational grade
between reference vector and comparative vectors, partial equation as shown in 3.3. The
Uaw,b
0 ,i
v is a partial utility of reference vector, and Uaw,b
k ,i
v is a partial utility of comparative
vector. If the grey relational grade value Γ0k is closer to 1, it means that Uaw,b
0 ,i
v and Uaw,b
k ,i
v
6 Mathematical Problems in Engineering
have high correlation. If the grey relational grade value Γ0k is closer to 0, it means that Uaw,b
0 ,i
v
and Uaw,b
k ,i
v have low correlation:
1/ρ
Δmax − Δ0k n
Γ0k Γ Uaw,b,i v, Ua
w,b
,i v , where Δ0k Δ0k v ρ
, 3.3
0 k
Δmax − Δmin k1
where ρ {1, 2, . . . , m}. Δmax is the largest value of Δ0k and Δmin is the smallest value of
Δ0k . Based on the grey relational grade value Γ, a threshold value is used to filter out the
low correlation actions, and the remainders are considered as candidate actions for solving a
specific problem situation.
In Sections 3.1 and 3.2, the worker will get the expected value of the issue of interest
from taking an action. Based on various issues of interest, how to select the reasonable action
from a large number of candidate actions requires a multicriteria decision analysis.
⎡ c,b c,b
⎤
v1,1 · · · v1,n
⎢ ⎥
⎢ . . .. ⎥
Q Qij m×n ⎢
⎢ .. . . . ⎥ ⎥. 3.4
⎣ ⎦
c,b c,b
vm,1 · · · vm,n
Step 1. To calculate the weighted normalization decision matrix, a weight for each QoS item
must be set to form a weighted matrix W, as shown in:
⎡ ⎤
W1 · · · 0
⎢ ⎥
⎢ .. ⎥ .
W ⎢ ... . . . . ⎥ 3.5
⎣ ⎦
0 · · · Wn n×n
V vij m×n . 3.6
Step 2. Compare arbitrary different row i and row j in the weighted normalization decision
matrix V to make sure of the concordance and discordance set. If value v of row i is higher
Mathematical Problems in Engineering 7
than value v of row j, the component k can be classified as the concordance set Cij or the
discordance set Dij . The concordance set Cij or the discordance set Dij is shown as:
Cij k | vik ≥ vjk , Dij k | vik < vjk . 3.7
Step 3. The sum of each component’s weight forms a concordance matrix C, as shown in:
k∈Cij wk
C cij m×m
, cij n . 3.8
k∈1 wk
Step 4. We use a formula to get the discordance matrix. S is the set including all QoS items,
S {1, 2, . . . , n}, as shown in:
maxk∈Dij vik − vjk
dij . 3.9
maxk∈S vik − vjk
Step 5. The reverse complementary value is used to modify D to get the modified discordance
matrix D . The calculation of D is shown as:
D dij , dij 1 − dij . 3.10
m×m
Step 6. To show the large component value of the candidate solution, when the expected
value is larger, we combine each component cij of the concordance set with the discordance
matrix to calculate the production and get the modified total matrix A (Hadamard product of
cij and dij ), as shown in:
A aij m×m , aij cij ◦ dij . 3.11
Step 7. Get the maximum value aj of each column from modified total matrix. The purpose
is to determine the modified superiority matrix, as shown in:
aj max aij | i 1, 2, . . . , m , j 1, 2, . . . , m. 3.12
To make sure to get a reasonable solution, we have to rank aj from small to large:
a1 , a2 , . . . , am . The threshold a is set behind the smallest value a1 and the next smallest value
a2 . If the value aij is smaller than threshold a, it is replaced as 0 or 1. Then we get the modified
total superiority matrix, as shown in:
⎧
⎨1, aij ≥ a,
E eij , eij 3.13
⎩0, aij < a.
8 Mathematical Problems in Engineering
Problem-solving process
Enterprise Data processing for
identification
knowledge knowledge discovery
(case-based reasoning)
bases
Step 8. Finally, we get eij 1 from the matrix E . It indicates that solution i is better than
solution j. We can eliminate solution j and show it as Ai → Aj .
From Steps 1 to 8, we get the relationship among the QoS items of the candidate actions
and get the optimal selection order for all candidate actions. The candidate action is the action
provided by an action provider. The worker can follow the selection order to take a reasonable
action.
5.1. Experiments
This paper uses the wafer manufacturing problem 4 as a useful example to illustrate the
experiment. A wafer manufacturing process in a semiconductor foundry is used to illustrate
the proposed approach. The process comprises the following steps: crystal growing, wafer
cutting, edge rounding, lapping, etching, polishing, cleaning, final inspection, packaging, and
shipping. The wafer cleaning step mainly uses DI deionized, ultrapure water to remove
debris left over from the mounting wax and/or polishing agent. A stable water supply
system to deliver ultrapure water for wafer cleaning is therefore vital in semiconductor
manufacturing. The knowledge retrieval technique is used to explore the knowledge base log
which includes 1,077 relevant data records of wafer cleaning step in a wafer manufacturing
process. The discovered data records involve with 72 situations from 7 interdatabases, 23
workers, and 238 actions. The 5 domain experts assist this experiment to carry out and
evaluate.
Table 2: Transformed property of QoS item and action for abnormal situation.
The weighted matrix W for each QoS item is shown as follows:
⎡ ⎤
0.4 0 0
⎢ ⎥
W ⎢
⎣ 0 0.35 0 ⎦.
⎥ 5.2
0 0 0.25
The multiplication of a normalization matrix Q and a weighted matrix W which gets the
weighted normalization decision matrix V V QW is shown as follows:
⎡ ⎤
0.136 0.112 0.055
⎢ ⎥
V ⎢ ⎥
⎣0.124 0.1225 0.0625⎦. 5.3
0.128 0.098 0.075
The concordance set Cij or the discordance set Dij is shown as follows:
C12 {1}, D12 {2, 3}, C13 {1, 2}, D13 {3},
C21 {2, 3}, D21 {1}, C23 {2}, D23 {1, 3}, 5.4
C31 {3}, D31 {1, 2}, C32 {1, 3}, D32 {2}.
Mathematical Problems in Engineering 11
k∈C13 wk W1 W 2
C13 3 0.75,
k1 wk W1 W2 W3
⎡ ⎤ 5.5
− 0.4 0.75
⎢ ⎥
C⎢
⎣ 0.6 − 0.35⎦.
⎥
0.25 0.65 −
⎡ ⎤ 5.6
− 0.875 1
⎢ ⎥
D⎢
⎣1 − 0.51⎥
⎦.
0.7 1 −
⎡ ⎤
− 0.125 0
⎢ ⎥
D ⎢
⎣0 − 0.49⎥
⎦. 5.7
0.3 0 −
⎡ ⎤
− 0.05 0
⎢ ⎥
A⎢
⎣ 0 − 0.1715⎥
⎦. 5.8
0.075 0 −
⎡ ⎤
− 1 0
⎢ ⎥
E ⎢ ⎥
⎣ 0 − 1 ⎦. 5.9
1 0 −
Finally, the optimal selection order is determined for all candidate actions. The
experiment results show that action B is better than action D and action D is better than
action A. The worker can follow the selection order to get a reasonable action.
12 Mathematical Problems in Engineering
Item
Method
Candidate actions Precision Recall
The method in 4 32 62.8% 86/137 75.4% 86/114
This paper’s method 26 68.3% 86/126 83.5% 86/103
6. Conclusion
In business enterprises, especially the manufacturing industry, various problem situations
may occur during the production process. In a problem-solving process, how to determine
an action needed to be taken to resolve the situation becomes an important issue. This
work proposes a selection approach for optimized problem-solving process to assist workers
in taking a reasonable action. A grey relational utility model and a multicriteria decision
analysis are used to determine the optimal selection order of candidate actions. The selection
order is presented to the worker as an adaptive recommended solution. The worker fetches
a reasonable problem-solving action based on the selection order. The contribution of this
research is in demonstrating a method which is easy to implement in a problem-solving
knowledge recommendation system for selecting a reasonable solution.
A high-tech company’s knowledge base log is used for an experiment. In the
experiment process and result analysis, this research found that weight value in a
multicriteria decision analysis task and worker feedback influenced the experimental results.
Future work should pay more attention to designing a worker feedback mechanism for
Mathematical Problems in Engineering 13
QoS item identification. The worker feedback would help the proposed selection approach
by intelligent tuning and learning to improve the service quality incrementally. The
recommended technique is to consider combining with more intelligent methods to enhance
the effect.
References
1 K. H. Chang, P. Raman, W. H. Carlisle, and J. H. Cross, “A self-improving helpdesk service system
using case-based reasoning techniques,” Computers in Industry, vol. 30, no. 2, pp. 113–125, 1996.
2 M. K. Park, I. Lee, and K. M. Shon, “Using case based reasoning for problem solving in a complex
production process,” Expert Systems with Applications, vol. 15, no. 1, pp. 69–75, 1998.
3 J. S. Heh, “Evaluation model of problem solving,” Mathematical and Computer Modelling, vol. 30, no.
11-12, pp. 197–211, 1999.
4 D. R. Liu and C. K. Ke, “Knowledge support for problem-solving in a production process: a hybrid of
knowledge discovery and case-based reasoning,” Expert Systems with Applications, vol. 33, no. 1, pp.
147–161, 2007.
5 G. C. Silaghi, A. E. Arenas, and L. M. Silva, “A Utility-based reputation model for service-oriented
computing,” Towards Next Generation Grids, pp. 63–72, 2007.
6 W. C. Chang, K. L. Wen, H. S. Chen, and T. C. Chang, “Selection model of pavement material via Grey
relational grade,” in Proceeding of the IEEE International Conference on Systems, Man and Cybernetics, pp.
3388–3391, October 2000.
7 J. Deng, “Grey information space,” The Journal of Grey System, vol. 1, pp. 103–117, 1989.
8 A. R. Afshari, M. Mojahed, R. M. Yusuff, T. S. Hong, and M. Y. Ismail, “Personnel selection using
ELECTRE,” Journal of Applied Sciences, vol. 10, no. 23, pp. 3068–3075, 2010.
9 J. Allen, N. Blaylock, and G. Ferguson, “A problem solving model for collaborative agents,” in
Proceedings of the 1st International Joint Conference on: Autonomous Agents adn Multiagent Systems, pp.
774–781, Bologna, Italy, July 2002.
10 B. S. Yang, T. Han, and Y. S. Kim, “Integration of ART-Kohonen neural network and case-based
reasoning for intelligent fault diagnosis,” Expert Systems with Applications, vol. 26, no. 3, pp. 387–395,
2004.
11 S. Guardati, “RBCShell: a tool for the construction of systems with case-based reasoning,” Expert
Systems with Applications, vol. 14, no. 1-2, pp. 63–70, 1998.
12 S. H. Liao, “Problem solving and knowledge inertia,” Expert Systems with Applications, vol. 22, no. 1,
pp. 21–31, 2002.
13 B. C. Jiang, S. L. Tasi, and C. C. Wang, “Machine vision-based gray relational theory applied to IC
marking inspection,” IEEE Transactions on Semiconductor Manufacturing, vol. 15, no. 4, pp. 531–539,
2002.
14 Y. T. Hsu, C. B. Lin, and S. F. Su, “High noise vehicle plate recognition using grey system,” The Journal
of Grey System, vol. 10, pp. 193–208, 1998.
15 S. J. Huang, N. H. Chiu, and L. W. Chen, “Integration of the grey relational analysis with genetic
algorithm for software effort estimation,” European Journal of Operational Research, vol. 188, no. 3, pp.
898–909, 2008.
16 L. Antonieta and S. S. Vasco, “Multi criteria decision making models: an overview on electre
methods,” Universidade Portucalense, Centro de Investigação em Gestão e Economia CIGE,
Working paper, 2011.
17 Y. L. Chi, C. W. Lee, and C. Y. Chen, “A selection approach for optimized web services compositions,”
Electronic Commerce Studies, vol. 2, no. 3, pp. 297–314, 2004.
18 T. H. Davenport and L. Prusak, Working Knowledge: how Organizations Manage What They Know,
Harvard Business School Press, Boston, Mass, USA, 1998.
19 P. H. Gray, “The impact of knowledge repositories on power and control in the workplace,”
Information Technology & People, vol. 14, no. 4, pp. 368–384, 2001.
20 K.D. Fenstermacher, “Process-aware knowledge retrieval,” in Proceedings of the 35th Hawaii
International Conference on System Sciences, pp. 209–217, Big Island, Hawaii, USA, 2002.
21 M. M. Kwan and P. Balasubramanian, “KnowledgeScope: managing knowledge in context,” Decision
Support Systems, vol. 35, no. 4, pp. 467–486, 2003.
22 R. Baeza-Yates and B. Ribeiro-Neto, Modern Information Retrieval, The ACM Press, New York, NY,
USA, 1999.
14 Mathematical Problems in Engineering
23 G. Salton and C. Buckley, “Term-weighting approaches in automatic text retrieval,” Information
Processing and Management, vol. 24, no. 5, pp. 513–523, 1988.
24 J. L. Herlocker and J. A. Konstan, “Content-independent task-focused recommendation,” IEEE
Internet Computing, vol. 5, no. 6, pp. 40–47, 2001.
25 S. E. Middleton, N. R. Shadbolt, and D. C. De Roure, “Ontological user profiling in recommender
systems,” ACM Transactions on Information Systems, vol. 22, no. 1, pp. 54–88, 2004.
26 M. Pazzani and D. Billsus, “Learning and revising user profiles: the identification of interesting web
sites,” Machine Learning, vol. 27, no. 3, pp. 313–331, 1997.
27 A. Abecker, A. Bernardi, H. Maus, M. Sintek, and C. Wenzel, “Information supply for business
processes: coupling workflow with document analysis and information retrieval,” Knowledge-Based
Systems, vol. 13, no. 5, pp. 271–284, 2000.
28 A. Agostini, S. Albolino, R. Boselli, G. De Michelis, F. De Paoli, and R. Dondi, “Stimulating knowledge
discovery and sharing,” in Proceedings of the International ACM Conference on Supporting Group Work,
pp. 248–257, Sanibel Island, Fla, USA, November 2003.
29 N. J. Davies, A. Duke, and A. Stonkus, “OntoShare: evolving ontologies in a knowledge sharing
system,” in Towards the Semantic Web—Ontology-Based Knowledge Management, N. J. Davies, D. Fensel,
and F. van Harmelen, Eds., pp. 161–176, Wiley, London, UK, 2003.
30 D. R. Liu, I. C. Wu, and K. S. Yang, “Task-based script K sign-Support system: disseminating and
sharing task-relevant knowledge,” Expert Systems with Applications, vol. 29, no. 2, pp. 408–423, 2005.
31 M. Nagai, D. Yamaguchi, and G. D. Li, “Grey structural modeling,” Journal of Grey System, vol. 8, no.
2, pp. 119–130, 2005.
32 D. Yamaguchi, G. D. Li, and M. Nagai, “Verification of effectiveness for grey relational analysis
models,” Journal of Grey System, vol. 10, no. 3, pp. 169–182, 2007.
33 D. Yamaguchi, G. D. Li, K. M. T. Akabane, and M. Kitaoka, “A realization algorithm of grey structural
modeling,” Journal of Grey System, vol. 10, no. 1, pp. 33–40, 2007.
Hindawi Publishing Corporation
Mathematical Problems in Engineering
Volume 2012, Article ID 376546, 26 pages
doi:10.1155/2012/376546
Research Article
Predictor-Corrector Primal-Dual Interior Point
Method for Solving Economic Dispatch Problems:
A Postoptimization Analysis
Copyright q 2012 Antonio Roberto Balbo et al. This is an open access article distributed under
the Creative Commons Attribution License, which permits unrestricted use, distribution, and
reproduction in any medium, provided the original work is properly cited.
This paper proposes a predictor-corrector primal-dual interior point method which introduces line
search procedures IPLS in both the predictor and corrector steps. The Fibonacci search technique
is used in the predictor step, while an Armijo line search is used in the corrector step. The method
is developed for application to the economic dispatch ED problem studied in the field of power
systems analysis. The theory of the method is examined for quadratic programming problems and
involves the analysis of iterative schemes, computational implementation, and issues concerning
the adaptation of the proposed algorithm to solve ED problems. Numerical results are presented,
which demonstrate improvements and the efficiency of the IPLS method when compared to
several other methods described in the literature. Finally, postoptimization analyses are performed
for the solution of ED problems.
1. Introduction
Since its introduction in 1984, the projective transformation algorithm proposed by
Karmarkar in 1 has proved to be a notable interior point method for solving linear
programming problems LPPs. This pioneer study caused an upheaval in research activities
in this area. Among all the variations of Karmarkar’s original algorithm, the first to attract
the attention of researchers was the one that uses a simple affine transformation to replace
Karmarkar’s original and highly complex projective transformation, enabling work with
2 Mathematical Problems in Engineering
LPP in its standard form. The affine algorithm was first introduced in 2 by Dikin, a
Soviet mathematician. Later, in 1986, the work was independently rediscovered by Barnes
in 3 and by Vanderbei et al. in 4. They proposed using the primal-affine algorithm
to solve LPP in standard form and also presented proof of the algorithm’s convergence.
A similar algorithm, called dual-affine, was developed and implemented by Adler et al.
5 to solve the LPP in the form of inequality. Compared to the relatively cumbersome
projective transformation, the implementation of the primal-affine and dual-affine algorithms
was simpler because of its direct relationship with the LPP. These two algorithms produced
promising results when applied to large problems 6, although the theoretical proof of
polynomial time complexity was not obtained from the affine transformation. Megiddo
and Shub’s study in 7 indicated that the trajectory that leads to the optimal solution
provided by affine algorithms depends on the initial solution. A poor initial solution, which
is close to a viable domain vertex, could result in an investigation that covers all vertex
problems.
Nevertheless, the polynomial time complexity of primal-affine and dual-affine
algorithms can be re-established by incorporating a logarithmic barrier function into the
objective function of the original LPP. The purpose of this procedure is to solve the
problem pointed out by Megiddo, that is, to prevent an interior solution from becoming
“trapped” at the border of the problem possibly a vertex. This procedure also provides
proof of the complexity of the method. The idea of using the logarithmic barrier function
method for convex programming problems was developed by Fiacco and McCormick
in 8, 9, based on the method proposed by Frisch in 10. After the introduction of
Karmarkar’s algorithm in 1984, the logarithmic barrier function method was reconsidered
to solve linear programming problems. Gill et al. used this method in 11 to develop
a projected Newton barrier method and demonstrated its equivalence with Kamarkar’s
projective algorithm. The methods proposed, among others, by Ye in 12, Renegar in
13, Vaidya in 14, and Megiddo in 15, as well as those of a central trajectory—called
path-following methods, which were proposed by Gonzaga in 16, 17 and Monteiro and
Adler in 18, use the objective function augmented by the logarithmic barrier function.
A third variation, the so-called primal-dual interior point algorithm, was introduced by
Monteiro et al. in 19 and by Kojima et al. in 20. This algorithm explores a potential
primal-dual function, a variation of the logarithmic barrier function, called the potential
function. The polynomial time complexity theory was successfully demonstrated by Kojima
et al. in 20 and Monteiro et al. in 19 based on Megiddo’s work, which provided
a theoretical analysis for the logarithmic barrier method and proposed the primal-dual
approach.
The predictor-corrector procedure initially defined by Mehrotra and Sun in 21 and
implemented in by Lustig et al. in 22 explored variant directions of the primal-affine interior
point methods in the predictor step. In the corrector step, the point previously obtained
in the predictor step was “centralized” to exploit the potential function related to the
logarithmic barrier function. This procedure significantly improved the performance of the
primal-dual interior point methods.
This strategy was reviewed and modified by Wu et al. in 23 and successfully applied
to solve optimal power flow problems. Wu used the logarithmic barrier penalization, the
Newton method, and first-order approximations in the predictor step to determine search
directions and the approximate solution. Second-order approximations were considered in
the corrector step to refine solutions obtained in the predictor step.
Mathematical Problems in Engineering 3
The methods related to the primal-dual interior point methodology, especially those
proposed by Kojima et al. in 20, Monteiro and Adler in 18, and Monteiro et al. in 19,
which were broadly investigated by Fang and Puthenpura in 24, have been explored in this
past decade to solve linear, nonlinear, and integer mathematical programming problems in
several fields of research.
This paper proposes a primal-dual interior point method that uses the predictor-
corrector strategy described by Lustig et al. in 22 and Wu et al. in 23 and incorporates
line search procedures in both the predictor and corrector steps. Theoretical aspects as well
as iterative schemes and computational implementation are investigated. A Fibonacci line
search procedure is carried out in the predictor step and an Armijo line search is used in the
corrector step to calculate the step sizes while taking into account constraints in the variables
of the problem. The line search procedures adopted are aimed at improving the overall
convergence of the proposed method for solving quadratic programming problems. The
method is applied to solve the economic dispatch ED problem, a classical quadratic problem
studied in the field of electrical power systems. The results obtained with the proposed
method are compared with those obtained by several others described in the literature
for solving ED problems, such as the primal-dual method described in 6, evolutionary
algorithms found in 25–27, genetic and coevolutionary genetic algorithms described by
Samed in 27, the cultural algorithm described by Rodrigues in 26, and a hybrid atavistic
genetic algorithm given in 25. This comparative investigation demonstrates the efficiency
of the proposed IPLS method.
This paper is organized as follows: Section 2 presents the ED problem; Section 3
develops the theory of the proposed IPLS method and presents its algorithm; Section 4
describes a computational implementation of the method, applying it to solve ED problems.
This section also includes a postoptimization analysis. Finally, conclusions are drawn in
Section 5.
n
Minimize CT Ci Pi ai Pi2 bi Pi ci
i1
n
2.1
subject to Pi D
i1
Pimin ≤ Pi ≤ Pimax ; i 1, . . . , n,
where CT : total cost function for generating units, n: number of generating units, Ci Pi : cost
of generation unit i without considering the valve-point effect, ai , bi , ci : coefficient of the cost
function for generating unit i, Pi : power output of generating unit i, D: total power demand,
Pimin : minimum power output limit for generating unit i, and Pimax : maximum power output
limit for generating unit i.
The objective function in 2.1 may also represent the so-called valve-point effects, as
in 28, which are associated with the opening of pressure valves at some specific operating
points. In such cases, Ci Pi is mathematically described as in
Ci Pi ai Pi2 bi Pi ci ei sin fi Pimin − Pi . 2.2
The cost function Ci Pi described in 2.2 is continuous but not differentiable.
Although it is more representative, function 2.2 makes ED a much more complex problem
to solve due to its characteristics of nondifferentiability.
The literature describes several different methodologies to solve ED problems. In those
studies, the methodology associated with evolutionary algorithms stands out, especially
when issues related to nondifferentiability such as those described in 2.2 are involved.
Evolutionary algorithms have been used to solve ED problems because they are able to find
optimal solutions even when the objective function and/or constraints are not continuous
or non-differentiable. Numerical problems related to evolutionary algorithms involve the
inability to verify the optimality conditions associated with the solutions obtained and also
the computational effort necessary to obtain the solutions, especially for large systems.
Optimal solutions to ED problems have also been investigated through traditional
nonlinear programming methods, including interior point methods 6, 23, 30. This paper
proposes a predictor-corrector primal-dual interior point method for solving ED problems,
which incorporates line search procedures in both the predictor and corrector steps. The
results presented here demonstrate that the proposed method improves the solution of ED
problems. The following section examines the application of the primal-dual interior point
method to solve quadratic programming problems.
Mathematical Problems in Engineering 5
1 T
Minimize x Qx cT x
2
subject to Ax b 3.1
l ≤ x ≤ u,
1 T
Minimize x Qx cT x
2
subject to Ax b
3.2
x z u; z≥0
x − r l; r ≥ 0.
1 T n n
Minimize Fμ x, r, z x Qx cT x − μ ln rj − μ ln zj
2 j1 j1
subject to Ax b 3.3
xzu
x − r l.
1
Lμ x, r, z, w, s, y xT Qx cT x wT b − Ax sT l − x r yT x z − u
2
n
n 3.4
−μ ln sj − μ ln yj ,
j1 j1
6 Mathematical Problems in Engineering
where the dual variables associated with the three equality constraints in 3.3 are,
respectively, w ∈ m , y ∈ n , s ∈ n . The optimal Karush-Kuhn-Tucker KKT conditions
for problem 3.1 are depicted in
∂Lμ
0 ⇐⇒ − Qx − c AT w s − y 0,
∂x
∂Lμ
0 ⇐⇒ Ax b,
∂w
∂Lμ
0 ⇐⇒ x z u,
∂y
3.5
∂Lμ
0 ⇐⇒ x − r l,
∂s
∂Lμ
0 ⇐⇒ ZY e − μe 0,
∂z
∂Lμ
0 ⇐⇒ RSe − μe 0,
∂r
where R, Z, S and Y are diagonal matrices whose diagonal elements are ri , zi , si , and yi ; i
1, . . . , n, respectively; e 1, . . . , 1T ; μ is a dual metric or adjustment parameter for the curve
defined by the central trajectory path-following parameter. The set Ω0 given in 3.6 is
defined to simplify to notation. This set describes interior points for problem 3.2 and its
corresponding dual problem:
Ω0 x, w, z, r, y, s ⊥ − Qx AT w s − y c,
3.6
Ax b, x z u, x − r l, z, r, y, s > 0 .
Equations 3.4, 3.5, and 3.6 are considered in the following sections to perform the
analysis of important issues concerning the proposed IPLS method, such as search directions,
step sizes, stopping criteria, and updating of the barrier parameter.
⎛ ⎞ ⎛ ⎞
xk1 xk dxk
⎜ k1 ⎟ ⎜ k ⎟
⎜ w ⎟ ⎜ w dw k ⎟
⎜ ⎟ ⎜ ⎟
⎜ k1 ⎟ ⎜ k ⎟
⎜z ⎟ ⎜ z dzk ⎟
⎜ ⎟ ⎜ ⎟
hk1 ⎜ ⎟⎜ ⎟. 3.7
⎜ r k1 ⎟ ⎜ r k dk ⎟
⎜ ⎟ ⎜ r ⎟
⎜ ⎟ ⎜ ⎟
⎜ yk1 ⎟ ⎜ yk dk ⎟
⎝ ⎠ ⎝ y ⎠
sk1 sk dsk
J hk dk −F hk , 3.8
⎛ ⎞⎛ ⎞ ⎛ ⎞
A 0 0 0 0 0 dxk tk
⎜ ⎟⎜ k ⎟ ⎜ k ⎟
⎜−Q AT 0 0 −I I ⎟ ⎜ ⎟ ⎜ ⎟
⎜ ⎟ ⎜ dw ⎟ ⎜ g ⎟
⎜ ⎟⎜ k ⎟ ⎜ k ⎟
⎜ I 0 I 0 0 0⎟ ⎜ ⎟ ⎜ ⎟
⎜ ⎟ ⎜ dz ⎟ ⎜ f ⎟
⎜ ⎟⎜ k ⎟ ⎜ k ⎟, 3.9
⎜ I 0 0 −I 0 0 ⎟ ⎜ ⎟ ⎜ ⎟
⎜ ⎟ ⎜ dr ⎟ ⎜ o ⎟
⎜ ⎟⎜ ⎟ ⎜ ⎟
⎜ 0 0 Y 0 Z 0⎟ ⎜ k ⎟ ⎜ k⎟
⎝ ⎠ ⎝ dy ⎠ ⎝ q ⎠
0 0 0 S 0 R dsk vk
where the residuals of 3.9 for the predictor step are expressed as in
tk b − Axk ; g k Qxk c − AT wk − sk yk ; f k u − xk − zk ;
3.10
ok l − xk r k ; qk μk e − Zk Yk e; vk μk e − Rk Sk e; e 1, 1, . . . , 1T .
8 Mathematical Problems in Engineering
Adxk tk , 3.11
−Qdxk AT dw
k
dsk − dyk g k , 3.12
Isolating dyk and dsk in 3.15 and 3.16, respectively, leads to 3.19 and 3.20, as
follows:
dyk Zk−1 qk − Yk dzk , 3.19
dsk R−1
k v k
− S k d k
r . 3.20
Combining the results found in 3.17, 3.18, 3.19, and 3.20 with those given in
3.11–3.16, and considering 3.21, yields the directions 3.22 and 3.23:
−1
θ R−1
k S k Zk
−1
Y k Q , 3.21
−1
k
dw AθAT Aθ g k pk tk , 3.22
dxk θ AT dw
k
− g k − pk , 3.23
where
pk R−1
k Sk o −v
k k
Zk−1 qk −Yk f k . 3.24
Note that, after calculating 3.23, the remaining components of the direction vector
dzk , drk , dyk , and dsk in 3.17, 3.18, 3.19, and 3.20, respectively, are easily calculated. Since
matrix AθAT is symmetrical and positive definite in 3.22 considering that Q is symmetrical
k
and positive definite, dw can be determined by using the Cholesky decomposition.
Mathematical Problems in Engineering 9
J hk dk −F hk , 3.25
⎛ ⎞⎛ dk ⎞ ⎛ k ⎞
A 0 0 0 0 0 x t
⎜ ⎟⎜⎜
⎟ ⎜ ⎟
⎟
⎜−Q AT I⎟ dw ⎟ ⎜
k k⎟
⎜ 0 0 −I ⎟⎜⎜ ⎟ ⎜g ⎟
⎜ ⎟⎜ k ⎟ ⎜ ⎟
⎜ I
⎜ 0 I 0 0 0⎟⎟⎜ dz ⎟ ⎜ fk ⎟
⎜ ⎟⎜ ⎟⎜
⎜ ⎟,
⎟
3.26
⎜ I ⎜ ⎟
⎜ 0 0 −I 0 ⎟ k ⎜
0 ⎟⎜ dr ⎟ ⎜ o ⎟ k
⎟
⎜ ⎟⎜⎜ k⎟ ⎜
⎟ ⎟
⎜ 0 0 Y 0 Z 0⎟ ⎜ ⎟ ⎜ qk ⎟
⎝ ⎠ ⎝ dy ⎠ ⎝ ⎠
0 0 0 S 0 R dk v k
s
where
Dxk , Dzk , Dyk and Dsk are diagonal matrices whose diagonal components are dxk i , dzk i , dyk i
and dsk i , i 1, . . . , n, respectively.
The calculation of the residuals tk , g k , f k , ok , qk , vk , described in 3.10, and of
t , g , fk , ok , qk , v k , described in 3.27, basically distinguishes the predictor and corrector
k k
steps in the proposed method. It is important to note that the corrector step procedure uses
direction values dxk , dzk , dyk , and dsk , which have already been calculated in the predictor step,
to redefine residuals qk and v k in 3.27. Using 3.25 and 3.26 and following the same steps
10 Mathematical Problems in Engineering
taken to determine the directions of the predictor step, as seen in Section 3.2, the components
of the direction vector dk can be calculated using the following:
−1
dw
k
AθAT Aθ g k pk tk , 3.28
dxk θ AT dw
k
− g k − pk , 3.29
where
pk R−1 k − v k Zk−1 qk − Yk fk ,
k Sk o 3.34
wk1 wk αDk dw
k
, 3.38
The step size αPk for the primal variables is calculated through
αPk Min αP , αQ , αLS , 3.41
where, for 0 < α < 1, the step sizes αP , αQ , αLS are determined as follows.
i The step size for primal variables αP is obtained without violating the nonnegativity
requirements of the primal variables:
αzki αrik
αP Min − ,− such that dzki , drki <0 . 3.42
dzki drki
ii The step size αQ is the maximum step size possible without increasing the objective
value:
t
dxk Qxk c
αQ − t . 3.43
k k
dx Qdx
iii The step size αLS is determined as in 3.44 from the Fibonacci line search strategy,
which is briefly summarized as follows:
αLS Min Fμ xk1 αk Min Fμ xk αk dxk , 3.44
αk αk
For notation simplicity, in the summary of the Fibonacci search, the following
T
identities are defined: xk1 αLS xk1 αLS , r k1 αLS , zk1 αLS xk αLS dxk , r k
αLS drk , zk αLS dzk T , and dxk dxk , drk , dzk ; so that dxk is defined in 3.23, drk is defined in
3.18, and dzk is defined in 3.17. Only primal variables are considered in the Fibonacci
T
search algorithm used by the IPLS method. Starting at a point xk xk , r k , zk , the
algorithm searches a new point xk1 in direction dxk using the function Fμ defined in 3.3. The
Fibonacci method calculates αLS so that the minimization of Fμ is ensured. αLS is determined
considering the Fibonacci sequence. The initial value of αLS is set taking into account the
interval of uncertainty 0, 1.
The step size αDk for the dual variables 3.38–3.40 is calculated through 3.45 for
0 < α < 1:
αski αyik
αDk Min 1, − ,− such that dski ; dyki <0 . 3.45
dski dyki
12 Mathematical Problems in Engineering
where
βPk Min βP , βQ , βLS , 3.52
αzki αrik
βP Min 1, − ,− k k
such that dzi , dri < 0 . 3.53
dzk i
drk i
ii The step size βQ is the maximum step size possible without increasing the objective
value:
t
dxk Qxk c
βQ − t . 3.54
dxk Qdxk
iii The step size βLS in 3.55 is determined from the Armijo line search:
βLS βk , 3.55
so that
T
Fμ xk1 βk ≤ Fμ xk βk ∇Fμ xk dxk , 3.56
Mathematical Problems in Engineering 13
T
where xk1 βk xk1 βk , r k1 βk , zk1 βk xk βk dxk , r k βk drk , zk βk dzk ; dxk
T
dk , dk , dk ; so that dk , dk , dk are defined in 3.29–3.31, respectively. In the proposed IPLS
x r z x r z
method, the Armijo search is also performed in the direction defined by primal variables.
Starting at a point xk xk , r k , zk , the method searches for a new point xk1 , in direction dxk ,
so that the function Fμ defined in 3.3 decreases. This search calculates βLS , thereby ensuring
the reduction of Fμ . To prevent oscillations in the iterative process, the initial choice for βk
should not be too high, and to prevent the process from stopping prematurely, it should not
be too low. Therefore, this value is generally adjusted to β0 1. If 3.44 is not satisfied, βk is
updated using the following sequence:
βk
βk1 , k 0, 1, . . . , n; with δ > 1, usually set to δ 2. 3.57
δ
T T
∇Fμ xk ∇Fμ xk , ∇Fμ r k , ∇Fμ zk Qxk c, − μk R−1
k e, − μ k Zk
−1
e 3.58
αski αyik
βDk Min 1, − ,− k k
such that dsi ; dyi < 0 . 3.59
dsk dyk
i i
The general principle used here to calculate αPk and βPk is to choose a step size that
reduces the quadratic objective by a maximum amount without violating the nonnegativity
requirements of the primal variables.
The basic idea for defining the line searches in both the predictor and corrector steps is
to use a more accurate search for the predictor step which uses first order approximation to
calculate the residuals and a simpler search, albeit more robust, for the corrector step which
uses second-order approximation to calculate the residuals. Therefore, the Fibonacci search
is used in the predictor step, since it is more accurate, and provides the minimum value for
the objective function in the predefined direction; the Armijo search is used in the corrector
step because it is simpler and more robust.
reduce scaling problems, as described in 32. Typical stopping rules are shown in 3.60–
3.65, as follows:
i primal feasibility:
k
t
≤ ε1 predictor step , 3.60
b 1
k
t
≤ ε1 corrector step ; 3.61
b 1
k
u
≤ ε2 predictor step , 3.62
Qxk c 1
k
u
≤ ε2 corrector step ; 3.63
Qx c 1
k
k k
v ≤ ε3 , q ≤ ε3 predictor step 3.64
k k
v ≤ ε3 ,
q ≤ ε3 corrector step , 3.65
where ε1 , ε2 , and ε3 are sufficiently small positive numbers. Eventually, other criteria can be
adopted according to the specific characteristics of each problem, as can be seen in 24, 32.
k T k T
r s zk yk 3.66
μ1k ; μ2k ,
n n
where the parameter σ is used to accelerate the convergence of the iterative process. This
procedure for updating the barrier parameter proposed by Wright in 32 helps in the
theoretical convergence proof and also in the complexity analysis of primal-dual methods.
Mathematical Problems in Engineering 15
Step 4 computing step size—predictor. Calculate Fibonacci step sizes αPk and αDk using
3.41–3.44.
Step 5 moving to a new solution—predictor. Update xk1 ; wk1 ; sk1 ; yk1 ; zk1 ;r k1
obtained from the predictor step, according to 3.35–3.44.
Step 6 checking for optimality—predictor. If the criteria defined in 3.60, 3.62, and 3.64
are satisfied, stop. The solution is optimal. Otherwise, go on to the following step.
Step 7 intermediate calculations—corrector. Calculate: tk ; g k ; fk ; ok ; qk ; v k using 3.27, μk
using 3.67 and matrix θ using 3.21.
Step 9 computing step size—corrector. Calculate the Armijo step size βPk and βDk using
3.52–3.58.
Step 10 moving to a new solution—corrector. Update xk1 ; wk1 ; sk1 ; yk1 ; zk1 ; r k1 using
3.46–3.48.
Step 11 checking for optimality—corrector. If the criteria defined in 3.61, 3.63, and 3.65
are satisfied, stop. The solution is optimal. Otherwise, go on to the next step.
The predictor Steps 2 through Step 6 are held in odd iterations, while the corrector
Steps 7 through Step 11 are performed in even iterations. The next section describes numerical
simulations involving the application of the proposed method to ED problems.
6 generators, and Tables 8 and 9 depict the results for the system with 13 generators. These
tables compare the solutions obtained by the proposed IPLS method against those calculated
by the following methods: the predictor-corrector primal-dual PCPD method described in
6, the hybrid genetic algorithm HGA, the coevolutionary genetic algorithm COEGA, the
hybrid atavistic genetic algorithm HAGA and the cultural algorithm CA. The solutions
determined by the HGA and COEGA methods are available in 27, while the solutions
obtained with the CA method are described in 26, and those obtained with HAGA method
are given in 25, but only for the system comprising 13 generators.
The comparison between the IPLS method and the evolutionary approaches cited
above were introduced in this section because these are traditional methods used in power
system to solve ED problems, especially when a more general nondifferentiable objective
function as shown in 2.2 is used. However, it is important to highlight that these
evolutionary approaches are heuristic procedures, which provide only approximate solutions
to ED problems. When ED is formulated as a quadratic problem, it can be solved by means
of exact methods, such as the IPLS and PCPD method.
Therefore, to better evaluate the performance of proposed IPLS method, this method
has been compared to the PCPD method in Table 10. The results in Table 10 have the
main purpose to show the reduction in the computational effort when the IPLS method is
compared to the PCPD method. Both methods were implemented using Borland Pascal 7.0
programming language.
The parameters related to the system’s total demand are set at D 850 MW and the
active power losses are neglected. The values adopted for ε1 , ε2 , and ε3 are 10−8 .
Table 2 shows the active power output calculated by the methods. The results for the
HAGA algorithm are not presented in this case study. Table 3 compares optimal values for
the objective functions obtained by each method. From the results presented in this table, it
is clear that the dispatches calculated by all the types of genetic algorithms cannot reach the
global optimum dispatch attained by the interior point methods PCPD and IPLS. As Table 3
indicates, the cost calculated by the IPLS and PCPD methods is lower than that obtained by
Mathematical Problems in Engineering 17
Table 2: Comparison of the power generation outputs obtained for the system with 3 generators.
the HGA and COEGA methods. As already discussed, this is an expected result, since the
evolutionary approaches provide only approximate solution to the problem.
x0 20, 30, 75, 75, 145, 155; w0 0, 0, 0, 0, 0, 0; y0 1, 1, 1, 1, 1, 1. 4.2
The parameters related to the system’s total demand are set at D 500 MW and the
active power losses are neglected. The values adopted for ε1 , ε2 , and ε3 are 10−8 .
Table 5 shows the active power dispatch calculated by the methods. The solutions
for COEGA and HAGA algorithms were not presented by their authors. Table 6 compares
optimal values for the objective function obtained by each method. Again, as expected, the
costs calculated by the IPLS and PCPD methods are lower than those obtained by the HGA
method.
18 Mathematical Problems in Engineering
Table 5: Comparison of the power generation outputs obtained for the system with 6 generators.
Table 6: Values of the objective function for the system with 6 generators.
x0 660, 320, 330, 150, 140, 155, 165, 150, 140, 70, 80, 75, 85;
y0 30, 30, 30, 30, 30, 30, 30, 30, 30, 30, 30, 30, 30.
The parameters related to the system’s total demand are set at D 2520 MW and the
active power losses are neglected. The values adopted for ε1 , ε2 , and ε3 are 10−8 .
Mathematical Problems in Engineering 19
Table 8: Comparison of the power generation outputs obtained for the system with 13 generators.
Table 9: Values of the objective function for the system with 13 generators.
Table 8 lists the power generation outputs obtained by the methods. The values of
the dispatch calculated by the HAGA are not given by 25, who provided only the optimal
value for the objective function. Once more, as Table 9 indicates, the total cost calculated by
the PCPD and IPLS methods is lower than that calculated by all the others, although CA and
HAGA approaches get close to the optimal solution point.
Table 10: Number of iterations for the PCPD and IPLS methods.
Simulated system PCPD IPLS
3 generating units 26 12
6 generating units 22 15
13 generating units 23 17
both methods tested. Therefore, the efficiency of the methods is compared only in terms of
number of iterations, as described in Table 10.
The basic difference between the PCPD and IPLS algorithms is that line searches
are incorporated in the corrector and predictor steps of the IPLS method. Therefore, these
results demonstrate that the introduction of line searches in the IPLS method greatly reduces
the number of iterations required to solve ED problems. Since line searches represent only
a minor additional computational effort in the overall process of finding a solution, the
computational effort for solving ED problems is significantly improved by the proposed IPLS
method.
Minimize CT
n
subject to Pi D 4.4
i1
Pi ∈ S,
where S {Pi ∈ ⊥ Pi − Pimin ≥ 0 and Pimax − Pi ≥ 0, i 1 . . . , n}, and CT ni1 Ci Pi
ai Pi2 bi Pi ci .
Problem 4.4 is subsequently used for the postoptimization analysis of the solutions
obtained by the systems studied in the previous section.
n
n
n
L Pi , w, si , yi CT Pi w D − Pi si Pimin − Pi yi Pi − Pimax , 4.5
i1 i1 i1
Mathematical Problems in Engineering 21
n
D− Pi 0, 4.6
i1
Pi − Pimin ≥ 0; si Pi − Pimin 0; i 1, . . . , n, 4.7
Pimax − Pi ≥ 0; yi Pimax − Pi 0; i 1, . . . , n, 4.8
λi Pi − w − si yi 0; i 1, . . . , n, 4.9
where
∂Ci Pi
λi Pi 2ai Pi bi ; i 1, . . . , n. 4.10
∂Pi
In the literature on power systems, the Lagrange multiplier w is called energy price
i.e., the price of 1 MWh, while λi Pi is called incremental or marginal cost of generating
unit i. Using expression 4.9, the energy price can be calculated by
w λi Pi − si yi ; i 1, . . . , n. 4.11
w λ∗i Pi∗ ; i 1, . . . , n, 4.12
that is, the energy price is equal to the incremental marginal costs for all the generating
units. This situation corresponds to the rule commonly used by power system operators,
which states that all marginal costs are equal. It is important to emphasize that this rule is
valid only for this case and should, therefore, be used cautiously.
active in the optimal solution of 4.4. To analyse this case, let the set Ωmax be defined with
the indices of the generating units that have been optimally dispatched in their maximum
power output. In this situation, s∗i 0 for i 1, . . . , n, yj∗ > 0 for j ∈ Ωmax , and yj∗ 0 for
j∈/ Ωmax . Starting from 4.11, one reaches
w λ∗j Pj∗ yj∗ ; j ∈ Ωmax ,
4.13
w λ∗j Pj∗ ; j∈
/ Ωmax .
Based on 4.13, it is possible to show that the marginal costs for the group of
generating units that belong to Ωmax are lower than the marginal costs for the group that
does not belong to Ωmax , since yj∗ > 0; j ∈ Ωmax .
w λ∗j Pj∗ − s∗j ; j ∈ Ωmin ,
4.14
w λ∗j Pj∗ ; j∈
/ Ωmin .
Based on 4.14, it can be shown that the marginal costs for the group of generating
units that belong to Ωmin are higher than the marginal costs for the group that does not belong
to Ωmin , since s∗j > 0; j ∈ Ωmin .
4.7.4. Active Constraints for Both Minimum and Maximum Power Output Capabilities
In this case, some generating units are dispatched in their upper limits, while others are
dispatched in their lower limits, or at some other operational point between their upper and
lower limits. Obviously, the same unit cannot simultaneously assume lower and upper limits.
The comments made in the preceding sections concerning marginal costs for the generating
units that are dispatched in their lower or upper limits also apply to this case.
lower or upper limit, so there is no active constraint in the optimal solution. This situation
coincides with the one described in Section 4.7.1. The energy price, marginal costs, and
Lagrange multipliers are shown in Table 11 for this case. As described in Section 4.7.1, the
values obtained by the IPLS method for s∗i and yi∗ ; i 1, . . . , n are all zero. Also, the energy
price is equivalent to the marginal costs, which are all equal, as expected. These results are in
accordance with the theory described in Section 4.7.1.
capabilities. The marginal costs for all the remaining units 4, 5, 6, 7, 8, and 9 are equal to the
energy price. These results are in accordance with the theory described in Sections 4.7.2 and
4.7.3 and also with 4.13 and 4.14.
As expected, in all the cases analysed in Tables 11, 12, and 13, the value of the
generation cost w λ∗i − s∗i yi∗ was determined univocally, which proves the equal
incremental cost criterion 29.
The analysed results indicate that the dispatches Pi∗ , i 1, . . . , n calculated,
respectively, in Tables 2, 5, and 8 and the results for λ∗i , w, s∗i , and yi∗ determined, respectively,
in Tables 11, 12, and 13, satisfy the complementary slackness conditions set forth in 4.7 and
4.8.
5. Conclusions
This paper proposes a predictor-corrector primal-dual interior point method which
introduces line search procedures IPLS in both the predictor and corrector steps. The
Fibonacci search is used in the predictor step, while an Armijo search is used in the corrector
step. The method is developed for application to the economic dispatch ED problem, which
is an example of a quadratic programming problem studied in the field of power systems
analysis. ED problems have already been solved through primal-dual interior point methods,
although the strategy adopted here to solve the problem has not yet been tested.
The proposed algorithm is applied to solve ED problems in case studies involving
systems with 3, 6, and 13 generating units. The results provided by the proposed IPLS method
are compared to those provided by several other methods described in the literature, such as
the predictor-corrector primal-dual PCPD interior point method, hybrid genetic algorithm,
coevolutionary genetic algorithm, hybrid atavistic genetic algorithm, and cultural algorithm.
The dispatches calculated by all the types of genetic algorithms could not reach the global
optimal dispatch attained by the interior point methods PCPD and IPLS.
The computational effort of the interior point methods tested PCPD and IPLS was
evaluated and measured in terms of the number of iterations required to find the optimal
solution. The results indicate that the introduction of line searches in the IPLS method
Mathematical Problems in Engineering 25
considerably reduces the number of iterations required for solving ED problems. Line
searches pose represent only a minor additional computational effort in the overall process
of finding a solution; hence, the computational effort for solving ED problems is also greatly
improved by the proposed IPLS method.
A theory for performing a postoptimization analysis was also analysed. This theory
highlights the relation among energy prices, incremental generation costs, and other
Lagrange multipliers. This mathematical relation is used to validate optimal solutions for
ED problems calculated by the IPLS method.
Further research could involve the representation of the so-called valve point loading
in the objective function of ED problems. In that case, the nondifferentiability of the objective
function should be treated appropriately.
Acknowledgment
The authors are grateful for the financial support of FAPESP and CNPq Brazil.
References
1 N. Karmarkar, “A new polynomial-time algorithm for linear programming,” Combinatorica, vol. 4, no.
4, pp. 373–395, 1984.
2 I. I. Dikin, “Iterative solution of problems of linear and quadratic programming,” Doklady Akademii
Nauk SSSR, vol. 174, pp. 747–748, 1967 Russian.
3 E. R. Barnes, “A variation on Karmarkar’s algorithm for solving linear programming problems,”
Mathematical Programming, vol. 36, no. 2, pp. 174–182, 1986.
4 R. J. Vanderbei, M. S. Meketon, and B. A. Freedman, “A modification of Karmarkar’s linear
programming algorithm,” Algorithmica, vol. 1, no. 4, pp. 395–407, 1986.
5 I. Adler, M. G. C. Resende, G. Veiga, and N. Karmarkar, “An implementation of Karmarkar’s
algorithm for linear programming,” Mathematical Programming, vol. 44, no. 1–3, pp. 297–335, 1989.
6 A. R. Balbo, M. A. S. Souza, and E. C. Baptista, “ Métodos primal-dual de pontos interiores aplicados à
resolução de problemas de despacho econômico: sobre a influência da solução inicial,” in Proceedings
of the Simpósio Brasileiro de Pesquisa Operacional, João Pessoa—Pb. Anais do XL SBPO (XL SBPO ’08), pp.
2074–2085, ILTC, Rio de Janeiro, RJ, Brazil, 2008.
7 N. Megiddo and M. Shub, “Boundary behavior of interior point algorithms in linear programming,”
Mathematics of Operations Research, vol. 14, no. 1, pp. 97–146, 1989.
8 A. V. Fiacco and G. P. McCormick, Nonlinear Programming: Sequential Unconstrained Minimization
Techniques, John Wiley & Sons, New York, NY, USA, 1968.
9 A. V. Fiacco and G. P. McCormick, Nonlinear Programming, vol. 4 of Classics in Applied Mathematics,
SIAM, Philadelphia, Pa, USA, 2nd edition, 1990.
10 K. R. Frisch, The Logarithmic Potential Method of Convex Programming, University Institute of
Economics, Oslo, Norway, 1955.
11 P. E. Gill, W. Murray, M. A. Saunders, J. A. Tomlin, and M. H. Wright, “On projected Newton
barrier methods for linear programming and an equivalence to Karmarkar’s projective method,”
Mathematical Programming, vol. 36, no. 2, pp. 183–209, 1986.
12 Y. Ye, “An On3 L potential reduction algorithm for linear programming,” in Contemporary
Mathematics, vol. 114, pp. 91–107, American Mathematical Society, Providence, RI, USA, 1990.
13 J. Renegar, “A polynomial-time algorithm, based on Newton’s method, for linear programming,”
Mathematical Programming, vol. 40, no. 1, pp. 59–93, 1988.
14 P. M. Vaidya, “An algorithm for linear programming which requires Om nn2 m n1.5 nL
arithmetic operations,” Mathematical Programming, vol. 47, no. 2, pp. 175–201, 1990.
15 N. Megiddo, “On the complexity of linear programming,” in Advances in Economic Theory, T. Bewley,
Ed., pp. 225–268, Cambridge University Press, Cambridge, UK, 1989.
26 Mathematical Problems in Engineering
16 C. C. Gonzaga, “An algorithm for solving linear programming problems in On3 L operations,” in
Progress in Mathematical Programming: Interior-Point and Related Methods, N. Megiddo, Ed., pp. 1–28,
Springer, New York, NY, USA, 1989.
17 C. C. Gonzaga, “Polynomial affine algorithms for linear programming,” Mathematical Programming,
vol. 49, no. 1, pp. 7–21, 1990.
18 R. D. C. Monteiro and I. Adler, “Interior path following primal-dual algorithms. I. Linear
programming,” Mathematical Programming, vol. 44, no. 1, pp. 27–41, 1989.
19 R. D. C. Monteiro, I. Adler, and M. G. C. Resende, “A polynomial-time primal-dual affine scaling
algorithm for linear and convex quadratic programming and its power series extension,” Mathematics
of Operations Research, vol. 15, no. 2, pp. 191–214, 1990.
20 M. Kojima, S. Mizuno, and A. Yoshise, “A primal-dual interior point algorithm for linear
programming,” in Progress in mathematical programming: Interior-Point and Related Methods, N.
Megiddo, Ed., pp. 29–47, Springer, New York, NY, USA, 1989.
21 S. Mehrotra and J. Sun, “An algorithm for convex quadratic programming that requires On3.5 L
arithmetic operations,” Mathematics of Operations Research, vol. 15, no. 2, pp. 342–363, 1990.
22 I. J. Lustig, R. E. Marsten, and D. F. Shanno, “On implementing Mehrotra’s predictor-corrector
interior-point method for linear programming,” SIAM Journal on Optimization, vol. 2, no. 3, pp. 435–
449, 1992.
23 Y. C. Wu, A. S. Debs, and R. E. Marsten, “Direct nonlinear predictor-corrector primal-dual interior
point algorithm for optimal power flows,” IEEE Transactions on Power Systems, vol. 9, no. 2, pp. 876–
883, 1994.
24 S. C. Fang and S. Puthenpura, Linear Optimization and Extensions: Theory and Algorithms, Prentice-Hall,
Englewood Cliffs, NJ, USA, 1993.
25 J. O. Kim, D. J. Shin, J. N. Park, and C. Singh, “Atavistic genetic algorithm for economic dispatch with
valve point effect,” Electric Power Systems Research, vol. 62, no. 3, pp. 201–207, 2002.
26 N. M. Rodrigues, Um algoritmo cultural para problemas de despacho de energia elétrica, Dissertação de
Mestrado, Universidade Estadual de Maringá, 2007.
27 M. M. A. Samed, Um algoritmo genético Hibrido co-evolutivo para resolver problemas de despacho, Tese de
Doutorado, UEM, Departamento de Engenharia Quı́mica, 2004.
28 H. H. Happ, “Optimal power dispatch—a comprehensive survey,” IEEE Transactions on Power
Apparatus and Systems, vol. 96, no. 3, pp. 841–854, 1977.
29 M. J. C. Steinberg and T. H. Smith, Economic Loading of Power Plants and Electric Systems, MacGraw-
Hill, 1943.
30 A. R. L. Oliveira, S. Soares, and L. Nepomuceno, “Optimal active power dispatch combining network
flow and interior point approaches,” IEEE Transactions on Power Systems, vol. 18, no. 4, pp. 1235–1240,
2003.
31 M. S. Bazaraa, H. D. Sherali, and C. M. Shetty, Non Linear Programming-Theory and Algorithm, John
Wiley & Sons, New York, NY, USA, 2nd edition, 1993.
32 S. J. Wright, Primal-Dual Interior-Point Methods, SIAM, Philadelphia, Pa, USA, 1997.
Hindawi Publishing Corporation
Mathematical Problems in Engineering
Volume 2012, Article ID 953848, 11 pages
doi:10.1155/2012/953848
Research Article
Applying Hierarchical Bayesian Neural Network in
Failure Time Prediction
Copyright q 2012 L.-J. Kao and H.-F. Chen. This is an open access article distributed under
the Creative Commons Attribution License, which permits unrestricted use, distribution, and
reproduction in any medium, provided the original work is properly cited.
With the rapid technology development and improvement, the product failure time prediction
becomes an even harder task because only few failures in the product life tests are recorded.
The classical statistical model relies on the asymptotic theory and cannot guarantee that the
estimator has the finite sample property. To solve this problem, we apply the hierarchical Bayesian
neural network HBNN approach to predict the failure time and utilize the Gibbs sampler of
Markov chain Monte Carlo MCMC to estimate model parameters. In this proposed method, the
hierarchical structure is specified to study the heterogeneity among products. Engineers can use
the heterogeneity estimates to identify the causes of the quality differences and further enhance the
product quality. In order to demonstrate the effectiveness of the proposed hierarchical Bayesian
neural network model, the prediction performance of the proposed model is evaluated using
multiple performance measurement criteria. Sensitivity analysis of the proposed model is also
conducted using different number of hidden nodes and training sample sizes. The result shows
that HBNN can provide not only the predictive distribution but also the heterogeneous parameter
estimates for each path.
1. Introduction
In this high technology era, the society operations highly depend on various machinery
and equipments. Once the machinery or equipment is broken down, enormous trouble and
economics cost will be brought to the entire society. To enhance the product reliability,
the methodologies to assess product reliability have received much discussion in both
academics and industries. Among several mature techniques, degradation testing provides
an efficient way for reliability assessment when product quality is associated with a time-
varying degradation process. Typically, degradation measures can provide more reliability
information, particularly in modeling the failure-causing mechanism, than time-to-failure
data in few or no failure situation.
2 Mathematical Problems in Engineering
depends on parameters whose values must be determined with a training set of inputs and
outputs. Network architecture is the organization of nodes and the types of connections
permitted. The nodes are arranged in a series of layers with connections between nodes in
different layers, but not between nodes in the same layer.
Several researchers also integrate neural network algorithm with Bayesian theory,
which has been known as Bayesian neural network, in prediction. For examples, Neal 12
applied Hybrid Markov chain Monte Carlo MCMC numerical integration techniques for
the implementation of Bayesian procedures. Müller and Rios Insua 13 proposed a MCMC
scheme based on a static or dynamic number of hidden nodes. In their subsequent paper,
they have extended their research results by releasing the constraint of number of hidden
nodes 13. Also, Holmes and Mallick 14 used Bayesian neural network modeling in the
regression context.
In this paper, we conduct a hierarchical Bayesian neural network analysis with MCMC
estimation procedure in the failure time prediction problem. Here, hierarchy means that the
coefficients in our constructed HBNN model are specified by random effect distributions.
We attempt to use this hierarchical structure to determine if the heterogeneity exists among
paths. The advantage of proposed HBNN model cannot only provide a better failure time
prediction by incorporating the heterogeneity of components and autocorrelated structure
of error term but also provide a predictive distribution for the target value. Different from
previous research, the proposed HBNN model can successfully offer the full information of
parameter estimation and covariance structure. Engineers can use the heterogeneity estimates
to identify the causes of the quality differences and further enhance the product quality.
The data of the fatigue crack growth from Bogdanoff and Kozin 15 is used to
illustrate the proposed model. To demonstrate the effectiveness of the proposed model, the
prediction performance of the proposed model is evaluated using multiple performance
measurement criteria. Sensitivity evaluation of the proposed model is also conducted using
different number of hidden nodes and training sample sizes. The result shows that HBNN can
provide not only the predictive distribution but also the heterogeneous parameter estimates
for each path.
The rest of this paper is organized as follows: Section 2 introduces the proposed HBNN
model for failure time prediction. In Section 3, the fatigue crack growth data from Bogdanoff
and Kozin 15 is illustrated, and the model estimation procedure is provided. Failure time
prediction and sensitivity analysis are demonstrated in Section 4. Concluding remarks are
offered in Section 5.
M
ti,j βi,m0 βi,m Γi,m,j εi,j ,
m1 2.1
Γi,m,j ψ i,k0,m i,k1,m yi,j i,k2,m ln yi,j ,
where yi,j is the jth crack length of the ith path and ti,j is the observed cycle time of the ith
path, where i 1, 2, . . . , N and j 1, 2, . . . , ni . In addition, βi,m are the weights of the ith path
4 Mathematical Problems in Engineering
attached to the hidden nodes m m 1, 2, . . . , M, M is the number of hidden nodes, Γi,m,j
is the output of the mth hidden node when the jth crack length of the ith path is presented,
i,k1,m are the weights from the first input, yi,j , to the hidden node m, and ψ is the activation
function. Typically, the choice of M depends upon the problem under consideration. The
testing results of neural network with combinations of different numbers of hidden nodes
have been investigated. In the present case, we have set the number of hidden nodes M
equal to 3 because it gives the best predicting result.
According to the above equation, we know that there are totally N paths from a given
population, and ni observations are available for path i at fixed crack lengths yi,1 , yi,2 , . . . , yi,ni
i.e., The observations at length yi,1 , yi,2 , . . . , yi,ni are ti,1 , ti,2 , . . . , ti,ni , resp.. Herein, we assume
that the conditional distribution of ti,j given yi,j is normally distributed as fti,j | yi,j ∼
M
m1 βi,m Γi,m,j , σ . It means that each value of yi,j produces a random value of ti,j
2
Nβi,0
M
m1 βi,m Γi,m,j and a variance σ . Moreover,
2
from a normal distribution with a mean of βi,0
from literature 17, 18, we understand that degradation signals are usually autocorrelated
in nature. We also noticed that the values of the first-order autocorrelation r1 of the residuals
in Lu and Meeker 1 are not exactly equal to 2.0. Therefore, we suspected that the error
term might be characterized as a first-order autoregressive process. Based on this finding, we
proposed a new parametric crack growth model with autocorrelated errors as the following
equations:
M
ti,j βi,0 βi,m Γi,m,j εi,j , 2.2
m1
Γi,m,j ψ i,0,m i,1,m yi,j i,2,m ln yi,j , 2.3
where ρi is the autocorrelation coefficient and Zi,j is a normal distributed error with N0, σ 2
form. Note that the elements ti,1 , ti,2 , . . . , ti,ni in 2.2 are independent given βi,m , i,k,m , σ 2 , ρi
and yi,j , where k is the number of inputs. The function ψ· is referred to as an activation
function in a neural network. Typically, the activation function is nonlinear. Some of the
most common choices of ψ· are the logistic and the hyperbolic tangent functions. In order
to describe the heterogeneity varying from path to path, we characterized βi by a 4-variate
normal distribution with mean vector β and covariance matrix Vβ , βi | β, Vβ ∼ N4 β, Vβ , i
1, 2, . . . , N, and i,m is characterized by a 3-variate normal distribution with mean vector
m , and covariance matrix Vm for m 1, 2, and 3. Equations 2.2–2.4 specify a general
model for studying when observed cycle time sensitivity to crack length may increase. The
heterogeneity among paths is captured by parameters βi , i,k,m , and the specification of
covariates Vβ and Vm .
According to the above setting, the likelihood function for the data can be written as
N ni
l βi , i,m , σ 2 , ρi | ti,j ∝ ti,j | Yi,j , βi , i,m , σ 2 , ρi . 2.5
i1 j1
300000
250000
200000
Cycle time
150000
100000
50000
10 20 30 40 50
Crack length
Figure 1: Thirty paths of fatigue crack growth data from Bogdanoff and Kozin 15.
critical level of degradation i.e., Df 49 mm and assumed the experiment was terminated
at 40 mm. In other words, based on the measurements of degradation from 9 mm to 40 mm,
we would like to model the degradation process and use the proposed model to predict the
failure time ti,j Df to the assumed critical level for the degradation path i.e., crack length
49 mm. As mentioned, because the fatigue experiment was conducted on paths with fixed
crack length, we are interested in the predicted failure time for the path when a specific crack
length i.e., 49 mm is reached.
Estimated parameter
β m1 m2 m3 σ2 ρi
⎡ ⎤
−0.0013
−8.0899 −38.3625 −26.610
Posterior Mean ⎣ 0.2925 ⎦
0.3218 0.1952 −0.6606 −0.1878 0.001521 0.7806321
0.2749 0.8385 19.3739 9.8855
⎡ ⎤
0.0153
0.0075 0.0093 0.0019
Posterior STD ⎣ 0.0119 ⎦ 0.0003 0.0008 0.0070 0.000313 0.073125
0.0178 0.0012 0.0093 0.0084
0.0152
28
27
25
24
23
22
0 5 10 15 20 25 30
Path
Actual
Predicted
Metrics Calculation
30 2
RMSE RMSE i1 Ti − Pi /30
30
MAD MAD i1 |Ti − Pi |/30
MAPE MAPE 30i1 |Ti − Pi /Ti |/30
30 2
RMSPE RMSPE i1 Ti − Pi /Ti /30
∗
Note that T and P represent the actual and predicted failure time, respectively.
5. Conclusion
In this paper, we applied the HBNN approach to model the degradation process and to make
the failure time prediction. In the process of developing the HBNN model, the MCMC was
utilized to estimate the parameters. Since the prediction of failure time made by HBNN model
can sufficiently represent the actual data, the time-to-failure distribution can also be obtained
successfully. In order to demonstrate the effectiveness of the proposed hierarchical Bayesian
neural network model, the prediction performance of the proposed model is evaluated using
multiple performance measurement criteria. Sensitivity evaluation of the proposed model is
also conducted using different number of hidden nodes and training sample sizes. As the
results reveal, using HBNN can provide not only the predictive distribution but also accurate
parameter estimate. By specifying the random effects on the coefficients βi and i,m in the
HBNN model, the heterogeneity varying across individual products can be studied. Based
on these heterogeneities, the engineers will be able to conduct a further investigation in the
manufacturing process and then to find out the causes of differences.
For the future research, statistical inferences of failure time based on degradation
measurement, such as failure rate and tolerance limits, can be further evaluated given the
predicted failure time. In addition, for some highly reliable products, it is not easy to obtain
the failure data even under the elevated stresses. In such case, accelerated degradation testing
ADT can be an alternative that provides an efficient channel for failure time prediction.
The proposed HBNN approach can also be applied to depict the stress-related degradation
process by including those stress factors as covariates in the model.
Appendix
The Full Conditional Probability of Estimated Parameters
βi | rest ∝ ti | Yi , βi , i,m , σ 2 , ρi · βi | β, Vβ ,
N
β | rest ∝ βi | β, V β · β | μβ , Vβ ,
i1
10 Mathematical Problems in Engineering
N
Vβ | rest ∝ βi | β, Vβ · Vβ | f0 , G0 ,
i1
ni
i,m | rest ∝ ti | Γi , βi , i,m , σ 2 , ρi · i,m | m , Vm ,
j1
N
m | rest ∝ i,m | m , Vm · m | p, Vp ,
i1
N
Vm | rest ∝ i,m | m , Vm · Vm | f1 , G1 ,
i1
N
σ 2 | rest ∝ ti | Yi , βi , i,m , σ 2 , ρi · σ 2 | r0 , s0 ,
i1
ρi | rest ∝ ti | Yi , βi , i,m , σ 2 , ρi · ρi .
A.1
References
1 C. J. Lu and W. Q. Meeker, “Using degradation measures to estimate a time-to-failure distribution,”
Technometrics, vol. 35, no. 2, pp. 161–174, 1993.
2 C. J. Lu, W. Q. Meeker, and L. A. Escobar, “A comparison of degradation and failure-time analysis
methods for estimating a time-to-failure distribution,” Statistica Sinica, vol. 6, no. 3, pp. 531–546, 1996.
3 W. Q. Meeker and M. Hamada, “Statistical tools for the rapid development & evaluation of high-
reliability products,” IEEE Transactions on Reliability, vol. 44, no. 2, pp. 187–198, 1995.
4 K. A. Doksum and A. Hoyland, “Models for variable-stress accelerated life testing experiments based
on Wiener processes and the inverse Gaussian distribution,” Technometrics, vol. 34, no. 1, pp. 74–82,
1992.
5 J. M. Bernardo and A. F. M. Smith, Bayesian Theory, John Wiley & Sons, 1994.
6 A. Gelman, J. B. Carlin, H. S. Stern, and D. B. Rubin, Bayesian Data Analysis, Chapman & Hall, London,
UK, 1995.
7 C. P. Robert and G. Casella, Monte Carlo Statistical Methods, Springer, New York, NY, USA, 1999.
8 S. I. Liu, “Bayesian model determination for binary-time-series data with applications,” Computational
Statistics & Data Analysis, vol. 36, no. 4, pp. 461–473, 2001.
9 R. Zheng and B. R. Ellingwood, “Stochastic fatigue crack growth in steel structures subjected to
random loading,” Structural Safety, vol. 20, no. 4, pp. 305–324, 1998.
10 R. Zhang and S. Mahadevan, “Model uncertainty and Bayesian updating in reliability-based
inspection,” Structural Safety, vol. 22, no. 2, pp. 145–160, 2000.
11 M. Akama, “Bayesian analysis for the results of fatigue test using full-scale models to obtain the
accurate failure probabilities of the Shinkansen vehicle axle,” Reliability Engineering and System Safety,
vol. 75, no. 3, pp. 321–332, 2002.
12 R. M. Neal, Bayesian Learning for Neural Networks, Springer, New York, NY, USA, 1996.
13 P. Müller and D. Rios Insua, “Issues in bayesian analysis of neural network models,” Neural
Computation, vol. 10, pp. 571–592, 1998.
14 C. C. Holmes and B. K. Mallick, “Bayesian wavelet networks for nonparametric regression,” IEEE
Transactions on Neural Networks, vol. 11, no. 1, pp. 27–35, 2000.
15 J. L. Bogdanoff and F. Kozin, Probabilistic Models of Cumulative Damage, John Wiely, New York, NY,
UA, 1984.
Mathematical Problems in Engineering 11
16 E. P. Liski and T. Nummi, “Prediction in repeated-measures models with engineering applications,”
Technometrics, vol. 38, no. 1, pp. 25–36, 1996.
17 R. B. Chinnam, “On-line reliability estimation for individual components using statistical degradation
signal models,” Quality and Reliability Engineering International, vol. 18, no. 1, pp. 53–73, 2002.
18 R. B. Chinnam and P. Baruah, “A neuro-fuzzy approach for on-line reliability estimation and
condition based-maintenance using degradation signals,” International Journal of Materials and Product
Technology, vol. 20, no. 1–3, pp. 166–179, 2004.
19 J. G. Ibrahim, L. M. Ryan, and M.-H. Chen, “Using historical controls to adjust for covariates in trend
tests for binary data,” Journal of the American Statistical Association, vol. 93, no. 444, pp. 1282–1293,
1998.
20 B. P. Carlin and T. A. Louis, Bayes and Empirical Bayes Methods for Data Analysis, CRC Press, London,
UK, 2000.
21 S. P. Brooks, “Markov chain Monte Carlo method and its application,” Journal of the Royal Statistical
Society D, vol. 47, part 1, pp. 69–100, 1998.
22 A. E. Gelfand and A. F. Smith, “Sampling-based approaches to calculating marginal densities,” Journal
of the American Statistical Association, vol. 85, no. 410, pp. 398–409, 1990.
23 H. J. Kushner and P. Dupuis, Numerical Methods for Stochastic Control Problems in Continuous Time, vol.
24, Springer, New York, NY, USA, 2001.
Hindawi Publishing Corporation
Mathematical Problems in Engineering
Volume 2012, Article ID 827137, 16 pages
doi:10.1155/2012/827137
Research Article
A Fuzzy Dropper for Proportional Loss Rate
Differentiation under Wireless Network with a
Multi-State Channel
Yu-Chin Szu
Department of Information Management, St. John’s University, 499, Section 4, Tam King Road,
Tamsui District, New Taipei City 25135, Taiwan
Copyright q 2012 Yu-Chin Szu. This is an open access article distributed under the Creative
Commons Attribution License, which permits unrestricted use, distribution, and reproduction in
any medium, provided the original work is properly cited.
The proportional loss rate differentiation PLD model was proposed to provide controllable
and predictable loss rate for different classes of wired network connections. However, these
algorithms cannot be directly applied to wireless networks, because of the location-dependent
and time-varying wireless channel capacity. This paper proposes a novel packet dropper for
fuzzy controlling of the proportional loss rate differentiation in a wireless network with multistate
channel. The proposed dropper, fuzzy proportional loss rate dropper FPLR, prefers to drop the
small packets destined to a poor condition channel to improve the network performance. The loss
rate debts of the poor channel will be compensated later to keep PLD. From simulation results,
FPLR does achieve accurate loss rate proportion, lower queuing delay and loss rate, and higher
throughput, compared with other methods in the wireless environment.
1. Introduction
The Internet becomes an important infrastructure of global communication. However, many
multimedia applications require some guarantee of the quality of services QoS. Thus
the best-effort service is not suitable to such applications because it cannot promise any
guarantee about packet loss, delay, or jitter. The Internet Engineering Task Force IETF
first proposed the Integrated Service IntServ as a QoS architecture for IP networks 1.
Because IntServ suffers from the scalability problem, the Differentiated Service DiffServ
2 was then proposed. Diffserv has proceeded in two directions—absolute differentiated
service and relative differentiated service. The absolute differentiated service ensures that
an admitted user can enjoy certain and steady performance, while the relative differentiated
service ensures that users with a higher service class experience better performance than the
2 Mathematical Problems in Engineering
users with a lower one. The proportional differentiation model, a branch of relative service
differentiation, has received a lot of attention because it can perform the controllable and
predictable differentiation 3, 4. That is, the proportional differentiation model offers the
network manager a means of varying quality spacing between service classes according to
the given pricing or policy criteria and ensures that the differentiation between classes is
consistent in any measured timescale.
The proportional services can be differentiated according to different performance
metrics, such as throughput, delay, loss rate, or jitter. When adopting packet loss rate as the
performance metric, the proportional differentiation model is referred to as the proportional
loss rate differentiation PLD model. To provide PLD, some methods, such as Proportional
Loss Rate PLR 5, Average Drop Distance ADD 6, Debt Aware 7, Weighted Random
Early Detection WRED 8, Weight Simple Adaptive Proportion WSAP 9, Hop-count
Probabilistic Packet Dropper HPPD 10, and RED with maximum drop probability
adjustment, were proposed.
As wireless technologies rapidly advance and lightweight portable computing devices
become popular, wireless networks have become pervasive. Accordingly, the PLD model
is also urgently required for wireless environments, just as it was for wired networks.
However, the above approaches designed in a wired network are not applicable in a wireless
environment, which has some distinct characteristics, such as high error rate, location-
dependent and time-varying capacity, and scarce bandwidth 11.
In a wireless network with a multirate channel, when many packets encountering a
good channel are dropped and many packets encountering a poor channel are kept in the
buffer, the system performance becomes poor because most packets are transmitted in a
poor-capacity channel. Actually, dropping the packet having the poorest channel will cause
the optimal performance, but this behavior completely ignores to maintain the PLD model.
Therefore, how to keep the PLD model and raise the overall performance in a wireless
network with a multirate channel is a challenge and will be investigated in this paper.
This paper proposes a novel algorithm, named fuzzy proportional loss rate dropper
FPLR. In the FPLR dropper, we utilize the concept of fuzzy control to make an optimal
decision for dropping the small packets destined to a poor channel and reserving the large
packets destined to a good channel, causing a lot of large packets to be transmitted via a good
channel. Therefore, FPLR can provide high performance and keep PLD in a wireless with a
multistate channel.
Organization of the remainder of the paper is as follows. Section 2 describes the
background, including the proportional differentiation model, some previous related works,
structure of fuzzy controller systems, and what problems occur when these works are directly
applied into a wireless network. Section 3 describes in more detail the FPLR and the design of
the fuzzy controller. In Section 4, some simulation results and their implications are exhibited.
The conclusion of this work is given in Section 5.
2. Background
2.1. Proportional Differentiation Model
The proportional differentiation model was proposed first by Dovrolis and Ramanathan 5
and Bobin et al. 6, and its structure is shown as Figure 1. The arrival traffic is classified
into N service classes where each class has a dedicated queue. Let qi denote the measured
Mathematical Problems in Engineering 3
Class 1
Class 2
Proportional Proportional Output Wired link
Input differentiation . differentiation
.
dropper . scheduler
Class N
performance of class i. For proportional differentiation model, the following equation should
be satisfied for all pairs of classes, i and j,
qi ci
1 ≤ i, j ≤ N , 2.1
qj cj
where c1 ,c2 ,. . .,cN are the quality differentiation parameters QDP. A network operator could
manipulate the service quality spacing between classes by adjusting QDPs.
Let Li be the average loss of the class-i packets with σi as the loss differentiation
parameters LDPs. For all pairs of service classes, i and j, the proportional loss differentiation
model is specified by
Li σi
1 ≤ i, j ≤ N . 2.2
Lj σj
Let Di be the average queuing delay of the class-i packets with δi as the delay
differentiation parameters DDPs. The proportional delay differentiation model has the
following constraint for any pair of classes:
Di δi
1 ≤ i, j ≤ N . 2.3
Dj δj
Li t
Li t 1 ≤ i ≤ N, 2.4
Ai t
4 Mathematical Problems in Engineering
Mobile
Wireless host
channel
bandwidth
Class 1 MH1
Class 2 C1 (t)
Proportional Proportional Output MH2
Input differentiation . differentiation C2 (t)
.
dropper . scheduler .
.
.
Class N
CM (t)
MHM
i t Li t
L 1 ≤ i ≤ N, 2.5
σi
i t
J arg min L 1 ≤ i ≤ N. 2.6
i∈Bt
PLR aims to maintain a unanimous normalized average loss rate among all classes.
According to the number of packets that PLR estimates, the calculated average loss rate is
different; so there are two kinds of algorithms, namely, PLR with infinite memory, PLR ∞,
and PLR with memory M, PLRM. When calculating the average loss rate, PLR∞ counts
packets from initial to present, while PLRM observes the last M packets of every class at
present. From the long-term observation, the result of PLR∞ is closer to targeted proportion
than that of PLRM. However, when the class load significantly oscillates, adopting PLRM
is preferred because of its adaptation, but determining an optimal M is difficult.
Fuzzy
Defuzzification
inference engine
Fuzzification
Fuzzy Fuzzy
Controller inputs outputs Controller
inputs outputs
Fuzzy
rule base
FPLR
P Si
FCS P outi
(fuzzy control system) Drop decision Drop/not drop
C(Hi )
L1 t
Depti
L2 t DD (debt decision )
.
.
.
LN t
Therefore, designing a dropper to simultaneously keep the PLD model and raise the overall
performance in a wireless network with a multirate link is a challenge.
VS S MS L VL
1
(byte)
50 250 500 1000 1500 2000
a The membership functions for the input P S
VB B MC G VG
1
(Mbps)
0 1 2 5.5 11
b The membership functions for the input CH
nd pnd m pd d
1
P
0 0.25 0.5 0.75 1
c The membership functions for the output of fuzzy inference engine
3.1. Fuzzification
In this paper, most of the membership functions for the fuzzy sets are chosen to be triangular,
because triangular membership functions are the most economic. All membership functions
are defined on common interval 0, 1. With fuzzy logic, we assign grade values to our two
variables, which are packet sizes and channel states. Thus the fuzzy set A consists of two
fuzzy variables, P S, CH, where P S is the fuzzy variable term for the packet size and
CH is the channel capacity encountered when the scheduler transmits the HOL packet.
The membership functions of P S and CH are shown in Figures 5a and 5b, respectively,
and they are the input of the FCS. The fuzzification component translates the inputs into
corresponding fuzzy values.
In Figure 5a, VS, S, MS, L, and VL denote very small, small, medium small, large,
and very large, respectively. In Figure 5b, VB, B, MC, G, and VG indicate very bad channel,
bad channel, medium channel, good channel, and very good channel, respectively.
Let the fuzzified output of the fuzzy inference engine be P , which uses singleton fuzzy
sets representing the drop probability of the packet, as shown in Figure 5c. In Figure 5c,
nd, pnd, m, pd, and d denote not drop, probably not drop, maybe drop, probably drop, and
drop, respectively.
Mathematical Problems in Engineering 7
3.3. Defuzzification
The center of gravity CoG technique, which computes the weighted average of the center
of gravity of each membership function, is used to compute the defuzzified output, Pout ,
of the FCS. That is Pouti nf1 μi,f Pi,f Pi,f / nf1 μi,f Pi,f , where Pi,f is the center of the
membership function recommended by the consequence of rule f of class i, and the height
is cut by the minimum vale after MAX-MIN inference engine 14. ui,f Pi,f is the degree of
membership of input variables, packet size and channel states, of class i. The n is number of
rules.
N
i t
L
J t >
If L i1
then debtJ −0.25,
N
N 3.1
i t
L
J t <
If L i1
then debtJ 0.25,
N
where L J t is the normalized average loss rate of class J at time t and N is the number of
service classes.
8 Mathematical Problems in Engineering
We would like to use the decision function of FPLR to decide that the packet should
be dropped and need not be dropped. We use a threshold θ to compare with Pout . Thus the
condition of dropping the packet of the candidate class J is as follows:
PoutJ debtJ > θ. 3.2
3.5. Algorithm
Algorithm 1 presents the pseudocode of the FPLR. When the buffer has empty space, the
dropper simply inserts the packet into a proper queue. When buffer overflow occurs, the
dropper selects an appropriate packet to drop. Let Li t and Ai t be the number of dropped
packets and the number of arrived packets of class i at time t, respectively.
When an arriving packet encounters a full buffer, the dropper selects a proper packet,
which may be this arriving packet or a packet in the buffer to be dropped for keeping
the proportional loss rate among classes. At determining which packet to be dropped, the
dropper first calculates the normalized average loss rate L i t Li t/Ai tσi for each class i.
Then the class with the smallest normalized average loss rate, that is, J arg mini∈Bt L i t,
is regarded as the candidate class.
The dropper considers three important factors, including channel state, packet size,
and debt degree, to determine whether the packet of the class J should be dropped.
Note the dropper drops the HOL packet, rather than the tail packet or arriving packet,
because using this method can reduce the queuing delay of queued packets.
If the candidate class J does not satisfy the above condition, the dropper will choose
the candidate class K, which has the next smallest normalized average loss rate. Judging
whether the HOL packet of candidate class K will be dropped is similar to 3.2, that is,
PoutK debtK > θ.
Some noticeable points about the FPLR are explained as follows.
1 FPLR drops the HOL packet, rather than the tail packet, of the selected class,
because using this method can reduce the queuing delay of queued packets. Also
the HOL packet will actually experience the current channel capacity, but the
tail packet may not be transmitted at this capacity because of the time-varying
bandwidth.
2 For FPLR, because the packets destined to a low-capacity channel are easily
dropped and the packets destined to a high-capacity channel are easily kept
in the buffer, most packets are transmitted in a good channel, generating high
performance.
3 The fuzzy logic is an effective tool for efficient buffer management. It can
easily handle several nonlinear factors and does not need detailed mathematic
descriptions for the system.
upon WPLR and FPLR. Also, the behaviors of these two droppers under different timescales
are explored. First, the average loss rate of each class i, Li , is obtained by using 2.4, and the
loss rate ratio of class i over class j is defined as Rij Li /Lj . The loss improvement of class i is
P F P P F
defined as Li − Li /Li where Li and Li are the loss rates of class i made by WPLR and
FPLR, respectively. Similarly, the throughput improvement and delay improvement of class i
P F P
are defined as TiF −TiP /TiP and Di −Di /Di , respectively, where Ti denotes the throughput
of class i and Di denotes the average queuing delay of class i.
10 Mathematical Problems in Engineering
1 2 3 4 5
⎡ ⎤
1 −a ap1 ap12 ap13 ap14
⎢ ⎥
2⎢
⎢ap2 −a ap2 ap22ap23 ⎥
⎥
⎢ ⎥
⎢ ⎥
3⎢ 2 2 ⎥,
−a ap3 ap3 ⎥ 4.1
⎢ap3 ap3
⎢ ⎥
⎢ 3 ⎥
4⎢
⎢ap4 ap42 ap4 −a ap4 ⎥ ⎥
⎢ ⎥
⎣ ⎦
5 ap54 ap53 ap5 ap5 −a
2
where a is the state transition rate to other states and pi is the probability of state i being
translated to its neighbor states when the transition occurs. The default value of a is 30, and
the values of p1 , p2 , p3 , p4 , and p5 can be calculated by letting the sum of each row equal
to 0. In each simulation, at least 500,000 packets for each class are generated for the sake of
stability.
5 40
3
20
2
10
1
0 0
100 110 120 130 140 150 100 110 120 130 140 150
Arrival rate (packets/s) Arrival rate (packets/s)
15 30
10 20
5 10
0 0
100 110 120 130 140 150 100 110 120 130 140 150
Arrival rate (packets/s) Arrival rate (packets/s)
From Figure 6c, two phenomena are observed. First, the throughputs improved by
FPLR increase as arrival rate increases. Second, the throughput improvements of all classes
are in the order class 3 > class 2 > class 1. The first phenomena is because the more loss
improvement, the more throughput improvement. The second phenomenon is not trivial and
explained in the following.
Let λi be the packet arrival rate of class i. Thus for WPLR, the throughput of class i,
P
TiP , can be expressed as TiP λi 1 − Li . Also let the loss improvement of class i, made by the
P F P
dropper FPLR, be denoted as fiF , which is equal to Li − Li /Li . Therefore, the throughput
12 Mathematical Problems in Engineering
F P
of class i can be easily obtained as TiF λi 1 − Li λi 1 − 1 − fiF Li . Hence, the throughput
improvement of class i is
P
TiF 1 − 1 − fiF Li
−1 − 1. 4.2
TiP P
1 − Li
From this equation, we can prove the more loss improvement, the more throughput
improvement.
For different classes, observed from Figure 6b, the loss improvements, fiF , of all
P P P
classes are the same. Also Figure 6a shows that L3 > L2 > L1 . Under these conditions,
from 4.2, the throughput improvements of all classes are in the order class 3 > class 2 >
class 1. For example, all λi 120 packets/sec, loss rates of classes 1, 2, and 3 for WPLR are
20%, 40%, and 80%, respectively, that is, their throughputs are 100, 80, and 40. When the
loss improvement is 50% for all classes, the loss rates of classes 1, 2, and 3 for our proposed
dropper reduce to 10%, 20%, and 40%, respectively, that is, their throughputs are 110, 100,
and 80. In this example, the throughput improvements for classes 1, 2, and 3 are 10%, 25%,
and 100%, respectively.
Figure 6d shows the delay improvements achieved by FPLR. The delay improve-
ments are caused by two reasons. First, because the packet destined to a poor channel is easier
to be dropped than that destined to a good channel, nondropped packets usually encounter a
good channel, causing that they have short transmission time and thus short queuing delay.
Second, FPLR dropping the HOL packet, rather than dropping the tail packet, reduces the
queuing delay of the queued packets. Also the delay improvements enhance as the arrival
rate increases because the opportunity of using the fuzzy mechanism becomes large. Finally,
the delay improvements of all classes are in the order class 1 > class 2 > class 3 since the more
throughput improvement, the less delay improvement.
4.3. Timescale
In this simulation, the loss rate ratios between successive classes are measured over five
time intervals—100, 500, 1000, 5000, and 10000 p-units, where a p-unit is the average packet
transmission time. During each time interval, the loss rate ratios of class 2 over class 1 and
class 3 over class 2 are measured and averaged.
Figure 7 shows five percentiles, 5%, 25%, 50% median, 75%, and 95%, of the average
loss rate ratio. FPLR has the broader ranges on short timescales than WPLR and similar ranges
on long timescales. The reason is that although using the fuzzy mechanism will achieve the
loss rate ratio temporarily away from target loss rate proportion in the short term, they will
let the opportunity of using the debt decision mechanism keep PLD on the long term.
3.5
PLR
FPLR
PLR
PLR
FPLR
FPLR
PLR
1.5
PLR
FPLR
FPLR
1
0.5
0
100 500 1000 5000 10000
Measured timescales (p-units)
5% 75%
25% 95%
50%
Figure 7: Five percentiles of average loss rate ratio on various measured timescales.
50
5
45
4.5
40
4
35
3.5
Improvement (%)
30
3
Loss rate ratio
25
2.5
20
2
15
1.5
10
1
5
0.5
0
0 5 10 15 20 25
5 10 15 20 25 Number of mobile hosts
Number of mobile hosts
Loss
PLR : R 31 FPLR : R 31 Delay
PLR : R 21 FPLR : R 21 Throughput
a b
50
5
45
4.5
40
4
35
3.5
Improvement (%)
30
3
Loss rate ratio
25
2.5
20
2
15
1.5
10
1
5
0.5
0
0 20 25 30 35 40
20 25 30 35 40 State transition rate
State transition rate
Loss
PLR : R 31 FPLR : R 31 Delay
PLR : R 21 FPLR : R 21 Throughput
a b
loss improvement, total throughput improvement, and total delay improvement. Figure 8a
shows that the number of mobile hosts does not affect the loss rate ratios achieved by
two droppers. Figure 8b reveals that the loss improvement, throughput improvement, and
delay improvement made by FPLR increase as the number of mobile hosts increases. When
there are only a few mobile hosts, for FPLR, few HOL packets destined to poor channels can
be selected as a substitution, so that the improvements are not high. As the number of mobile
hosts increases, since more small packets transmitted via poor channels can be chosen, the
improvements made by FPLR raise.
40
5
35
4.5
30 4
25 3.5
Improvement (%)
10 1.5
1
5
0.5
0
0.1 0.3 0.5 0.7 0.9 0
Threshold 0.1 0.3 0.5 0.7 0.9
Threshold
Loss
Delay FPLR : R 31
Throughput FPLR : R 21
a b
4.6. Threshold θ
Herein we investigate the effects of the parameter θ for FPLR. Figure 10a reveals that loss
improvement, throughput improvement, and delay improvement made by FPLR increase
as threshold θ increases. As the threshold θ increases, since the opportunity of considering
packet size and channel states becomes large, smaller packets transmitted via low-capacity
channels are dropped, and more large packets transmitted via high-capacity channels are
kept. Thus the improvements made by FPLR increase. However, Figure 10b shows when
threshold θ exceeds 0.5, the target loss rate ratios slide down as threshold θ increases. The
reason is that in this case the loss rate proportions have minor influence, so the target loss
rate proportions cannot be maintained.
5. Conclusions
This paper proposed a high-performance algorithm, FPLR, which can be used to provide
proportional loss differentiation in a wireless network with multi-state channel. FPLR uses
three processes to decide how to drop packets. A fuzzy controller system on the first process
takes advantage of fuzzy theory to PLD. The debt decision module in the second process is
to compensate for the loss rate debts. The third process is to determine an appropriate packet
to drop. FPLR not only considers the normalized average loss rate, but also considers the
channel state and packet sizes, in order to improve the performance of dropping.
16 Mathematical Problems in Engineering
References
1 R. Braden, D. Clark, and S. Shenker, “Integrated services in the internet architecture: an overview,”
IETF RFC 1633, June 1994.
2 S Blake, D. Black, M. Carlson, Z. Wang, and W. Weiss, “An architecture for differentiated services,”
IETF RFC 2475, December 1998.
3 C. Dovrolis, D. Stiliadis, and P. Ramanathan, “Proportional differentiated services: delay differenti-
ation and packet scheduling,” ACM SIGCOMM Computer Communication Review, vol. 29, no. 4, pp.
109–120, 1999.
4 C. Dovrolis, D. Stiliadis, and P. Ramanathan, “Proportional differentiated services: delay differentia-
tion and packet scheduling,” IEEE/ACM Transactions on Networking, vol. 10, no. 1, pp. 12–26, 2002.
5 C. Dovrolis and P. Ramanathan, “Proportional differentiated services, part II: loss rate differentiation
and packet dropping,” in Proceedings of the IEEE/IFIP International Workshop on Quality of Service
(IWQoS ’00), pp. 52–61, June 2000.
6 U. Bobin, A. Jonsson, and O. Schelen, “On creating proportional loss-rate differentiation: predictabil-
ity and performance,” Lecture Notes in Computer Science, vol. 2092, pp. 372–379, 2001.
7 J. Zeng and N. Ansari, “An enhanced dropping scheme for proportional differentiated services,” in
Proceedings of the International Conference on Communications (ICC ’03), pp. 11–15, 2003.
8 J. Aweya, M. Ouellette, and D. Y. Montuno, “Weighted proportional loss rate differentiation of TCP
traffic,” International Journal of Network Management, vol. 14, no. 4, pp. 257–272, 2004.
9 M. Zhang, J. Wu, C. Lin, and K. Xu, “WSAP: provide loss rate differentiation with active queue
management,” in Proceedings of the International Conference on Communication Technology (ICCT ’03),
pp. 385–391, July 2003.
10 X. Zhou, D. Ippoliti, and T. Boult, “HPPD: a hop-count probabilistic packet dropper,” in Proceedings
of the IEEE International Conference on Communications (ICC ’06), pp. 116–121, July 2006.
11 Y. Cao and V. O. K. Li, “Scheduling algorithms in broad-band wireless networks,” Proceedings of the
IEEE, vol. 89, no. 1, pp. 76–87, 2001.
12 E. H. Mamdani, “Application of fuzzy algorithms for simple dynamic plant,” Proceedings of the
Institution of Electrical Engineers, vol. 121, no. 12, pp. 1585–1588, 1974.
13 E. H. Mamdani and S. Assilian, “An experiment in liquistic synthesis with a fuzzy logic controller,”
International Journal of Man-Machine Studies, vol. 7, no. 1, pp. 1–13, 1975.
14 L. Zadeh, “Soft computing and fuzzy lgic,” IEEE Software, vol. 11, no. 6, pp. 48–56, 1994.
Hindawi Publishing Corporation
Mathematical Problems in Engineering
Volume 2012, Article ID 783476, 21 pages
doi:10.1155/2012/783476
Research Article
A Cost-Effective Planning Graph Approach for
Large-Scale Web Service Composition
Copyright q 2012 Szu-Yin Lin et al. This is an open access article distributed under the Creative
Commons Attribution License, which permits unrestricted use, distribution, and reproduction in
any medium, provided the original work is properly cited.
Web Service Composition WSC problems can be considered as a service matching problem,
which means that the output parameters of a Web service can be used as inputs of another one.
However, when a very large number of Web services are deployed in the environment, the service
composition has become sophisticated and complicated process. In this study, we proposed a novel
cost-effective Web service composition mechanism. It utilizes planning graph based on backward
search algorithm to find multiple feasible solutions and recommends a best composition solution
according to the lowest service cost. In other words, the proposed approach is a goal-driven
mechanism, which can recommend the approximate solutions, but it consumes fewer amounts
of Web services and less nested levels of composite service. Finally, we implement a simulation
platform to validate the proposed cost-effective planning graph mechanism in large-scale Web
services environment. The simulation results show that our proposed algorithm based on the
backward planning graph has reduced by 94% service cost in three different environments of
service composition that is compared with other existing service composition approaches which
are based on a forward planning graph.
1. Introduction
Research on Web Service Composition WSC has become increasingly important in recent
years due to the growing number of web services over the Internet and the challenge of
automating the process. Particularly with the development of cloud computing, there will be
more and more diverse Web services deployed and published on cloud environments. Web
services are Internet-based software components which have the capabilities of delivering
service cross-platforms and languages. The W3C organization has defined “Web Services”
that “a software system designed to support interoperable machine-to-machine interaction
over a network. It has an interface described in a machine-processable format specifically
2 Mathematical Problems in Engineering
Web Services Description Language WSDL. Other systems interact with the Web service
in a manner prescribed by its description using SOAP messages, typically conveyed using
HTTP with an XML serialization in conjunction with other Web-related standards 1”. The
W3C has also pointed out that “We can identify two major classes of Web services, REST-
compliant Web services, in which the primary purpose of the service is to manipulate XML
representations of Web resources using a uniform set of stateless operations; and arbitrary
Web services, in which the service may expose an arbitrary set of operations 2”.
Since the growth of Web services to a large number is happening and possible
interactions among them are huge, searching, analying, and processing them to find the
required services to achieve user goals is very difficult via a manual process. This also means
service composition problem has become increasingly sophisticated and complicated in the
real world 3. Therefore, the issue of finding solutions efficiently via composing services to
form a complex composited service is one of the important studies.
The process of combining and linking existing Web services to create new service is
known as WSC. In other words, WSC problems can be considered as a service matching
problem, which means that the output parameters of a Web service can be used as inputs
of another Web service. The aim of WSC is to provide a means for composing diverse
Web services to accomplish user request which cannot be satisfied by a single Web service.
The Web services composition approaches can be broadly classified as static or dynamic
based on the process and the way of composing services. A static Web service composition
is constructed to solve the particular problem through manually identifying Web services
by their capabilities. They are composed by a series of known Web services and a set of
known data in order to obtain the expected results. Dynamic Web service composition is to
automatically select Web services and compose those at the execution/run time. The aim is
to build and utilize an automated service discovery and its associated execution mechanism
to produce the required composite services. There have been numerous methods proposed
for solving the problem of service composition, such as workflow 4 and AI planning 5.
The Web service composition is commonly described by using the Web Services Business
Process Execution Language BPEL 6 which is an XML-based language that provides
particular functionalities for processes, such as define variables, create conditionals, design
loops, and handle exception. It utilizes Web services as the model for the decomposition and
the composition of the process. However, BPEL promotes the development of workflow and
the integration of business processes.
Nowadays, numerous researches focus on finding and developing new approaches to
fit in with WSC. The task of WSC usually assumed that the composition process generates a
composition workflow, which starts from the known variables from the requirements or the
related constraints to the expected goal. Therefore, many algorithms based on AI planning
techniques that can facilitate to automat Web service composition have been proposed
3, 5, 7–9, but it is still a great challenge for solving large-scale WSC problem to obtain
multiple flexible service composition solutions with acceptable service cost and execution
time. It can assume that Web services as actions, and the process of composing them to
produce the desired result as planning, so planning graph is one of the most suitable
techniques could be used for WSC problem. However, there are very few studies using
planning graph approach to achieve WSC problem, especially with considering both sides
of cost and effectiveness in large-scale Web services composition.
Therefore, this paper proposes a new cost-effective planning graph approach based
on backward strategy for large-scale Web service composition on cloud environment, which
can find multiple solutions and recommend a list of best composite services composition to
Mathematical Problems in Engineering 3
users. In addition, we can recommend the approximate match services which may not totally
meet to user requests, but the user may accept the services and it uses fewer amounts of Web
services and less nested levels of composite Web services. The main research objectives in
this paper are 1 to present a novel framework and composition processes for WSC on cloud
environment, 2 to design a cost-effective WSC algorithm which can obtain multiple service
composition solution using fewer number of Web services with low cost and in acceptable
execution time, 3 the proposed approach must process large-scale Web services which
amount over 10000, and 4 to provide an approximate solution when there is no composite
solution which exactly corresponds to the request.
The rest of this paper is structured as follows. Section 2 describes the related works.
Section 3 proposes the planning graph service composition algorithm based on the backward
strategy. Section 3 presents the details of experiment and its results, and we give a summary
discussion about the result. Finally, Section 4 concludes this study and proposes the future
work.
2. Related Works
The planning graph, which is a representation technique by AI planning, provides a very
powerful search technique in a space 10 to improve the efficiency of AI algorithms. A
WSC problem can be modeled as a planning graph. The input parameters of the composition
request are mapped to the initial state, and the output parameters of the composition request
are mapped to goal propositions. If a planning graph reaches a proposition level which
contains all required parameters, then it searches backward from the last level of the graph for
a solution. However, the disadvantage of this approach is the difficulty of designing a strategy
to trade off two key criteria that are cost and effectiveness. Therefore, there are always two
problems of time consuming and redundant actions existed in the solution.
The planning graph is a layered graph whose edges are only allowed to connect two
nodes from one layer to next layer. And the planning graph’s layers are with an alternating
sequence of action layer and proposition layer. The proposition layer contains a finite set
of states, and the action layer contains a finite set of actions the action has preconditions,
negative effects, and positive effects. For example, the first layer of planning graph, P0, is a
proposition layer which contains the initial states of the planning problem. The next layer, A1,
is an action layer which contains a set of actions which preconditions can be satisfied by P0,
and P1 is the union of the states of P0 and the effects of all A1’s actions. Those preconditions
of actions in A1 are connected to the state nodes in P0 by incoming arcs, and those positive
or negative effects in P1 are connected to the state nodes in P1 by outgoing arcs. The process
continues until it reaches the goal states or the fixed-point level of the graph.
The study conducted by 10 showed a Planning Graph planner technique called
GRAPHPLAN. The GRAPHPLAN algorithm is operated in two main steps which alternate
within a loop: graph expansion and solution extraction. The solution extraction can be
formulated as a constraint solving problem 11 or as a search problem 12. In Peer’s survey
9, GRAPHPLAN’s advantages include good performance, soundness, completeness,
generation of shortest plans, and termination on unsolvable problems. However, the original
GRAPHPLAN algorithm has some limitations: 1 its representation language is restricted to
pure STRIPS operators, so no conditional or universally quantied effects are allowed; 2 the
performance can decrease drastically if too much irrelevant information is contained in the
specification of a planning task 13.
4 Mathematical Problems in Engineering
Preprocessing
Goal
Cloud
web Web
service service
repository
Service Search
group Web service
optimal
matching composition solution
module module module
Semantic
WordNet similarity
module
Composition Feedback
repository
Zheng and Yan 7 also transformed the problem of service composition into the
problem of simplified planning graph based on forward search, which could be constructed
in polynomial time. In classical AI planning technique, generating final solutions by a
backward search is a popular approach, but it is the most time consuming technique.
Researches have been working on it to improve it. However, forward search could improve
efficiency, but the redundant Web services during the construction of the planning graph
lead to the increase of service cost. Zheng and Yan 7 put efforts into using forward search in
planning graph algorithm to solve WSC problem, and it shows a good result which can find
a solution in polynomial time but encounters some drawbacks: 1 there are many redundant
Web services existed in the solution of service composition, and 2 it is lack of flexible
search mechanism which can recommend multiple solutions for service composition when
few input unknown parameters occur. In other words, the composition algorithm based on
the forward strategy aims to minimize the search time, but there are many possible redundant
and unnecessary Web services included in the final solution.
I Preprocessing. There are two components involved in the first step. One is the Web
service repository, and the other is semantic similarity module. The Web Service
Repository will search Web services from distributed UDDIs on cloud and store
those services in a repository database and entries in the repository will be updated
regularly. Therefore, the input of this mechanism is Cloud Web Services. Semantic
Mathematical Problems in Engineering 5
Similarity Module precalculates the semantic similarity values between any two
concepts and stores the similarity values in a semantic similarity database for
retrieval.
II Service group matching module. It utilizes Web Service Repository and Semantic
Similarity Module to select the Web services that can satisfy the query, and group
them based on the degree of their similarity. It will provide a set of service groups
for service composition.
III Web Service Composition Module. It will query Service Group Matching Module
to get services which are required by composition algorithm. Web Service Com-
position Module will generate multiple service composition solutions according to
the goal which is described in final output parameters of each solution. With the
expansion of levels in backward planning graph composition algorithm, the goal
will be refined at each iteration.
IV Search Optimal Solution Module: It will calculate the score of each solution and
choose the most suitable solution of WSC from these identified solutions according
to the given goal.
3.1. Preprocessing
In large-scale WSC, the number of querying services could be large. Querying Web service
entries registered in distributed UDDIs at runtime in process of service composition, the
efficiency is likely lower than those entries stored in one centralized database. So, all Web
service entries to be used in this approach will be stored to a centralized structured Web
Service Repository. In addition, the calculation of semantic similarity between concepts is a
time consuming task which is not efficient to meet dynamic service composition, so it will
be preprocessed by Semantic Similarity Module. Service Group Matching Model according
to the repository and the relationships of concept similarity to respond the query. The
preprocessing task requires two components.
WS
UDDI
WS
WS
UDDI WS
WS Web service
WS
WS repository
WS
UDDI UDDI
WS UDDI
Entity
Physical Abstract
module is useful but optional in the proposed service composition approach, so that it is not
introduced to experiments in Section 4.
ABCD W1 IJ
ABC W2 IJ
AB W4 IJK
ABC W3 IJK
AB W4 IJK
EF W5 IJK EF W5 IJK
i ← 0, P0 ← s0 , G ← P0
repeat
G ← Expand Based On Backward G
S ← Extract Solutions G, S
S ← Reduce Solutions S
ii 1
Validate Solution G, S
Compute Score G, S
Output Search Solution S
Algorithm 1
extraction from a planning graph, which help to find the initial state. The main composition
algorithm is shown as follows.
for w ∈ W do
for wg ∈ WG do
if effect w ⊆ effect wg ∧ precond w ⊇ precond wg then
wg ← wg ∪ w
if effect w ⊇ effect wg ∧ precond w ⊆ precond wg then
wg ← wg ∪ w
effect wg ← effect w
precond wg ← precond w
if not be filled with service group then
WG ← WG ∪ new wg
Ai 1 ← WG
return G
Algorithm 2
for s ∈ S do
for a ∈ Ai do
available a ← true
S←S−s
do
required ← inputs s
ns ← new solution
parent ns ← s
for a ∈ Ai do
if available a ∧ required ∩ effect a /
NULL then
required ← required − required ∩ effect a
available a ← false
ns ← ns ∪ a
if required NULL then
S ← S ∪ ns
break
while required NULL
return S
Algorithm 3
for s ∈ S do
if count s > Service T hreshold then
S←S−s
for s ∈ S do
if initial s ⊃ {initial s2 | s2 ∈ S} then
S←S−s
return S
Algorithm 4
services used in any composition solution. It is the key to reduce redundant solutions and
facilitate the efficiency of composition algorithm. It can be determined by users according to
their server performance. The Service Threshold value is set to 3000 by default value which
is very huge in the experiments. The other is to remove the solutions which have too many
services and similar to other short solutions in each action level. The process helps to filter
large number of unwanted solutions and identified the most appropriate ones.
for s ∈ S do
intersection ← Count initial s ∩ initial g
union ← Count initial s ∪ initial g
precise s ← intersection/union
if Precise Threshold ≤ precise s then
Search Solution S ← Search Solution S ∪ s
Algorithm 5
parameters rout {M, N}, and there are nine Web services in our Web Service Repository.
Table 1 shows the details of the example of Web Service Repository.
Figure 5 shows the expanded planning graph result of the above given example.
{A, B, C, D} and {M, N} are the input and output parameters of the composition request. At
first, we search Web services which can output {M, N}, then we get {w7 , w8 , w9 } which can
support the proposition 4 P4, our goal. Those three Web services will be involved in action 3
A3. We collect input parameters of Web service in action 3 A3, and we will get proposition
3 P3. The rest of proposition and action are like this, and so forth. From the previous step, a
planning graph is generated to extract solutions. We utilize our proposed algorithm to extract
solutions which store them in a tree structure to form a solution tree for tracing. Every leaf
node in solution tree means that there is a solution from leaf node to root. The result is shown
in Figure 6. Many possible paths that can reach initial state of the tree have been discovered.
This is one of advantages of adopting the backward strategy, so that multiple solutions for
user request can be found.
{M, N} are the output parameters of user request. It is located in proposition 4 P4,
so we need to find the combinations of Web services in action 3 A3, which corresponds
to {M, N}. And we will get two combinations, which are {W7, W9} and {W8}. Those
combinations will be added to the root as its child. Now, there are two nodes at second level.
The solution node composited by {W7, W9} requires a set of input parameters {H, L}, so we
need to find the combinations of Web services in action 2 A2, which can correspond to
{H, L}. So we get {W2, W5} and {W2, W6}, and so forth.
12 Mathematical Problems in Engineering
W2
A A H W7
W3
B I M
B
W1 D J W8
W4 N
C E K
D F L W9
W5
G
W6
P1 A1 P2 A2 P3 A3 P4
M, N
W7, W9 W8
A, B, D A, B, D, G W1
A, B, C, D
Figure 6: The solution tree for the planning graph for the above example.
∩ s1 s2
Precise s1 , s2 . 3.1
∪ s1 s2
Equation 3.1 shows the precision. In this equation, ∩ s1 s2 represents the amount of
the same concepts and ∪ s1 s2 represents the amount of all different concepts, where S1 and
S2 both are lists of concepts. This equation evaluates the similarity between the solution’s
initial states and the user request’s inputs. It also means the difference between two lists of
concepts. After the previous step, we have the likelihood of the solutions to achieve the goal.
Mathematical Problems in Engineering 13
For calculation of the score for each solution, we need to calculate matching degree between
the levels of the solutions. The matching equation is illustrated in the following
Equation 3.2 shows the matching score. In the above equation, KM represents
classical Kuhn-Munkres algorithm which solved the assignment problem, and Sim represents
the similarity between any two concepts in s1 and s2 . This equation is to evaluate the
matching score of two solutions. With the previous two formulas, we can calculate the
solution score. It sums the matching scores between levels of the solutions and divides by
the number of levels to gain the average. Then, we get the average of matching scores, and
multiple by the matching precision of the solution and the request. The score equation is
shown as below.
slu
s1 ,s2 Mat s1 , s2
Score slu Precise slu · in, r · in . 3.3
n
Equation 3.3 shows the solution score. Solution slu is a list of nodes from node leaf
to the root, which represents each level of service composition, slu · in represents the input
parameters of the solution slu, r · in represents the input parameters of the request r, and the
number of levels is represented by n. The Precision calculates the matching precision between
the input parameters of the request r and the input parameters of the solution slu.
Briefly, our proposed mechanism is particularly suitable for large-scale service
composition due to the problems that service composition is an error-prone and complicated
process. In this study, we designed a cost-effective planning graph mechanism for finding
multiple service composition solutions with lower service cost to overcome the above
problems. Furthermore, in some cases, the exact solution does not exist, and our proposed
mechanism still can recommend approximate solutions. On the other hand, in order to
provide multiple solutions, it must need to trace each possible solution. Our proposed
mechanism can filter out and reduce the less appropriate solutions in order to avoid
exponential growth in complexity.
4. Experiments
In this section, the simulation design, assumption, performance metrics, and simulation
results are described. Moreover, there is a brief discussion in the later section.
Web service
Web service composition
repository
The architecture of simulation platform is shown in Figure 7. WSBen was set up with
the test data including a set of Web services and a set of feasible requests for testing WSC
algorithm. In the simulation, a program extracts the information of Web service generated by
WSBen and then stores it in the Web Service Repository which includes service name, input
parameters, and output parameters. The Service Group Matching Module is in charge of Web
services selection according to the service query which has been processed by the Web Service
Composition Module. In addition, it also groups the selected Web services as Web service
groups for the composition algorithm. The Web Service Composition Module including a
composition algorithm composes a composite service according to the composition request,
which interacts with the service matching module in the discovery and selection process until
a possible solution had been found, if there is any. Therefore, it will generate a list of candidate
composite Web services which can fulfill the required services. The Search Optimal Solution
Module calculates each candidate solution to obtain a score in order to generate an ordered
recommendation list for selection.
i Erdo-Renyi |J|, p. The model has such a simple generation approach that it
creates |J| nodes in graph and assign each edge in the graph with probability
p.
ii Newman-Watts-Strogatz |J|, k, p. The initialization is a ring graph with k
nodes. Each node adds to graph and constructs edge by connecting to others
with probability p. The process will iterate until there are |J| nodes in the
graph.
iii Barabasi-Albert |J|, m. There are m nodes with no edge in the initial graph.
Each node adds with m edges, which are preferentially attached to existing
nodes with high degrees.
III η donates the parameter condense rate. Users can control the density of partial
matching cases in generated Web services.
IV Mp donates the minimum number of parameters in a cluster. In other words, each
cluster has at least Mp parameters.
V |W| donates the total number of Web services in a test dataset.
I Random Network: Barabasi-Albert 100, 0.06. The model creates 100 nodes in the
graph and each edge in the graph is with probability 0.06 to be chosen.
II Small-World Network: Newman-Watts-Strogztz 100, 6, 0.1. The initialization is
a ring graph with 6 nodes. Each node adds to the graph and constructs edges
connected to each other with probability 0.1, until there are 100 nodes in this graph
III Scale-Free Network: Erdo-Reyi 100, 6. There are 6 nodes with no edge in the
initial graph. Each node adds 6 edges until reach 100 nodes. Each added edge is
preferentially attached to existing nodes with high degrees.
For each network, there are 10 different sizes in each of test data types, which sizes are
10,000 to 100,000, respectively. Thus, there are 30 test sets three frameworks multiplied by
ten different test sizes in our large-scale WSC simulation.
Case 1 Random Network. Table 2 shows the results of five requests of random network
with |W| 10, 000 Web services in a test dataset. The results of criterion #P mean that both
Backward and Forward strategies can find solutions in all cases #P 1. Regarding #L,
Backward and Forward strategies have no difference in finding solutions. In the criterion of
#T , the computational efficiency of Forward is better than Backward, because of generating
Mathematical Problems in Engineering 17
Backward Forward
Test request
#L #C #T #P #L #C #T #P
r1 8 18 2.725 1 8 262 1.033 1
r2 8 14 3.028 1 8 237 1.05 1
r3 7 9 1.916 1 7 220 1.066 1
r4 7 10 2.191 1 7 245 1.072 1
r5 9 16 4.141 1 9 241 1.062 1
Avg. 7.8 13.4 2.8 1 7.8 241 1.056 1
Backward Forward
Test request
#L #C #T #P #L #C #T #P
r1 14 14 1.599 1 14 183 0.863 1
r2 11 11 1.016 1 11 175 0.769 1
r3 12 12 1.985 1 12 156 0.746 1
r4 10 10 2.419 1 10 174 0.716 1
r5 16 16 1.854 1 16 181 0.903 1
Avg. 12.6 12.6 1.776 1 12.6 173.8 0.8 1
final solutions by a backward search to expand graph layers is very time consuming. It still
falls within an acceptable time frame in our proposed approach even in the large-scale 10000
Web services environment. In the experiment with #C Web services, our proposed Backward
approach outperforms the Forward. Backward strategy uses less 20 services to fulfill the
request in all cases, but the Forward strategy requires more than 200 Web services to find
a solution.
Case 2 Small World Network. Table 3 shows the results of five test requests of a small world
network with |W| 10, 000 Web services in a test dataset. The results of criterion #P mean
that both our proposed Backward and the Forward strategies still can find solutions in all
cases #P 1. Regarding #L, the result of Backward strategy is as good as the Forward
strategy. Our proposed Backward strategy takes more a little time than the Forward strategy
in #T , but processing time is still acceptable. The reason is to generate final solutions by a
backward search to expand graph layers is very time consuming. Finally, in the experiment
with #C, it shows it has produced much better performance than the Forward strategy.
Case 3 Scale-Free Network. Table 4 shows the result of five test requests of a scale-free
network which contains |W| 10, 000 Web services in a test dataset. The results of criterion #P
mean that Forward strategy still can satisfy all requests #P 1, but it requires more than 200
Web services to obtain the solution in service cost #C. In the more complex scale-free network,
although Backward strategy finds the fully matched solution is impossible in some cases, an
approximate solution can be found and replaced, which use much less services to satisfy the
request. Regarding #T and #L, the result shows that Backward strategy are a bit better than
Forward in #T , but both produce almost the same performance in these two criteria.
18 Mathematical Problems in Engineering
Backward Forward
Test request
#L #C #T #P #L #C #T #P
r1 4 11 1.232 0.933 4 244 2.48 1
r2 4 6 3.151 1 4 343 1.654 1
r3 5 11 2.886 0.778 5 356 2.824 1
r4 — — — — 4 313 1.414 1
r5 4 10 0.203 0.814 4 281 2.122 1
Avg. 4.25 9.5 1.868 0.8812 4.4 316 2.3808 1
450
400
350
300
No. of C
250
200
150
100
50
0
1 2 3 4 5 6 7 8 9 10
|W| ×104
Backward
Forward
As shown in Figures 8, 9, and 10, the proposed algorithm has much better usage
of Web services in all cases in terms of obtaining the solution. Because of the aim of the
forward algorithm, which is to reach the goal as quick as possible, it expands the search space
in planning graph for services composition problem no matter how many redundant Web
services are produced. However, the proposed backward algorithm has no redundant Web
services existed in the solution, because its backward strategy searches what it needs to reach
the initial state. Figure 10 shows average cost of searching solution in Scale-Free Network.
It can be observed that the forward algorithm represents an unstable circumstance in Scale-
Free Network, when the size of a test dataset becomes large. Nevertheless, our backward
algorithm is still to appear stable and effective results. On average, our algorithm reduces
to 94% service cost for finding the solutions. From the above experiment results, we have a
simple deduction that the backward search will continue to display the stable and smooth
results in different types of network topology, even if the sizes of Web services continue to
increase.
In experiments, three types of network topology and diverse sizes of Web services are
utilized to evaluate these two algorithms. The main findings about the proposed algorithm
from the experiments are given as follows. In effectiveness, the experiment results show
our proposed backward algorithm has 94% better effectiveness compared to the forward
algorithm in most cases. In few cases, it uses little more levels to obtain the solution, but it
Mathematical Problems in Engineering 19
450
400
350
300
No. of C
250
200
150
100
50
0
1 2 3 4 5 6 7 8 9 10
|W| ×104
Backward
Forward
450
400
350
300
No. of C
250
200
150
100
50
0
1 2 3 4 5 6 7 8 9 10
|W| ×104
Backward
Forward
can get low-cost solutions. As confirmed by the experiment results, our proposed algorithm
can also get very high-precise solutions in most cases. In efficiency, although the forward
algorithm has better performance than the backward algorithm, the cost of solution is very
high. The proposed backward algorithm is efficient when the sizes of Web service are less
than 50,000. Therefore, the effect of this cost-effective approach for large-scale WSC problem
is exhibited.
20 Mathematical Problems in Engineering
5. Conclusion
In this study, we proposed a backward planning graph mechanism for Web service
composition on cloud environment. It utilizes a planning graph based on a backward
search to find multiple feasible solutions and recommends the best composite solution
based on their service costs. We also validated that the proposed algorithm can improve
the error-prone problem of service composition and the redundant Web service involved in
large-scale service composition problem. Therefore, the algorithm based on the backward
planning graph search, which is capability of recommending multiple service composition
and remove the redundant services. As the experiment results in this paper, we proved that
our proposed backward algorithm had a better cost-effectiveness than the forward search
algorithm in terms of service cost. The proposed algorithm is able to recommend approximate
solutions of service composition using very few Web services, because it has higher quality
of relationships between services. In other words, we can decrease the amount of cost of Web
services and remain acceptable planning graph levels and execution time.
In the future, we will study how to improve a greedy algorithm in order to expand
the solution tree. To obtain right combinations is a very important issue for the design of
algorithm. It cannot only help to decrease wrong combinations, but also to improve the
effectiveness and efficiency of algorithm. Moreover, we can add more predictable restrictions
to prune the huge combination tree nodes for our algorithm efficiency. If there are some
more predictable restrictions and composition information, then that will help us to make
more appropriate decisions to find the solutions. Moreover, because the Semantic Similarity
Module in the proposed approach is optional, this module has been not analyzed in the
experiments. There is a need to have an environment that can help to validate semantic
association of service compositions. To design a semantic experiment environment should
be undertaken determining how the semantic can influence the effectiveness of service
composition.
Acknowledgment
This research work was supported by the National Science Council of the Republic of
China under the Grant NSC-99-2410-H-009-036-MY3. This research is carried out as a part
of GREENet which was supported by a Marie Curie International Research Staff Exchange
Scheme Fellowship within the 7th European Community Framework Programme under
grant agreement no. 269122.
References
1 H. Haas and A. Brown, “Web Services Glossary,” http://www.w3.org/TR/ws-gloss/.
2 D. Booth, H. Haas, F. McCabe et al., “Web Services Architecture,” http://www.w3.org/TR/ws-arch/.
3 M. Kuzu and N. K. Cicekli, “Dynamic planning approach to automated web service composition,”
Applied Intelligence, vol. 36, no. 1, pp. 1–28, 2010.
4 S. Wang, W. Shen, and Q. Hao, “Agent based workflow ontology for dynamic business process
composition,” in Proceedings of the 9th International Conference on Computer Supported Cooperative Work
in Design (CSCWD ’05), pp. 452–457, Coventry, UK, 2005.
5 L. Qiu, L. Chang, F. Lin, and Z. Shi, “Context optimization of AI planning for semantic Web services
composition,” Service Oriented Computing and Applications, vol. 1, no. 2, pp. 117–128, 2007.
6 A. Alves, A. Arkin, S. Askary et al., “Web Services Business Process Execution Language Version 2.0,”
OASIS WSBPEL TC, 2007.
Mathematical Problems in Engineering 21
7 X. Zheng and Y. Yan, “An efficient syntactic web service composition algorithm based on the planning
graph model,” in Proceedings of the IEEE International Conference on Web Services (ICWS ’08), pp. 691–
699, September 2008.
8 Y. Yan and X. Zheng, “A planning graph based algorithm for semantic web service composition,” in
Proceedings of the IEEE Joint Conference on E-Commerce Technology (CEC ’08) and Enterprise Computing,
E-Commerce and E-Services (EEE ’08), pp. 339–342, July 2008.
9 J. Peer, Web Service Composition as AI Planning—a Survey, University of St.Gallen, St. Gallen,
Switzerland, 2005.
10 A. L. Blum and M. L. Furst, “Fast planning through planning graph analysis,” Artificial Intelligence,
vol. 90, no. 1-2, pp. 281–300, 1997.
11 S. Kambhampati and M. B. Do, “Planning as constraint satisfaction: Solving the planning graph by
compiling it into CSP,” Artificial Intelligence, vol. 132, no. 2, pp. 151–182, 2001.
12 S. Russell and P. Norvig, Artificial Intelligence: A Modern Approach, Prentice Hall Press, Upper Saddle
River, NJ, USA, 2009.
13 Y. Dimopoulos, B. Nebel, and J. Koehler, “Encoding planning problems in nonmonotonic logic
programs,” in Proceedings of the 4th European Conference on Planning: Recent Advances in AI Planning
(ECP ’97), pp. 169–181, Toulouse, France, 1997.
14 S. C. Oh and D. Lee, “WSBen: A web services discovery and composition benchmark toolkit,”
International Journal of Web Services Research, vol. 6, no. 1, pp. 1–19, 2009.
15 X. F. Wang and G. Chen, “Complex networks: Small-world, scale-free and beyond,” IEEE Circuits and
Systems Magazine, vol. 3, no. 1, pp. 6–20, 2003.
Hindawi Publishing Corporation
Mathematical Problems in Engineering
Volume 2012, Article ID 398232, 17 pages
doi:10.1155/2012/398232
Research Article
An Optimal Classification Method for
Biological and Medical Data
Copyright q 2012 Yao-Huei Huang et al. This is an open access article distributed under the
Creative Commons Attribution License, which permits unrestricted use, distribution, and
reproduction in any medium, provided the original work is properly cited.
This paper proposes a union of hyperspheres by the mixed-integer nonlinear program to classify
biological and medical datasets. A classifying program with nonlinear terms uses piecewise
linearization technique to obtain a global optimum. The numerical examples illustrate that the
proposed method can obtain the global optimum more effectively than current methods.
1. Introduction
Classification techniques have been widely applied in the biological and medical research
domains 1–5. Either objects classification or patterns recognition for biological and medical
datasets necessarily demands an optimum accuracy for saving patients’ lives. However,
cancer identification with the supervised learning technique does not take a global view in
identifying species or predicting survivals. The improvement should cover the whole scope
to give implications instead of only considering the efficiency for diagnosis. This research
aims to extract features from whole datasets in terms of induction rules.
In the given dataset with several objects, in which each object has some attributes and
belongs to a specific class, classification techniques are used to find a rule of attributes that
appropriately describes the features of a specified class. The techniques have been studied
over the last four decades, including decision tree-based methods 6–11, hyperplane-based
methods 12–14, and machine learning-based methods 14–17.
2 Mathematical Problems in Engineering
To assess the effects of these classifying techniques, three criteria are used for
evaluating the quality of inducing rules based on the study of Li and Chen 3.
i Accuracy. The rule fitting a class should not cover the objects of other classes. The
accuracy of a rule should be the higher the better.
ii Support. A good rule of fitting a class should be supported by most of the objects of
the same class.
iii Compact. A good rule should be expressed in a compact way. That is, the fewer the
number of rules, the better the rules are.
This study proposes a novel method to induce rules with high rates of accuracy,
support, and compactness based on global optimization techniques, which have become
more and more useful in biological and medical researches.
The rest of this paper is organized as follows. Section 2 gives an overview of the related
literatures. Two types of mathematical models and a classification algorithm are proposed in
Section 3. The numerical examples demonstrate the effectiveness of the proposed method in
Section 4. Finally, the main conclusions of this study and future work are drawn in Section 5.
2. Literature Review
Currently, two well-known methods are used to induce classification rules. The first method
is the decision tree-based method, which has been developed in the last few decades 6–10.
It is widely applied to fault isolation of an induction motor 18 to classify normal or tumor
tissues 19, skeletal maturity assessment 20, proteomic mass spectra classification 21, and
other cases 22, 23. Although the decision tree-based method assumes that all classes can be
separated by linear operations, the inducing rules will suffer if the boundaries between the
classes are nonlinear. In fact, the linearity assumption prohibits practical applications because
many biological and medical datasets have complicated nonlinear interactions between
attributes and predicted classes.
Consider the classification problem with two attributes as shown in Figure 1, where
“” represents a first-class object, and “•” represents a second-class object. Figure 1 depicts a
situation in which a nonlinear relationship exists between the objects of two classes. Decision
tree method focuses on inducing classification rules for the objects, as shown in Figure 1b,
in which the decision tree method requires four rectangular regions to classify the objects.
The second is the support vector hyperplane method, which conducts feature selection
and rule extraction from the gene expression data of cancer tissue 24; it is also applied
in other applications 12–14, 25. The technique separates observations of different classes
by multiple hyperplanes. As the number of decision variables is required to express the
relationship between each training datum and hyperplane, and the separating hyperplane
is assumed a nonlinear programming problem, the training speed becomes slow for a large
number of training data. Additionally, similar hypersphere support vector methods have
been developed by Lin et al. 26, Wang et al. 27, Gu and Wu 28, and Hifi and M’Hallah
29 for classifying objects. In classification algorithms, they partition the sample space using
the sphere-structured support vector machine 14, 30. However, these methods need to form
a classification problem as a nonlinear nonconvex program, which makes reaching an optimal
solution difficult. Taking Figure 1 as an example, a hyperplane-based method requires four
hyperplanes to discriminate the objects, as shown in Figure 2.
Mathematical Problems in Engineering 3
i Decision tree-based methods are heuristic approaches that can only induce feasible
rules. Moreover, decision tree-based methods split the data into hyperrectangular
regions using a single variable, which may generate a large number of branches
i.e., low rates of compactness.
ii Hyperplane-based methods use numerous hyperplanes to separate objects of
different classes and divide the objects in a dataset into indistinct groups. The
method may generate a large number of hyperplanes and associated rules with
low rates of compactness.
4 Mathematical Problems in Engineering
m: number of attributes,
rt,k
rt,k
Notation 1. Normalization rescales all ai,j as ai,j . The following is a normalizing formula:
ai,j − aj
ai,j , 3.1
aj − aj
where 0 ≤ ai,j ≤ 1, aj is the largest value of attribute j, and aj is the smallest value of
attribute j.
xi ai,1 , ai,2 , . . . , ai,m ; ci , 3.2
Maximize ut,i,k
i∈I
m
2
subject to ai,j − ht,k,j ≤ 1 − ut,i,k M 2
rt,k , ∀i ∈ I ,
j1
3.4
m
2
ai ,j − ht,k,j 2
> rt,k , ∀i ∈ I − ,
j1
where I and I − are the two sets for all objects expressed, respectively, by
I i | i 1, 2, . . . , n, where object i ∈ class t , 3.5
I − i | i 1, 2, . . . , n, where object i ∈
/ class t . 3.6
Referring to Li and Chen 3, the rates of accuracy and support of Rt in Model 1 can be
specified by the following definitions.
Definition 3.2. The support rate of a rule Rt for Model 1 is specified as follows.
i If k∈K ut,i,k ≥ 1 for all i belonging to class t, then Ut,i 1; otherwise Ut,i 0, where
K indicates the hypersphere set for class t.
ii
i∈classt Ut,i
SRRt , 3.7
nt
The second model looks for an accuracy rate as high as possible while the support rate
is fixed to 1, as shown in Model 2.
m
2
subject to ai,j − ht,k,j ≤ r 2t,k , ∀i ∈ I ,
j1
Mathematical Problems in Engineering 7
m
2
ai ,j − ht,k,j > vt,i ,k − 1M r 2t,k , ∀i ∈ I − ,
j1
where I and I − are the two sets expressed by 3.5 and 3.6, respectively.
Similarly, the rates of accuracy and support of Rt in Model 2 can be considered as
follows.
Definition 3.3. The accuracy rate of a rule Rt of Model 2 is denoted as ARRt and is specified
as follows.
i If k∈K vt,i ,k 0 belongs to class t, then Vt,i 1 for all i ; otherwise, Vt,i 0, where
K represents the hypersphere set for class t.
ii
Rt − i ∈class t Vt,i
ARRt , 3.9
Rt
Definition 3.4. The support rate of a rule Rt of Model 2 is denoted as SRRt , and SRRt 1.
g
CR R1 , . . . , Rg g , 3.10
t1 USt
where USt means the number of hyperspheres and unions of hyperspheres for class t.A union
of hyperspheres indicates that the object is covered by different hyperspheres, as shown in
Figure 5. Take Figure 5 for an example, in which there are two classes. The objects of class
“” are covered by two unions of the circles i.e., S1,1 ∪ S1,2 ∪ S1,3 and S1,4 ∪ S1,5 , and the
objects of class “•” are covered by one circle i.e., S2,1 . Therefore, US1 2, US2 1, and
CRR1 , R2 2/3.
Moreover, Models 1 and 2 are separable nonlinear programs solvable to find an
optimal solution by linearizing the quadratic terms h2t,k,j . The piecewise linearization
technique is discussed as follows.
8 Mathematical Problems in Engineering
S1,5
S1,4
S1,1
S2,1
S1,2
S1,3
Proposition 3.6 referring to Beale and Forrest 31. Denote approximate function Lfx as a
piecewise linear function (i.e., linear convex combination) of fx, where bl , l 1, 2, . . . , q represents
the break points of Lfx. Lfx is expressed as follows:
q
fx ∼
L fx fbl wl , 3.11
l1
q
x wl bl , 3.12
l1
q
wl 1, 3.13
l1
where wl ≥ 0, and 3.13 is a special-ordered set of type 2 (SOS2) constraint (reference to Beale and
Forrest [31]).
Note that the SOS2 constraint is a set of variables in which at most two variables may
be nonzero. If two variables are nonzero, they must be adjacent in the set.
Notation 4. According to Proposition 3.6, let fx h2t,k,j . fx is linearized by the
Proposition 3.6 and is expressed as Lh2t,k,j .
Algorithm 3.7.
Step 1. Normalize all attributes i.e., rescale ai,j ai,j − aj /aj − aj to be 0 ≤ ai,j ≤ 1.
Step 0
Normalize (all attributes)
Step 1
t = 1 and k = 1
Step 2
Remove the objects
Solve model 1 for k ′ th
covered by St,k from
hypersphere of class t
the dataset, temporarily
Step 3
No
k =k+1 Satisfy stop
conditions?
Yes
Step 4
Are all classes t = t + 1; k = 1
processed?
Step 5
Check the unions of hypersphere
St,k in the same class t
Step 6
Calculate the number of unions
US t for all t
End
Step 3. Solve Model 1 or Model 2 to obtain the k th hypersphere of class t. Remove the
objects covered by St,k from the dataset temporarily.
Step 4. Let k k 1, and resolve Model 1 or Model 2 until all objects in class t are assigned
to the hyperspheres of same class.
Step 5. Let k 1 and t t 1, and reiterate Step 3 until all classes are processed.
Step 6. Check the independent hyperspheres and unions of hyperspheres St,k in the same
class t.
10 Mathematical Problems in Engineering
1 15 1 15
0.75 0.75
2 2
5 5
7 8 7 S2,1 8 S1,2
a2 0.5 a2 0.5
9 9 6
6
4 4
10 S1,1 10
0.25 0.25
1 3 S3,1 13 S2,2
1 3 14 13 14 11
11 12
0 12 0
0 0.25 0.5 0.75 1 0 0.25 0.5 0.75 1
a1 a1
a Normalized data for Example 1 b Classified by the hypersphere method for Example 1
Step 7. Calculate and record the number of independent hyperspheres and unions of
hyperspheres in USt , and iterate t until all classes are done.
According to this algorithm, we can obtain the optimal rules to classify objects most
efficiently. The process of the algorithm is depicted in Figure 6.
Step 1. Normalize all attributes i.e., ai,1 ai,1 − 6/33.5 − 6 and ai,2 ai,2 − 3.5/30 − 3.5.
Table 2: Centroid points for the Iris data set by the proposed method.
Rule
Union of spheres St,k ht,k,1 ht,k,2 ht,k,3 ht,k,4 rt,k
number
R1 S1,1 S1,1 0.366 0.807 0.000 0.000 0.6205
S2,1 0.557 0.320 0.205 0.460 0.2540
R2 S2,1 ∪ S2,2 ∪ S2,3 S2,2 0.575 0.581 0.626 0.515 0.0612
S2,3 0.423 0.261 0.352 0.490 0.1388
S3,1 0.248 0.000 2.226 2.151 4.8087
R3 S3,1 ∪ S3,2
S3,2 0.329 0.187 0.650 0.613 0.0330
Table 3: Comparing results for the Iris flower data set R1 ,R2 ,R3 .
Hyperplane
Items Proposed method Decision tree
support vector
ARR1 ,R2 ,R3 1,1,1 1,0.98,0.98 1,0.98,0.96
SRR1 ,R2 ,R3 1,0.98,0.98 1,0.98,0.98 1,0.96,0.98
CR 1 0.5 0.1875
m
ai,j − 2ai,j ht,k,j
2
subject to L h2t,k,j ≤ 1 − ut,i,k M 2
rt,k , ∀i ∈ I ,
j1
3.14
m
ai ,j − 2ai ,j ht,k,j ∀i ∈ I − ,
2
L h2t,k,j 2
> rt,k ,
j1
Step 4. k k 1: the optimal solution of the ht,k,1 , ht,k,2 , rt,k 0.736,0.5,0.0138 for S1,2 , where
S1,2 covers objects 5-6. Class 1 is then done.
Step 5. As t t 1, k 1, and Steps 3 and 4 are iterated, we then, respectively, have optimal
solutions for St,k as follows. The results are shown in Figure 7b.
i h2,1,1 , h2,1,2 , r2,1 0.514, 0.469, 0.0127, where S2,1 covers objects 7–9.
ii h2,2,1 , h2,2,2 , r2,2 0.929, 0.210, 0.0251, where S2,2 covers objects 10-11.
iii h3,1,1 , h3,1,2 , r3,1 0.583, 0.188, 0.0436, where S3,1 covers objects 12–14.
Step 6. Check and calculate the unions of hypersphere St,k for all k in class t i.e., Initial t 1.
Step 7. As t t 1, mark the number of unions of class t into USt and iterate Step 6 until t g.
12 Mathematical Problems in Engineering
4. Numerical Examples
This study shows how the experimental results evaluate the performance, including accuracy,
support, and compactness rates, and compares the proposed model with different methods
using CPLEX 32. All tests were run on a PC equipped with an Intel Pentium D 2.8 GHz CPU
and 2 GMB RAM. Three datasets were tested in our experiments as follows:
i Iris Flower dataset introduced by Sir Ronald Aylmer Fisher 1936,
ii European barn swallow Hirundo rustica dataset obtained by trapping individual
swallows in Stirlingshire, Scotland, between May and July 1997 1, 3,
iii the highly selective vagotomy HSV patient dataset of F. Raszeja Memorial
Hospital in Poland 3, 33, 34.
i “if ai,1 − 0.3662 ai,2 − 0.8072 ai,3 − 02 ai,4 − 02 ≤ 0.6205, then object xi
belongs to class 1.”
Rule R2 in Table 2 contains a union of three hyperspheres i.e., S2,1 ∪ S2,2 ∪ S2,3 which
implies that
i “if ai,1 − 0.5572 ai,2 − 0.322 ai,3 − 0.2052 ai,4 − 0.462 ≤ 0.254, then object xi
belongs to class 2,” or
ii “if ai,1 − 0.5752 ai,2 − 0.5812 ai,3 − 0.6262 ai,4 − 0.5152 ≤ 0.0612, then object
xi belongs to class 2,” or
iii “if ai,1 − 0.4232 ai,2 − 0.2612 ai,3 − 0.3522 ai,4 − 0.492 ≤ 0.1388, then object
xi belongs to class 2.”
Rule R3 in Table 2 contains a union of two hyperspheres i.e., S3,1 ∪ S3,2 , which implies
that
i “if ai,1 − 0.2482 ai,2 − 02 ai,3 − 2.2262 ai,4 − 2.1512 ≤ 4.8087, then object xi
belongs to class 3,” or
ii “if ai,1 − 0.3292 ai,2 − 0.1872 ai,3 − 0.652 ai,4 − 0.6132 ≤ 0.033, then object
xi belongs to class 3.”
Comparing the proposed method with both decision tree 3 and hyperplane methods
35 in deducing the classification rules for the Iris Flower dataset, Table 3 lists the
experimental result.
Mathematical Problems in Engineering 13
Table 4: Centroid points for the Swallow data set by the proposed method.
Rule Union of
St,k ht,k,1 ht,k,2 ht,k,3 ht,k,4 ht,k,5 ht,k,6 ht,k,7 ht,k,8 rt,k
number spheres
S1,1 0.607 0.110 0.806 0.000 0.000 0.406 1.077 1.163 1.780
S1,1 ∪S1,2 ∪ S1,2 0.483 0.000 0.236 0.328 0.000 0.793 0.280 0.931 0.948
R1
S1,3 ∪ S1,4
S1,3 0.588 0.000 1.179 0.000 0.000 0.653 1.037 0.982 1.879
S1,4 0.414 0.000 0.658 0.000 0.227 0.085 0.563 1.224 1.197
S2,1 0.358 0.667 0.371 1.403 0.360 1.530 0.634 0.114 2.825
S2,1 ∪ S2,2 0.198 1.102 0.697 0.000 0.000 1.734 0.110 0.595 2.632
R2 S2,2 ∪S2,3 ∪ S2,3 0.532 0.422 0.323 1.676 1.009 0.000 0.180 0.020 2.458
S2,4 ∪ S2,5
S2,4 0.528 0.000 0.430 0.876 0.579 0.461 0.435 0.040 0.694
S2,5 0.659 1.408 0.516 0.000 0.382 0.000 0.173 0.064 1.570
Table 5: Comparing results for the Swallow data set R1 ,R2 .
Hyperplane support
Items Proposed method Decision tree
vector
ARR1 ,R2 1, 1 0.97,1 0.97,1
SRR1 ,R2 1, 0.97 0.97,1 0.97,1
CR 1 0.3 0.1
The accuracy rates of R1 , R2 , R3 in the proposed method are 1,1,1, as Model 1 has
been solved. This finding indicates that none of objects in class 2 or class 3 are covered by S1,1 ,
none of objects in classes 1 or 3 are covered by S2,1 ∪S2,2 ∪S2,3 , and none of the objects in classes
1 or 2 are covered by S3,1 ∪ S3,2 . The support rate of R1 ,R2 ,R3 in the proposed method is
1,0.98,0.98, indicating that all objects in class 1 are covered by S1,1 , 98% of the objects in class
2 are covered by S2,1 , S2,2 , and S2,3 , and 98% of the objects in class 3 are covered by S3,1 and
S3,2 . The compactness rate of rules R1 , R2 , and R3 is computed as CRR1 , R2 , R3 3/3 1.
Finally, we determine the following.
i Although all three methods perform very well in the rates of accuracy and support,
the proposed method has the best performance for the accuracy of classes 2 and 3
i.e., R2 and R3 .
ii The proposed method has the best compactness rate.
The result of the hyperplane method, referred to in Chang and Lin 35, is also listed
in Table 5, whereARR1 , R2 0.97, 1, SRR1 , R2 0.97, 1, and CR 0.1.
We compared the three methods in Table 5 to show that the proposed method can
induce rules with better or equivalent values of AR and SR. In fact, the proposed method
also has the best compactness rate.
Rule Union of
Sk,l ht,k,1 ht,k,2 ht,k,3 ht,k,4 ht,k,5 ht,k,6 ht,k,7 ht,k,8 ht,k,9 ht,k,10 ht,k,11 rt,k
numbers spheres
S1,1 0.504 0.528 0.366 0.288 0.850 0.848 0.351 −0.282 0.405 0.234 0.313 1.576
S1,2 0.469 0.395 0.312 0.133 −0.274 −0.309 0.134 0.590 0.813 1.000 −0.847 3.044
S1,3 0.287 0.867 0.458 0.189 0.977 0.267 0.017 −0.846 0.462 −1.000 1.000 4.249
S1,1 ∪
R1 S1,2 . . . ∪ S1,4 0.586 −0.400 0.605 0.467 −0.167 −0.631 0.268 1.000 0.526 1.000 0.120 3.365
S1,8 S1,5 0.296 −0.511 −1.000 0.536 −0.263 0.552 0.954 0.635 −0.117 −1.000 0.600 4.608
S1,6 0.775 0.678 −0.092 0.340 0.366 0.455 0.451 1.000 0.422 0.295 0.470 1.560
Mathematical Problems in Engineering
S1,7 0.525 0.296 0.194 0.124 −0.945 −0.467 0.127 1.000 0.097 −0.897 0.654 4.113
S1,8 0.000 0.140 0.109 0.250 0.193 0.168 0.063 0.142 0.181 0.119 0.082 0.089
S2,1 S2,1 0.513 0.056 0.143 0.078 0.533 0.439 0.180 0.167 −0.284 −0.464 0.749 1.450
R2 S2,2 S2,2 0.533 0.381 0.139 0.568 −0.373 −0.005 0.080 0.831 0.293 0.215 0.369 1.293
S2,3 S2,3 0.000 −0.450 0.709 0.553 0.896 1.000 0.157 −0.447 0.068 −0.580 1.000 3.579
S2,4 S2,4 0.000 0.862 0.543 −0.066 0.389 0.451 0.106 0.394 −0.004 0.042 −0.014 0.409
S3,1 0.624 0.507 −1.000 0.147 0.840 1.000 0.827 −0.585 0.199 0.831 −0.380 4.277
S3,1 ∪ S3,2
R3 S3,2 0.534 0.483 −0.088 0.365 −0.364 0.270 0.437 0.785 0.786 0.682 −1.000 3.003
S3,3 S3,3 0.000 0.210 0.630 0.750 −1.000 0.244 0.388 1.000 0.474 −0.604 0.475 3.162
S4,1 0.551 0.374 0.637 0.256 0.865 0.944 0.315 −0.831 −0.485 −0.676 0.979 4.330
S4,1 ∪
S4,2 ∪ S4,3 S4,2 0.717 0.254 0.548 −0.718 0.118 0.730 −0.547 0.498 −0.580 −0.464 0.821 3.287
R4 S4,3 0.527 1.000 0.533 0.089 −0.152 −1.000 0.366 0.046 0.547 0.947 −0.436 2.943
S4,4 S4,4 0.489 0.625 −0.209 0.522 0.644 0.582 0.615 1.000 0.306 −0.401 0.931 2.059
S4,5 S4,5 −0.011 0.491 −0.409 0.155 −0.766 −0.227 −0.107 1.000 0.303 −0.680 −0.185 2.843
15
16 Mathematical Problems in Engineering
Table 7: Comparing results for the HSV data set R1 ,R2 ,R3 ,R4 .
Hyperplane support
Items Proposed method Decision tree
vector
ARR1 ,R2 ,R3 ,R4 1,1,1,1 0.93,0.81,0.7,0.71 0.9,1,1,0.9
SRR1 ,R2 ,R3 ,R4 0.99,1,1,1 0.93,0.72,0.78,0.71 0.9,0.72,0.67,0.69
CR 0.4 0.17 0.09
Acknowledgments
The authors wish to thank the editor and the anonymous referees for providing insightful
comments and suggestions, which have helped them improving the quality of the paper.
This work was supported by the National Science Council of Taiwan under Grants NSC 100-
2811-E-009-040-, NSC 99-2221-E-030-005-, and NSC 100-2221-E-030-009-.
References
1 M. J. Beynon and K. L. Buchanan, “An illustration of variable precision rough set theory: the gender
classification of the European barn swallow Hirundo rustica,” Bulletin of Mathematical Biology, vol.
65, no. 5, pp. 835–858, 2003.
2 H. L. Li and C. J. Fu, “A linear programming approach for identifying a consensus sequence on DNA
sequences,” Bioinformatics, vol. 21, no. 9, pp. 1838–1845, 2005.
3 H. L. Li and M. H. Chen, “Induction of multiple criteria optimal classification rules for biological and
medical data,” Computers in Biology and Medicine, vol. 38, no. 1, pp. 42–52, 2008.
4 C. W. Chu, G. S. Liang, and C. T. Liao, “Controlling inventory by combining ABC analysis and fuzzy
classification,” Computers and Industrial Engineering, vol. 55, no. 4, pp. 841–851, 2008.
5 J.-X. Chen, “Peer-estimation for multiple criteria ABC inventory classification,” Computers &
Operations Research, vol. 38, no. 12, pp. 1784–1791, 2011.
6 E. B. Hunt, J. Marin, and P. J. Stone, Experiments in Induction, Academic Press, New York, NY, USA,
1966.
7 L. Breiman, J. H. Friedman, R. A. Olshen, and C. J. Stone, Classification and Regression Trees, Wadsworth
Statistics/Probability Series, Wadsworth Advanced Books and Software, Belmont, Calif, USA, 1984.
8 J. R. Quinlan, “Induction of decision trees,” Machine Learning, vol. 1, no. 1, pp. 81–106, 1986.
9 J. R. Quinlan, “Simplifying decision trees,” International Journal of Man-Machine Studies, vol. 27, no. 3,
pp. 221–234, 1987.
10 J. R. Quinlan, C4.5: Programs for Machine Learning, Morgan Kaufman, San Mateo, Calif, USA, 1993.
11 H. Kim and W. Y. Loh, “Classification Trees with Unbiased Multiway Splits,” Journal of the American
Statistical Association, vol. 96, no. 454, pp. 589–604, 2001.
12 V. N. Vapnik, The Nature of Statistical Learning Theory, Springer, New York, NY, USA, 1995.
13 R. M. Rifkin, Everything old is new again: a fresh look at historical approaches in machine learning, Ph.D.
thesis, Massachusetts Institute of Technology, Ann Arbor, MI, USA, 2002.
14 S. Katagiri and S. Abe, “Incremental training of support vector machines using hyperspheres,” Pattern
Recognition Letters, vol. 27, no. 13, pp. 1495–1507, 2006.
15 D. E. Rumelhart, G. E. Hinton, and R. J. Williams, “Learning representations by back-propagating
errors,” Nature, vol. 323, no. 6088, pp. 533–536, 1986.
Mathematical Problems in Engineering 17
16 C. H. Chu and D. Widjaja, “Neural network system for forecasting method selection,” Decision Support
Systems, vol. 12, no. 1, pp. 13–24, 1994.
17 C. Giraud-Carrier, R. Vilalta, and P. Brazdil, “Guest editorial: introduction to the special issue on
meta-learning,” Machine Learning, vol. 54, no. 3, pp. 187–193, 2004.
18 D. Pomorski and P. B. Perche, “Inductive learning of decision trees: application to fault isolation of an
induction motor,” Engineering Applications of Artificial Intelligence, vol. 14, no. 2, pp. 155–166, 2001.
19 H. Zhang, C. N. Yu, B. Singer, and M. Xiong, “Recursive partitioning for tumor classification with
gene expression microarray data,” Proceedings of the National Academy of Sciences of the United States of
America, vol. 98, no. 12, pp. 6730–6735, 2001.
20 S. Aja-Fernández, R. De Luis-Garcı́a, M. Á. Martı́n-Fernández, and C. Alberola-López, “A
computational TW3 classifier for skeletal maturity assessment. A Computing with Words approach,”
Journal of Biomedical Informatics, vol. 37, no. 2, pp. 99–107, 2004.
21 P. Geurts, M. Fillet, D. de Seny et al., “Proteomic mass spectra classification using decision tree based
ensemble methods,” Bioinformatics, vol. 21, no. 14, pp. 3138–3145, 2005.
22 C. Olaru and L. Wehenkel, “A complete fuzzy decision tree technique,” Fuzzy Sets and Systems, vol.
138, no. 2, pp. 221–254, 2003.
23 M. Pal and P. M. Mather, “An assessment of the effectiveness of decision tree methods for land cover
classification,” Remote Sensing of Environment, vol. 86, no. 4, pp. 554–565, 2003.
24 Z. Chen, J. Li, and L. Wei, “A multiple kernel support vector machine scheme for feature selection
and rule extraction from gene expression data of cancer tissue,” Artificial Intelligence in Medicine, vol.
41, no. 2, pp. 161–175, 2007.
25 H. Nunez, C. Angulo, and A. Catala, “Rule-extraction from support vector machines,” in Proceedings
of the European Symposium on Artificial Neural Networks, pp. 107–112, 2002.
26 Y. M. Lin, X. Wang, W. W. Y. Ng, Q. Chang, D. S. Yeung, and X. L. Wang, “Sphere classification for
ambiguous data,” in Proceedings of the International Conference on Machine Learning and Cybernetics, pp.
2571–2574, August 2006.
27 J. Wang, P. Neskovic, and L. N. Cooper, “Bayes classification based on minimum bounding spheres,”
Neurocomputing, vol. 70, no. 4–6, pp. 801–808, 2007.
28 L. Gu and H. Z. Wu, “A kernel-based fuzzy greedy multiple hyperspheres covering algorithm for
pattern classification,” Neurocomputing, vol. 72, no. 1–3, pp. 313–320, 2008.
29 M. Hifi and R. M’Hallah, “A literature review on circle and sphere packing problems: models and
methodologies,” Advances in Operations Research, vol. 2009, Article ID 150624, 22 pages, 2009.
30 M. Zhu, Y. Wang, S. Chen, and X. Liu, “Sphere-structured support vector machines for multi-class
pattern recognition,” in Proceedings of the 9th International Conference (RSFDGrC ’03), vol. 2639 of
Lecture Notes in Computer Science, pp. 589–593, May 2003.
31 E. M. L. Beale and J. J. H. Forrest, “Global optimization using special ordered sets,” Mathematical
Programming, vol. 10, no. 1, pp. 52–69, 1976.
32 IBM/ILOG, “Cplex 12.0 reference manual,” 2010, http://www.ilog.com/products/cplex/.
33 D. C. Dunn, W. E. G. Thomas, and J. O. Hunter, “An evaluation of highly selective vagotomy in the
treatment of chronic duodenal ulcer,” Surgery Gynecology and Obstetrics, vol. 150, no. 6, pp. 845–849,
1980.
34 K. Slowinski, “Rough classification of HSV patients,” in Intelligent Decision Support-Handbook of
Applications and Advances of the Rough Sets Theory, R. Slowinski, Ed., pp. 77–94, Kluwer Academic
Publishers, Dordrecht, The Netherlands, 1992.
35 C. C. Chang and C. J. Lin, “LIBSVM: a library for support vector machines,” 2010, http://www
.csie.ntu.edu.tw/∼cjlin/libsvm/index.html.
36 C. Than, R. Sugino, H. Innan, and L. Nakhleh, “Efficient inference of bacterial strain trees from
genome-scale multilocus data,” Bioinformatics, vol. 24, no. 13, pp. i123–i131, 2008.
37 M. Rockville, “Large scale computing and storage requirements for biological and environmental
research,” Tech. Rep., Ernest Orlando Lawrence Berkeley National Laboratory, Berkeley, Calif, USA,
2009.
38 X. Xie, X. Fang, S. Hu, and D. Wu, “Evolution of supercomputers,” Frontiers of Computer Science in
China, vol. 4, no. 4, pp. 428–436, 2010.
Hindawi Publishing Corporation
Mathematical Problems in Engineering
Volume 2012, Article ID 365697, 18 pages
doi:10.1155/2012/365697
Research Article
A Modified PSO Algorithm for Minimizing
the Total Costs of Resources in MRCPSP
Copyright q 2012 Mohammad Khalilzadeh et al. This is an open access article distributed under
the Creative Commons Attribution License, which permits unrestricted use, distribution, and
reproduction in any medium, provided the original work is properly cited.
1. Introduction
The resource-constrained project scheduling problem RCPSP is the scheduling of project
activities subject to precedence relations as well as renewable resource constraints with the
objective of minimizing the makespan of the project. Each nonpreemptive activity in RCPSP
can be done in a single mode. For more information on RCPSP and solution methods, we
refer to Demeulemeester and Herroelen 1. In the multimode RCPSP MRCPSP, a set of
2 Mathematical Problems in Engineering
allowable modes can be defined for each activity which is characterized by a constant
duration and associated resource requirements. In this paper we consider MRCPSP with
the objective of minimizing total costs of all resources. Two types of resources, renewable
and nonrenewable, are considered. Nonrenewable resource cost of an activity is a function
of its resource requirements, determined by its modes. The limited renewable resources are
rented and each renewable resource is available in a predetermined sequential time period
specified by its ready time and due date and is not available before the ready time. However,
each renewable resource can be used after its due date with tardiness penalty cost. As the
cost of renting for each renewable resource is fixed, there is no need to incorporate it into
the objective function and only tardiness penalty cost is considered for each renewable re-
source. The MRCPSP under minimization of total costs of resources RCPSP-TWRTPC is an
applicable problem and a modified version of the MRCPSP in which all assumptions and
constraints of the MRCPSP are held, but the objective function is different. We assume that
there are a few renewable resources such as very expert human resources with high skill
levels, particular types of cranes, and tunnel boring machines that should be leased from
other companies providing these types of resources. Since these limited renewable resources
are employed in other projects, there is a dictated ready date as well as a due date for each
of them such that no resource can be accessible before its ready date, but these resources are
allowed to be used after their due dates by paying penalty cost, depending on the resource
type. Also, we suppose that there are a few nonrenewable resources like budget, materials,
energy, or other resources which are consumed during the project.
Ranjbar et al. 2 studied this problem with single mode for each activity and availab-
ility of one unit for each type of renewable resource, without considering nonrenewable re-
sources. They called this problem resource-constrained project scheduling problem, minimization
of total weighted resource tardiness penalty cost RCPSP-TWRTPC, which is an extended form of
resource-constrained project scheduling problem RCPSP. They developed a metaheuristic-based
GRASP algorithm together with a branch and bound procedure to solve the problem.
The problem we have studied here is a generalization of the problem introduced by
Ranjbar et al. 2 with more realistic viewpoint of resource costs by considering both renew-
able and nonrenewable resources cost. We call this problem multimode resource-constrained
project scheduling problem, minimization of total weighted resource tardiness penalty cost MRCPSP-
TWRTPC.
Several exact and heuristic methods have been presented for MRCPSP. For instance,
we can point to branch and cut method introduced by Heilmann 3 and branch and bound
method developed by Zhu et al. 4 as two of the most powerful exact methods. Zhang et al.
5 presented classical particle swarm optimization PSO methods for multimode resource-
constrained project scheduling problems to minimize the duration of construction projects.
Lova et al. 6 suggested a heuristic algorithm based on priority rules and thereafter a hy-
brid genetic algorithm 7 to solve MRCPSP. Jarboui et al. 8 applied a combinatorial pa-
rticle swarm optimization for solving multimode resource-constrained project scheduling
problem. Ranjbar et al. 9 developed scatter search algorithm to tackle this problem, using
path relinking methodology as its local search method. Van Peteghem and Vanhoucke 10
proposed genetic algorithm to solve preemptive and nonpreemptive multimode resource-
constrained project scheduling problems. Kazemi and Tavakkoli-Moghaddam 11 devel-
oped a multimode particle swarm optimization which combines with genetic operator to
solve a biobjective multimode resource-constrained project scheduling problem with positive
and negative cash flows. The difference between this study and the previous papers is that
they all considered the minimization of project makespan as objective function, but, in our
Mathematical Problems in Engineering 3
problem, the objective function is minimization of total costs of renewable and nonrenewable
resources.
MRCPSP-TWRTPC is a generalization of the RCPSP problem, and, considering the
NP-hardness of RCPSP 12, 13, the MRCPSP-TWRTPC problem is NP-hard as well, and
hence, metaheuristic method is the practical approach. In the remainder of this paper, we
introduce a metaheuristic-based PSO algorithm for solving this problem. PSO, in its present
form, dates back to 1990s; however, in this short period, PSO has shown good performance
in a variety of application domains, particularly in the constrained optimization problems.
Many researchers studied PSO widely and proposed several modifications. In this paper, we
use a modified PSO algorithm developed and used by Tchomté and Gourgand for solving
RCPSP efficiently 14.
The rest of this paper is organized as follows. In the next section, MRCPSP-TWRTPC
is described in detail and is formulated in a mathematical model. In Section 3, a description
of the PSO algorithm and its modifications is presented, and in Section 4 an algorithm based
on modified PSO introduced by Tchomté and Gourgand 14 is explained. Section 5 is the
experimental analysis. Finally, Section 6 concludes the work.
2. Problem Description
In MRCPSP-TWRTPC, a project is to be scheduled in order to minimize its total costs.
Resources available for completing project activities can be classified as either renewable
or nonrenewable. Activity j may have a number of execution modes Mj . Each activity
mode specifies the activity duration and the activity requirements for the certain amount of
renewable and nonrenewable resources. Each type of limited renewable resource is rented
for a fixed time interval, starting from its ready time and ending with its due date, and is
not available before its ready time but can be used after its due date with tardiness penalty
cost. Nonrenewable resources are not limited. All activities are ready at the beginning of
the project, and no preemption is permitted. If an activity is started under a specific mode,
the activity mode cannot be changed. Activity j executed in mode m has duration djm and
requires rjmk units of renewable resource k and njmk units of nonrenewable resource k. The
project network is depicted by an activity on node AON representation with finish-to-start
precedence relations and zero time lag. Dummy activities 1 and n correspond to start and
completion of the project. The list of activities is topologically numbered; that is, each
predecessor of every activity has a smaller number than the number of activity itself. Also,
we define the earliest and latest start time of activity j by ESTj and LSTj , respectively. ESTj s
and LSTj s are computed by CPM forward and backward passes using the mode with shortest
duration for each activity and assigning LSTn LFTn T , where T is an upper bound for
project makespan determined by any valid method, such as the sum of the longest duration
of entire project activities plus the ready times of renewable resources. Consequently, each
activity j can only be performed in time period ESTj , LSTj .
We define problem parameters as follows:
⎧
⎨1, if activity j is started under mode m in period τ,
xjmτ
⎩0, otherwise,
⎧ 2.1
⎨1, if renewable resource k is used in period τ,
ykτ
⎩0, otherwise.
⎛ ⎞
NR
n
Mj
LSTj
R
Min ck ⎝ njmk xjmτ ⎠ pk · lk 2.2
k1 j1 m1 τESTj k1
Mj LSTj
S.t. xjmτ 1, j 1, 2, . . . , n, 2.3
m1 τESTj
Mathematical Problems in Engineering 5
i
Mi LST
Mj LSTj
τ dim ximt ≤ τxjmt , j 1, 2, . . . , n, i ∈ Pj , 2.4
m1 τESTi m1 τESTj
n
Mj
τ
rjmk xjmz ≤ Rk · ykτ , k 1, 2, . . . , R, τ 1, 2, . . . , T, 2.5
j1 m1 zτ−djm 1
r
k −1
ykτ 0, k 1, 2, . . . , R, 2.6
τ1
lk ≥ 0, k 1, 2, . . . , R. 2.10
In the above model, objective function 2.2 is project cost minimization in which the
first and second terms are total costs of using nonrenewable resources and total penalty costs
of renewable resources tardiness, respectively. Constraint set 2.3 ensures that each activity j
is started under one of its modes within its specified start time periods, that is, ESTj , LSTj .
Constraint set 2.4 forces precedence relationship between activities. Constrain 2.5 limit
renewable resource usage. According to constraint 2.6, renewable resources cannot be used
before their ready times and their tardiness periods are determined by constraint 2.7.
Finally, constraint sets 2.8, 2.9, and 2.10 are nonfunctional ones.
This vector is a function of three components: previous velocity of the particle, the best
experience of the particle, also, the entire swarm’s best experiences up to the current iteration
which are called inertia, cognition part, and social part, respectively 16.
Updating process continues until the termination criterion is met which usually is the
maximal number of generations, processing time, or the best particle position of the whole
swarm that cannot improve further after a predefined number of generations.
In 3.2, r2 and r3 are real random numbers with uniform distribution which are
usually selected from the interval 0, 1. c2 and c3 are known constants as learning factors,
showing the significance of local and global best experiences, respectively. Also, c1 is defined
as a positive inertia weight which was first introduced by Shi and Eberhart 17. This
parameter can be specific for each particle 18. Liu et al. 19 and Shi and Eberhart 20
introduced time-decreasing inertia weight.
The PSO parameters analyses have been the subject of several researches. For instance,
Tchomté and Gourgand 14 determined some conditions for parameters to ensure that each
particle converges to some equilibrium point after enough number of iterations.
Although PSO has been originally designed for solving continuous problems, it can
be used for solving discrete problems as well. Different techniques have been designed to
use this algorithm for combinatorial optimization problems such as the ones introduced by
Jarboui et al. 8, Clerc 21, and Kennedy and Eberhart 22.
In this paper, we use the modified PSO approach which was introduced by Tchomté
and Gourgand 14 as an extension of PSO that integrates a new displacement rule of the
particles. The computational results of their algorithm showed that their PSO algorithm
outperformed all state-of-the-art algorithms in solving RCPSP, and this is the reason for
selecting their approach for our problem. We describe this modified PSO method in the
following.
A metaheuristic algorithm should be able to explore search space effectively and
efficiently. PSO algorithm should be intelligent enough to both intensively explore regions of
the search space with high-quality solutions and to diversely move to unexplored regions of
the search space. These two techniques that were introduced by Glover and Laguna 23 are
known as intensification and diversification methods, respectively. Tchomté and Gourgand
14 analyzed particle trajectories and modified particle position updating rules. The idea
originated from the PSO applications in which particles basically move from their current
positions toward the best local and global positions Pi,t , Gt , but the particles do not get close
enough to Pi,t and Gt . As a result, diversification is performed well, but intensification is
not. Consequently, Tchomté and Gourgand 14 proposed a new particle displacement rule
to improve the intensification process by letting each particle visit two positions Si,t and Ti,t
before moving from current position, Xi,t , to the next position Xi,t 1 . First, the inertia has
influences on the position by making the particle move from Xi,t to Si,t . Then the cognition
part moves the particle to Ti,t and finally under social part affect the particle to reach its
new position, Xi,t 1 , at the next iteration. Adapted from Tchomté and Gourgand 14, particle
displacement in the classical PSO and this modified PSO has been shown in Figures 1a and
1b, respectively.
Mathematical Problems in Engineering 7
Pit
Vit Pit
Vit
Tit
Xi,t+1
Sit
Xi,t+1
Xit Xit
Gt Gt
a b
Figure 1: Particle displacement in the swarm: a classical PSO, b modified PSO.
Tchomté and Gourgand 14 showed that the necessary conditions for coefficients so
that the particles converge to the equilibriums are satisfying 3.6 plus 3.7 or 3.6 plus 3.8:
1 Do Preprocessing
2 Generate initial particle swarm
3 While termination criterion is met do
4 While all particles have been evaluated do
5 Determine activities priorities
6 Schedule activities based on their modes and priorities using the parallel
schedule generation and delay local search
7 While schedule is improved do
8 Improve schedule by Mode Assignment Modification—Part I
9 Improve schedule by Local Left Shift
10 End while
11 Improve schedule by Mode Assignment Modification—Part II
12 Compute corresponding cost of the generated schedule
13 End while
14 Update the local and global best solutions if necessary
15 Update position and velocity of each particle according to 3.3 and 3.4, respectively
16 End while
17 Report the global best solution
Xi,t xi1
t t
, xi2 t
, . . . , xin in which xijt , j 1, 2, 3, . . . , n is the mode assignment to the jth activity
at iteration t and is an integer in the interval 1, Mj . A feasible schedule of the project is
constructed from each Xi,t . For this purpose, first the activities are prioritized, see Section 4.3.
Then, using single-pass parallel schedule generation scheme, the activities are scheduled, see
Section 4.4. Certain local search and improvement procedures are applied on this solution to
reach a better schedule, see Sections 4.5, 4.6, 4.7, and 4.8.
Each particle’s fitness value is determined by calculating total cost of the final sched-
ule. Then, if necessary, local and/or global best positions are updated, see Section 4.9. If ter-
mination criterion is not met, particle positions and velocity vectors are updated by 3.3 and
3.4, respectively, for the next iteration, see Section 4.10. Different parts of this algorithm are
described in more details as follows.
4.1. Preprocessing
Sprecher et al. 24 introduced several preprocessing rules in order to reduce feasible space of
MRCPSP. Later, these rules have been used in other articles such as Lova et al. 6, Peteghem
and Van Vanhouck 10, and Hartmann and Briskorn 25. Considering the similarities
between MRCPSP-TWRTPC and MRCPSP, we apply two of these rules to our proposed
problem. One is the nonexecutable mode elimination rule for an activity. For a nonexecutable
mode, the amount of the resource needed for executing the activity is more than the resource
availability. Another method is inefficient mode elimination method. A given mode is inefficient
for an activity if there is another mode for which the activity duration is less, and that activity
can be accomplished with less total amount of both renewable and nonrenewable resources.
Therefore, activities modes are analyzed one by one and nonexecutable and inefficient modes
are deleted.
3 For each j ∈ FFA, as we go through forward pass, consider the least nonrenewable
resource cost mode of activity j. If it is not its current mode, reschedule activity
j in this mode using its free float and considering resource constraints, without
increasing its finish time. If this rescheduling is not possible, check the next least
nonrenewable resource cost mode of activity j.
1 Let NSA {j | activity j is the direct predecessor of dummy activity n}.
2 For each j ∈ NSA, consider some modes for activity j, in which nonrenewable re-
sources requirement cost is less and this cost saving is more than the probable in-
crease in the penalty cost of renewable resources. Compute Δc for each of them.
Considering renewable resource constraints, if the least Δc is negative, its cor-
responding mode replaces the current mode of activity j.
5. Experimental Analysis
In this section, we present experimental analysis of the algorithm. All programs have been
coded and executed on C#.NET 2008 platform on a PC with Core 2 Duo 2.53 GHz CPU and
3 GB RAM.
Parameter Levels
q0 0.2–0.4–0.6
P 10–30–60
c1 0.2–0.45–0.7
Satisfying relations 3.6 and 3.7 Satisfying relations 3.6 and 3.8
c2 0.5–1–1.5 2.5–3–3.5
c3 0.5–1–1.5 2.5–3–3.5
Each test instance was run for all 35 × 2 486 permutations of parameter values, a total
of 48,600 problems. The tuned values of the parameters are q0 0.4, P 60, c1 0.2, c2 0.5,
c3 1.
Table 5 shows the effect of deleting mode assignment improvement method-part II. We
can see that in most cases the performance of the algorithm deteriorates, and in all cases the
average CPU time increases remarkably. In the following, we explain the reason for increasing
the average CPU time.
6. Conclusions
In this paper, we introduced MRCPSP-TWRTPC problem as a resource-oriented cost min-
imization project scheduling problem considering both renewable and nonrenewable re-
source costs. We formulated and mathematically modeled this problem as mixed integer pro-
gramming model and discussed its NP-hardness. Subsequently, we developed a metaheuris-
tic algorithm to tackle the proposed project scheduling problem. We briefly reviewed the ap-
plications of the PSO algorithm for solving combinatorial and constrained optimization prob-
lems. Thereafter, we applied a modified PSO algorithm including modified updating rules
for particles velocity and position. In order to generate feasible schedules, we used the
PSO algorithm for activity mode assignment and developed a novel heuristic technique to
prioritize activities for parallel scheduling scheme. Two improvement heuristics, delay local
search and local left shift, in line with two mode assignment modification methods, were
implemented to improve the solutions. The computational results revealed proper algorithm
robustness in solving different instances especially with high number of iterations. Also, the
validity analysis showed small deviations from the optimal solutions for the test instances in
reasonable solving time. Finally, we assessed two improvement methods used in our algo-
rithm to demonstrate their good performance.
References
1 E. Demeulemeester and W. S. Herroelen, Project Scheduling, A Research Handbook, Kluwer Academic,
Dordrecht, The Netherlands, 2001.
2 M. Ranjbar, M. Khalilzadeh, F. Kianfar, and K. Etminani, “An optimal procedure for minimizing total
weighted resource tardiness penalty costs in the resource-constrained project scheduling problem,”
Computers and Industrial Engineering, vol. 62, no. 1, pp. 264–270, 2012.
3 R. Heilmann, “A branch-and-bound procedure for the multi-mode resource-constrained project
scheduling problem with minimum and maximum time lags,” European Journal of Operational Research,
vol. 144, no. 2, pp. 348–365, 2003.
4 G. Zhu, J. F. Bard, and G. Yu, “A branch-and-cut procedure for the multimode resource-constrained
project-scheduling problem,” INFORMS Journal on Computing, vol. 18, no. 3, pp. 377–390, 2006.
5 H. Zhang, C. M. Tam, and H. Li, “Multimode project scheduling based on particle swarm optimiza-
tion,” Computer-Aided Civil and Infrastructure Engineering, vol. 21, no. 2, pp. 93–103, 2006.
6 A. Lova, P. Tormos, and F. Barber, “Multi-mode resource constrained project scheduling: scheduling
schemes, priority rules and mode selection rules,” Inteligencia Artificial, vol. 10, no. 30, pp. 69–86, 2006.
7 A. Lova, P. Tormos, M. Cervantes, and F. Barber, “An efficient hybrid genetic algorithm for scheduling
projects with resource constraints and multiple execution modes,” International Journal of Production
Economics, vol. 117, no. 2, pp. 302–316, 2009.
8 B. Jarboui, N. Damak, P. Siarry, and A. Rebai, “A combinatorial particle swarm optimization for
solving multi-mode resource-constrained project scheduling problems,” Applied Mathematics and
Computation, vol. 195, no. 1, pp. 299–308, 2008.
9 M. Ranjbar, B. De Reyck, and F. Kianfar, “A hybrid scatter search for the discrete time/resource trade-
off problem in project scheduling,” European Journal of Operational Research, vol. 193, no. 1, pp. 35–48,
2009.
10 V. Van Peteghem and M. Vanhoucke, “A genetic algorithm for the preemptive and non-preemptive
multi-mode resource-constrained project scheduling problem,” European Journal of Operational
Research, vol. 201, no. 2, pp. 409–418, 2010.
18 Mathematical Problems in Engineering
Copyright q 2012 Baiqing Hu et al. This is an open access article distributed under the Creative
Commons Attribution License, which permits unrestricted use, distribution, and reproduction in
any medium, provided the original work is properly cited.
In light of the intuition that a better symmetrical structure can further increase the numerical
accuracy, the paper by Fan and Zeng 2009 developed a new sigma point construction strategy
for the unscented Kalman filter UKF, namely, geometric simplex sigma points GSSP. This
comment presents a different perspective from the standpoint of the numerical integration. In this
respect, the GSSP constitutes an integration formula of degree 2 with equal weights. Then, we
demonstrate that the GSSP can be derived through the orthogonal transformation from the basic
points set of degree 2. Moreover, the method presented in this comment can be used to construct
more accurate sigma points set for certain dynamic problems.
With the intuition that a better symmetry property provides a better numerical behavior 1,
addressed the construction strategies to make the best symmetric structure in simplex sigma
point set and derived the so-called geometric simplex sigma points GSSP for Euclidean
geometric space. As compared with the previously exiting simplex sigma points set, the GSSP
has a symmetric structure and a lower computational expense, is numerically more accurate,
and can be used in a variety of 3-dimensional modeled dynamic problems.
In this comment we will show that the GSSP can also be derived from the integration
rule of degree 2. Embedding the Gaussian assumption in the Bayesian filter we can reach
the idea that the functional recursion of the Bayesian filter reduces to an algebraic recursion
operating only on conditional means and covariances which share the same structure of
Gaussian weighted integrals whose integrands are all of the form nonlinear function ×
Gaussian density. The multidimensional integrals are usually intractable for systems involving
nonlinearity, so the recursive estimation problem boils down to how to compute the integrals
using approximate methods. There are many well-known numerical integration methods
such as Gauss-Hermite quadrature, cubature rules, fully symmetric integration rule, and
central-difference-based methods that can be used to handle such integrals 2–4. The
unscented transformation UT used in the traditional unscented Kalman filter UKF can
be interpreted as either fully symmetric integration rule or cubature rule of degree 3. The
2 Mathematical Problems in Engineering
simplex UT can also be interpreted as a numerical integration formula of degree 2 3. Next
we will focus on the numerical integration formula of degree 2 in order to derive the GSSP.
Before getting involved in further details, we first introduce some definitions when
constructing the exact monomials rule as follows 3, 4.
Definition 1. Consider the monomials of the form di1 xiαi , where the powers αi are nonnega-
tive integers and di1 αi ≤ p, a rule said to have precision p if it can integrate such monomials
accurately and it is not exact for monomials of degree p 1.
The numerical integration formulas are conducted by approximating the integrals
with the weighted sum of an elaborately chosen set of points as follows 5–7:
gx · Wx ≈ αk g χk , 1
Rn k
gx · N x; μ, Σ dx g Aξ μ · Nξ; 0, Idξ, 2
where AAT Σ and I is the identity matrix, we can start by considering the multidi-
mensional unit Gaussian integral. Based on Definition 1, we can construct a rule, which is
exact up to degree 2 by determining the weighted points set χk such that it is exact for
selections g i ξ 1, gi ξ ξi , gi,j ξ ξi ξj , i
/ j, and gi,i ξ ξi2 . The true values of the
integrals are
I0 1 · Nξ; 0, Idξ 1, 3
I1 ξi · Nξ; 0, Idξ 0, i 1, 2, . . . n,
I2 ξi2 · Nξ; 0, Idξ 1, i 1, 2, . . . n, 4
I1 × 1 ξi ξj · Nξ; 0, Idξ 0, i
/ j 1, 2, . . . n.
In 5 Stroud had proved that n 1 is the minimum number of points for equally weighted
degree 2 formulas. Let us define n 1 equally weighted points
χ χ1 , χ2 , . . . , χn1 , 5
Mathematical Problems in Engineering 3
1 n1
χk 0,
n 1 k1
6
1 n1
χk · χTk In ,
n 1 k1
where I n is the n-dimensional identity matrix. Any equally weighted points set that fulfills
6 can approximate the unit Gaussian integral accurately up to degree 2. References 6, 7
have presented a basic points set that fulfills such conditions with the form as
√ 2rkπ
χk,2r−1 2 cos ,
n1 7
√ 2rkπ
χk,2r 2 sin ,
n1
r 1, 2, . . . , n/2, and if n is odd, χk,n −1k . n/2 is the greatest integer not
exceeding n/2. When n 3, the basic points set is
⎡ √ √ ⎤
0 − 2 0√ 2
√
⎢ ⎥
S1 χ1 | χ2 | χ3 | χ4 ⎣ 2 0 − 2 0 ⎦. 8
−1 1 −1 1
Next we will give a theorem through which we can get the GSSP from the basic points set of
degree 2.
Theorem 1. Assume that n 1 equally weighted points set as that in 5 constitutes an integration
formula of degree 2. A is an n×n orthogonal matrix. Then, Aχ also constitutes an integration formula
of degree 2.
For the 3-dimensional Euclidean space, there are many orthogonal matrixes. Here, we
use the direction cosine matrix DCM which is widely used in the practical systems such as
guidance and navigation 8. The DCM that rotates an angle φ about u 0 0 1T is
⎡ ⎤
cos φ sin φ 0
C φ ⎣− sin φ cos φ 0⎦. 12
0 0 1
Since all the points share equal weight, S2 is virtually just the GSSP derived in 1.
Up to this point we have derived the GSSP through the numerical integration formulas
method. Compared with the intuitionistic method in 1, our method is more principled
in mathematical terms. Although Theorem 1 is proposed for integration formula of degree
2, it can be generalized for different degrees, that is, the orthogonal transformation on
the numerical integration formula will not change its accurate degree. Reference 7 also
presented the points set of degree 3, that is,
γ γ1 , γ2 , . . . , γ 2n , 14
√
2r − 1kπ
γk,2r−1 2 cos ,
n
√ 15
2r − 1kπ
γk,2r 2 sin ,
n
r 1, 2, . . . , n/2, and if n is odd, γk,n −1k . n/2 is the greatest integer not
exceeding n/2. It can be proven that the points set 14 can be derived through orthogonal
transformation on the cubature points set with the form 4
√ √
λ nek − nek , k 1, . . . , n, 16
References
1 C.-S. Fan and Y. Zheng, “Highly efficient sigma point filter for spacecraft attitude and rate estimation,”
Mathematical Problems in Engineering, vol. 2009, Article ID 507370, 23 pages, 2009.
2 K. Ito and K. Xiong, “Gaussian filters for nonlinear filtering problems,” IEEE Transactions on Automatic
Control, vol. 45, no. 5, pp. 910–927, 2000.
3 Y.-X. Wu, D.-W. Hu, M.-P. Wu, and X.-P. Hu, “A numerical-integration perspective on Gaussian
filters,” IEEE Transactions on Signal Processing, vol. 56, no. 8, pp. 2910–2921, 2006.
4 I. Arasaratnam and S. Haykin, “Cubature Kalman filters,” IEEE Transactions on Automatic Control, vol.
54, no. 6, pp. 1254–1269, 2009.
5 A. H. Stroud, “Numerical integration formulas of degree two,” Mathematics and Computers in
Simulation, vol. 14, pp. 21–26, 1960.
6 A. H. Stroud, “Remarks on the disposition of points in numerical integration formulas,” Mathematical
Tables and Other Aids to Computation, vol. 11, pp. 257–261, 1957.
7 D. Xiu, “Numerical integration formulas of degree two,” Applied Numerical Mathematics, vol. 58, no.
10, pp. 1515–1520, 2008.
8 D. H. Titterton and J. L. Weston, Strapdown Inertial Navigation Technology, The Institution of Electrical
Engineers, 2nd edition, 2004.
9 S. J. Julier, “The scaled unscented transformation,” in Proceedings of the American Control Conference,
vol. 6, pp. 4555–4559, Anchorage, Alaska, USA, May 2002.
10 S. Chen, H. Tong, Z. Wang, S. Liu, M. Li, and B. Zhang, “Improved generalized belief propagation for
vision processing,” Mathematical Problems in Engineering, vol. 2011, Article ID 416963, 12 pages, 2011.
11 S. Y. Chen, “Kalman filter for robot vision: a survey,” IEEE Transactions on Industrial Electronics, vol.
59, 2012.
Hindawi Publishing Corporation
Mathematical Problems in Engineering
Volume 2012, Article ID 349178, 20 pages
doi:10.1155/2012/349178
Research Article
A Filter Algorithm with Inexact Line Search
Copyright q 2012 Meiling Liu et al. This is an open access article distributed under the Creative
Commons Attribution License, which permits unrestricted use, distribution, and reproduction in
any medium, provided the original work is properly cited.
A filter algorithm with inexact line search is proposed for solving nonlinear programming prob-
lems. The filter is constructed by employing the norm of the gradient of the Lagrangian function to
the infeasibility measure. Transition to superlinear local convergence is showed for the proposed
filter algorithm without second-order correction. Under mild conditions, the global convergence
can also be derived. Numerical experiments show the efficiency of the algorithm.
1. Introduction
Fletcher and Leyffer 1 proposed filter methods in 2002 offering an alternative to traditional
merit functions in solving nonlinear programming problems NLPs. The underlying concept
is that a trial point is accepted provided that there is a sufficient decrease of the objective
function or the constraint violation. Filter methods avoid the difficulty of determining a
suitable value of the penalty parameter in the merit function. The promising numerical
results in 1 led to a growing interest in filter methods in recent years. Two variants of trust-
region filter sequential quadratic programming SQP method were proposed by Fletcher et
al. 2, 3. Chin and Fletcher 4 developed filter method to sequential linear programming
strategy that takes equality-constrained quadratic programming steps. Ribeiro et al. 5
proposed a general filter algorithm that does not depend on the particular method used for
the step of computation. Ulbrich 6 argued superlinear local convergence of a filter SQP
method. Ulbrich et al. 7 and Wächter and Biegler 8 applied filter technique to interior
method and achieved the global convergence to first-order critical point. Wächter and Biegler
9, 10 proposed a line-search filter method and applied it to different algorithm framework.
Gould et al. 11 and Shen et al. 12 developed new multidimensional filter technique.
2 Mathematical Problems in Engineering
Su and Pu 13 extended the monotonicity of the filter technique. Nie 14 applied filter
method to solve nonlinear complementarity problems. In this paper, the global convergence
is analyzed widely. However, it has been noted by Fletcher and Leyffer 1 that the filter
approach can suffer from the Maratos effect as that of a penalty function approach. By the
Maratos effect, a full step can lead to an increase of both infeasibility measure and objective
function in filter components even if arbitrarily close to a regular minimizer. This makes the
full step unacceptable for the filter and can prohibit fast local convergence.
In this paper, we propose a filter algorithm with inexact line-search for nonlinear
programming problems that ensures superlinear local convergence without second-order
correction steps. We use the norm of the gradient of the Lagrangian function in the infea-
sibility measure in the filter components. Moreover, the new filter algorithm has the same
global convergence properties as that of the previous works 2, 3, 9. In addition, since the
sufficient decrease conditions in an SQP framework can usually make the algorithm complex
and time-consuming, the presented method is a line-search method without using SQP steps.
An inexact line-search criterion is used as the sufficient reduction conditions. In the end,
numerical experiences also show the efficiency of the new filter algorithm.
This paper is organized as follows. For the main part of the paper, the presented tech-
niques will be applied to general NLP. In Section 2, we state the algorithm mechanism. The
convergent properties are shown in Section 3. The global and superlinear convergence are
proved. Furthermore, the Maratos effect is avoided. Finally, Section 4 shows the effectiveness
of our method under some numerical experiences.
minn fx
x∈Ê
2.1
subject to ci x 0, i ∈ E,
where the objective function f : Ên → Ê and the constraints ci are assumed to be continu-
ously differentiable, and E {i | i 1, 2, . . . , m}.
The corresponding Lagrangian function is
where the vector λ corresponds to the Lagrange multiplier. The Karush-Kuhn-Tucker KKT
conditions for 2.1 are
∇fx λT ∇cx 0,
2.3
cx 0.
Mathematical Problems in Engineering 3
For a given initial estimate x0 , the line-search algorithm generates a sequence of iter-
ates xk by xk 1 xk αk dk as the estimates of the solution for 2.1. Here, the search direction
dk is computed from the linearization at xk of the KKT conditions 2.3:
Wk ∇cxk dk ∇fxk
− , 2.4
∇cxk 0 λ cxk
where the symmetric matrix Wk denotes the Hessian ∇2xx Lxk , λk of 2.2 or a positive
definite approximation to it.
After a search direction dk has been computed, the step size αk ∈ 0, 1 is determined
by a backtracking line-search procedure, where a decreasing sequence of step size αk is tried
until some acceptable criteria are satisfied. Generally, the acceptable criteria are constructed
by a condition that if the current trial point xk can provide sufficient reduction of a merit
function. The filter method proposed by Fletcher and Leyffer 1 offers an alternative to merit
functions. In this paper, the filter notion is defined as follows.
Definition 2.1. A pair Vk , fk is said to dominate another pair Vl , fl if and only if both Vk ≤ Vl
and fk ≤ fl .
Here, we define V x : cx2 ∇Lx2 as the l2 norm of the infeasibility measure.
That is, we modify the infeasibility measure in filter, with this modification, the superlinear
convergence is possibly derived. Strictly, it is a stationarity measure. However, we still call
it infeasibility measure according to its function. In the rest of paper, the norm is always
computed by l2 norm excepting special noting.
Definition 2.2. A filter is a list of pairs Vl , fl such that no pair dominates any other. A point
Vk , fk is said to be acceptable for inclusion in the filter if it is not dominated by any point in
the filter.
When a pair Vk , fk is said to be acceptable to the filter, we also say the iterate xk is
acceptable to the filter. In filter notion, a trial point xk 1 is accepted if it improves feasibility
or improves the objective function. So, it is noted that filter criteria is less demanding than
traditional penalty function. When improving optimality, the norm of the gradient of the
Lagrangian function will tend to zero, so it offers a more precise analysis for the objective
function.
However, this simple filter concept is not sufficient to guarantee global convergence.
Fletcher et al. 3 replace this condition by requiring that the next iterate provides at least as
much progress in one of the measure V or f that corresponds to a small fraction of the current
infeasibility measure. Here, we use the similar technique to our filter. Formally, we say that a
trial point can be accepted to the current iterate xk or the filter if
or
for some fixed constants β, γ ∈ 0, 1, and V xl , fxl are points in current filter. In practical
implementation, the constants β close to 1 and γ close to 0. However, the criteria 2.5a and
2.5b may make a trial point always provides sufficient reduction of the infeasibility measure
alone, and not the objective function. To prevent this, we apply a technique proposed in
Wächter and Biegler 10 to a different sufficient reduction criteria. The switching condition
is
e1
∇fkT dk < 0, −αk ∇fkT dk > δV xk e2 , 2.6
where δ > 0, e2 > 1, e1 ≥ 2e2 . If the condition 2.6 holds, we replace the filter condition 2.5b
as an inexact line-search condition, that is, the Armijo type condition
where η ∈ 0, 1/2 is a constant. If 2.6 holds but not 2.7, the trial points are still determined
by 2.5a and 2.5b.
If a trial point xk can be accepted at a step size by 2.7, we refer to xk as an f type
iterate and the corresponding αk as an f step size.
⎧
⎨min 1 − β, − γ V xk , δV xk
e2
⎪
e , if ∇fkT dk < 0,
αmin −∇fkT dk −∇fkT dk 1 2.8
k ⎪
⎩
1 − β, otherwise.
By αmin
k
, it is ensured that the algorithm does not switch to the feasibility restoration
phase as long as 2.6 holds for a step size α < αk and that the backtracking line-search
procedure is finite. Thus, for a trial point xk , the algorithm eventually either delivers a new
iterate xk 1 or reverts to the feasibility restoration phase. Once finding a feasible direction, the
algorithm still implements the normal algorithm.
Of course, the feasibility restoration phase may not always be possible to find a point
satisfying the filter-accepted criteria and the compatible condition. It may converge to a
nonzero local minimizer of the infeasibility measure and indicate that the algorithm fails. In
this paper, we do not specify the particular procedure for the feasibility restoration phase.
Any method for dealing with a nonlinear algebraic system can be used to implement a
feasibility restoration phase.
Algorithm 2.3.
Step 1. Given: starting point x0 , constants V max max{104 , 1.2V x0 }, β, γ ∈ 0, 1, η ∈
0, 1/2, δ > 0, e1 ≥ 2e2 , e2 > 1, τ ∈ 0, 1.
Step 4. Compute the search direction dk from 2.4. If the system 2.4 is incompatible, go to
the feasibility restoration phase in Step 7.
Step 7. Feasibility restoration phase: by decreasing the infeasibility of V to find a new iterate
xk 1 such that 2.4 is compatible. And if 2.7 holds at xk 1 , continue with the normal
algorithm in Step 6; if 2.5a and 2.5b hold at xk 1 , augment the filter by Fk 1 Fk ∪
{V xk 1 , fxk 1 }, and then continue with the normal algorithm in Step 6; if the feasibility
restoration phase cannot find such a point, stop with insuccess.
Remark 2.4. In contrast to SQP method with trust-region technique, the actual step does not
necessarily satisfy the linearization of the constraints.
6 Mathematical Problems in Engineering
Remark 2.5. Practical experience shows that the filter allows a large degree of nonmonotonic-
ity and this can be advantageous to some problems.
Remark 2.6. To prevent the situation in which a sequence of points for which are f type
iterative point with Vk → ∞ is accepted, we set an upper bound V max on the infeasibility
measure function V .
3. Convergence Analysis
3.1. Global Convergence
In this section, we give a global convergence analysis of Algorithm 2.3. We refer to the global
convergence analysis of Wächter and Biegler 10 in some places. First, state the necessary
assumptions.
Assumption A1. Let all iterates xk are in a nonempty closed and bounded set S of Ên .
Assumption A 2. The functions f and c are twice continuously differentiable on an open set
containing S.
Assumption A 3. The matrix Wk is positive definite on the null space of the Jacobian ∇cxk
and uniformly bounded for all k, and the Lagrange multiply λ is bounded for all k.
Lemma 3.1. Suppose Assumptions A1–A3 hold, if {xki } is a subsequence of iterates for which dki >
1 with a constant 1 > 0 independent of ki , then for constant 2 mW /2 1 , if V xki ≤
mW /2Mλ 1 , then
≤ Mλ V xki − mW dki 2
Mathematical Problems in Engineering 7
mW
≤ Mλ 1 − mW 1
2Mλ
− 2.
3.3
Lemma 3.1 shows that the search direction is a descent direction for the objective
function when the trial points are sufficiently close to feasible region.
Lemma 3.2. Suppose Assumptions A1–A3 hold, and that there exists an infinite subsequence {xki }
of {xk } such that conditions 2.6 and 2.7 hold. Then
Proof. From Assumptions A1 and A2, we know that ∇f is bounded. Hence, it has with 3.1
that there exists a constant Mm > 0 such that
T
∇fk dk ≤ Mm . 3.5
By 2.6 it has
e1
δV xki e2 < −αki ∇fkTi dki ≤ Mm
e1
αki . 3.6
e1 1−1/e1
Hence, for c1 : ηδ1/e1 δ/Mm , an integer K and all j 1, 2, . . .,
K
j−1 K
j−1
f xK j fxK fxki 1 − fxki < fxK − c1 V xki e2 . 3.9
ki K ki K
Since fxK j is bounded below as j → ∞, the series on the right hand side in the last line of
3.8 is bounded, then implies the conclusion.
8 Mathematical Problems in Engineering
Lemma 3.3. Let {xki } ⊂ {xk } be an infinite subsequence of iterates so that V xki , fxki is entered
into the filter. Then
Proof. Here, we refer to the proof of 2, Lemma 3.3. If the conclusion is not true, there exists
an infinite subsequence {kj } ⊂ {ki } ⊂ J such that
V xkj ≥ , 3.11
for all j and for some > 0. This means that no other V, f pair can be added to the filter at a
later stage within the region
Vkj − 1 − β , Vkj × fkj − γ , fkj , 3.12
S0 0, V max × f min , ∞ 3.13
for some constants f min ≤ fxk . Now, the area of each of these regions is 1−βγ 2 . Hence, the
set S0 ∪ {V, f | f ≤ Mf } is completely covered by at most a finite number of such regions,
for any Mf ≥ f min . Since the pairs Vkj , fkj keep on being added to the filter, fkj tends to
infinity when i tends to infinity. Without loss of generality, assume that fkj 1 ≥ fkj for all j is
sufficiently large. But 2.5a and 3.11 imply that
so Vkj → 0, which contradicts 3.11. Then, this latter assumption is not true and the
conclusion follows.
Next, we show that if {xk } is bounded, there exists at least one limit point of the
iterative points is a first-order optimal point for 2.1.
Lemma 3.4. Suppose Assumptions A1–A3 hold. Let {xki } be a subsequence with ∇fkTi dki < − 2 for
a constant 2 > 0 independent of ki . Then, there exists a constant α > 0 so that for all ki and αki < α,
Proof. From Assumptions A1 and A2, dT ∇2 fxd ≤ cf d2 for some constant cf > 0. Thus, it
follows from the Taylor Theorem and 3.1 that
if αki ≤ αki : 1 − η 2 /cf Md2 , where y1 denotes some point on the line segment from xki to
xki αki dki . Then the conclusion follows.
Lemma 3.5. Suppose Assumptions A1–A3 hold, and the filter is augmented only a finite number of
times, then
Proof. Since the filter is augmented only a finite number of times, there exists an integer K1
so that for all iterates {xk }k>K1 the filter is not augmented. If the claim is not true, there must
exist a subsequence {xki } and a constant > 0 so that dki ≥ for all i. Then by Lemma 3.1,
it has ∇fkTi dki ≤ − 2 for all ki ≥ K2 , K2 is some integer and K2 > K1 . And from Lemmas 3.2
and 3.4, it has V xki ≤ and
Since fxki is bounded below and monotonically decreasing for all k ≥ K2 , one can conclude
that limi → ∞ αki 0. This means that for ki > K2 the step size α 1 has not been accepted. So,
we can get a αki < 1 such that a trial point xki 1 xki αki dki satisfies
V xki 1 , fxki 1 ∈/ Fki 3.19
or
T
V xki 1 V xki αki ∇V y2 dki
≤ V xki αki ∇V y2 dki 3.21
≤ V xki cV αki Md ,
for some constant cV , where y2 denotes some point on the line segment from xki to xki αdki .
Since limi → ∞ αki 0 and limi → ∞ V xki 0 by Lemmas 3.2 and 3.3, it has V xki 1 < V min
10 Mathematical Problems in Engineering
for ki sufficiently large, so 3.19 is not true. In case 3.20, since αki → 0 for sufficiently
large ki , we have αki ≤ α with α from Lemma 3.4, that is, 3.20 can not be satisfied. Then the
conclusion follows.
Lemma 3.6. Suppose Assumptions A1–A3 hold. Let {xki } be a subsequence of {xk } with ∇fkTi dki ≤
− 2 for a constant 2 > 0 independent of ki . Then, there exists trial points can be accepted to the filter.
Proof. The mechanisms of Algorithm 2.3 ensure that the first iterate can be accepted to
the filter. Next, we can assume that V xk , fxk is acceptable to the kth filter and
V xl , fxl ∈ Fk , l < k. If αki ≤ c2 : 2 /Md2 cf , it has
Similarly, if αki ≤ c3 V xki ≤ V xki /d2 cV , with c3 : 1/Md2 cV and cV from Lemma 3.5, it
has
Hence, we have
The last Lemma 3.6 shows, for case V xk > 0, Algorithm 2.3 either accepts a new
iterate to the filter or switches to the feasibility restoration phase. For case V xk 0 and the
algorithm does not stop at a KKT point, then ∇fkT dk < 0, αmin k
0, and the Armijo condition
2.7 is satisfied for sufficiently small step size αk , so an f type iterate is accepted. Hence, the
inner loop in Step 5 always terminates in a finite number of trial steps, and Algorithm 2.3 is
well defined.
Lemma 3.7. Suppose Assumptions A1–A3 hold. Let {xki } be a subsequence with dki ≥ for a
constant > 0 independent of ki . Then, there exists an sufficient large integer K such that for all
ki > K the algorithm can generate some trial points either be accepted to the filter or be f type steps.
Proof. By Lemmas 3.1, 3.2, and 3.3, there exist constants 1, 2 > 0 so that
If V xki 0, the f type switching condition 2.6 is true, there must exist iterates for
which are f type iterates. For the remaining iterates with V xki > 0, if
⎧ 1/e2 −1 ⎫
⎨α c τc3 e1 ⎬
2 2
V xki < min , , , 3.27
⎩ c3 c3 δ ⎭
with α from Lemma 3.4 and c2 , c3 from Lemma 3.6, it implies with e2 > 1
as well as
Lemmas 3.4 and 3.6 then imply that a trial step size αki ≤ c4 satisfies both
Since α > ταk > ταmin k by the definition of αmink , the method does not switch to the
feasibility restoration phase for those trial step sizes. Then the claim follows.
Based on the above lemmas, we can give the main global convergence result.
that is, there exits a limit point x of {xk } which is a first-order optimal point for 2.1.
can generate a f type iterate, that is, the filter is not augmented, this contradicts the choice of
{xkj }, so that 3.33 holds.
V x 0, ∇x L x, λ 0, ∇xx L x, λ is positive definite on d : ∇cxT d 0 ,
Assumption A 5. There is a neighborhood Nx of x such that Wk ∇xx Lxk , λk , for all
xk ∈ Nx.
Remark 3.9. Under Assumption A4, the point x is a strict local minimum of 2.1.
Remark 3.10. Under Assumptions A4 and A5, it is well known that with the choice xk 1
xk dk , the sequence {xk } converges q-superlinearly to x and that the convergence is q-
quadratic if ∇xx f and ∇xx ci are lipschitz continuous in a neighborhood of x. That is, for any
given ζ ∈ 0, 1, xj ∈ Nx, j k, k 1, . . ., and xj 1 xj dj , it has
dj 1 ≤ ζdj , xj 1 − x ≤ ζxj − x. 3.35
We use the proof techniques of local convergence in 6. In proof, define lρ x, λ
Lx, λ ρ/2cx22 and lρ x, λ fx ρ/2V x with ρ is a parameter.
Lemma 3.11. Suppose Assumptions A1–A3 hold. Let x satisfy the Assumption A4. Then, there exist
constants 0 < r < t, ρ0 > 0 and a neighborhood Nσ x {x : x − x < σ} ⊂ Nx such that
r ρ − ρ0
x − x2 cx2 ≤ lρ x, λ − lρ x, λ ≤ t 1 ρ x − x2 , 3.36
2 2
r ρ − ρ0
x − x2 V x ≤ lρ x, λ − lρ x, λ ≤ t 1 ρ x − x2 . 3.37
2 2
Mathematical Problems in Engineering 13
Proof. Let x ∈ Nx. Using Taylor’s Theorem and ∇x lρ x, λ 0, ∇λlρ x, λ 0, we have with
some x , λ on the line segment between x, λ and x, λ
1
lρ x, λ − lρ x, λ x − xT ∇xx lρ x , λ x − x
2
3.38
x − xT ∇xλ lρ x , λ λ − λ .
Obviously, it has
ρ ρ T
∇xx lρ x , λ ∇xx lρ/2 x , λ c x ∇c x ∇c x . 3.39
2 2
Under Assumption A4, there exists ρ > 0 such that for all ρ ≥ ρ,
dT ∇xx lρ/2 x, λ d ≥ 4rd2 , ∀d ∈ Ên , 3.40
that is the left inequality in 3.36. For the right inequality in 3.36, it is obvious from 3.38
that for all x ∈ Nx,
lρ x, λ − lρ x, λ ≤ t 1 ρ x − x2 , 3.46
By an analogue of 3.45 holds for lρ , this proves the left inequality in 3.37.
On the other hand, it has
ρ
lρ x, λ − lρ x, λ fx V x − fx. 3.48
2
Since fx and cx are twice continuously differentiable on closed set S, we have fx −
fx Ox − x2 and V x Ox − x2 . This shows the right inequality in 3.37 possibly
after increasing t.
Lemma 3.12. Let x satisfy Assumptions A4 and A5. Then for any ζ ∈ 0, 1 and M ≥ 1, there is an
index K such that for all k ≥ K, with
Proof. Let Nx as in Assumption A5, ζ 1/2, ρ0 and Nσ x ⊂ Nx be given by
Lemma 3.11. For all k ≥ K1 , K1 is a sufficient large integer, and choose ρ ≥ ρ0 so large that
ρ
ρ ρ − ρ0 β ρ
β −γ ≥β , ≥ 1− . 3.51
2 4 2 4 2
Let
V : min Vj . 3.53
Vj ,fj ∈FK1
Mathematical Problems in Engineering 15
By V > 0 and continuity there exists 0 < σ1 < σ such that V x ≤ βV for all x ∈ Nσ1 x, so
the point x is acceptable to FK1 . Since xk → x, xk ∈ Nσ1 x for all k ≥ K2 > K1 , K2 is an
integer. By 3.35, we can choose σ1 so small that for all k ≥ K2 , the sequence {xj }j≥k with
αk 1 converges linearly with a contraction factor of at least
1 1 − β/2 r 1
xj
1−x ≤ 1− xj − x ≤ xj − x. 3.54
2 1 − β/4 2M2 t 1 ρ 2
Suppose an arbitrary k2 ≥ K2 such that 3.49 holds, and set σ2 : xk − x. By 3.54, it has
xk 1 ∈ Nσ2 x and xk 1 is acceptable to FK1 . Next, it is a need to show that xk is acceptable
with respect to Vl , fl ∈ Fk ∪ Vk , fk for K1 ≤ l ≤ k. By 3.49, it has xl ∈ Nσ x \ Nσ2 /M x,
so by 3.52
lρ xl , λl − lρ x, λ ≥ r β ρ
σ2 1− Vl . 3.55
2M2 2 4 2
1 1 − β/2 r
lρ xk 1 , λk 1 − lρ x, λ ≤ t 1 ρ xk 1 − x2 ≤ 1− σ 2. 3.56
2 1 − β/4 2M2 2
Next, suppose x, λ with x ∈ Nσ2 x is not acceptable to Vl , fl then
ρ ρ
lρ x, λ fx V x > fl − γ V x
2 2
ρ ρ
> fl β − γ Vl ≥ fl β Vl
2 4
3.58
β ρ
lρ xl , λl − 1 − Vl
2 2
1 − β/2
≥ lρ xl , λl − lρ xl , λl − lρ x, λ .
1 − β/4
lρ x, λ − lρ x, λ > 1 − 1 − β/2 lρ xl , λl − lρ x, λ
1 − β/4
3.59
1 − β/2 r
≥ 1− σ 2.
1 − β/4 2M2 2
16 Mathematical Problems in Engineering
Next, we show that the sequence {xj }j≥k with αk 1 can make the sufficient decreasing
condition 2.7 hold.
Lemma 3.13. Suppose Assumptions A1–A3 hold. Let x satisfy Assumptions A4 and A5 and let K
be as in Lemma 3.11. Then for all k > K the sequence {xj }j≥K with αj 1 satisfies
f xj 1 ≤ f xj ηαj ∇fjT dj . 3.60
Proof. Suppose αj ∇fjT dj < 0 and αj ∇fjT dj e1 < −δV xj e2 hold. By αj 1, thus
e2 /e1
∇fjT dj < −δ1/e1 V xj . 3.61
On the other hand, with αj 1 the assertion fxj 1 ≤ fxj η∇fjT dj yields
1 T 2
η∇fjT dj ≥ ∇fjT dj d ∇ f x dj . 3.62
2 j xx
Thus,
1
∇fjT dj ≤ − djT ∇2xx f x dj , 3.63
2 1−η
e2 /e1 1
−δ1/e1 V xj ≤− djT ∇2xx f x dj . 3.64
2 1−η
Since cj ∇cjT dj 0, dj −∇cjT ∇cj ∇cjT −1 cj . By Assumption A4, ∇cj has full-row
rank, there exists cd > 0 such that
dj ≤ cd V xj , 3.65
thus,
2
dj ≤ c2 V xj 2 . 3.66
d
Choose K large enough such that V xj ≤ 1 for all j ≥ k ≥ K. By e1 > 2e2 and 3.66 it
has
2
dj ≤ c2 c5 V xj e2 /e1
, 3.67
d
Mathematical Problems in Engineering 17
Header Description
Problem The name of the CUTE problem being solved
n Number of variables of the problem
m The total number of constraints
mnl Number of nonlinear constraints
NIT1 Number of iterations of Algorithm 2.3
NIT2 Number of iterations of algorithm Tri-filter
NIT3 Number of iterations of algorithm SNOPT
NF1 Number of f evaluations of Algorithm 2.3
NF2 Number of f evaluations of algorithm Tri-filter
NF3 Number of f evaluations of algorithm SNOPT
for a constant c5 : 2δ1/e1 1 − η/cf cd2 , we can choose suitable parameters such that the last
inequality holds. Thus,
1 2 1
cf dj ≥
e2 /e1
δ1/e1 V xj ≥ djT ∇2xx f x dj . 3.68
2 1−η 2 1−η
Theorem 3.14. Suppose Assumptions A1–A5 hold. Then, there exists K > 0 such that Algorithm 2.3
takes steps with αk 1 for all k ≥ K, that is,
xk 1 xk dk . 3.69
Proof. Since Assumptions A4 and A5 hold, xk → x with x satisfying A4. By Lemmas 3.12 and
3.13, the iterate xk 1 , λk 1 xk dk , λxk dk is acceptable to the filter Fk ∪ Vk , fk and
satisfies the sufficient decreasing condition 2.7. Thus, the trial iterate xk dk , λxk dk is
accepted by the algorithm and it has
xk 1 xk dk , λk 1 λxk dk . 3.70
That is in both cases the algorithm takes the steps with αk 1. And according to Remark 3.10,
{xk } converges q-superlinearly to x.
4. Numerical Experience
In this section, we give some numerical results of Algorithm 2.3. We take some CUTE prob-
lems 17, which are available freely on NEOS, to test our algorithm. The test codes are edited
in MATLAB. The details about the implementation are described as follows.
18 Mathematical Problems in Engineering
a The parameters are set to β 0.99, γ 0.0001, η 0.4, δ 1, τ 0.6, e1 2.3,
e2 1.1, the termination tolerance 1E − 6.
b The optimal residual is defined as
!
res max ∇fxk ∇cxk λ, V xk . 4.1
Acknowledgment
This research is partially supported by the National Natural Science Foundation of China
Grant nos. 10771162 and 91024021.
References
1 R. Fletcher and S. Leyffer, “Nonlinear programming without a penalty function,” Mathematical
Programming, vol. 91, no. 2, pp. 239–269, 2002.
2 R. Fletcher, N. I. M. Gould, S. Leyffer, P. L. Toint, and A. Wächter, “Global convergence of a trust-
region SQP-filter algorithm for general nonlinear programming,” SIAM Journal on Optimization, vol.
13, no. 3, pp. 635–659, 2002.
3 R. Fletcher, S. Leyffer, and P. L. Toint, “On the global convergence of a filter-SQP algorithm,” SIAM
Journal on Optimization, vol. 13, no. 1, pp. 44–59, 2002.
4 C. M. Chin and R. Fletcher, “On the global convergence of an SLP-filter algorithm that takes EQP
steps,” Mathematical Programming, vol. 96, no. 1, Ser. A, pp. 161–177, 2003.
5 A. A. Ribeiro, E. W. Karas, and C. C. Gonzaga, “Global convergence of filter methods for nonlinear
programming,” SIAM Journal on Optimization, vol. 19, no. 3, pp. 1231–1249, 2008.
6 S. Ulbrich, “On the superlinear local convergence of a filter-SQP method,” Mathematical Programming,
vol. 100, no. 1, pp. 217–245, 2004.
7 M. Ulbrich, S. Ulbrich, and L. N. Vicente, “A globally convergent primal-dual interior-point filter
method for nonlinear programming,” Mathematical Programming, vol. 100, no. 2, pp. 379–410, 2003.
8 A. Wächter and L. T. Biegler, “On the implementation of an interior-point filter line-search algorithm
for large-scale nonlinear programming,” Mathematical Programming, vol. 106, no. 1, pp. 25–57, 2006.
9 A. Wächter and L. T. Biegler, “Line search filter methods for nonlinear programming: motivation and
global convergence,” SIAM Journal on Optimization, vol. 16, no. 1, pp. 1–31, 2005.
10 A. Wächter and L. T. Biegler, “Line search filter methods for nonlinear programming: local
convergence,” SIAM Journal on Optimization, vol. 16, no. 1, pp. 32–48, 2005.
20 Mathematical Problems in Engineering
Research Article
Optimal Incentive Pricing on Relaying
Services for Maximizing Connection Availability in
Multihop Cellular Networks
Copyright q 2012 M.-H. Lin and H.-J. Hsu. This is an open access article distributed under the
Creative Commons Attribution License, which permits unrestricted use, distribution, and
reproduction in any medium, provided the original work is properly cited.
This paper investigates an incentive pricing problem for relaying services in multihop cellular
networks. Providing incentives to encourage mobile nodes to relay data is a critical factor in
building successful multihop cellular networks. Most existing approaches adopt fixed-rate or
location-based pricing on rewarding packets forwarding. This study applies a mathematical
programming model to determine an optimal incentive price for each intermediate node that
provides relaying services. Under the obtained incentive price, the connection availability of
the networks is maximized by using the same relaying costs as other pricing schemes. A
signomial geometric programming problem is constructed, and a deterministic optimization
approach is employed to solve the problem. Besides, quality-of-service constraints are added in the
proposed model to mitigate the unfairness between connection availabilities of individual nodes.
Computational results demonstrate that the proposed model obtains the optimal incentive price
on relaying services to maximize connection availability of the networks.
1. Introduction
Over the past few years, wireless networks and wireless devices have rapidly developed and
undergone significant advances. More and more services that dramatically affect personal
and business communications are provided by wireless access networks. How to build a
seamless wireless network has received increasing attention from the practitioners and the
researchers. Most wireless networks are based on cellular architecture, which means that
a mobile host is handled by a central base station in a limited range. Cellular networks
have inherent limitations on cell coverage and the dead spot problem. Traditionally, the
network providers utilize more infrastructure equipments such as base stations, to solve
2 Mathematical Problems in Engineering
these problems. However, this method is expensive. Therefore relaying technology has been
developed to solve this problem. In the last decade, multihop cellular networks have been
proposed to harness the benefits of conventional cellular networks and emerging multihop
ad hoc networks. In cellular networks, a mobile device directly connects with the base
station; in multihop networks, a mobile device communicates with others over peer-to-peer
connections. Figure 1 indicates the scenario of general multihop cellular networks. Adopting
hop-by-hop connections can extend the service area at the boundaries of the network and
eliminate dead spots, including indoor environments and basements. Much research has
evaluated and summarized the advantages of multihop cellular networks over existing
single-hop cellular networks as follows 1–5.
Cooperation among nodes is a critical factor for ensuring the success of the relaying
ad hoc networks 2, 6. In recent years, a number of approaches have been proposed to
encourage mobile nodes to relay data for others in ad hoc networks. Most of existing
motivation-based approaches focus on a charging protocol and use fixed-rate pricing that
gives identical reward level on per unit of packet forwarded. Although the major advantage
of fixed-rate pricing is that billing and accounting processes are simple, providing identical
reward level to all mobile nodes neglects the distinct importance of each mobile node
in the networks. Lo and Lin 2 developed a location-based incentive pricing scheme
rewarding each mobile node based on its degree of contribution to successful hop-by-
hop connections. Simulation results indicate that their method provides higher connection
availability compared to the fixed-rate pricing scheme.
This paper constructs a mathematical programming model to the problem of optimal
pricing on relaying services provided by the mobile nodes in the multihop cellular networks.
The formulated model that maximizes connection availability of the networks under identical
relaying costs used by the fixed-rate pricing scheme and the location-based pricing scheme
2 is a signomial geometric programming SGP problem. Convexification strategies and
piecewise linearization techniques are employed to reformulate the problem into a convex
mixed-integer nonlinear programming MINLP problem that can be solved by conventional
MINLP methods to reach a global solution. We also add quality-of-service QoS constraints
in the constructed model to guarantee each mobile node with a minimum successful
connection probability, therefore mitigating the unfairness between connection availabilities
of individual nodes. Computational experiments are conducted to compare the proposed
method with existing pricing schemes. Simulation results indicate that the proposed method
obtains higher connection availability of the networks than the existing pricing methods
without additional relaying costs.
The rest of the paper is organized as follows. Section 2 reviews existing multihop
cellular networking models and incentive pricing models. In Section 3, an incentive pricing
model for maximizing connection availability is proposed to determine the optimal price on
Mathematical Problems in Engineering 3
Building
Indoor
environment Base station
Basement
relaying services provided by mobile nodes. Section 4 provides a solution approach based on
variable transformations and piecewise linearization techniques. In Section 5, we present the
computational experiments, and finally concluding remarks are made in Section 6.
2. Literature Review
Opportunity-driven multiple access ODMA is an ad hoc multihop protocol where the
transmissions from the mobile hosts to the base station are broken into multiple wireless
hops, thereby reducing transmission power 4. The high-data-rate coverage of the cell can
be increased by adopting relaying technologies for the mobile nodes outside the high-data-
rate coverage area. The Ad Hoc GSM A-GSM system is a network protocol platform that
accommodates relaying capability in GSM cellular networks. Although the GSM system aims
to provide global roaming, the dead spot problem still exists, for example, in subway stations
and basements. Since installing additional base stations at each dead spot location is not
economical, the A-GSM system extends the data communication through the mobile nodes
7. Qiao and Wu 3 presented an integrated cellular and ad hoc relay iCAR system to
combine the cellular infrastructure with ad hoc relaying technologies. All cellular networks
have problems of limited capacity and unbalanced traffic. Cellular networks probably cannot
provide the connection service because some of the cells are heavily congested, but at the
same time other cells still have available channels. This kind of centralized obstruction
makes the system unable to establish successful communication, even though the number
of required channels does not reach the maximum capacity of the entire system. Utilizing
relaying technologies a mobile host in a congested cell obtains a free channel in another
cell and establishes a new call successfully. Wu et al. 8 proposed a scheme called mobile-
assisted data forwarding MADF to add an ad hoc overlay to the fixed cellular infrastructure,
and special channels are assigned to connect users in a hot cell to its neighboring cold
cells without going through the base station in the hot cell. An intermediate forwarding
4 Mathematical Problems in Engineering
agent, such as a repeater or another mobile terminal, in the cold cell is required to relay
the data to that cell. Wu et al. 8 observed that under a certain delay requirement, the
throughput can be greatly improved. Luo et al. 9 proposed a unified cellular and ad hoc
network UCAN architecture that considers the balanced traffic and network throughput.
The UCAN architecture uses relaying technologies to send information to the mobile device
if the transmission quality of the channel is poor. Each mobile device in the UCAN model has
both a 3G cellular link and IEEE 802.11-based peer-to-peer links. The 3G base station forwards
packets for destination clients with poor channel quality to proxy clients with better channel
quality. Multihop cellular network is considered as a promising candidate of 4G wireless
network for future mobile communications. The complete surveys of technologies advances
and economic perspective on the deployment of multihop cellular networks are provided by
Li et al. 1 and Manoj et al. 10.
Since forwarding data for others consumes battery energy and delays its own data,
providing incentives for mobile nodes to cooperate as relaying entries is necessary. The
existing incentive schemes can be classified into detection-based and motivation-based
approaches. The detection-based approach finds out the misbehaving nodes and reduces
their impact in the networks. Marti et al. 11 developed two methods that find the
misbehaving nodes and avoid routing packets through these nodes. Michiardi and Molva
12 proposed a mechanism to enforce cooperation among nodes based on reputation and
to prevent denial of service attacks because of selfishness. Buchegger and Le Boudec 13
developed a protocol to detect and isolate misconduct nodes, therefore making it unattractive
to deny cooperation.
Instead of discouraging misbehavior by punishing misbehavior node, the motivation-
based approach encourages positive cooperation by rewarding incentives for relaying
packets. Buttyán and Hubaux 6, 14 developed different approaches to provide incentives
to cooperative nodes, therefore simulating packet forwarding. Buttyán and Hubaux 14 did
not discuss the reward level, and Buttyán and Hubaux 6 suggested to reward the relaying
service based on the number of forwarding packets. Jakobsson et al. 15 developed a micro-
payment scheme to encourage collaboration and discourage dishonest behavior in multihop
cellular networks. A subject reward level is determined according to the importance of the
packet. Lamparter et al. 16 proposed a charging scheme using volume-based pricing. A
fixed price per unit of data is rewarded for forwarding traffic in ad hoc stub networks. The
rewarding mechanisms mentioned above adopt fixed-rate pricing and do not consider the
importance of each mobile node in the routing topology.
Lo and Lin 2 proposed a location-based incentive pricing scheme that adjusts
the price of incentives for packet forwarding based on the degree of each mobile nodes
contributing to successful hop-by-hop connections. Since the willingness of the mobile node
to relay packets has a significant impact on the success of the multihop connections from
all nodes in its subtree to the base station, the importance of a mobile node depends on the
number of mobile nodes in its subtree. They defined the location index LI v of a mobile node v
as the number of nodes residing in the tree rooted at node v. Let N be the set of intermediate
nodes providing relaying services for the mobile nodes that require hop-by-hop connections
to the base station, and ALI be the average location index of all nodes in N; the price of the
feedback incentives for node v, pv , is defined as follows 2:
Rp 1
pv p0 LI v − ALI · · , 2.1
RLI LI v
Mathematical Problems in Engineering 5
where Rp min{p0 , Pmax − p0 }, RLI max{ALI − minv∈N {LI v }, maxv∈N {LI v } − ALI}, p0 is
the price used in the fixed-rate pricing method, and Pmax is the maximum price the network
provider can reward to an intermediate mobile node. Equation 2.1 employs p0 as a basic
price and gives a higher incentive price on relaying services for the node with a higher
location index. Because some incentive rewards are shifted from the nodes of low importance
to the node of high importance, the Lo and Lin 2 pricing scheme results in higher connection
availability but does not generate higher relaying costs compared to the fixed-rate pricing
scheme. However, their method does not provide an optimal incentive pricing solution that
maximizes connection availability of the networks.
pv
S pv , 0 ≤ pv ≤ Pmax , 3.1
Pmax
where pv is the incentive price on per unit of relayed data, Pmax is the maximum price the
network provider can reward to an intermediate mobile node per unit of relayed data, and
Spv is the willingness of forwarding packets under the incentive price pv . Spv is the
supply function representing the reaction of mobile nodes to the change in the price of the
incentives. Spv 0 0 means that node v will not relay traffic for others if no feedback
is provided for relaying services. The willingness of forwarding packets linearly increases as
the incentive price on relaying services increases. Spv Pmax 1 means the maximum price
is acceptable for all mobile nodes to provide relaying services.
In multihop cellular networks, data packets must be relayed hop by hop from a given
mobile node to a base station; thus the connection availability of node i depends on the
willingness of all intermediate mobile nodes on the routing path to forward packets. Let CAi
be the connection availability of node i, that is, the successful connection probability from
node i to the base station. CAi can be expressed as 2
CAi S pv , 3.2
v∈Mi
where Mi is the set of intermediate nodes in the path from node i to the base station, and all
the other variables are the same as defined before.
6 Mathematical Problems in Engineering
w
CAi
i1
Maximize 3.3
w
pv
subject to CAi , i 1, . . . , w, 3.4
v∈Mi
Pmax
w
w
Ti · pv ≤ Ti · Pfixed , 3.5
i1 v∈Mi i1 v∈Mi
0 ≤ pv ≤ Pmax , 3.6
where w is the number of nodes requiring hop-by-hop connections to the base station in
the networks, Ti is the units of traffic sent by node i, and Pfixed is the fixed incentive price
on relaying services used by the fixed-rate pricing scheme. The objective function aims
to maximize the connection availability of the networks, that is, the average connection
availability of all mobile nodes using hop-by-hop connections to the base station. Lo and
Lin 2 refers the objective function as service availability. Constraint 3.5 indicates the total
relaying costs of the proposed method are not greater than the total relaying costs of the
fixed-rate pricing scheme.
Maximize 3.3
3.7
subject to 3.4 ∼ 3.6,
where Min CALB represents the minimal connection availability of all mobile nodes requiring
hop-by-hop connections to the base station in the location-based pricing scheme. Constraint
3.8 indicates that the connection availability of each mobile node satisfies the required QoS
level QoSCA . If the minimal connection availability of all mobile nodes in the location-based
pricing scheme is greater than zero, then the required QoS level is set as Min CALB . Otherwise,
the required QoS level is set as 0.01.
Mathematical Problems in Engineering 7
4. Problem-Solving Approach
Since the problem described in the previous section is an SGP problem, that is, a class
of nonconvex programming problems. SGP problems generally possess multiple local
optima and experience much more theoretical and computational difficulties. This study
uses variable transformations and piecewise linearization techniques to reformulate the
problem into a convex MINLP problem that can be globally solved by conventional
MINLP methods. Much research has proposed variable transformation techniques to
solve optimization problems including signomial functions to global optimality 19–22.
For convexifying positive signomial terms, Lundell and Westerlund 23 proved that the
exponential transformation always results in a tighter underestimator than the negative
power transformation. This study applies the exponential transformation to convexify a
positive signomial function ni1 xiαi by the following remark 19, 24.
Remark 4.1. If αj > 0 for some j, j∈ / I, I {k | αk < 0, k 1, 2, . . . , n}, then we convert ni1 xiαi
into another function i∈I xiαi e j ∈/I αj yj , where yj Lln xj and Lln xj , is a piecewise linear
function of ln xj . Then i∈I xiαi e j ∈/I αj yj where xi > 0, i ∈ I, yj ∈ , j ∈
/ I is a convex function.
Herein the concept of special ordered set of type 2 SOS-2 constraints can be utilized to
formulate the piecewise linear function 25, 26. This study adopts the piecewise linearization
technique introduced by Vielma and Nemhauser 27 that uses a logarithmic number of
binary variables and extra constraints. The computational results in 27 show that their
piecewise linearization technique outperforms other piecewise linearization formulations.
Tsai and Lin 28 applied the piecewise linearization technique developed by Vielma and
Nemhauser 27 to efficiently solve posynomial geometric programming problems.
The original model has one nonconvex objective function, one constraint, and n
variables, where n is the number of intermediate nodes providing relaying services. The
reformulated model has one convex objective function, one constraint, n variables, and
several piecewise linear functions. The number of piecewise linear functions depends on
convexification process in reformulating the problem. For each piecewise linear function,
log2 m binary variables, m 1 continuous variables, and 3 2log2 m constraints are required
to express a piecewise linear function with m line segments.
The following example is used to illustrate how the proposed method discussed
previously determines the incentive price on relaying services provided by each mobile node.
Example 4.3. Consider an example taken from Lo and Lin 2 with twelve mobile nodes
distributed in the multihop cellular networks indicated in Figure 2. Nine nodes nodes 2,
3, 4, 5, 7, 8, 9, 11, 12 require hop-by-hop connections to reach the central base station.
8 Mathematical Problems in Engineering
11
12
10
8 9
7
1
5
2 3
Base station
Mobile node
Assume each mobile node has identical traffic load u units, and the fixed-rate pricing
scheme uses 0.5Pmax as the incentive price for relaying per unit of data, that is, Pfixed 0.5Pmax .
A mathematical programming model can be constructed for the connection availability
maximization problem as follows:
0 ≤ pi ≤ Pmax , i 1, 3, 6, 7, 8, 10,
4.1
where CAi , i 2, 3, 4, 5, 7, 8, 9, 11, 12, represents the connection availability of mobile node
i, and Pmax and u are constants. This program is a nonconvex SGP problem. Applying the
method mentioned previously, the connection maximization problem of Example 4.3 can be
converted into a convex MINLP problem as follows:
4.2
subject to 2u · p1 2u · p1 p3 u · p6 u · p6 p7 u · p6 p7 p8 2u · p10
≤ 7uPmax ,
y1 L p12 , y3 L p32 , y6 L p63 , y7 L p73 , y8 L p83 ,
0 ≤ pi ≤ Pmax , i 1, 3, 6, 7, 8, 10,
4.3
where Lp12 , Lp32 , Lp63 , Lp73 , and Lp83 are piecewise linear functions of p12 , p32 , p63 , p73 ,
and p83 , respectively. By using the efficient piecewise linearization technique introduced by
Vielma and Nemhauser 27, this program is reformulated as a convex MINLP problem that
can be solved on LINGO 29 to obtain a global solution p1 , p3 , p6 , p7 , p8 , p10 0.8763Pmax ,
0.7474Pmax , 0, 0, 0, Pmax . Table 1 compares the incentive price on relaying services and
connection availability of the networks under different pricing schemes. Herein the fixed-
rate pricing scheme rewards each mobile node 0.5Pmax for relaying per unit of data. From
the data listed in Table 1, we find the proposed pricing scheme provides higher connection
availability of the networks than the fixed-rate pricing scheme and the location-based pricing
scheme do. We also observe that the three methods use approximately the same relaying
costs.
From Table 1, we find out that although the proposed method has better connection
availability of the networks than the other two methods, the connection availabilities of
some mobile nodes CA7 , CA8 , CA9 are zero; that is, these nodes cannot connect to the
base station. This study adds the QoS constraints CAi ≥ 0.026, i 2, 3, 4, 5, 7, 8, 9, 11, 12,
to guarantee each mobile node with a minimum successful connection probability. The
required QoS level 0.026 is the minimal individual connection availability obtained from
10 Mathematical Problems in Engineering
Table 1: Comparison between the fixed-rate pricing scheme, the location-based pricing scheme, and the
proposed pricing scheme of Example 4.3.
the location-based pricing scheme. The proposed model with QoS requirements becomes as
follows:
Minimize 4.2
subject to 4.3
p1 y10.5 y30.5 p6
− < −0.026, − · < −0.026, − < −0.026,
Pmax Pmax Pmax Pmax
Solving this problem can obtain a globally optimal solution p1 , p3 , p6 , p7 , p8 , p10
0.6539Pmax , 0.3085Pmax , 0.3165Pmax , 0.2313Pmax , 0.3550Pmax , Pmax . Table 2 shows comparison
between the location-based pricing scheme and the proposed pricing scheme with QoS
requirements of Example 4.3. We find the minimal individual connection availability from
the proposed pricing scheme is equal to that from the location-based pricing scheme, and the
Mathematical Problems in Engineering 11
Table 2: Comparison between the location-based pricing scheme and the proposed pricing scheme with
QoS requirements of Example 4.3.
connection availability of the networks from the proposed pricing scheme is still higher than
that from the location-based pricing scheme.
5. Numerical Experiments
5.1. Fixed-Rate Method versus Location-Based Method versus
Proposed Method
This section describes the simulation results for verifying the advantages of the proposed
pricing scheme. We design our simulation tests by C language. All simulations are run
on a Notebook with an Intel CPU P8700 and 4 GB RAM. The simulation environment is a
rectangular region of 100 units width and 100 units height with a single base station of 30
units radius located in the central point. The radius of each mobile node is 20 units. In this
study, a shortest path tree is built such that each mobile node connects to the base station
with a minimum number of hops.
In the experiments 32, 64, and 128, mobile nodes, respectively, are randomly
distributed in the rectangular region. 10 simulations are run for each set of parameter
settings. Table 3 compares average connection availability of the networks of 10 simulations
by different incentive pricing schemes. Figure 3 indicates that the proposed pricing scheme
obtains higher average connection availability than the fixed-rate pricing scheme and the
12 Mathematical Problems in Engineering
Table 3: Comparison of connection availability of the networks by three methods in the simulation space
of width, height 100 units, 100 units.
0.6
Connection availability of networks
0.55
0.5
0.45
0.4
0.35
0.3
0.25
0.2
32 64 128
Number of mobile nodes
Fixed-rate pricing
Location-based pricing
Proposed pricing
Figure 3: Comparison of connection availability of the networks by three methods in the simulation space
of width, height 100 units, 100 units.
location-based pricing scheme under different number of mobile nodes in the simulation
environment. As the number of mobile nodes in the networks increases, the mobile nodes
are easier to find a shorter hop-by-hop path for connecting to the base station. Therefore
the average path length decreases when the number of mobile nodes increases. The effect
on improving connection availability of the networks by the proposed method is more
significant when the average path is longer.
Table 4: Comparison of connection availability of the networks by three methods in the simulation space
of width, height 200 units, 200 units.
0.6
Connection availability of networks
0.55
0.5
0.45
0.4
0.35
0.3
0.25
0.2
32 64 128
Number of mobile nodes
Fixed-rate pricing
Location-based pricing
Proposed pricing
Figure 4: Comparison of connection availability of the networks by three methods in the simulation space
of width, height 200 units, 200 units.
models. Figure 4 indicates that in the simulation space of 200 units width and 200 units
height, the difference in the connection availability of the networks obtained by the fixed-
rate pricing scheme and the location-based pricing scheme is not obvious. However, the
connection availability of the networks by the proposed method is much higher than that
by the other two methods if a longer path is required to reach the base station.
Table 5: Comparison of connection availability by two methods in the simulation space of width, height
100 units, 100 units.
Number of mobile nodes CALB CAP QoS CAP QoS − CALB /CALB
32 0.48730129 0.49575548 1.74%
64 0.49130662 0.49617013 1.00%
128 0.49131695 0.49246111 0.24%
CALB : connection availability of the networks from the location-based pricing scheme.
CAP QoS : connection availability of the networks from the proposed pricing scheme with QoS requirements.
0.6
Connection availability of networks
0.55
0.5
0.45
0.4
0.35
0.3
0.25
0.2
32 64 128
Number of mobile nodes
Location-based pricing
Proposed pricing with QoS
Figure 5: Comparison of connection availability of the networks by two methods in the simulation space
of width, height 100 units, 100 units.
height 200 units, 200 units, respectively. The comparisons are also indicated in Figures 5
and 6. Compared to the results from the proposed pricing method without QoS constraints in
Tables 3 and 4, although the connection availability of the networks decreases, the proposed
pricing method with QoS requirements still performs better than the location-based pricing
method. Since the minimum individual node’s connection availability in the proposed
pricing method is greater than or equal to that in the location-based pricing method, adding
QoS constraints makes the proposed pricing method consider both connection availability of
the networks and fairness between individual node’s connection availabilities.
6. Conclusions
Cost savings and connection availability are two crucial issues of a network provider
adopting multihop cellular networking technology. This paper determines the optimal
incentive price on relaying services for each mobile node by constructing a mathematical
programming model that maximizes connection availability without extra relaying costs.
A deterministic optimization approach based on variable transformations and piecewise
linearization techniques is utilized to solve the formulated problem. Simulation results
demonstrate that the proposed pricing model results in higher connection availability
than the fixed-rate pricing scheme and the location-based pricing scheme. In addition, a
mathematical programming model involving QoS requirements in connection availability of
Mathematical Problems in Engineering 15
Table 6: Comparison of connection availability of the networks by two methods in the simulation space of
width, height 200 units, 200 units.
Number of mobile nodes CALB CAP QoS CAP QoS− CALB /CALB
32 0.23610142 0.3183623 34.84%
64 0.24306587 0.31280795 28.69%
128 0.24986204 0.29797671 19.26%
CALB : connection availability of the networks from the location-based pricing scheme.
CAP QoS : connection availability of the networks from the proposed pricing scheme with QoS requirements.
0.6
Connection availability of networks
0.55
0.5
0.45
0.4
0.35
0.3
0.25
0.2
32 64 128
Number of mobile nodes
Location-based pricing
Proposed pricing with QoS
Figure 6: Comparison of connection availability of the networks by two methods in simulation space of
width, height 200 units, 200 units.
each individual mobile node is developed to eliminate the unfairness situation in the original
model.
Acknowledgments
The authors thank the anonymous referees for contributing their valuable comments
regarding this paper and thus significantly improving the quality of this paper. The research
is supported by Taiwan NSC Grant NSC 99-2410-H-158-010-MY2.
References
1 X. J. Li, B. C. Seet, and P. H. J. Chong, “Multihop cellular networks: technology and economics,”
Computer Networks, vol. 52, no. 9, pp. 1825–1837, 2008.
2 C. C. Lo and M. H. Lin, “Location-based incentive pricing for tree-based relaying in multi-hop cellular
networks,” Computer Communications, vol. 29, no. 16, pp. 3148–3157, 2006.
3 C. Qiao and H. Wu, “ICAR: an intelligent cellular and ad-hoc relay system,” in Proceedings of the IEEE
International Conference on Computer Communications and Networks (IC3N ’00), pp. 154–161, Las Vegas,
Nev, USA, October 2000.
4 T. Rouse, I. Band, and S. McLaughlin, “Capacity and power investigation of Opportunity Driven
Multiple Access ODMA networks in TDD-CDMA based systems,” in Proceedings of the International
Conference on Communications (ICC 2002), pp. 3202–3206, New York, NY, USA, May 2002.
16 Mathematical Problems in Engineering
5 N. Ben Salem, L. Buttyán, J. P. Hubaux, and M. Jakobsson, “A charging and rewarding scheme
for packet forwarding in multi-hop cellular networks,” in Proceedings of the 4th ACM international
symposium on Mobile ad hoc networking and Computing (MobiHoc ’03), pp. 13–24, Annapolis, Md, USA,
June 2003.
6 L. Buttyán and J. P. Hubaux, “Stimulating cooperation in self-organizing mobile ad hoc networks,”
Mobile Networks and Applications, vol. 8, no. 5, pp. 579–592, 2003.
7 G. N. Aggélou and R. Tafazolli, “On the relaying capability of next-generation GSM cellular
networks,” IEEE Personal Communications, vol. 8, no. 1, pp. 40–47, 2001.
8 X. Wu, S. -H.G. Chan, and B. Mukherjee, “MADF: a novel approach to add an ad-hoc overlay on
a fixed cellular infrastructure,” in Proceedings of the IEEE Wireless Communications and Networking
Conference (WCNC ’02), vol. 2, pp. 549–554, Chicago, Ill, USA, September 2002.
9 H. Luo, R. Ramjee, P. Sinha, L. Li, and S. Lu, “UCAN: a unified cellular and ad-hoc network
architecture,” in Proceedings of the 9th Annual International Conference on Mobile Computing and
Networking (MobiCom ’03), pp. 353–367, San Diego, Calif, USA, September 2003.
10 B. S. Manoj, R. Ananthapadmanabha, and C. S. R. Murthy, “Multi-hop cellular networks: architecture
and protocols for best-effort and real-time communication,” Journal of Parallel and Distributed
Computing, vol. 65, no. 6, pp. 767–791, 2005.
11 S. Marti, T. J. Giuli, K. Lai, and M. Baker, “Mitigating routing misbehavior in mobile ad hoc networks,”
in Proceedings of the 6th Annual International Conference on Mobile Computing and Networking (MobiCom
’00), pp. 255–265, Boston, Mass, USA, August 2000.
12 P. Michiardi and R. Molva, “Core: a collaborative reputation mechanism to enforce node cooperation
in mobile ad hoc networks,” in Proceedings of the IFIP TC6/TC11 6th Joint Working Conference on
Communications and Multimedia Security, pp. 107–121, Portorož, Slovenia, September 2002.
13 S. Buchegger and J. Y. Le Boudec, “Performance analysis of the CONFIDANT protocol Cooperation
of nodes: fairness in dynamic ad-hoc networks,” in Proceedings of the 3rd ACM international symposium
on Mobile ad hoc networking and Computing (MobiHoc ’02), pp. 226–236, Lausanne, Switzerland, June
2002.
14 L. Buttyán and J.-P. Hubaux, “Enforcing service availability in mobile ad-hoc WANs,” in Proceedings
of the 1st ACM international symposium on Mobile ad hoc networking and Computing (MobiHoc ’00), pp.
87–96, Boston, Mass, USA, August 2000.
15 M. Jakobsson, J. P. Hubaux, and L. Buttyán, “A micro-payment scheme encouraging collaboration in
multi-hop cellular networks,” Lecture Notes in Computer Science, vol. 2742, pp. 15–33, 2003.
16 B. Lamparter, K. Paul, and D. Westhoff, “Charging support for ad hoc stub networks,” Computer
Communications, vol. 26, no. 13, pp. 1504–1514, 2003.
17 J. Hou, J. Yang, and S. Papavassiliou, “Integration of pricing with call admission control to meet QoS
requirements in cellular networks,” IEEE Transactions on Parallel and Distributed Systems, vol. 13, no. 9,
pp. 898–910, 2002.
18 P. C. Fishburn and A. M. Odlyzko, “Dynamic behavior of differential pricing and quality of service
options for the internet,” in Proceedings of the 1st International Conference on Information and Computation
Economies, pp. 128–139, Charleston, SC, USA, October 1998.
19 A. Lundell and T. Westerlund, “Convex underestimation strategies for signomial functions,”
Optimization Methods and Software, vol. 24, no. 4-5, pp. 505–522, 2009.
20 A. Lundell, J. Westerlund, and T. Westerlund, “Some transformation techniques with applications in
global optimization,” Journal of Global Optimization, vol. 43, no. 2-3, pp. 391–405, 2009.
21 J.-F. Tsai and M.-H. Lin, “Global optimization of signomial mixed-integer nonlinear programming
problems with free variables,” Journal of Global Optimization, vol. 42, no. 1, pp. 39–49, 2008.
22 T. Westerlund, “Some transformation techniques in global optimization,” in Global Optimization, from
Theory to Implementations, L. Liberti and N. Maculan, Eds., vol. 84, pp. 45–74, Springer, New York, NY,
USA, 2007.
23 A. Lundell and T. Westerlund, “On the relationship between power and exponential transformations
for positive signomial functions,” Chemical Engineering Transactions, vol. 17, pp. 1287–1292, 2009.
24 R. Pörn, K.-M. Björk, and T. Westerlund, “Global solution of optimization problems with signomial
parts,” Discrete Optimization, vol. 5, no. 1, pp. 108–120, 2008.
25 E. L. M. Beale and J. A. Tomlin, “Special facilities in a general mathematical programming system for
nonconvex problems using ordered sets of variables,” in Proceedings of 5th International Conference on
OR, pp. 447–454, Venice, Italy, June 1969.
26 A. Martin, M. Möller, and S. Moritz, “Mixed integer models for the stationary case of gas network
optimization,” Mathematical Programming, vol. 105, no. 2-3, pp. 563–582, 2006.
Mathematical Problems in Engineering 17
27 J. P. Vielma and G. L. Nemhauser, “Modeling disjunctive constraints with a logarithmic number of
binary variables and constraints,” Mathematical Programming, vol. 128, no. 1-2, pp. 49–72, 2011.
28 J. -F. Tsai and M. -H. Lin, “An efficient global approach for posynomial geometric programming
problems,” INFORMS Journal on Computing, vol. 23, no. 3, pp. 483–492, 2011.
29 Lindo, LINGO, Release 11, Lindo System, Chicago, Ill, USA, 2004.