Machine Learning Approaches For Improving Condition-Based Maintenance of Naval Propulsion Plants
Machine Learning Approaches For Improving Condition-Based Maintenance of Naval Propulsion Plants
itime Environment
Machine Learning Approaches for 0(0):1–24
©The Author(s) 2010
Abstract
Availability, reliability and economic sustainability of naval propulsion plants is a key element to cope with, because
maintenance costs represent a large slice of total operational expenses. Depending on the adopted strategy, impact of
maintenance on overall expenses can remarkably vary: for example, letting an asset running up until breakdown can lead to
unaffordable costs. As a matter of fact, a desideratum is to progress maintenance technology of ship propulsion systems from
Breakdown (BM) or Preventive Maintenance (PM) up to more effective Condition Based Maintenance (CBM) approaches.
The central idea in CBM is to monitor the propulsion equipment by exploiting heterogeneous sensors, enabling diagnosis
and, most of all, prognosis of the propulsion system’s components and of their potential future failures. The success of
CBM clearly hinges on the capability of developing effective predictive models: for this purpose, in this paper is proposed
to use effective Machine Learning (ML) methods. In particular, authors take into consideration an application of CBM
to Gas Turbines (GTs) used for vessel propulsion, where the performance and advantages of exploiting ML methods in
modeling the degradation of the propulsion plant over time are tested. Experiments, conducted on data generated from a
sophisticated simulator of a GT, mounted on a Frigate characterized by a COmbined Diesel eLectric And Gas (CODLAG)
propulsion plant type, will allow to show the effectiveness of the proposed ML approaches and to benchmark them in a
realistic maritime application.
Keywords
Condition-Based Maintenance, Gas Turbine, COmbined Diesel eLectric And Gas (CODLAG) propulsion plant, Machine
Learning, Asset decay forecast
Disclaimer
The experiments have been carried out by means of parts of a complex numerical simulator of a naval vessel (Frigate),
characterized by a Gas Turbine propulsion plant. The different blocks forming the complete simulator have been developed
and fine tuned throughout the years on several similar real propulsion plants: in view of these observations, the data that
will be used are arguably in agreement with a possible real vessel, although they do not match exactly with any active
unit in service. Moreover, this work and any results and conclusions, therein obtained and drawn, only reflect the personal
views of the authors and does not represent the views, policies or interests of any government or governmental authority
of any type or country.
1. Introduction
According to the British Standard (1), maintenance includes all actions necessary to retain a system or an item in, or
restoring it to, a state in which it can perform its required functions. The most common way of inflecting such concepts in
practice has been always deployed according to a “fix it when it breaks” approach (2). However, this has been becoming
an unaffordable and gold-brick methodology, since data gathering from the field is ever cheaper and costs related to a
breakdown may overcome the asset value (3; 4). Indeed, the last decades going-smart technology and cross-industry needs
in maintenance, e.g. ranging from manufacturing (5; 6) to the transportation domain (7; 8), has engendered a pivotal change
from a reactive to a proactive perspective (9), trespassing the original focus on repairing-replacing actions towards more
sophisticated preventive and prescriptive activities. In particular, maintenance actions can be framed into a taxonomy, which
includes three categories (10): Corrective, Preventive, and Condition-Based.
In Corrective Maintenance (CM), the equipment or asset is run down to its breaking-down point, and maintenance
activities are carried-out afterwards with the purpose of restoring the system at earliest. In this case, maintenance activities
are triggered by an unscheduled event (e.g. failure), and no a-priori strategies can be deployed. Consequently, costs related
to such approach are usually high: they comprise direct costs, e.g. due to potentially concatenated failures of other parts of
the asset, and indirect costs, e.g. related to potential losses in (environmental, worker, etc.) safety and integrity, and asset
unavailability.
Preventive Maintenance (PM), instead, is carried-out before breakdowns in order to avoid them and minimizing the
possibility of potential issues related to failures. Several variations exist, such as adjustments, replacements, renewals
and inspections, which take place subject to a predetermined planning and schedule: this allows to establish time-slots of
unavailability of an asset (or part of it), oppositely to the unpredictability which characterize random failure patterns in
CM. With particular reference to systematic PM, parts are replaced independently of their actual status, as a safety-level
life-time is established for them: in such a way, the probability of failures for a system decreases, as popularly proved
through adverting to the so-called bath-tub curve (11).
Condition-Based Maintenance (CBM), instead, refers to triggering maintenance activities as they are necessitated by
the condition of the target system (2). This approach enables determining the conditions of in-service assets to predict
potential degradations and to plan, consequently, when maintenance activities will be needed and should be performed to
minimize disruptions. In other words, CBM switch maintenance view from pure diagnosis to high-valued prognosis of
faults. CBM thus entails the diagnosis of the target system and timely, accurate, and valuable identification of potential
issues, by exploiting effective approaches from the predictive analytics domain (2; 3; 12).
While CM is usually avoided due to the high costs characterizing this approach, CBM is gaining popularity because
of the competitive advantages over PM, related to the extreme conservativeness of preventive methods in several domains
(leading, once again, to often unjustified remarkable costs). As a matter of fact, assets are typically featured by complex
interrelations among the parts of which they consist: such correlations are not straightforward to grasp and identify, thus
leading to the unfeasibility of a simple approach like the above-cited bath-tub model (2; 11). CBM, instead, proved to be
effective to maximize availability, efficiency, sustainability, and safety of assets, enabling the upgrading of maintenance
role from being seen as evil up to a way for creating added value (13; 14).
The CBM approach to maintenance is very generic, depicting an horizontal view to a crosscut problem in different
heterogeneous domains. When a more vertical view is taken into consideration, focussing on the maritime domain, repair
Running head right side 3
maintenance expenses for conventional ships can amount up to about 20% of total operability costs, including manning
expenses, insurance and administration costs (15). For naval applications, maintenance optimization depicts as a key
task, focused to reduce operations costs while getting the optimal availability of the ship for the intended service: such
optimization is the result of a trade-off between excessive maintenance and machinery down-time, where CBM helps
opening the door towards best balancing costs and availability. In other words, CBM enables a just-in-time deployment of
ship maintenance, by allowing to plan and execute maintenance activities only when needed.
The Condition-Based Maintenance approach requires that a reliable and effective diagnostic policy on one (or more)
naval asset(s) is implemented, in order to enable further refinement towards identifying potential failures in advance. Gas
Turbines (GTs), used for naval propulsion, represent a key example of how gathering data from components can be a driver
for optimizing maintenance strategies (16; 17). Diagnostic methods for GTs (18; 19) allow to obtain information on the
health status of the components without the need to inspect the machine, thus causing an interruption of operations with
consequent disruptions of services. Diagnosis represents the starting milestone to describe the status of the turbine, in order
to evolve such systems towards prognostic models, able to predict potential future issues with a reasonable advance. In
this sense, the use of these approaches assume a central importance when applied to propulsion systems in naval vessels:
the implementation of CBM strategies allows to improve the reliability and efficiency of the asset and to provide support
to maintenance operations in a non-invasive way and by limiting interventions, which are carried-out in accordance with a
when needed fashion.
The cornerstone of applying CBM consists in the adoption of a proactive approach to maintenance in place of more
conventional reactive interventions: to implement such key change, predictive modeling assumes a central importance to
allow automatic triggering of alarms without the need of designing complex, often unknown and, in some sense, unknowable
models. For this purpose, the use of robust Machine Learning (ML) models (20; 21) is here proposed and tested. These
models shown appealing modeling and predictive capabilities even in very complex and heterogeneous problem domains.
In particular, the authors focus on Regularized Least Square (RLS) (21) and Support Vector Machine (SVM) (20), i.e.
two state-of-the-art methods whose effectiveness in coping with real-world problems is also confirmed by the inclusion
in commercial software packages (e.g. it is the case of SVM, included in the ORACLE Data Mining suite (22)). RLS and
SVM models are induced only from data collected from the field on the assets that need to be modeled, i.e. the predictive
model does not require any a-priori knowledge on the turbines, their architecture/configuration or on the system overall.
In this paper, the performance and potentialities of RLS and SVM models are benchmarked by training them on
data to forecast the performance decay of Gas Turbines, installed on a naval vessel. The aim of this work is to pave the
way towards designing a prognostic CBM methodology to naval propulsion plants equipped with GTs. In particular, the
propulsive performance degradation of a naval vessel powered by a GT has been simulated, between two drydocks, in
order to make available and to collect a dataset of experiments. Such data are then exploited to derive a predictive model
to automatically detect the system decay state, targeting effective CBM in the considered domain.
For these purposes, the paper is organized as follows. Section 2 introduces the model and the simulator of the naval
vessel, considered for presenting the ML approach to CBM in the subsequent analysis. Then, Section 3 focuses on the
Machine Learning (ML) framework, by introducing the main approaches that will be exploited for this work’s targets:
actionable algorithms will be proposed, which can be implemented for forecasting purposes in CBM applications. With
reference to the characteristics of the simulator of the naval vessel and of the needs, emerging from the ML framework, a
dataset has been created, according to the setup described in Section 4: this dataset will be released for the use to the research
community on the widespread well-known dataset repository of the University of California in Irvine (UCI). Section 5
subsequently presents some experimental results, derived by applying the proposed ML approaches to the dataset, while
Section 6 will finally draw the conclusions.
4 Journal name 0(0)
Gas Generator
Fuel
C HP Brake
LP
Shaft Shaft Torque
Air
Inlet
Telegraph
Ship speed (v)
Rt Tp
Gearbox
n Engine
Qp Qe
Fig. 3. One degree of freedom ship dynamics
where where Tp is the propulsor thrust, p is the number of propulsors, Rt is the total ship resistance, M is the ship mass,
Qe is the engine torque, i is the gear ratio, e is the number of engines per shaft, Qp is the propulsor torque and Jp is the
polar moment of the rotating masses.
The first equation represents the longitudinal motion of the ship, the second equation represents the rotational motion
of the shaft line , finally the third equations represents the control system behaviour. Forces and torques in Equation 1 can
be represented in different ways depending on: the type of the propulsion system, the type of analysis and the expected
degree of accuracy. Among the possible available approaches, in order to reduce the calculation time, the inter-component
volumes method has been adopted, preferring it to more sophisticated models for the GT dynamic simulation. For the
simulation of the General Electric LM 2500 gas turbine this procedure has been applied (25) and in order to validate the
model, steady state and transient results obtained by the simulator have been compared with experimental data provided
by the gas turbine manufacturer.
The gas turbine mathematical model is structured in a modular arrangement, as shown in Figure 4 where the Simulink®
top level scheme of the considered gas turbine is reported. In the same figure the input and output variables of the different
plant components such as compressor, high pressure turbine, power turbine, combustor and so on are shown. It is worth
mentioning that each module, representing a specific engine component, has been modelled by means of steady state
performance maps, time dependent momentum, energy and mass equations and non linear algebraic equations.
The main differential equations to be solved in the time domain refer to both dynamic and thermodynamic engine
parameters. In particular the working fluid processes are governed by the continuity equation:
d(ρ · V ) · ρ
= Mi − M o (2)
dt
6 Journal name 0(0)
Mf% Mf% 1
2 TIC % Pexh
T6bl T6bl
TIC % p6 p6
Mf% Mf% p6 p6
GGrpm GGrpm
T48 T48 M6 M6
1 GT rpm p5 p5
GT rpm OUTLET
M5 M5 1
T2r T2r
GT_T GT_T GT Torque
2 TCS_TIC
Mblc Mblc
T48 TCS_TIC Torque Torque Torque TOT GT_T
T48
GG rpm GG rpm
Mf% Mf% p2
Mfl Mfl p4 10
8 T2r T2r T3 T3
Mf% Mf%
T4bl T4bl
5 M2 p3 p3
p2 T2r T2r M2
T48 T48
M3 mf T3 M4 M4
M3 4 T3 M4 M4
M2 M2 p1 p1 M2 M2
mf
p3 p3
COMBUSTOR M5 M5 p5 p5
3 Ta T1 T1 Mblc Mblc T2r T2r M3 M3
Ta
3 3 HP - LP
INLET Mblc Mblc
T1 GGrpm GGrpm C Torque C_T GGrpm VOLUME
HP_T HP_T
GGrpm GGrpm Hp Torque HP_T
GGrpm GGrpm
COMPRESSOR
C_T C_T
HP TURBINE
GAS GENERATOR
SHAFT DYNAMICS
where Mi and Mo are the inlet and outlet mass flow rates, V is the GT component volume, h and u are enthalpy and internal
energy, while P and Q0 represent power and heat flow respectively.
The dynamics of the rotating shafts is determined by the dynamic momentum equation:
dω 1
= (Pm − Pbr ) (4)
dt J ·ω
being Pm and Pbr the motor and brake powers and ω the angular velocity. The propulsion system is managed by a typical
combinator law whose general layout is presented in Figure 5. The control system has been designed to provide a linear
relationship between lever position and ship speed. In this application the ship propulsion control system dynamics has
Running head right side 7
where: TIC is the control signal (i.e. % of fuel flow) for the propulsion engine, n and nsp are the actual and the set
point propeller shaft speed respectively, while Kp and Ki represent the PI control loop constants. This is only a rough
approximation, but can be considered satisfactory for the most important ship maneuvers. Finally it worth noting that model
validation for a ferry, a naval corvette and an aircraft carrier can be found in (29; 30; 31).
The hull and propellers degradation here are not take into account. An improvement of the model is under development for
introducing this phenomena in the propulsion plant simulator.
For this condition based maintenance application let us consider the propulsion system identified by the followings
parameters:
1. Ship speed v: this parameters allows to take into account different GT load conditions. The authors choose to consider
different probability density function (PDF) for the speed distribution as there are three variants for this kind of naval
vessel: Anti-submarine Warfare (ASW) variant, general-purpose variant (GP) and Anti-aircraft warfare (AAW) variant.
In particular three different probability function density has been take into account: uniform PDF for the GP variant,
bimodal PDF for the AAW and ASW variant. A further analysis has been performed by means of a unimodal distribution
for the cruising speed.
2. GT Compressor degradation coefficient kMc : this parameters describe the reduction, over service hours, of the numerical
values of the airflow rate Mc and of the isentropic efficiency ηc .
3. Gas Turbine degradation coefficient kMt : this parameters describe the gas flow rate reduction factor over service hours.
In Table 1 the main simulation outputs are reported. This subset of model’s outputs are the same quantities that the
automation system installed onboard can acquire and store.
As far as numerical model is concerned it is worth noting that the ship dynamics part of the model has been validated
and fine tuned comparing the simulation results with sea trial campaign data, measured during the full scale test campaign
of the vessel, as reported in (36).
1. during the training phase, a set of data is used to induce a model that best fits them, according to some criteria;
2. then, the trained model is set in motion into the real application during the feed-forward phase.
Running head right side 9
In this paper, authors focus on the problem of identifying the decay trend in Gas Turbines performance, when GTs
are used for naval propulsion. In this sense, a regression problem must be solved, where the decay trend over time
must be estimated based on data, collected from GTs and, more broadly speaking, from the vessel. In particular, in the
conventional
n oregression framework in ML (20; 37; 38), a set of training data X = {(xi , yi )}, i = 1, ..., n, where in general
(k)
x i = xi ∈ Rd (k = 1, ..., d) and yi ∈ R is considered. As the authors are targeting a regression problem, the purpose
is to find the best approximating function h(x), where h : Rd → R, which should be close (in some sense) to the unknown
target function f (x).
During the training phase, the quality of the learned regressor h(x) is usually measured according to a loss function
`(h(x), y) (39), which calculates the discrepancy between the true output y and the estimated one. The empirical error then
computes the average discrepancy per sample, reported by a model on the training patterns:
n
1X
L̂n (h) = `(h(xi ), yi ). (10)
n i=1
A simple criterion for selecting the final model during the training phase could then consist in simply choosing the
approximating function that minimizes the empirical error L̂n (h): this approach is known as Empirical Risk Minimization
(ERM) (40). However, ERM is usually avoided in ML as it leads to severe overfitting of the model on the training dataset: as
a matter of fact, in this case the training process chooses a model, complicated enough to perfectly describe all the training
samples (including noise, which afflicts them). In other words, ERM implies memorization of data rather than learning
from them.
A more effective approach is represented by minimizing a cost function where the tradeoff between accuracy on the
training data and a measure of the complexity of the selected approximating function is implemented (41; 42)1 :
In other words, the best approximating function h∗ is chosen as the one that is complicated enough to learn from data
without overfitting them. In particular, C(·) is a complexity measure: depending on the exploited ML approach, different
measures are realized2 . Instead, λ is a hyperparameter, that must be aprioristically set and is not obtained as an output of
the optimization procedure: it regulates the trade-off between the overfitting tendency, related to the minimization of the
empirical error, and the underfitting tendency, related to the minimization of C(·). The optimal value for λ is problem-
dependent, and tuning this hyperparameter is a non-trivial task: the authors will devote a subsection in the following for
this purpose.
Note that, when monitoring the decay of complex environments and systems like in CBM applications, the authors
could have to deal with Multiple Output (MO) problems instead of the Single Output (SO) ones, considered so far: in the
MO framework, the objective value of the decay is an array y i ∈ Rm , with m > 1. Although several approaches, such as the
ones based on Multi-Task Learning (44) or Transfer Learning (45), can be applied to training models in the MO framework,
the most effective solution often consists in decomposing the MO dataset into m SO problems (46; 47). In addition to allow
simplifying the presentation of the ML method, this approach enables grasping the phenomena, characterizing the decay
of components for CBM purposes, whenever the considered outputs are not very strongly correlated one to each other: as
this hypothesis sounds reasonable in the considered application, the authors opt for following this path.
where a non-linear mapping φ : Rd → RD , D d, is applied in such a way non-linearity is pursued while still coping
with linear models.
For RLS, Problem (11) is configured as follows. The complexity of the approximation function is measured as
i.e. the Euclidean norm of the set of weights describing the regressor, which is a quite standard complexity measure in ML
(e.g. (20)). Regarding the loss function, the Mean Squared Error (MSE) loss is adopted:
n n
1X 1X 2
L̂n (h) = `(h(xi ), yi ) = [h(xi ) − yi ] . (14)
n i=1 n i=1
By exploiting the Representer Theorem (50), the solution h∗ of the RLS Problem (15) can be expressed as a linear
combination of the samples projected in the space defined by φ:
n
X
∗
h (x) = αi φ(xi )T φ(x). (16)
i=1
It is worth underlining that, according to the kernel trick (51; 37), it is possible to reformulate h∗ (x) without an explicit
knowledge of φ by using a proper kernel function K(xi , x) = φ(xi )T φ(x):
n
X
h∗ (x) = αi K(xi , x). (17)
i=1
Several kernel functions can be retrieved in literature (51; 37), but the Gaussian kernel is often used as it enables learning
every possible function (52):
2
K(xi , xj ) = e−γkxi −xj k2 , (18)
where γ is a hyperparameter, that must be aprioristically tuned (analogously to λ). In particular, for small values of γ,
simpler functions are realized (note that for γ → 0 the authors derive a linear regressor); instead, if large values of γ are
employed, more complicated functions can be realized.
Running head right side 11
As a final issue, the RLS Problem (15) can be reformulated by exploiting kernels:
2
n n n X n
1 X X X
α∗ : min αj K(xj , xi ) − yi + λ αi αj K(xj , xi ). (19)
α n i=1 j=1 i=1 j=1
Let us define y = [y1 , . . . , yn ]T , α = [α1 , . . . , αn ]T , the matrix K such that Ki,j = Kji = K(xj , xi ), and the Identity
matrix I ∈ Rn×n , a matrix-based formulation of Problem (19) can be obtained:
1 2
α∗ : min kKα − yk2 + λαT Kα (20)
α n
By setting equal to zero the derivative with respect to α, α can be found as follows:
(K + nλI) α∗ = y, (21)
that is a linear system for which effective solvers have been developed throughout the years, allowing to cope with even
very large sets of training data (53).
where |·|ε = max (0, |·| − ε). On the other hand, the same complexity measure, exploited by RLS, is used by SVR.
As a consequence, Problem (11) can be reformulated as follows:
n
1 X
w∗ : min kwk22 + C ξˆi + ξˇi (23)
w,ξ̂,ξ̌ 2 i=1
1
where C = λnis conventionally adopted as hyperparameter in place of λ. The two constraints of Eqns. (24) and (25)
are added to force learning of approximation functions which are close to the target one: note, in fact, that ξˆi , ξˇi are slack
variables.
3 Note that the originally proposed function for SVR is h(x) = wT φ(x) + b, where b ∈ R is a bias term. However, the introduction of the bias would
unnecessary complicate the analysis of this case: in fact, it is possible to prove that the theoretical framework of SVM is maintained also if b = 0 as
far as a positive definite kernel, like the Gaussian one of Eq. (18), is used (54).
12 Journal name 0(0)
Note that Problem (23) is convex (20). Nevertheless, for computational reasons, the dual formulation of the SVR
problem above is usually adopted, which can be derived by introducing the Lagrange Multipliers α̂, α̌, η̂, η̌ (55):
1 Xn
L w, ξ̂, ξ̌, α̂, α̌, η̂, η̌ = kwk22 + C ξˆi + ξˇi (26)
2 i=1
n
X h i
+ α̂i wT φ(xi ) − yi − ε − ξˆi
i=1
n
X
α̌i yi − wT φ(xi ) − ε − ξˇi
+
i=1
n
X n
X
− η̂i ξˆi − η̌i ξˇi
i=1 i=1
By exploiting these relationships in Eq. (26), the dual formulation of Problem (23) is derived, for which several efficient
solvers have been developed throughout the last years (56; 57):
n n
1 XX
α̂∗ , α̌∗ : min (α̌i − α̂i ) (α̌j − α̂j ) φ(xi )T φ(xj )
α̂,α̌ 2 i=1 j=1
n
X n
X
+ α̂i [yi + ε] − α̌i [yi − ε] (36)
i=1 i=1
0 ≤ α̌i ≤ C, ∀i ∈ {1, . . . , n}
0 ≤ α̂i ≤ C, ∀i ∈ {1, . . . , n}
The trained approximating function h∗ (x) can be expressed through α̂∗ , α̌∗
n
X
h∗ (x) = (α̌i∗ − α̂i∗ ) φ(xi )T φ(x). (37)
i=1
Running head right side 13
It is worth noting that, if |wT φ(xi ) − yi | ≤ ε, it derives from Eqns. (33) and (34) that α̂i = α̌i = 0: as a consequence, the
solution of SVR is usually sparse, i.e. it can be described by exploiting only a limited subset of parameters (and, obviously,
of training data). In particular, the less data are outside the ε–insensitive tube, the more sparse the SVR solution will be
(20; 37; 38). In other words, let Ssv = {i : α̂i = α̌i = 0} be the set of indexes, and |Ssv | ≤ n (and, usually, |Ssv | n)
its cardinality; then, in order to compute h∗ (x), only data indexed by Ssv can be contemplated:
X
h∗ (x) = (α̌i∗ − α̂i∗ ) φ(xi )T φ(x). (38)
i∈Ssv
This allows to reduce the computational burden with respect to RLS, for which the solution is dense (i.e. almost all αi 6= 0
in Eq. (21)).
As a final issue, analogously to what done for RLS, the kernel trick can be exploited in order to reformulate the SVR
training problem by exploiting a matrix formulation:
" #T " #" # " #T " #
∗ ∗ 1 α̌ K −K α̌ α̌ −y + ε
α̂ , α̌ : min + (39)
α̂,α̌ 2 α̂ −K K α̂ α̂ y+ε
0 ≤ α̌, α̂ ≤ C
set of indexes of the samples in X; then, k sets of indexes S (t) , t = 1, ..., k, are created. These subsets are such that:
k
[ k
\
(t)
S =S ∨ S (t) = ∅. (40)
t=1 t=1
The best error for the KCV procedure is initialized to L∗KCV = +∞. Then, for every possible set of values in the grid for the
hyperparameters tuple, (λ, γ) for RLS or (γ, C, ε) for SVR, a KCV stage is run. At the t-th KCV step, the patterns in the
validation set are indexed by S (t) , while the training set consists of the remaining samples, whose indexes are S \ S (t) .
A model is then trained on the training set and its performance is verified on the validation patterns by computing the
empirical error
(t) 1 X (t)
L̂n/k (h(t) ) = ` h (xi ), yi . (41)
n/k (t)
i∈S
After the k KCV steps are over, the estimated generalization error for the hyperparameters tuple can be computed by
averaging the t empirical errors on the validation sets:
k
1 X (t)
LKCV = L̂ (h). (42)
k t=1 n/k
If the obtained error is lower than the best L∗KCV reported so far, the hyperparameters tuple is saved as potential best
configuration for the learning procedure.
As a final issue the authors want to point out that, in order to select the best hyperparameters tuple, k models are created
at every KCV step. Once the best tuple is found, the final model is trained on the whole set X by running the learning
procedure where hyperparameters are set to the identified best values.
4. Dataset Creation
In this paper, the authors aim is to develop a predictive model for CBM purposes with reference to Gas Turbines mounted
on a Frigate naval vessel. Since the COmbined Diesel eLectric And Gas (CODLAG) propulsion plant, mounted on the
Frigate, is of recent installation, no real-world data gathered from ship’s automation system are currently available, which
could help studying and validating a Machine Learning solution to the Condition-Based Maintenance needs of this asset.
Nevertheless, much of the effort has been spent in realizing the physical model and, consequently, an accurate simulator
of the system, as introduced in Section 2 as well as in previous works (23; 24). As a consequence, the model of Section
2 can be exploited in order to create a virtual (but ground-truth complying (23; 24)) experimental datasets, to be used for
ML modeling purposes.
As described in Section 2.3, the propulsion system is depicted through three parameters:
1. Speed: this parameter is controlled via the control lever. The latter can only assume a finite number of position lpk , k ∈
{0, . . . , 9}, which in turn correspond to a finite set of possible configurations for fuel flow and blade position. Each set
point is designed to reach a desired speed vk , k ∈ {0, · · · , 9}:
needed by the learning procedure. Nevertheless, some practical suggestions can be retrieved in literature (e.g. (66)), which help finding good trade-offs
between performance and computational time.
Running head right side 15
Algorithm 1 The k–Fold Cross Validation (KCV) model selection procedure for RLS.
Require: X, k, a grid for every hyperparameter involved Gλ,γ
1: L∗KCV = +∞
2: (λ∗ , γ ∗ ) = (−∞, −∞)
3: S (t) = split(S), t = 1, ..., k
4: for all λ ∈ Gλ , γ ∈ Gγ do
5: Let λ(p) , γ (q) be the current hyperparameters tuple
6: LKCV = 0
7: for t = 1 → k do
8: Train h(t) with RLS Problem (21) on the set of data indexed by S \ S (t) , by using the hyperparameter tuple
λ(p) , γ (q)
(t)
9: Compute L̂n/k (h(t) ) according to Eq. (41)
10: end for
11: Compute LKCV according to Eq. (42)
12: if LKCV < L∗KCV then
13: L∗KCV = LKCV
(λ∗ , γ ∗ ) = λ(p) , γ (q)
14:
15: end if
16: end for
17: Train h∗ with RLS Problem (21) by using the whole set X and the tuple (λ∗ , γ ∗ )
18: return h∗
Algorithm 2 The k–Fold Cross Validation (KCV) model selection procedure for SVR.
Require: X, k, a grid for every hyperparameter involved Gγ,C,ε
1: L∗KCV = +∞
2: (C ∗ , γ ∗ , ε∗ ) = (−∞, −∞, −∞)
3: S (t) = split(S), t = 1, ..., k
4: for all C ∈ GC , γ ∈ Gγ , ε ∈ Gε do
5: Let C (p) , γ (q) , ε(s) be the current hyperparameters tuple
6: LKCV = 0
7: for t = 1 → k do
8: Train h(t) with SVR Problem (39) on the set of data indexed by S \ S (t) , by using the hyperparameter tuple
C (p) , γ (q) , ε(s)
(t)
9: Compute L̂n/k (h(t) ) according to Eq. (41)
10: end for
11: Compute LKCV according to Eq. (42)
12: if LKCV < L∗KCV then
13: L∗KCV = LKCV
(C ∗ , γ ∗ , ε∗ ) = C (p) , γ (q) , ε(s)
14:
15: end if
16: end for
17: Train h∗ with SVR Problem (39) by using the whole set X and the tuple (C ∗ , γ ∗ , ε∗ )
18: return h∗
16 Journal name 0(0)
Note that, if the transients is not taken into account, lpi and vi are deterministically related by a linear law. In the
presented analysis the transients between different speeds are not considered since, in the operational life of a ship,
these phenomena cover a negligible time with respect to steady states.
2. Compressor decay:
The decay rate has been sampled through a uniform grid with resolution 10−3 , so to reach a good granularity when
representing the GT Compressor decay state:
Analogously to the case of compressor, the decay rate is sampled through a uniform grid with resolution 10−3 . The GT
decay state can be then represented as:
Once these quantities are fixed, several heterogeneous measures, related to the propulsion plant and listed in Table 1, can
be derived by exploiting the simulator, punctually describing the state of the system. The purpose of the analysis is to train
a model on such measures, in order to estimate kMc and kMt from them. It is worthwhile noting that the measures, derived
by the simulator, can also be obtained by the automation system equipping the ship, and could be used in a real-world
application scenario as well.
Given the above premises, the evolution of the system can exhaustively explored by simulating all its possible states.
The space of possible states is described via the triples:
since:
It is then possible to run the simulation for the 9 · 51 · 26 = 11934 conditions considered in the Matlab® Simulink®
environment for each triple [lp, kMc , kMt ]i . Every model is run until the steady state is reached: then, the above-cited 16
measures listed in Table 1 can be extracted from the simulation and finally organized into a dataset. The latter defines a
conventional Multiple Output (MO) problem, where every sample xi ∈ R16 and every output is y i ∈ R2 , as it includes
the values of kMc and kMt .
The authors intend to publish this dataset for free use by the scientific community through the well-known UCI repository
(67). While waiting for final approval for publication, news and preliminary releases of the dataset will be soon available
at the dedicated webpage http://cbm.smartlab.ws.
Running head right side 17
In order to train an effective model and to test its performance in a real-world-like scenario, the dataset X of available
experiments is divided into two subsets:
• Xtrain , consisting of ntrain samples, which is used for the training phase (thus including model selection and creation)
• Xtest = X \ Xtrain , consisting of ntest = n − ntrain data. The feedforward phase of the model h∗ , trained on Xtrain , is run
on these samples analogously to an application setting of the regressor. As the target outputs are available for Xtest , it
is always possible to compute the error that h∗ would obtain in the application scenario, which is defined as:
∗ 1 X
L̂test
ntest (h ) = ` (h∗ (xi ), yi ) . (56)
ntest
xi ∈Xtest
In particular, the authors opt for exploiting the relative error (68; 69) as loss function, which can be expressed as:
∗
∗
h (xi ) − yi
` (h (xi ), yi ) =
(57)
yi
The splitting of the dataset into training and test sets is a key point to deal with. In particular, two aspects must be considered:
(i) the choice of ntrain , i.e. the amount of data for training purposes; (ii) the way data are picked from X to be included into
Xtrain . Both points can be analyzed from the modeling and application point of view. Regarding issue (i) above:
• From the Machine Learning point of view, choosing the most proper number of data for training and test purposes has a
non-negligible impact on the quality of the final model: if ntrain is too large, the model will generally be more accurate,
but the test assessment probably will not be relevant from a statistical point of view; on the contrary, if too samples are
reserved for testing purposes, the regressor could become too inaccurate, being trained on a limited knowledge-base.
Usually, a feasible approach consists in performing some trials, by using different increasing values for ntrain , in order
to find the best trade-off.
• From the naval application point of view, it is worth underlining that the role of the simulator would be played by
an expert operator, who checks the decay state of the component and reports any measures, coming from the ship’s
automation, into a database. The purpose of the ML predictive approach for CBM, targeted by this paper, is to support
18 Journal name 0(0)
the expert in such monitoring operations. Determining the value of ntrain is equivalent to quantifying the number of
manual inspection the expert has to complete before the model can be trained, or, in other words, before he is supported
in his activities by the ML predictive model for CBM. Too large values for ntrain would require a remarkable effort by
the expert and/or long waiting time before operating the model, reducing the potential impact of such technology in
the maritime domain.
• From the ML point of view, data can be picked either in a random way or through a more rigorous procedure, which
exploits some information on the observed phenomenon. Random selection is obviously easier to implement, but
usually leads to poorer quality models, as the space of possible values for measures and decay rates is not exhaustively
explored. Approaches, allowing to explore the topology representation of data, usually enable deriving more effective
regressors at the expenses of a small supplementary delta-effort: it is the case, for example, of the Nearly Homogeneous
Sampling method (70), that can be exploited in the experiments.
• The way data are sampled in this virtual framework maps into the protocol for data gathering for maintenance purposes
in the real-world application. In other words, a random sampling corresponds to an expert operator that checks for
the health status of the vessel components without following any predetermined and/or rigorous time-schedule and/or
procedure; on the contrary, using more sophisticated sampling techniques is equivalent to an operator, who follows
a planned schedule on predetermined operational conditions for monitoring the vessel. Though this second scenario
seems to be more plausible, both settings will be analyzed in the following.
As described in Section 2.3, the Frigate can be characterized by different mission profiles (AAW, ASW, GP), which
correspond to different operational profiles: they are distinguished by the different probability density function, used for
depicting the distribution of the speed. Such information cannot be straightforwardly retrieved in the dataset, created as
described in the previous section. Nevertheless, the whole space of potential operating conditions is mapped, it is always
possible to resample the dataset according to different speed distributions, designed with reference to the different mission
profiles. Unfortunately, such profiles are confidential information not available in public literature: however, it is always
possible to design some reasonable distributions, based on authors’ expertise. In particular the following scenarios will be
considered:
1. a uniform distribution (UD) (Figure 6), characterizing a general purpose mission of the Frigate;
2. a bimodal distribution (BD) (Figure 7), where the two modes are centered, respectively, on the cruising speed and on
the maximum speed (while other speed values are neglected);
3. a monomodal distribution (MD) (Figure 8), where the mode is centered over the cruise speed and the time, spent by the
vessel at different speed values, is considered negligible.
For the sake of completeness, it is worth noting that the ASW mission profile can be characterized by a trimodal distribution:
in this case, another mode is introduced in addition to the ones of the BD profile, centered on the optimal speed for the
antisubmarine operations. However, in this analysis, the last case is not taken into account, since the vessel exploits the
electrical prolusion mode, which is not contemplated by the Matlab® Simulink® model, for reaching the first peak (71; 72)
in place of the Gas Turbine mode. Consequently, the ASW mission profile can still be reasonably represented by BD, as
introduced above.
Table 2 shows the results for the experiments, where only GT compressor decay is taken into account. In particular,
∗
the error on the independent test set L̂test
ntest (h ) is reported. The error is computed for different experimental setups, listed
below:
• different experiments are performed by varying the amount of patterns used for training purposes (refer to issue (i),
introduced above). In particular, ntrain ∈ {10, 25, 50, 100, 200} were tested.
Running head right side 19
0.8
0.6
PDF
0.4
0.2
Fig. 6. Uniform distribution (UD), used as probability density function of the speed (percentage respect to the design speed equal to 15
Knots) of the Frigate.
0.8
0.6
PDF
0.4
0.2
Fig. 7. Bimodal distribution (BD), used as probability density function of the speed (percentage respect to the design speed equal to 15
Knots) of the Frigate.
20 Journal name 0(0)
0.8
0.6
0.4
0.2
Fig. 8. Monomodal distribution (MD), used as probability density function of the speed (percentage respect to the design speed equal
to 15 Knots) of the Frigate.
• the experiments has been replicated for different values of ntrain , by opting for different sampling strategies (refer to
issue (ii) above). In particular a random sampling and the Nearly Homogeneous Sampling (70) strategies is exploited.
Note, moreover, that the splitting procedure for creating Xtrain and Xtest from X is replicated 30 times per each
experiment, so to derive statistically relevant results
• the experiments were replicated by varying the probability distribution of vessel speed according to the three
distributions, previously introduced (UD, BD, MD)
• the RLS and SVR learning procedures are used, respectively presented in Algorithms 1 and 2, to derive the model h∗ .
In particular, the authors set:
– RLS: the grids for hyperparameters tuning purposes are fixed to Gλ ∈ [10−6 , 104 ] and Gγ ∈ [10−6 , 104 ], consisting
of 30 values equally spaced in logarithmic scale
– SVR: the grids are fixed to GC ∈ [10−6 , 104 ] and Gγ ∈ [10−6 , 104 ], consisting of 30 values equally spaced in loga-
rithmic scale. Note that SVR requires that the hyperparameter ε is tuned as well. Nevertheless, this hyperparameter
is application-dependent and can be usually aprioristically set to a fixed value without compromising performance
while reducing the computational burden. The values of ε is set to ε = 10−3 , i.e. equal to the considered granularity
when modeling the system behavior.
Table 3 presents the results for the set of experiments, performed on the data: in particular, the average error and its
variance, as recorded on the different performed analyses, are shown. From them, it is worthwhile deriving that:
• Broadly speaking, SVR is characterized by better performance than RLS: as a matter of fact, SVR average error is
smaller, and the variance values for SVR show that this approach is more stable than RLS as the training set is changed
(e.g. because of different speed distributions or of different sampling strategies)
• Model derived from datasets, sampled according to the Nearly Homogeneous Sampling (70) approach, are usually
more accurate. However, the difference with respect to the performance of regressors, trained on randomly sampled
training sets, is negligible: such result allows to derive precious information for designing the procedures for CBM
data gathering on the Frigate
Running head right side 21
• On equal terms (e.g. value of ntrain ), if the uniform distribution (UD) or the vessel speed is contemplated, that takes
into account a broad range of potential speed values, the forecasting is less accurate than in those experiments, where
more concentrated distributions are exploited (e.g. monomodal and bimodal distribution – MD and BD, respectively).
This is expected: in fact, the use of MD and BD simplifies the problem by neglecting several different speeds, that are
considered, on the contrary, by UD. This allows to highlight the importance of correctly and informatively depicting
the speed distribution of the vessel, as a basic condition for deriving more effective models with a smaller amount of
gathered data
• Finally, ntrain ≈ 50 samples are usually enough, in the proposed approaches, to derive an accurate model for CBM
purposes, in this preliminary study framework.
The experimental setup is not changed from the one, depicted in the previous section; however, as a very partial
exception, the case ntrain = 200 is negletted due to the more limited amount of available data (234 vs 459). In fact, when
ntrain = 200, only 60 samples are available for testing purposes, which could be too few to target a reliable analysis.
Table 3 shows the results, obtained from the performed tests: the quantities, presented in the table, are analogous to the
ones of Table 2, and so are the conclusions that can be drawn from them.
5.3. Modeling and predicting both the GT compressor and the GT decays
As a final issue, the case where both decay factors are taken into account are analyzed, so depicting a Multiple Output
(MO) problem with m = 2 (refer to Section 3 for more details). By exploiting the approach proposed by (46; 47), i.e.
by decomposing the MO problem into m Single Output (SO) sub-problems, a models to estimate kMc and kMt in the
general CBM scenario is created. The objective is to verify whether the models are able to accurately track variations
in GT parameters, due to decay, even when two components are degrading their performance (and not necessarily in a
synchronous fashion).
The experimental setup is kept unmodified with reference to the one, applied for previous experiments. Clearly, at this
stage, the authors considered the whole dataset generated according to the considerations, reported in Section 4: having
11934 samples available, larger cardinalities for the training set are tested, that is ntrain ∈ {10, 25, 50, 100, 205, 500}. The
results are shown in Tables 4 and 5, which are respectively referred to the Gas Turbine compressor and the Gas Turbine
decay. Note that:
• In spite of the contemporaneous variations of two decay factors, the trend of the error, performed by the trained models,
is qualitatively the same as the one, observed when contemplating one decay at a time. This validates the choice of
opting for the decomposition of the MO problem into separate SO ones
• Nevertheless, slightly larger number of training samples is generally needed to create effective models. As a matter of
fact, around 100 data are typically required to gain low error rates, which can rise up to 100 ÷ 200 when a UD is used
to describe the vessel speed
22 Journal name 0(0)
Table 2. Test error rates (in percentage) when only the Gas Turbine compressor decay is considered.
Random Sampling
UD BD MD
RLS SVR RLS SVR RLS SVR
∗ ∗ ∗ ∗ ∗ ∗
ntrain L̂test
ntest (h ) L̂test
ntest (h ) L̂test
ntest (h ) L̂test
ntest (h ) L̂test
ntest (h ) L̂test
ntest (h )
10 30.23 ± 5.57 27.57 ± 5.18 5.36 ± 1.95 4.81 ± 1.70 0.96 ± 0.02 0.87 ± 0.02
25 12.39 ± 2.50 11.16 ± 2.28 1.08 ± 0.22 0.98 ± 0.20 0.44 ± 0.00 0.39 ± 0.00
50 5.13 ± 0.57 4.67 ± 0.52 0.34 ± 0.01 0.31 ± 0.01 0.28 ± 0.00 0.24 ± 0.00
100 1.59 ± 0.15 1.42 ± 0.13 0.17 ± 0.00 0.15 ± 0.00 0.13 ± 0.00 0.12 ± 0.00
200 0.50 ± 0.01 0.45 ± 0.01 0.08 ± 0.00 0.08 ± 0.00 0.04 ± 0.00 0.04 ± 0.00
UD BD MD
RLS SVR RLS SVR RLS SVR
∗ ∗ ∗ ∗ ∗ ∗
ntrain L̂test
ntest (h ) L̂test
ntest (h ) L̂test
ntest (h ) L̂test
ntest (h ) L̂test
ntest (h ) L̂test
ntest (h )
10 25.96 ± 4.93 23.59 ± 4.56 4.61 ± 1.60 4.28 ± 1.43 0.83 ± 0.02 0.74 ± 0.02
25 10.55 ± 2.16 9.35 ± 1.99 0.91 ± 0.19 0.82 ± 0.17 0.37 ± 0.00 0.33 ± 0.00
50 4.47 ± 0.49 4.10 ± 0.44 0.30 ± 0.01 0.27 ± 0.01 0.23 ± 0.00 0.21 ± 0.00
100 1.35 ± 0.13 1.20 ± 0.11 0.14 ± 0.00 0.13 ± 0.00 0.12 ± 0.00 0.10 ± 0.00
200 0.42 ± 0.01 0.37 ± 0.01 0.07 ± 0.00 0.07 ± 0.00 0.04 ± 0.00 0.03 ± 0.00
• Finally, similar conclusions to the ones, drawn in the previous subsections, can be made.
6. Conclusions
The impact of maintenance cost in the naval domain is remarkable: as a consequence, this enables the necessity of designing
more effective approaches, for example based on Condition-Based Maintenance (CBM). Relying on the need of constantly
monitoring the amounts of data, gathered from the field, CBM strategies require that proper Machine Learning (ML) models
are implemented, able to highlight potential safety issues and/or intervention requests with a reasonable advance. In this
paper, in particular, the authors focussed on an application of CBM to Gas Turbines (GTs) used for naval vessel propulsion
to propose an innovative modeling strategy based on effective ML approaches, targeted towards forecasting the degradation
of the propulsion plant over time.
In order to test the proposed techniques, a sophisticated and realistic simulator of a naval propulsion system is exploited,
mounted on a Frigate characterized by a COmbined Diesel eLectric And Gas (CODLAG), to generate a large dataset, that
will be soon released to the research community for benchmarking purposes. The proposed approach has been tested on
this dataset, showing its effectiveness for CBM purposes in a real-world maritime application, even in the challenging
scenario where more than one decay factor is varied over time.
The approach and results, shown in this paper, clearly represent a first step to pave the way towards effective and
actionable CBM systems, powered by ML capabilities. Future works will mostly concern the extension of the approach
to contemplate the decay of a larger number of components, which drive the maintenance needs of the propulsion plant;
moreover, validation with data, gathered on-field, will also allow to validate the ground-truth compliancy of the estimation
with real-world evidences. Such activities will open the door to further evolutions of the proposed CBM approach, so to
predict, with a reasonable advance, potential breakdowns without requiring a preliminary and propaedeutic labeling activity
by the operators (i.e. exploiting an unsupervised method (73)).
Running head right side 23
Table 3. Test error rates (in percentage) when only the Gas Turbine decay is considered.
Random Sampling
UD BD MD
RLS SVR RLS SVR RLS SVR
∗ ∗ ∗ ∗ ∗ ∗
ntrain L̂test
ntest (h ) L̂test
ntest (h ) L̂test
ntest (h ) L̂test
ntest (h ) L̂test
ntest (h ) L̂test
ntest (h )
10 29.77 ± 5.26 27.07 ± 4.84 4.14 ± 1.44 3.77 ± 1.29 0.50 ± 0.00 0.45 ± 0.00
25 15.41 ± 4.45 13.94 ± 4.08 0.40 ± 0.01 0.36 ± 0.01 0.38 ± 0.00 0.34 ± 0.00
50 4.50 ± 1.35 4.07 ± 1.24 0.28 ± 0.03 0.25 ± 0.03 0.21 ± 0.00 0.18 ± 0.00
100 1.60 ± 0.17 1.44 ± 0.16 0.11 ± 0.00 0.09 ± 0.00 0.04 ± 0.00 0.03 ± 0.00
UD BD MD
RLS SVR RLS SVR RLS SVR
∗ ∗ ∗ ∗ ∗ ∗
ntrain L̂test
ntest (h ) L̂test
ntest (h ) L̂test
ntest (h ) L̂test
ntest (h ) L̂test
ntest (h ) L̂test
ntest (h )
10 25.52 ± 4.60 23.25 ± 4.23 3.63 ± 1.20 3.38 ± 1.12 0.43 ± 0.00 0.38 ± 0.00
25 13.20 ± 3.88 11.67 ± 3.50 0.34 ± 0.00 0.31 ± 0.00 0.32 ± 0.00 0.29 ± 0.00
50 3.90 ± 1.18 3.57 ± 1.07 0.23 ± 0.03 0.22 ± 0.02 0.17 ± 0.00 0.15 ± 0.00
100 1.36 ± 0.15 1.22 ± 0.13 0.09 ± 0.00 0.08 ± 0.00 0.03 ± 0.00 0.03 ± 0.00
Table 4. Test error rate (in percentage) when forecasting the Compressor decay in the Multiple Output setup.
Random Sampling
UD BD MD
RLS SVR RLS SVR RLS SVR
∗ ∗ ∗ ∗ ∗ ∗
ntrain L̂test
ntest (h ) L̂test
ntest (h ) L̂test
ntest (h ) L̂test
ntest (h ) L̂test
ntest (h ) L̂test
ntest (h )
10 28.95 ± 4.31 26.32 ± 4.02 6.98 ± 2.34 6.20 ± 2.06 1.98 ± 0.05 1.79 ± 0.05
25 15.80 ± 2.45 14.26 ± 2.24 1.50 ± 0.12 1.38 ± 0.11 0.69 ± 0.01 0.61 ± 0.00
50 6.21 ± 0.41 5.62 ± 0.37 0.55 ± 0.01 0.49 ± 0.01 0.44 ± 0.00 0.38 ± 0.00
100 2.72 ± 0.19 2.43 ± 0.17 0.32 ± 0.00 0.29 ± 0.00 0.27 ± 0.00 0.25 ± 0.00
250 0.78 ± 0.03 0.69 ± 0.02 0.20 ± 0.00 0.18 ± 0.00 0.16 ± 0.00 0.14 ± 0.00
500 0.40 ± 0.00 0.35 ± 0.00 0.09 ± 0.00 0.08 ± 0.00 0.09 ± 0.00 0.08 ± 0.00
UD BD MD
RLS SVR RLS SVR RLS SVR
∗ ∗ ∗ ∗ ∗ ∗
ntrain L̂test
ntest (h ) L̂test
ntest (h ) L̂test
ntest (h ) L̂test
ntest (h ) L̂test
ntest (h ) L̂test
ntest (h )
10 24.78 ± 3.83 22.54 ± 3.48 5.97 ± 1.95 5.56 ± 1.79 1.68 ± 0.04 1.51 ± 0.04
25 13.49 ± 2.14 12.01 ± 1.93 1.30 ± 0.10 1.16 ± 0.09 0.58 ± 0.00 0.53 ± 0.00
50 5.40 ± 0.35 4.91 ± 0.32 0.46 ± 0.01 0.42 ± 0.01 0.37 ± 0.00 0.33 ± 0.00
100 2.30 ± 0.16 2.06 ± 0.15 0.28 ± 0.00 0.25 ± 0.00 0.24 ± 0.00 0.21 ± 0.00
250 0.65 ± 0.02 0.58 ± 0.02 0.17 ± 0.00 0.15 ± 0.00 0.14 ± 0.00 0.12 ± 0.00
500 0.33 ± 0.00 0.29 ± 0.00 0.08 ± 0.00 0.07 ± 0.00 0.08 ± 0.00 0.07 ± 0.00
24 Journal name 0(0)
Table 5. Test error rate (in percentage) when forecasting the Turbine decay in the Multiple Output setup.
Random Sampling
UD BD MD
RLS SVR RLS SVR RLS SVR
∗ ∗ ∗ ∗ ∗ ∗
ntrain L̂test
ntest (h ) L̂test
ntest (h ) L̂test
ntest (h ) L̂test
ntest (h ) L̂test
ntest (h ) L̂test
ntest (h )
10 32.00 ± 4.97 29.31 ± 4.61 9.41 ± 2.34 8.52 ± 2.03 1.77 ± 0.04 1.60 ± 0.04
25 25.27 ± 3.56 22.83 ± 3.23 1.96 ± 0.36 1.78 ± 0.32 0.65 ± 0.00 0.58 ± 0.00
50 12.71 ± 2.23 11.50 ± 2.01 0.67 ± 0.02 0.60 ± 0.02 0.46 ± 0.00 0.40 ± 0.00
100 4.95 ± 0.93 4.43 ± 0.85 0.39 ± 0.00 0.35 ± 0.00 0.33 ± 0.00 0.30 ± 0.00
250 1.34 ± 0.10 1.19 ± 0.09 0.24 ± 0.00 0.22 ± 0.00 0.18 ± 0.00 0.16 ± 0.00
500 0.63 ± 0.02 0.55 ± 0.02 0.17 ± 0.00 0.15 ± 0.00 0.10 ± 0.00 0.09 ± 0.00
UD BD MD
RLS SVR RLS SVR RLS SVR
∗ ∗ ∗ ∗ ∗ ∗
ntrain L̂test
ntest (h ) L̂test
ntest (h ) L̂test
ntest (h ) L̂test
ntest (h ) L̂test
ntest (h ) L̂test
ntest (h )
10 27.62 ± 4.40 25.26 ± 3.97 8.20 ± 1.93 7.54 ± 1.75 1.50 ± 0.04 1.36 ± 0.03
25 21.59 ± 3.09 19.19 ± 2.79 1.67 ± 0.32 1.49 ± 0.28 0.55 ± 0.00 0.49 ± 0.00
50 11.02 ± 1.92 10.05 ± 1.72 0.57 ± 0.02 0.52 ± 0.01 0.39 ± 0.00 0.34 ± 0.00
100 4.18 ± 0.80 3.75 ± 0.70 0.34 ± 0.00 0.30 ± 0.00 0.29 ± 0.00 0.26 ± 0.00
250 1.12 ± 0.08 1.00 ± 0.08 0.20 ± 0.00 0.19 ± 0.00 0.15 ± 0.00 0.14 ± 0.00
500 0.52 ± 0.02 0.46 ± 0.01 0.14 ± 0.00 0.13 ± 0.00 0.09 ± 0.00 0.08 ± 0.00
References
Theory. 1998;44(5):1974–1980.
[40] Lugosi G, Zeger K. Nonparametric estimation via empirical risk minimization. Information Theory, IEEE Transactions on.
1995;41(3):677–687.
[41] Tikhonov A, Arsenin VY. Methods for solving ill-posed problems. Nauka, Moscow; 1979.
[42] Engl HW, Hanke M, Neubauer A. Regularization of inverse problems. Springer; 1996.
[43] Wolpert DH, Macready WG. No free lunch theorems for optimization. IEEE Transactions on Evolutionary Computation. 1997;1(1):67–82.
[44] Evgeniou T, Micchelli CA, Pontil M. Learning Multiple Tasks with Kernel Methods. Journal of Machine Learning Research. 2005;6:615–
637.
[45] Pan SJ, Yang Q. A survey on transfer learning. IEEE Transactions on Knowledge and Data Engineering. 2010;22(10):1345–1359.
[46] Zhang W, Liu X, Ding Y, Shi D. Multi-output LS-SVR machine in extended feature space. In: Computational Intelligence for Measurement
Systems and Applications (CIMSA), 2012 IEEE International Conference on; 2012. .
[47] Baldassarre L, Rosasco L, Barla A, Verri A. Multi-output learning via spectral filtering. Machine learning. 2012;87(3):259–301.
[48] Pollard D. Empirical processes: theory and applications. In: NSF-CBMS regional conference series in probability and statistics; 1990. .
[49] Caponnetto A, De Vito E. Optimal rates for the regularized least-squares algorithm. Foundations of Computational Mathematics.
2007;7(3):331–368.
[50] Schölkopf B, Herbrich R, Smola AJ. A generalized representer theorem. In: Computational learning theory; 2001. .
[51] Scholkopf B. The kernel trick for distances. In: Neural information processing systems; 2001. .
[52] Keerthi SS, Lin CJ. Asymptotic behaviors of support vector machines with Gaussian kernel. Neural computation. 2003;15(7):1667–1689.
[53] Young DM. Iterative solution of large linear systems. DoverPublications. com; 2003.
[54] Poggio T, Mukherjee S, Rifkin R, Raklin A, Verri A. b. In: Uncertainty in Geometric Computations; 2002. .
[55] Boyd SP, Vandenberghe L. Convex optimization. Cambridge university press; 2004.
[56] Shawe-Taylor J, Sun S. A review of optimization methodologies in support vector machines. Neurocomputing. 2011;74(17):3609–3618.
[57] Keerthi SS, Shevade SK, Bhattacharyya C, Murthy KRK. Improvements to Platt’s SMO algorithm for SVM classifier design. Neural
Computation. 2001;13(3):637–649.
[58] De Vito E, Caponnetto A, Rosasco L. Model selection for regularized least-squares algorithm in learning theory. Foundations of
Computational Mathematics. 2005;5(1):59–85.
[59] Bartlett PL, Boucheron S, Lugosi G. Model selection and error estimation. Machine Learning. 2002;48(1-3):85–113.
[60] Anguita D, Ghio A, Oneto L, Ridella S. In-Sample and Out-of-Sample Model Selection and Error Estimation for Support Vector Machines.
IEEE Transactions on Neural Networks and Learning Systems. 2012;23(9):1390–1406.
[61] Guyon I, Saffari A, Dror G, Cawley G. Model selection: Beyond the bayesian/frequentist divide. The Journal of Machine Learning
Research. 2010;11:61–87.
[62] Kohavi R, et al. A study of cross-validation and bootstrap for accuracy estimation and model selection. In: International Joint Conference
on Artificial Intelligence; 1995. .
[63] Anguita D, A G, Ridella S, Sterpi D. K-Fold Cross Validation for Error Rate Estimate in Support Vector Machines. In: International
Conference on Data Mining; 2009. .
[64] Dietterich TG. Approximate statistical tests for comparing supervised classification learning algorithms. Neural computation.
1998;10(7):1895–1923.
[65] Anguita D, Ghelardoni L, Ghio A, Oneto L, Ridella S. The K in K-fold Cross Validation. In: European Symposium on Artificial Neural
Networks; 2012. .
[66] Hsu CW, Chang CC, Lin CJ. A practical guide to support vector classification, available for download at http://www.csie.ntu.
edu.tw/~cjlin/papers/guide/guide.pdf; 2009.
[67] Bache K, Lichman M. UCI Machine Learning Repository; 2013. Available from: http://archive.ics.uci.edu/ml.
[68] Narula SC, Wellington JF. Prediction, Linear Regression and the Minimum Sum of Relative Errors. Technometrics. 1977;19(2):185–190.
[69] Chen K, Guo S, Lin Y, Ying Z. Least absolute relative error estimation. Journal of the American Statistical Association.
Running head right side 27
2010;105(491):1104–1112.
[70] Aupetit M. Nearly homogeneous multi-partitioning with a deterministic generator. Neurocomputing. 2009;72(7):1379–1389.
[71] Coraddu A, Figari M. Ship electric propulsion: Analyses through modeling and simulation marine 2011. In: Computational Methods in
Marine Engineering; 2011. .
[72] Li Z, Yan X, Peng Z. Ship electric propulsion with a sensorless permanent magnet synchronous motor: A simulation study. Proceedings
of the Institution of Mechanical Engineers, Part M: Journal of Engineering for the Maritime Environment. 2012;226(4):378–386.
[73] Weber M, Welling M, Perona P. Unsupervised learning of models for recognition. Springer; 2000.