Uncertainty Analysis in Ship Performance Monitoring Draft Under Review
Uncertainty Analysis in Ship Performance Monitoring Draft Under Review
2014
Energy Institute, University College London, 14 Upper Woburn Place, London, WC1H 0NN,
UK
c
Department of Mechanical Engineering, University College London, Torrington Place, London,
WC1E 7JE, UK
d
BMT Group Ltd, Goodrich House, 1 Waldegrave Road, Teddington, TW11 8LZ, UK
Corresponding author: L. Aldous, email [email protected], phone number +44 (0)7887 441349
Keywords: Uncertainty, Noon report, Ship performance, Measurement, Monte Carlo, Continuous
monitoring
1
Abstract
There are increasing economic and environmental incentives for ship owners and operators
to develop tools to optimise operational decisions, particularly with the aim of reducing
fuel consumption and/or maximising profit. Examples include real time operational
optimisation (e.g. ship speed and route choice), maintenance triggers and evaluating
technological interventions. Performance monitoring is also relevant to fault analysis,
charter party analysis, vessel benchmarking and to better inform policy decision making.
The ship onboard systems (propulsion, power generation etc.) and the systems in which it
operates are complex and its common for data modelling and analysis techniques to be
employed to help extract trends. All datasets and modelling procedures have an inherent
uncertainty and to aid the decision maker, the uncertainty can be quantified in order to
fully understand the economic risk of a decision, if this risk is deemed unacceptable then it
makes sense to re-evaluate investment in data quality and data analysis techniques. This
paper details and categorises the relevant sources of uncertainty in performance
measurement data, and presents a method to quantify the overall uncertainty in a ship
performance indicator based on the framework of the Guide to Uncertainty in
Measurement using Monte Carlo Methods. The method involves a simulation of a ships
operational profile and performance and is validated using 4 datasets of measurements of
performance collected from onboard in-service ships. A sensitivity analysis conducted on
the sources of uncertainty highlight the relative importance of each. The two major data
acquisition strategies, continuous monitoring (CM) and noon reported (NR), are compared.
Abbreviations
CM: Continuous Monitoring
NR: Noon Reports
2014
1. Introduction
Ship operational performance is a complex subject, not least because of the various
systems and their interactions in which a ship operates, the major factors are presented in
Figure 1. At the ship level the ship design, machinery configurations and their efficiencies
determine the onboard mechanical, thermal and electrical energy flows which, despite
automation built in to the configuration mode settings at the ship design phase, there is still
an appreciable level of human interaction during day to day operations. The environmental
conditions (sea state, wind speed, sea/air temperature etc.) are dynamic, unpredictable and
complicated to quantify, due in part to the characteristics of the turbulent flow fields by
which they are determined. These environmental conditions exert an influence on the
ships resistance and therefore the ship power requirements in differing relative quantities.
The rate of deterioration in ship performance (engine, hull and propeller) is dependent on a
vast array of variables; including the quality and type of hull coating and the frequency of
hull and propeller cleaning which are also dependent on the ocean currents, temperature
and salinity in which the ship operates. Further, the shipping industry operates in an
economic sphere in which the global consumption of goods and global energy demand, and
conditions in the various shipping markets determine operating profiles, costs and prices
(see for example Lindstad (2013) which also explores environmental effects). In addition,
technological investment, fuel efficiency and savings are complicated by the interactions
between ship owner-charterer-manager [Agnolucci (2014)].
Data collection, either through daily noon reporting procedures or high frequency,
automatic data acquisition systems, and data processing techniques such as filtering and/or
modelling have so far proven to be useful tools in capturing and quantifying some of the
intricacies and nuances of these interactions to better understand the consequences of
operational decisions. These datasets and modelling outputs are applied in areas such as
real time operational optimisation including trim adjustments, maintenance triggers,
predicting and evaluating the performance of new technologies or interventions,
2
2014
particularly for cost benefit analysis, fault analysis, charter party analysis, vessel
benchmarking and to better inform policy decision making.
The need to conduct uncertainty analysis is linked to the amplitude of the noise or scatter
in the data relative to the underlying, longer term trends that are to be extracted. The ship
system interactions give rise to scatter in the data, not only from inherent sensor precision
but also from unobservable and/or unmeasurable variables. According to the central limit
theorem (assuming independent, identical distributions), over time the scatter will tend to a
normal distribution with zero mean. This time period is dependent on the data acquisition
and processing strategy; the temporal resolution of sensors and data collection frequency,
the sensor precisions and human interactions in the collection process and the modelling or
filtering methods all play a part. There are also uncertainties in the data that will introduce
a potentially significant bias in the results and this too needs to be understood and
evaluated. The magnitude of the underlying trends to be identified are a function of the
modelling application; for evaluating the performance of new technologies the signal delta,
i.e. the improvement in ship performance, may be a step change of the order of 1-3% (as in
the case of propeller boss cap fins) or up to 10-15% as in the case of hull cleaning or new
coating applications where analysis of trends in the time domain is also necessary.
Therefore, the required measurement uncertainty depends on the application and this drives
the desired acquisition strategy which is of course constrained by costs; economic, time
and resources.
Acquisition strategies are broadly separated into two dominant dataset types. Noon report
(NR) datasets are coarse but cheap to compile and readily available since they are currently
in widespread use across the global fleet. The frequency of recording is once every 24
hours (time zone changes allowing) and the fields reported are limited, generally included
as a minimum are ship speed and position, fuel consumption, shaft rotational speed, wind
speed derived Beaufort number, date/time and draught. Given the economic and regulatory
climate there has been a shift towards more complete, automatic measurement systems
referred to in this paper as continuous monitoring (CM) systems. The uptake of these has
been limited by installation costs in service while improved data accuracy, speed of
acquisition, high sampling frequency (5minutes) and repeatability are cited as the key
drivers. All datasets and modelling procedures have an inherent uncertainty associated and
as a prerequisite the uncertainty must be quantified in order to fully understand the
economic risk of the decision. The benefit of reducing uncertainty is a function of the
uncertainty of the data relative to the change in ship performance, the former depends on
the data acquisition strategy (noon reports vs continuous monitoring, for example) and the
latter depends on the application (the technological / operational decision being made) and
the cost-benefit of both determine the overall economic risk of the decision. If the
economic risk is deemed unacceptable then it makes sense to re-evaluate investment in
data quality and data analysis techniques.
The uncertainty is also important because of the risks and costs that are associated with the
decision derived from the measured ships performance. The desire to quantify these (in
other industries) has led to the field of research into methods for risk based decision
making. The application of these methods to the shipping industry is also important, for
example, measurement and verification is cited as a key barrier to market uptake in fuel
efficient technologies and retrofit. In order to secure capital, investment projects must be
3
2014
2014
2014
4. Method
The simulation of the performance indicator is based on the difference between the
expected power and the measured power. The measured power is equal to the expected
power plus some linearly incremented power increase over time to simulate the effect of
ship degradation (hull / propeller), the simulation also allows for the inclusion of some
instrument error (precision, drift and bias) and model error. A sampling algorithm is able to
simulate the effect of sample averaging frequency and the change in the number of
observations. Further details regarding the nature and magnitude of these data acquisition
variables are provided in sections 5, 6 and 7.
The method is summarised by the following steps (Figure 2):
6
2014
2014
2014
The sources of uncertainty presented in Figure 3 are discussed in detail in the sections that
follow. Table 1 presents the relevant data acquisition variables (pdf and sampling
algorithm parameters) input to the MC simulation and which are categorised as two
different data acquisition strategies; noon reports (NR) and continuous monitoring (CM).
DAQ decision variable
Baseline input,
NR
90
270
24
72
5.00
1.00
1.00
1.00
1.74
9.00
Baseline input, CM
90
2160
270
6480
0.51
1.00
1
96
1.74
9
5.1
Sampling
Sampling error arises from the fact that a finite sample is taken from an infinite temporal
population and because variable factors are averaged between samples. Both of these
effects are investigated through a sampling algorithm that samples from the underlying
ship performance simulation which is based on a maximum temporal resolution of 4
samples per hour.
The number of observations in Table 1 are based on a sample rate which is, realistically,
less than the sample averaging frequency because of periods when the ship is at anchor,
9
2014
manoeuvring, waiting or in port and because of the filtering criteria. The days spent at sea
depends on ship type and size, figures from the latest IMO GHG study (Smith 2014) show
that for 2012 the average number of days at sea for all sizes of tankers, bulk carriers,
containers and general cargo ships was 53%. This is similar to the results found from the
397 day CM dataset analysed in this study (51.5% at sea days), the results of which are
presented in Table 2.
Number of Observations
38112
19618
12799
12798
9570
% of All
100.0
51.5
33.6
33.6
25.1
Table 2: Effect of filtering on the number of observations of a CM dataset. *See equation [1]
The proportion of the results that are removed due to filtering depends on the
environmental / operational conditions. From the CM dataset used in this study, after
filtering, 25% of the observations for use in the ship performance quantification remained.
The NR datasets that were used in this study showed a high variation in the data filtered for
environmental conditions with 50% being the average, meaning that of the possible
number of samples 26.5% were used in the ship performance quantification (generally NR
data is only recorded while at sea).
The averaging frequency is either equal to the underlying temporal resolution of the true
ship performance parameters (96/day) as in the CM baseline, or 96 samples are averaged
for a daily frequency (NR baseline). To simulate the effect of daily averaging of noon
reported variables, a daily speed variability is included at a level of 1.74% of the ships
speed. This was the outcome (median) of a study of the CM dataset covering a 396 day
time period after filtering for outliers and wind speed > BF 4 (not sea water depth since
this data is not generally recorded in the NR data) . The variability is due to the effects of
ship accelerations, rudder angle alterations, wind speed fluctuations within 0 < BF < 4
range, and crew behaviour patterns (slow steaming at night for example), which arent
included in the ship performance model or cant be filtered out in the NR data set. The
variability is introduced into the simulation through the addition of normally distributed
noise to the underlying true ship speed. The effect of averaging daily draught variability
(i.e. due to intentional changes in trim or due to influences on trim due to fuel
consumption) is not included but assumed to be negligible.
The number of observations is also affected by the evaluation period length and the
sensitivities of the overall uncertainty to this are investigated by evaluating both 90 day
and 270 day periods. There is seen in some datasets the effect of seasonality, whereby
weather fluctuations (even within the BF filtering criteria) cause discontinuities in the
measured power and therefore a minimum of one years data collection is required. The
presence of this is dependent on the cyclic nature and the global location of the ships
operations and is difficult to generalise, it is therefore not included in this analysis.
10
5.2
2014
Instrument Uncertainty
Instrument uncertainties arise from variations attributable to the basic properties of the
measurement system, these properties widely recognised among practitioners are
repeatability, reproducibility, linearity, bias, stability (absence of drift), consistency, and
resolution (ASTM 2011). Since the truth is unknown then bias, and by extension drift
(change in bias over time), are epistemic uncertainties, and are assumed to be zero for all
measurements in the baseline. Their effect on the overall uncertainty is investigated in the
sensitivity analysis by an elementary interval analysis.
Some manufacturers quote sensor precisions and/or accuracy which are assumed to reflect
manufacturing tolerances, however these are not well documented or publicly available
and also they are more likely to quote sensor repeatability calculated under laboratory
conditions (same measurement system, operator, equipment) or reproducibility (different
operator) and therefore likely to under estimate the actual precision in service conditions
where different operators, equipment/facility and measurement system may increase the
overall precision. Quoted shaft power sensor accuracies range from 0.25% (Seatechnik,
http://www.seatechnik.com/) to 0.1% Datum Electronics (http://www.datumelectronics.co.uk/). The IMO resolution Performance Standards for Devices to Indicate
Speed and Distance (IMO 1995) stipulate that errors in the indicated speed, when the ship
is operating free from shallow water effect and from the effects of wind, current and tide,
should not exceed 2% of the speed of the ship, or 0.2knots, whichever is greater. The shaft
power in the CM is the addition by quadrature of an RPM sensor error of 0.1% (1) and a
torque sensor error of 0.5% (1). Discussions with industry experts and experience of ship
performance datasets suggest that the instrument precisions quoted in Table 1 are
appropriate, however they are estimates and the sensitivity of the performance indicator to
changes in these is explored in section 7. All sensors are assumed to be consistent during
the simulation (no change in repeatability over time) and linear (absence of change in bias
over the operating range of the measurement instrument). Sensor resolution is assumed to
be reflected in the quoted sensor precision and not a restricting factor in the uncertainty of
ship performance measurement.
In the NR dataset, fuel consumption is generally recorded rather than shaft power; fuel
flow meter accuracy ranges between 0.05 per cent and 3 per cent depending on the type,
the manufacturer, the flow characteristics and the installation. Fuel readings by tank
soundings are estimated to have an accuracy of 2-5% [Faber (2013)]. This uncertainty
estimate does not include however additional uncertainties associated with the fuel
consumption, for example fuel oil quality variation (grade and calorific value), pulsations
in the circulation system, back flushing filters, the effect of cat-fine, waste sludge from
onboard processing and inaccuracies in ullage measurements due to weather, fuel
temperature etc. The use of fuel consumption as a proxy for shaft power is included by
assuming a power sensor precision of 5% for the NR strategy.
The draught uncertainty associated with draught gauges is of the order 0.1m, in noon
report entries the draught marks at the perpendiculars may be read by eye and, depending
on sea conditions, a reading error of 2 cm can be assumed (Insel 2008). However, noon
report entries are often not altered during the voyage to record variability due to trim
11
2014
and/or the reduction in fuel consumption and therefore a greater 1.0 m uncertainty is
assumed. Often the draught field in CM datasets is manually input and therefore the same
uncertainty applies.
5.3
Model Uncertainty
12
5.4
2014
Human Error
Human error (which is often categorised as instrument uncertainty) may occur in any
measurements when operating, reading or recording sensor values if the report completion
is not automated. For example the noon data entry may not occur at exactly the same time
each day, the recording of time spent steaming may not be adjusted to compensate for
crossing time zones and it is possible that different sensors are used to populate the same
field, for example, some crew may report speed from the propeller RPM and others report
speed through water. The measurement of wind speed through crew observations may also
be subject to uncertainties. Human error is not included in this current analysis.
(1)
where
Table 3 shows the comparison of the methods, the simulation method through error
propagation of elemental uncertainties tends to over-estimate the uncertainty.
13
Evaluation
period length
(days)
200
265
377
373
CM 1
NR 1
NR 2
NR 3
2014
Statistical
uncertainty
calculation*, %
1.4
8.78
6.00
4.95
Table 3: Statistical uncertainty calculation vs elemental uncertainty propagation for method 2, error
reported as the confidence interval of P (95% level) as a percentage of mean shaft power (CM) or
mean fuel consumption (NR)
Figure 4 shows the experimental precision from the continuous monitoring dataset and the
simulation results. The y-axis is the change in the number of samples for the same
evaluation time period. The performance indicator uncertainties (plotted as a percentage of
mean shaft power over the evaluation period) are similar, although again, slightly over
estimated in the simulation. However the difference is again minimal and the comparison
provides evidence in support of the use of the simulation and propagation of elemental
uncertainty in this way.
0
0
10
-10
Change
samples
per
day,
experimental
precision
-20
-30
-40
-50
Change
samples
per
day,
simulaAon
-60
-70
-80
-90
-100
Figure 4: Simulation and experimental precision comparison for varying evaluation periods and
sampling frequencies
14
2014
7. Sensitivity Analysis
The sensitivity indices relate to the precision of the uncertainty in the overall performance
indicator. The bias in the result is investigated and graphically represented also.
7.1
A local sensitivity analysis was conducted by computing partial derivatives of the output
function (y) with respect to the input factors (xi). The output function, y is the precision of
the measurand; twice the standard error of the P at time = teval (see section 4). The input
factor is one of the data acquisition parameters of Table 4. Each factor is altered onefactor-at-a-time (OAT) and a dimensionless expression of sensitivity, the sensitivity index,
SIi is used to measure changes relative to the baseline.
! =
!
.
!
The input parameter interval bounds reflect results from studies of actual data where
possible or, in the absence of data, estimates of realistic values are made given the source
of the uncertainties as detailed in section 5. Sensor precisions are induced by the addition
of a normally distributed error, and in the baseline this is assumed to have a mean of zero
and a standard deviation according to Table 4. Sensor bias is introduced by a deterministic
offset to the mean of the normal distribution of the error.
DAQ decision variable
Evaluation Period (days), teval
Number of observations, N
Power sensor precision (1) %
Power sensor bias %
Draught sensor precision (1) m
Draught sensor bias m
Ship speed sensor precision (1) %
Ship speed sensor bias %
Ship speed sensor drift %/270 days (linear
increase)
Averaging frequency, fave, sample/day
Daily speed variability, %
Model precision error, (1) %
Baseline input
90
270
24
72
1
0
0.1
0
1
0
90
45
SA input
270
135
5
3
1.5
1
5
3
1
1.74
1.5
96
3.76
9
The increase in the sample size from 25.0% or 26.5% to 50% (presented in the table by the
increase in number of samples, N) is to investigate the effect of increasing the sample size
to that of a dataset that is normalised rather than filtered by increasing the number of
observations to the total observations before filtering. The daily speed variability SA input
represents the upper interquartile of the results of an investigation conducted on a CM
dataset of time period length of 396 days. Model precision error represents the results of
15
2014
the study by (Insel 2008) who found 7-9% of shaft power to be the precision error in the
sea trial data of 12 sister ships studied. If the ship reporting is speed over ground (SOG)
rather than speed through the water (STW) then this will increase the uncertainty; a study
of the difference between STW and SOG for 20 ships over 3 to 5 years indicates a standard
deviation of 0.95knots, (Smith 2014). The effect of using SOG as a proxy for STW is
therefore investigated by increasing the speed sensor precision to 5%. Likewise, power
sensor precision is to reflect the effect of fuel consumption as a proxy for shaft power.
7.2
The second study is to compare the uncertainty for the data acquisition parameters of a
standard CM dataset with the data acquisition parameters of a standard NR dataset. The
measurand is the uncertainty (SE at 95% confidence level) reported as a percentage of the
parameter of interest, for the purposes of this investigation, the parameter of interest is the
change in ship performance over time, i.e. increase in shaft power for a given speed and
draught (e.g. due to hull fouling), this was set to linearly increase by 18% over 6months.
This is a practical metric since it puts the uncertainty in a useful context and is of greater
interest when comparing the data acquisition strategies however it is worth highlighting
that the uncertainty measured in this way is not only representative of the impact of all the
elemental uncertainties but it is also a function of the magnitude of the performance
indicator itself and consequently of the time period of evaluation. For example, if the rate
of deterioration was actually 10%/year and not 5% or if the evaluation time period was set
to increase from 90 days to 270 days then in both cases the uncertainty would naturally
decrease even if the elemental uncertainties were unchanged
This was carried out for each of the two baselines (CM and NR, Table 1) and for two time
periods of evaluation; 90 days and 270 days. This experiment is repeated once with an
increase in number of observations (N) for both, and once with a reduction in model
precision error Table 5.
DAQ decision variable
Evaluation Period days
Number of observations
N
Shaft power sensor
precision (1) %
Ship speed sensor
precision (1) %
Ship draught sensor
precision (1) m
Averaging frequency
/day
Model precision error
(1) %
144
4320 12960
24
72
Alternative
2 CM
90
270
2160
6480
0.51
0.51
96
96
1.5
1.5
16
2014
This section looks at sensitivities of the performance indicator uncertainty relative to the
variations in the data acquisition decision parameter uncertainties (see section 7.1). The
sensitivity index indicates the change in the parameter of interest (the precision of overall
uncertainty quantified by 2) relative to the percent change in the input quantity
examined (the data acquisition variable) and then weights it according to the relative size
of each. For example, an SI of unity, similar to the case of the speed sensor precision, SI =
0.96 (Figure 5), indicates that the change in overall uncertainty due to a change in the
precision of the speed sensor is almost exactly offset by the magnitude of the ratio of its
absolute precision (5%) to the resultant overall uncertainty.
Speed
driL
Speed
bias
Daily
speed
variability
Power
bias
Draught
bias
Daily
averaging
Power
precision
Sample
frequency
Draught
precision
Model
precision
Speed
precision
0
0.1
0.2
0.3
90
days
0.4
0.5
270
days
0.6
Figure 5: Sensitivities of the performance indicator uncertainty to model, instrument and sampling
input uncertainties (evaluation time period: 90days)
The exponent in the relationship between speed and power means that the performance
indicator uncertainty is more sensitive to the speed sensor precision than either of the
draught or shaft power sensor precisions. This highlights the importance of investment in
high precision speed sensor, and the criticality of using speed through the water sensor
rather than speed over ground (which may cause a precision of 5%, see section 7.1) which
will have dramatic consequences for obtaining meaningful information from the data. CM
systems that augment core data acquisition with additional measurement of GPS, tides and
currents would provide a means to independently calculate and verify the speed through
water measurement from speed over ground. The results emphasise the importance of a
precise draught measurement which may be manually input, even in continuous monitoring
systems, the overall uncertainty would benefit from investment in a draught gauge which
also records potentially significant variations during the voyage (including in trim) rather
than recording the draught at the last port call. If fuel consumption is used as a proxy for
shaft power, which may cause an inaccuracy of 5% (due to uncertainties in the fuel
17
2014
consumption measurement) then there could be a significant effect also on the overall
uncertainty. The effect of sensor precision reduces over a longer evaluation period (270
days relative to 90 days) because over time the variability cancels and stabilises towards
the mean.
Sensor bias has a lesser relative effect on the overall precision, the order of significance is
the same as for precision; speed, draught then power. For the speed sensor, the bias and
drift cause the precision in the overall PI to increase because the increased Vmeas is
compounded by the cubic relationship which causes P to drift (see next section, Figure 6)
and therefore decreases the precision in the linear trend line of the PI. The drift in P
means that the effect of bias and drift in Vmeas increases for a longer evaluation period. It is
worth highlighting however that the presented magnitude of the effect is influenced by the
input parameters of the ships operational profile and the rate of speed reduction over the
time period. There is a similar but reduced effect occurring due to the draught and power
bias also causing P to drift as the evaluation period increases but this is to a much lesser
extent and small relative to the counteracting effect of increasing sample size which
reduces the overall precision as the evaluation period increase.
The uncertainty in the performance indicator is second most sensitive to changes in the
model precision error, the model is the calibration or correction factor applied to correct
for the ships speed as established from sea trial data which will have some level of
uncertainty. The effect on overall uncertainty is significant because of how this translates
to Pmeas through a cubic relationship. Instead of sea trial data, a continuous monitoring
calibration/reference dataset representing a short period of operation may be used to derive
the ship speed/power performance curve (the ship performance should be as close to
stationary as practically possible during the time period). The advantage of the latter
method is the increased number of samples in the dataset, even after filtering for
environmental / operational conditions, which reduces the uncertainty. The CM dataset
will also include the effect of technological enhancements made since the sea trial dataset
was compiled; the effects of these more recent interventions may not vary linearly with the
model input parameters and may therefore be incorrectly attributed to ship performance
changes as measured during the evaluation period. The effect of model precision reduces
over a longer evaluation period because over time the variability cancels and stabilises
towards the mean.
Increasing the sampling frequency is also significant because the overall standard error is
inversely proportional to the square root of the sample size, the absolute values, when
increased on a samples per day basis cause the effect over 270days to be more significant
relative to 90 days. An increased sample size may be achieved by addressing root causes of
outliers or missing data in the datasets (i.e. due to stuck sensors or human reporting errors);
although this is also limited by the ships at-sea days and the days the ship operates in
environmental/operational conditions outside the bounds of the filtering algorithm. There
is therefore a considered trade-off to be made between the sample size and the model error
which reflects the advantages of filtering over normalising (see follow up paper).
The other sampling effect is related to the daily averaging frequency, the impact of this on
the overall uncertainty is because the daily environmental fluctuations are captured;
averaging a range of speeds will cause power due to higher speeds to be mistakenly
attributed to deterioration in ship performance that in reality is not present. The actual
18
2014
influence on uncertainty is of course a function of the daily speed variability, and this
interaction is not explicitly studied here however the 1.78% found from the data is realistic
and the SA provides the relative significance under the assumption of linearity. The daily
averaging effect is independent of evaluation period length because the daily speed
variability is assumed to be constant over time.
8.2
Generally, the effects of the data acquisition factors studied do not affect the absolute mean
of the performance indicator (the bias component of the overall uncertainty). Instrument
bias however not only affects the precision component (as demonstrated in the previous
section) but also biases the result; this is observable in Figure 6.
Performance
Indicator
Bias,
90
day
evalua>on
period
Monte
Carlo
Simula>on:
Average
Performance
Indicator,
kW
4000
3000
2000
1000
-1000
-2000
Figure 6: Performance indicator bias with error bars indicating the confidence interval (95% level).
The baseline performance indicator is highlighted by the red line
The effect of increasing the STW sensor bias from 0 to 5% is to increase the average
performance indicator value from 790 kW to 1196 kW. The graphic is indicative only
since the exact magnitude of the effect is to some extent dependent on the operational
profile but it is clear that in cases of sensor bias and drift the actual underlying
deterioration trend is difficult to identify.
The error bars in Figure 6 shows how results based on daily averaging for a short 90 day
time period may be inconclusive; if for example, speed sensor precision is 5% then it might
19
2014
not be possible to conclude (with 95% confidence) if the ships performance has improved
or deteriorated.
8.3
This sensitivity index is detailed in section 7.2. Figure 7 shows the results from the MC
methods deployment to calculate uncertainty for a range of changes to the baseline data
acquisition variables. As described previously, the uncertainty measured in this way is not
only representative of the impact of all the elemental uncertainties but it is also a function
of the magnitude of the performance indicator itself and consequently of the time period of
evaluation.
2
4
2
3
Baseline
0.00
26
16
159
25
22
165
36
235
50.00
100.00
150.00
200.00
250.00
CM 90 days
NR 270 days
NR 90 days
2014
effective as reducing the model uncertainty from the sea trial data from 9% to 1.5% as in
alternative 2, but arguably more achievable given resources.
The extent to which the uncertainty is improved depends on the quality of the model used;
it should be emphasised that every algorithm/model introduced will also increase the
model form uncertainty as well as the number of data fields with a consequent possible
increase in instrument uncertainty, this needs to be carefully considered to ensure that the
addition of the algorithm produces an overall reduction in uncertainty. This analysis also
assumes there is no bias and so if the sensors, in particular the speed sensor, are not
calibrated or maintained then the positive effect of the increased time period on the overall
uncertainty will be reduced (see section 8.1).
There is no uncertainty included to model the influence of human error, therefore the
results for the noon report model is perhaps optimistic, although the overestimation from
the simulation, even for noon report data, in the experimental precision comparison of
section 6 indicates that this might not always be an issue if procedures are in place to
ensure careful reporting.
9. Conclusions
This paper proposes and describes the development of a rigorous and robust method for
assessing the uncertainty in ship performance quantifications. The method has been
deployed in order to understand the uncertainty in estimated trends of ship performance
resulting in the use of different types of data (noon report and continuous monitoring) and
different ways in which that data is collected and processed. This is of high relevance to
the shipping industry, which regularly uses performance quantifications for operational
decision making, and the method and results in this paper can both inform the decision
making process and provide insight into how lower uncertainty in some of the key decision
variables could be achieved. The desired level should be appropriate to the particular
application (the technological / operational decision being made); this informs the
appropriate data acquisition strategy as does an analysis of the cost-benefit of reducing
uncertainty which should also be weighed against the economic, environmental or social
risk of an incorrect decision.
The results indicate the significant uncertainty benefit of CM data over NR data; this is of
the order of 90% decrease in uncertainty, and is especially relevant to shorter term
analysis. It has been shown in this analysis that the uncertainty of the 90 day CM baseline
is similar to the uncertainty achievable from a 270 day NR dataset.
The precision of the speed sensor is of fundamental importance when it comes to achieving
low uncertainty and using SOG rather than STW is likely to have a dramatic effect on the
overall uncertainty. The sensor precision however only affects the aleatory uncertainty of
the performance indicator and the effect decreases over longer evaluation periods; the
additional confounding factor in the uncertainty analysis is epistemic uncertainty which
may be introduced through sensor bias and drift. Speed sensor bias and drift causes
significant changes in the performance indicator, it is difficult to extract the underlying
performance trend when bias or drift is present, after 90 days the effect on the precision of
21
2014
Acknowledgements
The research presented here was carried out as part of a UCL Impact studentship with the
industrial partner BMT group. The authors gratefully acknowledge advice and support
from colleagues within the UCL Mechanical Engineering department and within UCL
Energy Institute, in particular Dr David Shipworth, Energy & the Built Environment.
References
Agnolucci, P., Smith, T. W. P., Rehmatulla, N. (2014). "Energy efficiency and time charter
rates: Energy efficiency savings recovered by ship owners in the Panamax market."
Transportation Research Part A: Policy and Practice 66: 173-184.
AIAA (1999). S-071A-1999 Assessment of Experimental Uncertainty With Application to
Wind Tunnel Testing, American Institute of Astronautics and Aeronautics.
ASTM (2011). American National Standard Standard E2782 - 11: Guide for Measurement
Systems Analysis (MSA). ASTM Committee on Quality and Statistics, ASTM. E2782.
Coleman, H. W., Steele, W. G. (1990). Experimentation and Uncertainty Analysis for
Engineers. US, John Wiley & Sons.
Cox, M. G., Harris, P. M. (2006). Software Support for Metrology Best Practice Guide No.
6: Uncertainty Evaluation. UK, National Physics Laboratory.
Faber, J., Nelissen, D., Smit, M. (2013). Monitoring of bunker fuel consumption, Delft, CE
Delft.
IMO, A. R. (1995). Performance Standards for Devices to Indicate Speed and Distance.
22
2014
Insel, M. (2008). "Uncertainty in the analysis of speed and powering trials." Ocean
Engineering 35(11-12): 1183-1193.
ITTC (2008). Guide to the Expression of Uncertainty in Experimental Hydrodynamics.
ITTC - Recommended Procedures and Guidelines. 7.5-02-01-01
ITTC (2008). Testing and Extrapolation Methods, General Guidelines for Uncertainty
Analysis in Resistance Towing Tank Tests. ITTC - Recommended Procedures and
Guidelines. 7.5-02-02-02.
ITTC (2011). Uncertainty Analysis - Example for Waterjet Propulsion Test. ITTC Recommended Procedures and Guidelines. 7.5-02-05-03.3.
JCGM100:2008 (2008). "Evaluation of measurement data - Guide to the expression of
uncertainty in measurement (GUM 1995 with minor corrections)."
Lindstad, H., Asbjrnslett, B. E., Jullumstr, E. (2013). "Assessment of profit, cost and
emissions by varying speed as a function of sea conditions and freight market."
Transportation Research Part D: Transport and Environment 19: 5.
Longo, J., Stern, F. (2005). "Uncertainty Assessment for Towing Tank Tests With
Example for Surface Combatant DTMB Model 5415." Journal of Ship Research 49(1): 5568.
Loucks, D. P., Van Beek, E., Stedinger, J. R., Dijkman, J. P.M., Villars, M. T. (2005).
Water Resources Systems Planning and Management: CH 9 Model Sensitivity and
Uncertainty Analysis. Paris, UNESCO.
MAN (2011). Basic Principles of Ship Design. Denmark.
Smith, T. W. P., Jalkanen, J. P., Anderson, B. A., Corbett, J. J., Faber, J., Hanayama, S.;
OKeeffe, E., Parker, S., Johansson, L., Aldous, L., Raucci, C., Traut, M., Ettinger, S.,
Nelissen, D., Lee, D. S., Ng, S., Agrawal, A., Winebrake, J. J., Hoen, M., Chesworth, S.,
Pandey, A. (2014). Reduction of GHG Emissions from Ships: Third IMO GHG Study
2014. I. M. O. (IMO). London, UK.
Stulgis, V., Smith, T. W. P., Rehmatulla, N., Powers, J., Hoppe, J. (2014). Hidden
Treasure: Financial Models for Retrofits. T. L. H. MCMAHON.
23