algorithms-16-00061-v2
algorithms-16-00061-v2
Article
A Real-Time Novelty Recognition Framework Based on
Machine Learning for Fault Detection
Umberto Albertin 1, * , Giuseppe Pedone 2 , Matilde Brossa 3 , Giovanni Squillero 2 and Marcello Chiaberge 1
1 Department of Electronics and Telecommunications, Politecnico di Torino, Corso Duca degli Abruzzi, 24,
10129 Torino, Italy
2 Department of Control and Computer Engineering, Politecnico di Torino, Corso Duca degli Abruzzi, 24,
10129 Torino, Italy
3 Research and Development Department, Spea Spa, Via Torino, 16, 10088 Volpiano, Italy
* Correspondence: [email protected]
Abstract: New technologies are developed inside today’s companies with the ascent of Industry 4.0
paradigm; Artificial Intelligence applied to Predictive Maintenance is one of these, helping factories
automate their systems in detecting anomalies. The deviation of statistical features from standard
operating conditions computed on collected data is a common investigation technique that companies
use. The information loss due to transformation from raw data to extracted features is a problem
of this approach. Furthermore, a common Predictive Maintenance framework requires historical
data about failures that often do not exist, neglecting the possibility of applying it. This paper
uses Artificial Intelligence as Machine Learning models to recognize when something changes in
the data’s behavior collected up to that moment, also helping companies to gather a preliminary
dataset for future Predictive Maintenance implementation. The aim concerns a framework in which
several sensors are used to collect data by adopting a sensor fusion approach. The architecture is
composed of an optimized software system able to enhance the computation scalability and the
response time regarding novelty detection. This article analyzes the proposed architecture, then
explains a proof-of-concept development using a digital model; finally, two real cases are studied
to show how the framework behaves in a real environment. The analysis done in this paper has an
application-oriented approach; hence a company can directly use the framework in its systems.
Citation: Albertin, U.; Pedone, G.;
Brossa, M.; Squillero, G.; Chiaberge, Keywords: predictive maintenance; prognostic; novelty detection; machine learning; fault detection;
M. A Real-Time Novelty Recognition anomaly detection; condition monitoring; industrial real case study
Framework Based on Machine
Learning for Fault Detection.
Algorithms 2023, 16, 61. https://
doi.org/10.3390/a16020061 1. Introduction
Academic Editor: Mustafa Demetgül Since the 2010s, Artificial Intelligence (AI) and Machine Learning (ML) have become
standard ways to tackle complex problems relating to Industry, Medical, Manufacturing, or
Received: 14 November 2022
Economics. Flexibility, generality, and efficiency are only some of the qualities permitted by
Revised: 5 January 2023
Accepted: 10 January 2023
AI to be adopted into multiple scenarios. Furthermore, academics have recently enhanced
Published: 17 January 2023
learning programs, research programs, and collaboration projects with companies to re-
spond to the industrial world’s increasing AI and ML needs. With the upcoming Industry
5.0 paradigm, the term Predictive Maintenance (PdM) is becoming more common in the
language of companies today. This type of maintenance plays an interesting role in prevent-
Copyright: © 2023 by the authors. ing machine failures and reducing the downtime cost in a preponderant way. The cost of
Licensee MDPI, Basel, Switzerland. unplanned failure occurring in industrial machines is estimated at USD 50 billion annually
This article is an open access article as reported in the analysis by Ref. [1]. An explanation of how predictive maintenance is
distributed under the terms and able to reduce the overall costs by 5–10% and the standard maintenance costs by 20–50% is
conditions of the Creative Commons reported in Ref. [2], highlighting its impact on the company’s economic aspect.
Attribution (CC BY) license (https:// Throughout the years, the term “Maintenance” has taken different shapes and termi-
creativecommons.org/licenses/by/ nologies, as shown in (Figure 1) [3]:
4.0/).
Since the first industrial revolution, the predictive maintenance branch has increased
companies’ and scientific community’s interest in an exponential way [4]. Predictive
Maintenance monitors the real status of the machine by employing events, states, and
sensor data. With this kind of maintenance, component substitution is done when it is truly
necessary. This approach represents the future trend of maintenance thanks to its ability to
estimate when a failure will occur, reducing the cost of useless substitution and harmful
downtime. Its main drawback concerns the need to collect numerous data to perform an
estimation with adequate accuracy.
Often, the term “Predictive Maintenance” is used improperly by engineers, blurring
it with Condition Monitoring (CM). There is a substantial difference between these two
methodologies [5]: CM is based on the overcome of specific thresholds computed by a
domain expert or on the datasheet basis; hence, if a warning appears, the maintainer
has to intervene in solving the anomaly. With CM, there is no prediction about future
issues, and an alarm is emitted when a problem has already appeared. CM is deeply
explored in different application domains, as reported in Refs. [6–8]. The PdM’s approach
is different from CM’s; using the collected data, PdM can predict when an issue will appear
in the future, giving time to the maintainer to intervene before the breakdown. These
two techniques are also related to two other common words: diagnostics and prognostics.
Diagnostics is related to analyzing the past collected data, studying which damage occurred
and the cause of that; similar to the medical environment, prognostics analyzes the data
collected in the past together with the present ones to predict issues and failures in the
future [9]. In detail, according to the definition of ISO 13381-1 [10], prognostics has the
quirk to predict the Remaining Useful Life (RUL) of the components analyzed. Different
models exist to perform a prognosis, as reported in Ref. [11], in which three different
variants are well explained. Today’s PdM methodologies use more complex algorithms
Algorithms 2023, 16, 61 3 of 26
than the statistical ones used in the past. Among these methodologies, Machine Learning
and Deep Learning (DL) cover part of today’s algorithms used to recognize anomalies
and predict the Remaining Useful Life (RUL) of components analyzed. The main and
common problem that companies have when working with standard algorithms concerns
the complete absence of data collection about failures encountered in the past (e.g., a new
machine). Hence, a standard ML algorithm cannot be applied due to the proper model
training unfeasibility. This approach requires a labeled dataset that is already collected and
well-known by an expert, in which each label represents the machine’s status; hence, the
label gives information about the component’s RUL. When a gathered dataset does not
exist (the failures labeled dataset is not present), a standard approach cannot be applied.
A recent research field—called “Novelty Detection” (ND)—responds to this demand by
searching for a solution to apply the PdM without a dataset collected. This paper concerns
a fast and easy implementation of the Novelty Detection Framework (NDF). The paper is
organized into different chapters: Section 2 presents the state-of-art and the “Related Works”
about this topic; Section 3 reports the framework’s architecture with the analysis performed
during the study; Section 4 concerns the Proof-of-Concept in which the framework is
applied to a digital model, and at the end, two real cases are analyzed in Section 5.
Hence, more columns are present in the data frame used depending on the time and
the sampling frequency.
3.3. AI Unit
The AI Unit is shown in Figure 3, the framework’s “brain”. As a standard ML approach,
the parts that compose the process flow are represented in Figure 4:
As mentioned before, the idea of this paper is to give a different use of ML compared to
a normal flow. A common approach follows the standard way that a dataset is divided into
Training Set, Testing Set, and Validation Set; in this case, the framework continuously works
in testing mode to compute the prediction error for every dataset that arrives in the database.
Monitoring this error makes it possible to understand if a deviation occurs in the collected
dataset. Every time that data collected are similar to the ones used for Training, the error
obtained is low. In contrast, if an issue happens in the machine, the error tends to increase
due to different behavior in the data pattern collected. This type of detection does not
predict only problems; it can also understand when a novelty related to a change appears
in the dataset (typical in a real situation, an example can be environmental changes).
For this reason, the trained dataset must represent the whole machine’s behavior. The
number of datasets chosen for the training depends on the framework’s application: by
increasing the number of training datasets, the prediction’s accuracy increases too (the
risk of a false positive due to a change is low); however, training the framework with
a small number of datasets allows using it earlier, but the maintainer may work more
during the first days by updating it (the risk of detecting a change is higher). A trade-off
concerning the duration of the training process has to be defined. Generally, companies
Algorithms 2023, 16, 61 7 of 26
The feature extraction and labeling phase follow the preprocessing stage. This block is
responsible for preparing data for the following ML model training. With a multi-threading
approach, a feature is extracted from the dataset by converting it to a label; this procedure
is done for each window/feature, creating a number of labels equal to the number of
windows and predictors. This rule is schematized in the following figure:
At the end of the preprocessing stage, a number of trainsets equal to the number of
windows are created; each trainset is used for the related ML model training. Then, the
testing phase follows the training one. The same preprocessing approach is applied to each
new dataset; in the end, the label related to ith model is predicted and compared with the
real one, computing the error used for novelty detection. Hence, the AI Unit can be divided
into two different macro-stages:
• The first part is related to the training process in which the system waits for a prede-
fined number of data to train the models;
• The second part is the testing process in which incoming data are continuously tested
to compute the predictions’ error results.
For the Training phase, three regressors are tested to visualize the potential differences
in the prediction behavior. The models used during the evaluation are:
• Linear Regressor;
• Decision Tree Regressor;
• Random Forest Regressor.
The “Low Lubrication Case” of the “Digital Model” (better explained in the following
paragraph) is used as a representative example to see how the mentioned models work
in order to compare them. Generally, a system works always in the same way: at the
beginning of life, when it is healthy, the error computed by the model’s prediction is low;
when the failure appears, the error’s prediction starts to increase. A further confirmation
that a general system works in this manner is shown in Section 5 in which the real cases
behave exactly like the digital model used for the proof-of-concept, starting with a low
error that increases at the end of the life when the failure appears. This general behavior
is shown in Figures 6 and 7: the former concerns the models’ selection, while the latter
represents the errors’ selection.
The results obtained by comparing the three models are almost identical. In the
Linear Regressor model, the y-scale has a lower variance in terms of the range in which
the error can span, while in the other two models, the y-scale range is almost the same.
The user can choose one of the three models with no particular weakness; generally, the
tree models have better accuracy than the easier Linear Regressor, but this one has the
advantage of being faster when the training process is performed. Decision Tree Regressor
and Random Forest Regressor work similarly from a general point of view: the former
works only with decisions while the latter works with decision trees. The main difference
between them concerns the accuracy obtained and the computation’s complexity: a Decision
Tree Regressor has lower computational complexity and accuracy than a Random Forest
Regressor. The operator can choose one of the three models depending on the application
in which the framework operates: for example, if there are no “computational effort
thresholds”, such as a workstation or a computer, a Random Forest Regressor can be used
to obtain an accurate prediction, while if the aim is to deploy the framework in a stand-
alone embedded device, a Linear Regressor or a Decision Tree Regressor can be used in
order to reduce the overall computational effort. The paper’s framework is uploaded in a
workstation to compute the predictions and results; hence, a Random Forest Regressor is
chosen as the reference model.
Furthermore, different errors were also evaluated during the study: Squared Error
(SE), Mean Squared Error (MSE), Root Mean Squared Error (RMSE), and Mean Absolute
Error (MAE), with the shape reported below:
MSE(y, ŷ) = 1
n ∑in=−01 (yi − ŷi )2
q
RMSE(y, ŷ) = 1
n ∑in=−01 (yi − ŷi )2
MAE(y, ŷ) = 1
n ∑in=−01 |yi − ŷi |
where yi is the real value while ŷi is the predicted one. Figure 7 shows different error
pictures computed on the same predictor using a Random Forest Regressor (RFR).
As shown in Figure 7 there is a substantial difference between the obtained results; the
last two graphs are more disrupted than the first two because they follow the prediction’s
behavior faster. As in the AI model selection case, there is no particular rule for the error
Algorithms 2023, 16, 61 10 of 26
model selection. A possible solution could be to configure all four error models to monitor
the deviation by creating dedicated predictor agents to perform the computation, but this
creates waste from a computational energy point of view. The latter two error models show
very similar behavior because of the huge likeness in the formula√ behind them; with the
same number of samples, the MAE has 1/n while RMSE has 1/n, multiply the square
root member. The similarity in the graphs above is due√to the number of data used for
the computation: when the number is low, the 1/n and 1/n are very comparable, also
bringing a similarity in the results obtained. In this case, the operator can again choose
which error model or which error models the framework has to use to monitor the behavior
of the machine. In this paper and the following case studies, the error chosen for the
evaluation is the MAE. Figure 8a,b show two examples in which the error is computed
with two sample ranges n; Figure 8a shows the last two records while Figure 8b shows the
last five records:
(a) (b)
Figure 8. (a) Error computation on the last two records; (b) Error computation on the last five records.
A two-sample error is chosen for the following experimental executions because of its
ability to predict quickly. Another ability of the main framework concerns the possibility of
updating or retraining the already trained models: as explained above, Novelty Detection
can understand when something new happens in the system; this change can be related to
an issue or a change does not create any problem. For this reason, when a novelty appears,
the maintainer has to examine if the problem truly exists in order to choose one of the
following procedures:
• Models Update: in the case in which the data collected are different from the trained
ones, but this change is not related to an issue in the machine, the maintainer can
update the framework with the last parametrized n data, considering the incoming
data as good ones. Thanks to the models’ update, a substantial reduction of the error
value is performed;
• Models Retrain: when a problem or a component substitution is performed in the
machine, the retraining process is necessary due to the inevitable data change in
the prediction (a common change could be a shift error in the data collected). The
maintainer can decide to retrain the models by neglecting all previous ones and
returning to the beginning phase.
An example of the update/retrain function is shown in the following “Digital Model”
paragraph. The model update function also collects consistent datasets of the component
under study to help the beginning data collection, which is useful to create the machine’s
behavior history for future PdM implementation.
Recapping the AI Unit’s phases (Figure 3):
Algorithms 2023, 16, 61 11 of 26
• Data Fetching Phase: in a continuous loop, the Unit waits for new data uploaded in
the local/remote database to take them for the following Training or Testing phase;
• Training Phase: when a predefined number of datasets are uploaded to the database
(depending on how long the train dataset collection lasts), the framework understands
that the training process can start. After the Training process, several models are
created as a function of the features extracted (hence depending on the number of
windows of the windowing function);
• Testing Phase: after the training phase, every dataset uploaded is processed by the
framework with a number of predictors equal to the number of models previously
created. Then, every predictor estimates its label with an error; the upcoming failures
conduct the dataset to change its shape, increasing the error in the prediction. The
number of predictors equal to the number of features can help in error recognition by
covering all the data distribution.
• Retraining/Updating Phase: in some cases, the model error deviates without a real
problem on the machine; in this case, the maintainer can update the models trained,
considering incoming new data as “normal”, resetting the error value. The retraining
process is necessary when a maintenance operation is performed; then, past data are
neglected, and only new data are used for future predictions.
Figure 10. The Agent Manager main process that schedules and executes all ready agents.
optimization and extendability, this MAS allows the development of new applications
easily through new agents.
The framework’s code will be available at the following link: https://github.com/
PIC4SeR/PdM_NoveltyDetectionFramework.git (accessed on 1 November 2022).
The digital model is voluntarily simplified because the aim concerns understanding
how NDF works during a change in the data pattern. A realistic mechanical model runs
away from this paper. The vibration propagation is a fundamental property that the model
must possess; this property allows sensing vibrations from other components far from
the noise source (an example can be a vibration felt by the right bearing with the left
bearing subjected to the noise source). For this scope, “virtual accelerometers” are used to
understand how vibrations affect the shaft and bearings to recognize failure conditions or
problems during the rotation. The material used for all the components is the same, which
is steel with the following properties (Table 2):
4.1. Shaft
The shaft is divided into different components representing each section diameter
change; hence, a rigid connection is used to link all the “fragments” together, assembling
the whole shaft. A discretization of two finite elements reduces the model simulation
complexity. The gear affects the shaft with a radial force and a torque; bearings are
Algorithms 2023, 16, 61 15 of 26
responsible for the noise generation represented by a random uniform force impulse, which
affects radial x and y directions.
4.2. Bearings
Each bearing is divided into two parts: an internal ring linked to the shaft by means
of a rigid joint and an outer ring rigidly connected to the fixed reference frame of the
“environment”. The representation of the roller vibration is performed using a planar joint
with three degrees of freedom (dof) that gives the possibility to move in x and y directions
but also around z with predefined stiffness and damping values. Hence, x and y directions
represent the vibration directions generated by each sphere during rotation designed as
small impulses acting on the inner ring; the motion around z represents the rotational
viscous damping due to the bearing’s lubricant. The remaining dof are neglected to simplify
the computations done by the simulator.
4.3. Simulations
Numerous simulations are performed to represent different fault cases. The chosen
tests performed on the framework are the following:
• Low Lubrication: a lubrication degradation condition is simulated inside bearings.
This situation is represented by increasing the variance range of the impulse’s force
acting on the shaft. Two accelerometers are placed on the bearings and on the shaft
to measure vibrations emitted and transmitted. Hence, in every simulation, the
accelerometers measure an increase in general noise with a reduction of rotational
viscous damping around z. The parameters change in each simulation in the same
manner for the left and the right bearing. It can imagine that every simulation has
different time instants in which a maintainer takes data from the machine to monitor;
for this reason, each simulation is delayed by one day. This case is represented in
Figure 12a,b in which the lubricated and not completely lubricated bearings are shown.
The vibrations shown are measured on the right bearing:
• Inner Race Faults: as reported in Ref. [27], a general rolling bearing has four standard
types of fault that can occur over time; these faults are “translated” as frequencies
with the following names and shapes:
– Ballpass frequency outer race (BPFO), a fail occurs in the bearing’s outer race:
n fr d
BPFO = 2 (1 − D cosφ )
– Ballpass frequency inner race (BPFI): a fail occurs in the bearing’s inner race:
n fr d
BPFI = 2 (1 + D cosφ )
– Fundamental train frequency (FTF): a fail occurs in the bearing’s cage:
fr d
FTF = 2 (1 − D cosφ )
– Ball spin frequency (BSF): a fail occurs in the bearing’s rolling elements:
D d 2
BSF = 2d (1 − ( D cosφ ) )
where f r is the shaft speed, n is the number of spheres, φ is the angle of the load with
respect to the radial plane, d is the sphere diameter, and D is the bearing’s radius
between two rings.
For the proof-of-concept of this paper and to test the framework in different situations,
an Inner Race Fault is simulated, as shown in Figure 13a,b. Hence, this kind of
simulation works similarly to the low lubrication one with the difference in the force
application. In the low lubrication case, the force’s impulse acts with a random
amplitude at high frequency, while in this case, it acts with a period depending on the
BPFI computed with the above formula. Similar to the low lubrication case, the force’s
amplitude increases in each simulation, reproducing a failure’s worsening. The initial
and final simulation gives the following results:
Algorithms 2023, 16, 61 16 of 26
(a) (b)
Figure 12. (a) Lubricated bearing noise; (b) Low lubricated bearing noise.
(a) (b)
Figure 13. (a) Bearing without fault; (b) Bearing with Inner Race Fault.
The framework is tested for both simulations to understand how it works. Figure 14a
represents the framework application on the low lubrication data, while Figure 14b repre-
sents the application on the BPFI fault:
(a) (b)
Figure 14. (a) Prediction with low lubrication fault; (b) Prediction with BPFI bearing fault.
Algorithms 2023, 16, 61 17 of 26
The blue dashed line represents the result obtained from the AIU, while the red lines
represent the predictions computed by the MCU. Starting from Figure 14a, the MCU is
applied to the low lubrication case. An increase in the noise generated by the model
corresponds to an increase in the error computed by the AIU, subjecting this to a substantial
difference compared to the elements trained at the beginning. The parameters set for this
evaluation are the following (Table 3):
Windows 20
Average 4
Lowest Bound 5
Highest Bound 40
The average value gives the “speed” of the prediction: an MCU with a small average
creates predictions that are able to better follow the real system’s volatility; a high number
of the average corresponds to a slower prediction response, compensating for the volatility
and outliers of the real measures. The configuration used for the digital model allows
for fast predictions that are able to follow the error trend. As reported in the previous
paragraph, each predictor monitors a window computed after the preprocessing stage.
When the error in a window crosses the lowest bound, the related predictor is triggered;
hence, a prediction is performed every time the obtained error is higher than the lowest
bound. Each predictor computes when the error will reach the highest bound using
the Linear Regressor explained in the previous paragraph. If the error keeps rising, the
predictions will accordingly fit even better. According to Figure 14a, this window crosses
the lowest bound 20 days before the failure, and the framework gives a warning to the
maintainer that something has changed compared to the training phase; 10 days before the
failure, the prediction performed by the MCU fits well with the real behavior, alerting the
maintainer that the change will become too relevant. In Figure 14a, the framework is tested
on a bearing inner race defect type. In this case, the bearing behaves differently compared
to the low lubrication one; as explained, the fault manifests itself as periodical impulses
with high amplitude in this second configuration. This fault might not be detected from all
AIUPs because it occurs only in specific time ranges. In this case, the configuration set is
the following (Table 4):
Windows 20
Average 4
Lowest Bound 1
Highest Bound 5
The error computed is quite disrupted, but the framework accurately predicts when
the bearing’s change becomes relevant in error deviation. As before, the MCU triggers its
agents 20 days before the final failure with low accuracy; the prediction fits well about
10 days before the final failure.
An example of the update/retrain function is shown in Figure 15 in which, around
July 2021, the update function is executed by the user; the error decreases drastically,
reducing itself to a value close to zero, expanding the trained data and transferring the
information that the last k datasets acquired are healthy and the framework should not
send any warning to alert the maintainer.
Algorithms 2023, 16, 61 18 of 26
5.2. Bearings
The dataset is collected using High Sensitivity Quartz ICP accelerometers sampling
data at 20 kHz. The dataset is composed of files in which every file corresponds to
20,480 measures sampled every 10 min (except for he first 43 files taken every 5 min).
After 100 million revolutions, an inner race defect occurred in bearing 3.
As in the case of the Digital Model, the parameters are set to have a good trade-off
between responsiveness and prediction accuracy. AIU’s models are ready and trained after
30 h by using this configuration; then, the framework starts the error evaluation for each
model to monitor the bearings’ status. Concerning the MCU Configuration, the settings
used are the following (Table 6):
Average 16
Lowest Bound 10
Highest Bound 150
The result obtained after the framework application is reported in Figure 17a,b in
which the blue dashed line corresponds to the x and y error degradation of bearing 3 com-
puted on acceleration data and red lines represent the predictions performed by the MCU.
As reported on the graphs, the error obtained at the beginning is low; at the end, when the
bearing failure appears, the error also increases accordingly. The system recognizes the
relevant change 1/2 days before; the trigger could be set lower to wake up the framework
earlier, but the number of “positive false” increases accordingly. Predictions fit better over
time, as represented by red lines, in which the monotony increases close to the complete
breakdown. The advantage of having different predictors as the features number allows
for compensating possible “positive false”; the framework can be programmed to send
Algorithms 2023, 16, 61 20 of 26
a warning only if more than k predictors detect a change. In the same figures, the long
straight lines in the graph represent data missing during that period; using a line plot
instead of a scatter plot can be useful to understand if a shift appears in the pattern. For
example, a shift could be related to a maintenance operation or a component substitution
performed on the machine. The framework’s update function is shown in Figure 18 to
see how it behaves. The difference can be noted by comparing Figure 18 with Figure 17b
around 11/21. The trainset is updated with 200 other datasets, reaching 400 as the total
number of datasets collected. After the model update, the error computed by the AIU
decreases, reaching a value close to zero. The maintainer can perform this procedure when
the model detects a change without a real problem on the machine/component.
(a) (b)
Figure 17. (a) Predictions with BPFI fault [ x acc ]; (b) Predictions with BPFI fault [y acc ].
The machine chosen is a multi-purpose machine with x/y/z motion capabilities, used by
the company to perform different tasks, from MEMS to spring probes tests. This latter
ability regards the main focus of the data collection and framework implementation. The
company designs a stress program to test the probe’s wear over time. Usually, probes
are considered spoiled after a given number of cycles using a preventive maintenance
approach. Spea chose this threshold considering the past probes’ electrical resistance value
degradation. If the resistance exceeds the minimum threshold, the probe is intended to be
worn. The company has wanted to change this approach by applying the framework to
evaluate when the probe deviates in an anticipated manner following a predictive method.
The Spea’s machine keeps a set of probes as the main tool. It measures the electrical
resistance of probes by knocking them repeatedly on a conducting test surface. The
machine executes this operation continuously, collecting raw data during the process.
A customized raw dataset template has been built. It is composed of sampling each
probe’s resistance and the related time references. The AIU preprocessing task trans-
forms the custom raw dataset into the standard AIU data format treated by AIUP. Each
dataset comprises 10 resistance values collected for each probe over time. The same frame-
work’s storage is maintained for this experiment: MongoDB as Storage and Data Source
Unit database.
The AIU Configuration used for this evaluation is the following (Table 7):
Train Elements 50
Test Elements 2
Windows 5
The MCU configuration is set with the following parameters (Table 8):
Average 2
Lowest Bound 0.20
Highest Bound 1
As for the previous experiments, AIU trains its models on a set of initial raw data
coming from the machine, producing results related to the model’s errors. The result
obtained after this framework application is reported in Figure 19.
The result of AIU computation is the blue dashed line in the graphs above, correspond-
ing to the MAE error degradation of a probe computed on electrical resistance data over
time. MAE is stable and fixed around zero; apart from a few low amplitude disturbances,
no degradation appeared on the dataset. Hence, there are no significant changes from the
initial trained model, meaning that the probe is not worn. As explained in the real case, the
blue straight dashed lines that appear in the data pattern are due to data missing during
that period. The data acquisition lasts a few hours, while the idle period without acquisi-
tions lasts numerous days; this different scale range creates long lines. The graph reported
in Figure 19 shows that no shift appears in the data pattern; finally, no general deviation
appears, which is a symptom that no relevant changes or problems are happening. This
prototypical application confirms that incoming data does not change from the training
ones. Hence, future works regard the ongoing collection of other data to identify possible
future changes related to failures.
Algorithms 2023, 16, 61 22 of 26
6. Discussion
In order to have an idea about the framework’s performances and better see the
scalability property, two tables are compiled with Train, Test, and Preprocess time
with different configurations. This analysis is performed for both the AIU and the MCU.
Tables 9 and 10 report the results of this analysis.
Table 9. AIU time performance indicator W, strain , stest and [min, max] time statistics for Preprocess,
Train, Test.
Table 10. MCU time performance indicator. It shows the statistics µ and σ calculated over all MCDAs
Execution, Train, Predict operations.
The table results lead to the idea of how the system works efficiently. The training
stage requires a very small amount of time, thanks to the easy models used. This time also
gives the possibility to maintain the framework in real-time when other sensors are added
to the machine. Indeed, the order of magnitude about time always remains around seconds
when performing the computation. The AIU requires more time than the MCU due to the
higher computational effort; the training process is the hardest computation processed by
the AIU. All the performances strictly depend on the number of windows elaborated: a
high number of windows corresponds to high requirements in terms of time performance.
Concerning the MCU, Table 10 refers to the agent’s mean and variance time performances
during the execution of its task on a single window. Each agent requires a small amount
of time to perform a prediction compared to AIU. MCU works each agent one at a time,
following a time scheduling approach (when it concludes with one agent, it can start with
the following one); for this reason, the number of windows influences the MCU’s time
performance in a preponderant way. This computation can be performed by multiplying
the execution time of each agent by the total number of windows:
Tm = Te · w
Tm is the Total MCU Time, Te is the execution time, and w is the windows number.
The different domain examples proposed in this paper prove the flexibility and adapt-
ability of the framework in any domain where it is adopted. Bearing degradations are
analyzed for both Simulink simulations and real case, in which the framework was able
to understand when the system deviates excessively, alerting the maintainer before the
breakdown. Different types of faults are used to test the framework’s capability: low
lubrication increases the system’s noise in terms of vibration captured by the sensors and
processed by the framework, giving an estimation of when this vibration becomes too
high. Then, a ball-pass-inner-race-fault is tested to prove the framework’s flexibility, in
which a force impulse acts on the bearing with a specific period; hence, these impulses
can be seen only in some of the windows that include them. The predictors related to
such windows recognized that something was changing from the trained data, alerting
the maintainer. While in the first case, all predictors showed a general error increase, only
some showed this behavior in the second test, as wanted. Finally, the framework is tested
in a completely different real environment, a resistance test, always to show flexibility and
adaptability properties. In this case, the framework has to work with another physical
variable, understanding if some probes’ resistance is changing from the beginning and
giving an alert to the maintainer when the probe will be worn. As shown, the framework
works well again in finding that probes are healthy till that moment without a false positive
warning sent to the maintainer. As reported in Section 2, thanks to the algorithm’s ease and
adaptability, it can be used in numerous domains by comparing the trained data with the
actual ones. To summarize, the framework’s advantages are as follows:
• Experience and data history is not necessary: this framework can detect novelties in
observed data such as errors, faults, problems, and changes related to machines, as
well as environmental data such as vibrations or noises, temperature, or magnetic field
comparing to the trained ones. Furthermore, this property can help collect datasets
and maintenance warnings for future definitive PdM applications;
• Flexibility/Scalability: the framework is very flexible and scalable; a maintainer can
add a sensor in the machine anytime by updating or retraining the framework creating
a new starting point, but always recording past data for a future PdM application;
• Low computational effort required: the framework uses straightforward ML models
to estimate when and how the trained and current data differ with low computational
effort, facilitating a future implementation on embedded devices.
hence lightening the overall payload of the MCU, reducing the time to perform prediction,
and consequently, the framework’s real-time efficiency. Another consideration to develop
as a next step can be related to the application of this framework in a multi-domain dataset
with features sampled with different sampling times (an example can be the vibration mea-
sured by an accelerometer and the temperature measured by a thermometer sampled with
less order of magnitude). Adapting this framework not only to the same motion but also to
a system’s random/general motion will be an important enhancement. Furthermore, future
research concerns deploying this framework on real embedded devices by optimizing the
framework developed, evaluating how tiny ML models work, and obtaining an intelligent
edge device to mount directly on the component to analyze. As a final future step, after
implementing the mentioned framework’s enhancements, a baseline comparison will be
performed to better understand the differences compared to standard and well-known
algorithms concerning speed and accuracy.
Author Contributions: Conceptualization, U.A., G.P., G.S., M.C. and M.B.; methodology, U.A., G.P.
and G.S.; software, U.A. and G.P.; validation, U.A. and G.P.; data curation, U.A.; writing, U.A., G.P.
and M.B.; supervision, G.S. and M.C. All authors have read and agreed to the published version of
the manuscript.
Funding: This research received no external funding.
Data Availability Statement: Data will be available at this repository: https://github.com/PIC4
SeR/PdM_NoveltyDetectionFramework.git, accessed on 1 November 2022.
Acknowledgments: This work has been developed with the contribution of the Politecnico di Torino
Interdepartmental Center for Service Robotics PIC4SeR (https://pic4ser.polito.it, accessed on 1
January 2023) and Spea Spa (https://www.spea.com, accessed on 1 January 2023).
Conflicts of Interest: The authors declare no conflict of interest.
Abbreviations
The following abbreviations are used in this manuscript:
AI Artificial Intelligence
AIU Artificial Intelligence Unit
AIUP Artificial Intelligence Unit Predictor
AM Agent Manager
AP Agent Program
AS Agent System
BPFO Ballpass frequency Outer
BPFI Ballpass frequency Inner
BSF Ball Spin Frequency
CM Condition Monitoring
CNN Convolutional Neural Network
DL Deep Learning
dof Degrees of Freedom
FTF Fundamental Train Frequency
LSTM Long Short-Term Memory
LR Linear Regressor
MAE Mean Absolute Error
MAS Multi-Agent system
MCU Maintenance Control Unit
MCDA Model Change Detector Agent
ML Machine Learning
MSE Mean Squared Error
MWTE Maximum Wait Time for Execution
ND Novelty Detection
Algorithms 2023, 16, 61 25 of 26
References
1. Industry Week & Emerson. How Manufacturers Achieve Top Quartile Performance. Available online: https://partners.wsj.com/
emerson/unlocking-performance/how-manufacturers-can-achieve-top-quartile-performance/ (accessed on 31 May 2017).
2. Deloitte. Predictive Maintenance and the Smart Factory. 2021. Available online: https://www2.deloitte.com/content/dam/
Deloitte/us/Documents/process-and-operations/us-cons-predictive-maintenance.pdf (accessed on 10 June 2021).
3. Swanson, L. Linking maintenance strategies to performance. Int. J. Prod. Econ. 2001, 70, 237–244. [CrossRef]
4. Poór, P.; Ženíšek, D.; Basl, J. Historical overview of maintenance management strategies: Development from breakdown
maintenance to predictive maintenance in accordance with four industrial revolutions. In Proceedings of the International
Conference on Industrial Engineering and Operations Management, Pilsen, Czech Republic, 23–26 July 2019; pp. 495–504.
5. Calabrese, F.; Regattieri, A.; Bortolini, M.; Gamberi, M.; Pilati, F. Predictive Maintenance: A Novel Framework for a Data-Driven,
Semi-Supervised, and Partially Online Prognostic Health Management Application in Industries. Appl. Sci. 2021, 11, 3380.
[CrossRef]
6. Ahmad, R.; Kamaruddin, S. An overview of time-based and condition-based maintenance in industrial application. Comput. Ind.
Eng. 2012, 63, 135–149. [CrossRef]
7. Tsunashima, H. Condition Monitoring of Railway Tracks from Car-Body Vibration Using a Machine Learning Technique. Appl.
Sci. 2019, 9, 2734. [CrossRef]
8. Stetco, A.; Dinmohammadi, F.; Zhao, X.; Robu, V.; Flynn, D.; Barnes, M.; Keane, J.; Nenadic, G. Machine learning methods for
wind turbine condition monitoring: A review. Renew. Energy 2019, 133, 620–635. [CrossRef]
9. Sikorska, J.Z.; Hodkiewicz, M.; Ma, L. Prognostic modelling options for remaining useful life estimation by industry. Mech. Syst.
Signal Process. 2011, 25, 1803–1836. [CrossRef]
10. ISO 13381-1; Condition Monitoring and Diagnostics of Machines—Prognostics—Part 1: General Guidelines. International
Standards Organization: Geneva, Switzerland, 2015.
11. Liao, L.; Köttig, F. Review of Hybrid Prognostics Approaches for Remaining Useful Life Prediction of Engineered Systems, and
an Application to Battery Life Prediction. IEEE Trans. Reliab. 2014, 63, 191–207. [CrossRef]
12. Si, X.; Wang, W.; Hu, C.; Zhou, D. Remaining useful life estimation—A review on the statistical data driven approaches. Eur. J.
Oper. Res. 2011, 213, 1–14. [CrossRef]
13. He, D.; Li, R.; Zhu, J. Plastic Bearing Fault Diagnosis Based on a Two-Step Data Mining Approach. IEEE Trans. Ind. Electron. 2013,
60, 3429–3440. [CrossRef]
14. Wang, J.; Liu, S.; Gao, R.X.; Yan, R. Current envelope analysis for defect identification and diagnosis in induction motors. J. Manuf.
Syst. 2012, 31, 380–387. [CrossRef]
15. Saimurugan, M.; Ramachandran, K.I.; Sugumaran, V.; Sakthivel, N.R. Multi component fault diagnosis of rotational mechanical
system based on decision tree and support vector machine. Expert Syst. Appl. 2011, 38, 3819–3826.[CrossRef]
16. Bezerra, C.G.; Costa, B.S.J.; Guedes, L.A.; Angelov, P.P. An evolving approach to unsupervised and Real-Time fault detection in
industrial processes. Expert Syst. Appl. 2016, 63, 134–144. [CrossRef]
17. Del Buono, F.; Calabrese, F.; Baraldi, A.; Paganelli, M.; Guerra, F. Novelty Detection with Autoencoders for System Health
Monitoring in Industrial Environments. Appl. Sci. 2022, 12, 4931. [CrossRef]
18. Liu, R.; Yang, B.; Zio, E.; Chen, X. Artificial intelligence for fault diagnosis of rotating machinery: A review. Mech. Syst. Signal
Process. 2018, 108, 33–47. [CrossRef]
19. Zhang, B.; Zhang, S.; Li, W. Bearing performance degradation assessment using long short-term memory recurrent network.
Comput. Ind. 2019, 106, 14–29. [CrossRef]
20. Souza, R.M.; Nascimento, E.G.S.; Miranda, U.A.; Silva, W.J.D.; Lepikson, H.A. Deep learning for diagnosis and classification of
faults in industrial rotating machinery. Comput. Ind. Eng. 2021, 153, 107060. [CrossRef]
21. Zhao, R.; Wang, D.; Yan, R.; Mao, K.; Shen, F.; Wang, J. Machine Health Monitoring Using Local Feature-Based Gated Recurrent
Unit Networks. IEEE Trans. Ind. Electron. 2018, 65, 1539–1548. [CrossRef]
22. Wei, D.; Han, T.; Chu, F.; Jian Zuo, M. Weighted domain adaptation networks for machinery fault diagnosis. Mech. Syst. Signal
Process. 2021, 158, 107744. [CrossRef]
23. Zhang, C.; Gupta, C.; Farahat, A.; Ristovski, K.; Ghosh, D. Equipment Health Indicator Learning Using Deep Reinforcement
Learning. In Machine Learning and Knowledge Discovery in Databases. ECML PKDD 2018; Lecture Notes in Computer Science;
Springer: Cham, Switzerland, 2019; Volume 11053. [CrossRef]
Algorithms 2023, 16, 61 26 of 26
24. Miljković, D. Review of novelty detection methods. In Proceedings of the 33rd International Convention MIPRO, Opatija, Croatia,
24–28 May 2010; pp. 593–598.
25. Oliveira, M.A.; Filho, E.F.S.; Albuquerque, M.C.S.; Santos, Y.T.B.; da Silva, I.C.; Farias, C.T.T. Ultrasound-based identification of
damage in wind turbine blades using novelty detection. Ultrasonics 2020, 108, 106166. [CrossRef]
26. Pimentel, M.A.F.; Clifton, D.A.; Clifton, L.; Tarassenko, L. A review of novelty detection. Signal Process. 2014, 99, 215–249.
[CrossRef]
27. Randall, R.; Jérôme Antoni, J. Rolling element bearing diagnostics—A tutorial. Mech. Syst. Signal Process. 2011, 25, 485–520.
[CrossRef]
28. Qiu, H.; Lee, J.; Lin, J.; Yu, G. Wavelet filter-based weak signature detection method and its application on rolling element bearing
prognostics. J. Sound Vib. 2006, 289, 1066–1090. [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual
author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to
people or property resulting from any ideas, methods, instructions or products referred to in the content.