11.HistoryBased Fault CNN-ResNet
11.HistoryBased Fault CNN-ResNet
Farzaneh Abdollahi
CNN-Based Approaches
Convolution Stage
Pooling
ResNet-Based Approaches
Further Discussion
Performance Evaluation Indices
Convolution NN [1, 2]
Convolution Stage
3 0 1 2 7 4
1 5 8 9 3 1
2 7 2 5 3 1
0 1 3 1 7 8
4 2 1 6 2 8
2 4 5 2 3 9
Convolution Stage
3 0 1 2 7 4
1 5 8 9 3 1
1 0 -1
2 7 2 5 3 1
∗ 1 0 -1
0 1 3 1 7 8
1 0 -1
4 2 1 6 2 8
2 4 5 2 3 9
Convolution Stage
31 00 1−1 2 7 4
11 50 8−1 9 3 1 -5
1 0 -1
21 70 2−1 5 3 1
∗ 1 0 -1 =
0 1 3 1 7 8
1 0 -1
4 2 1 6 2 8
2 4 5 2 3 9
Convolution Stage
3 01 10 2−1 7 4
1 51 80 9−1 3 1 -5 -4
1 0 -1
2 71 20 5−1 3 1
∗ 1 0 -1 =
0 1 3 1 7 8
1 0 -1
4 2 1 6 2 8
2 4 5 2 3 9
Convolution Stage
▶ Conseider a 6 × 6 gray scale image
▶ Convolve it by a 3 × 3 filter
▶ It results to a 4 × 4 matrix
3 0 1 2 7 4
1 5 8 9 3 1 -5 -4 0 8
1 0 -1
2 7 2 5 3 1 -10 -2 2 3
∗ 1 0 -1 =
0 1 3 1 7 8 0 -2 -4 -7
1 0 -1
4 2 1 6 2 8 -3 -2 -3 -16
2 4 5 2 3 9
▶ n × n∗f × f = (n − f + 1) × (n − f + 1)
▶ f is usually an odd number
w1 w2 w3
w4 w5 w6
w7 w8 w9
▶ In Convolution stage:
▶ The features which is similar to the defined kernel is captured.
▶ Makes connection sparse( make problems simple)
▶ Each output value depends only on small number of inputs (local)
▶ There is no condensed connection between two layers they are sparse
▶ Share parameters
▶ Filter can be useful in different parts of the input(image)
∗ =
▶ To avoid these problems we can Pad the image with additional borders
∗ =
▶ Padding by 1 (p=1)
Farzaneh Abdollahi Intelligent Fault Diagnosis Lecture 9 6/41
Outline CNN-Based Approaches ResNet-Based Approaches Further Discussion
Padding
▶ Strided Convolution
23 34 74 4 6 2 9
61 60 92 8 7 4 3
3−1 40 83 3 8 9 7 3 4 4 91
7 8 3 6 6 3 4 ∗ 1 0 2 =
4 2 1 8 3 4 6 -1 0 3
3 2 4 1 9 8 3
0 1 3 9 2 1 4
▶ Strided Convolution
▶ Stride by 2
2 3 73 44 64 2 9
6 6 91 80 72 4 3
3 4 8−1 30 83 9 7 3 4 4 91 100
7 8 3 6 6 3 4 ∗ 1 0 2 =
4 2 1 8 3 4 6 -1 0 3
3 2 4 1 9 8 3
0 1 3 9 2 1 4
▶ Strided Convolution
▶ Stride by 2
2 3 7 4 6 2 9
6 6 9 8 7 4 3
33 44 84 3 8 9 7 3 4 4 91 100 83
71 80 32 6 6 3 4 ∗ 1 0 2 = 69
4−1 20 13 8 3 4 6 -1 0 3
3 2 4 1 9 8 3
0 1 3 9 2 1 4
▶ Strided Convolution
2 3 7 4 6 2 9
6 6 9 8 7 4 3
3 4 8 3 8 9 7 3 4 4 91 100 83
7 8 3 6 6 3 4 ∗ 1 0 2 = 69 91 127
4 2 1 8 3 4 6 -1 0 3 44 72 74
3 2 4 1 9 8 3
0 1 3 9 2 1 4
Strided Convolution
▶ n × n × nc ∗f × f × nc = (n − f + 1) × (n − f + 1) × nc′
▶ nc : number of Channels; nc′ : number of filters
Farzaneh Abdollahi Intelligent Fault Diagnosis Lecture 9 11/41
Outline CNN-Based Approaches ResNet-Based Approaches Further Discussion
▶ Convolution (Conv)
▶ Pooling (Pool)
▶ Fully Connected (FC)
▶ It is a simple single layer NN with bias
▶ Pooling reduces the size of the feature map and training parameters
▶ It overcome with over fitting
▶ It speeds up computation
▶ Pooling makes translation invariant since it behaves like taking a
summery of the info(image)
▶ Pooling can be considered as down sampling
▶ Good for the images that are not centered
▶ Length of pooling the dimension could be dynamic
f=2
s=2 9 2
6 3
f=2
s=2 3.75 1.25
4 2
Pooling Layer
A Convolutional NN
▶ nH , nW ↓, nC ↑
▶ Each Conv and pooling is considered as a layer
▶ At the end we have FC
▶ Last layer is softmax
▶ Consider the training set: (x (1) , y (1) )...(x (m) , y (m) )
▶ Using BP alg. optimize the params to min. cost fcn:
J = m1 m (i) (i)
P
i=1 L(ŷ , y )
Farzaneh Abdollahi Intelligent Fault Diagnosis Lecture 9 17/41
Outline CNN-Based Approaches ResNet-Based Approaches Further Discussion
Why Convolutional NN
Example [3]
▶ An unsupervised anomaly detection convolutional AE (CAE) is
proposed
▶ They claimed that combining these two approaches
▶ exponentially reduces the computational cost
▶ decreases the required training data by its extraction capabilities of
essential features in spatial input data.
▶ identifies errors larger than usual pre-trained, reconstructed errors
▶ 1-D conv is applied
[3]
Farzaneh Abdollahi Intelligent Fault Diagnosis Lecture 9 21/41
Outline CNN-Based Approaches ResNet-Based Approaches Further Discussion
Data Set
Data Set
[3]
Farzaneh Abdollahi Intelligent Fault Diagnosis Lecture 9 23/41
Outline CNN-Based Approaches ResNet-Based Approaches Further Discussion
Pre-Processing
▶ The data includes sudden spike or dip patterns under the threshold
alarm level.
▶ They had no labels
▶ the interquartile range (IQR) is used as a measure of the dispersion
in the data.
▶ (IQR) contains the second and third quartiles
▶ Any values fall below the lower quartile Q1 – 1.5 IQR, or above the
upper quartile Q3 + 1.5 IQR, are considered as anomalies.
▶ another column is added to the data (14th column)
▶ any rows include anomalies are labeled to 1
▶ without anomalies are labeled to 0
▶ This labeling is only considered for comparing with other supervised
approaches
▶ CAE model the 14th column is not considered
▶ The set has been normalized to a range of 0 and 1
Farzaneh Abdollahi Intelligent Fault Diagnosis Lecture 9 24/41
Outline CNN-Based Approaches ResNet-Based Approaches Further Discussion
Hyperparameters set up
[4]
▶ Resnet is made by stacking several residual blocks.
▶ The residual blocks consist of:
▶ Forward channels are made by stacking some convolutional layers.
▶ For example, two convolutional layers handle the input features by
θl = {kl1 , kl2 , bl1 , bl2 }: training params in the lth residual block
▶ Shortcut connection calculates the sum of forward channels output
and input features
▶ The output of the residual block is
xli = σr (g (xl−1
i |θl ) + xl−1
i )
Farzaneh Abdollahi Intelligent Fault Diagnosis Lecture 9 26/41
Outline CNN-Based Approaches ResNet-Based Approaches Further Discussion
▶ Indicates how the quality of the model has estimated the correct
value, in the fraction between 0 and 1.
▶ The better estimation the closer to 0
TP + TN
A=
TP + FP + TN + FN
Box plot
Further Discussion
Further Discussion
References I
References II
References III