Itimes Investigating Semi-Supervised Time Series Classification Via Irregular Time Sampling
Itimes Investigating Semi-Supervised Time Series Classification Via Irregular Time Sampling
This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/TII.2022.3199374
Dataset
Abstract— Semi-supervised learning (SSL) provides a
Labeled
Full
Data
powerful paradigm to mitigate the reliance on large labeled
data by leveraging unlabeled data during model training.
However, for time series data, few SSL models focus on the
Self-supervised Transformation Type Prediction Supervised Classification
underlying temporal structure of time series, which results
in a sub-optimal representation learning quality on unla- Label 1:
beled time series. In this paper, we propose a framework of
semi-supervised time series classification by investigating Label 2:
© 2022 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission.See https://www.ieee.org/publications/rights/index.html for more information.
Authorized licensed use limited to: Harbin University of Science and Technology. Downloaded on October 29,2022 at 04:18:54 UTC from IEEE Xplore. Restrictions apply.
This article has been accepted for publication in IEEE Transactions on Industrial Informatics. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/TII.2022.3199374
2 IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, VOL. XX, NO. XX, XXXX
we propose a framework of semi-supervised time series clas- to the different input perturbations, e.g., apply different data
sification by investigating irregular time sampling (iTimes), augmentations to the same input. For example, Π-Model [14]
which learns the underlying temporal structure of unlabeled uses self-ensembling to form a consistent prediction of un-
time series in a self-supervised manner to benefit semi- labeled data under different augmentation and regularization.
supervised time series classification. Specifically, as shown in MixMatch [15] predicts low-entropy labels for data-augmented
Fig. 1, for labeled data, iTimes classify the time series directly unlabeled examples and mixes labeled and unlabeled data
under the supervision of the annotated label. For unlabeled using MixUp for training on unlabeled data. FixMatch [16]
data, iTimes follows the self-supervised paradigm, and is tries to generate high-confident pseudo-labels of the weakly
proposed to learn the temporal structure by training ConvNets augmented unlabeled input and then use the pseudo-label to
to recognizing the temporal transformations applied to the time supervise the model for learning on the strongly-augmented
series used as input. More specifically, we first define a set version of the same input. For time-series data, dynamic time
of time sampling functions, and then apply these sampling warping distance [10] based methods are used as distance
functions to each time series in the dataset. The resulting metrics in K-NN classifier to measure the similarity between
transformations are then feed to a ConvNets to recognize the labeled and unlabeled time series, and a phase-consistency
transformations of each time series. It is the set time sampling regularization [7] is applied to forbearing time series. More
functions that define the classification pretext tasks that iTimes recently, SemiTime [9] takes a past-future relation prediction
has to learn. Finally, the underlying temporal structure pattern of time segments as the pretext task to train the model
of unlabeled time series can be captured by completing the on unlabeled time series. However, the past-future relation
pretext task in the self-supervised module. The feature spaces prediction tries to capture a coarse-grained temporal structure,
between labeled data and unlabeled data can be aligned by which results in a sub-optimal representation quality on time
jointly training the supervised and self-supervised modules series. Therefore, how to design an efficient self-supervised
which boost the ability of model learning and the repre- pretext task that captures the temporal structure of unlabeled
sentation quality. Experimental results on multiple real-world time series is still an open problem.
datasets of different sources show that iTimes consistently
outperforms the state-of-the-art baselines, which demonstrates B. Self-supervised Learning
the effectiveness of the proposed method.
Self-supervised learning aims to learn the underlying struc-
In summary, the main contributions of this paper are as
tural features within the data under the supervision of the self-
follows: (1) We propose a framework of semi-supervised time
generated labels from data, which has achieved great success
series classification by investigating irregular time sampling
in natural language processing and computer vision domains
(iTimes), which learns the underlying temporal structure of
by defining and solving various pretext tasks such as restoring
unlabeled time series in a self-supervised manner to benefit
aspect ratio distortion [17], rotation prediction [18] and visual
semi-supervised time series classification. (2) We introduce
contrastive learning [19] for image, and frame order validation
four different irregular time sampling methods to transform
[20], playback rate perception [21] and pace prediction [22]
the original time series into different transformations, based
for video. For audio/speech data, there are also different self-
on which the fine-grained temporal structure can be captured
supervised techniques such as multi-task learning from raw
by predicting the types of different transformations. (3) We
audio by predicting a number of handcrafted features such as
conduct extensive experiments on multiple real-world time-
MFCCs, prosody, and waveform [23], and contrastive learning
series datasets, and the results show that our method achieves
on audio/speech data [24]. In [25], a inter-intra relational
the state-of-the-art on semi-supervised time series classifica-
reasoning based method is proposed to learn the representation
tion task.
of time series by identifying short, medium, and long term
relationships between segments and predicting positive and
II. R ELATED W ORK negative inter-sample relationships jointly. More recently, in
A. Semi-supervised Learning [26], a self-supervised learning model based on the wavelet
In the past few years, semi-supervised learning has gained transform (WT) for scalogram-signal correspondence learning
massive research in the domain of machine learning, which is proposed. In [27], self-supervised representation learning
provides an effective way to leverage unlabeled data to im- using heart rate (HR) signals as self-supervised signals for the
prove model performance [9]. Existing semi-supervised learn- activity data is proposed. In [28], six different transformations
ing approaches can be divided into two categories: entropy applied to ECG signals are proposed and the recognition of
minimization based methods [12], [13] and consistency reg- signal transformation types is used as a pretext task.
ularization based methods [14]–[16]. Entropy minimization-
based methods assume that the decision boundary should not III. M ETHOD
cross the high-density region of marginal distribution, and In this section, we show the proposed iTimes in detail. As
therefore try to train a classifier that can make low-entropy shown in Fig. 2, iTimes consists of three modules includ-
predictions on unlabeled data. For example, Pseudo-Label ing irregular time sampling module, supervised classification
[13] generates pseudo-labels of unlabeled data to supervise module, and self-supervised transformation prediction module.
model training on unlabeled data. Consistency regularization- The inputs of iTimes are a set of labeled input-target pairs
based methods assume that the model should be invariant DL = {(ti , yi )|ti = (t(i,1) , ...t(i,T ) ), yi ∈ {1, 2, ..., C}}M
i=1 ,
© 2022 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission.See https://www.ieee.org/publications/rights/index.html for more information.
Authorized licensed use limited to: Harbin University of Science and Technology. Downloaded on October 29,2022 at 04:18:54 UTC from IEEE Xplore. Restrictions apply.
This article has been accepted for publication in IEEE Transactions on Industrial Informatics. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/TII.2022.3199374
AUTHOR et al.: ITIMES: INVESTIGATING SEMI-SUPERVISED TIME SERIES CLASSIFICATION VIA IRREGULAR TIME SAMPLING 3
© 2022 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission.See https://www.ieee.org/publications/rights/index.html for more information.
Authorized licensed use limited to: Harbin University of Science and Technology. Downloaded on October 29,2022 at 04:18:54 UTC from IEEE Xplore. Restrictions apply.
This article has been accepted for publication in IEEE Transactions on Industrial Informatics. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/TII.2022.3199374
4 IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, VOL. XX, NO. XX, XXXX
Obviously, the number of solutions can be divide into 2 the sampled time series t′i,j to the backbone encoder fθ to get
cases, multiple solutions and unique solutions. For example, feature representation Zi,j = fθ (t′i,j ). Finally, the prediction
when l = 2, T = 5, all possible solutions will result in pi,j = hµ (Zi,j ) is made based on the obtained feature Zi,j .
a maximum or minimum value of ρ(g). If l = 4, T = 5 The supervised training loss Eq. (4) is defined as a cross-
maximizing ρ(g) or minimizing ρ(g) both have only one entropy loss:
solution, where ρ(g) obtains the maximum 0.774597 when the
|DL | |S|
solution g = {1, 1, 1, 2} and the minimum -0.774597 when 1 XX
g = {2, 1, 1, 1}. For the multi-solution case, we randomly Lsup = − yi · log(pi,j ) (4)
|DL | · |S| i=1 j=1
choose one of the solutions. After obtaining the time span g =
{g1 , g2 , · · · , gl } by maximizing or minimizing Eq. (1), the
sampling points {p1 , p2 , · · · , pl } are subsequently calculated
by pj = pj−1 + gj , p0 = 0, and finally the transformed time C. Training on Unlabeled Data
series t′i = (t(i,p1 ) , t(i,p2 ) , · · · , t(i,pl ) ) is obtained. To explore unlabeled data in temporal structure learning,
we use the self-generated time sampling function as super-
Algorithm 1 Irregular Time Sampling. vised learning signals and conduct transformation prediction
Require: tasks on unlabeled sampled data. Given unlabeled time series
ti : A T -length time series, α: Sampling rate datasets DU , set S of time sampling functions with α sampling
g: time spans rate. Each unlabeled time series ti will be sampled by the time
Ensure: sampling function sj ∈ S to obtain the sampled time series
t′i : Transformation of ti from irregular time sampling t′i,j . Input the sampled time series into the backbone encoder
1: function sampling(g) ▷ Generate sampling points fθ with shared weights to extract the feature embedding
2: p0 = 0 ▷ Starting sampling point Zi,j = fθ (t′i,j ), and then the obtained features are fed to the
3: for each gi in g do self-supervised transformation prediction head hϕ to get the
4: pi = pi−1 + gi ▷ Calculation of sampling points final prediction pi,j = hϕ (Zi,j ). Self-supervised training loss
5: end for Eq. (5) is defined as follow:
6: return (p1 , p2 , · · · , pl )
|DU | |S|
7: end function 1 XX
Lusp = − j · log(pi,j ) (5)
|DU | · |S| i=1 j=1
8: function s/ (ti , α)
9: l = ⌈T · α⌉
where j is the sampling function label. The model training
10: p = sampling(g) ▷ g obtained by Eq.2
algorithm is shown in Algorithm 2.
11: return (t(i,p1 ) , t(i,p2 ) , · · · t(i,pl ) )
12: end function
IV. E XPERIMENTS
13: function s\ (ti , α)
14: l = ⌈T · α⌉ In this section, we describe the experimental settings includ-
15: p = sampling(g) ▷ g obtained by Eq.3 ing datasets, baseline methods, and implementation. Then, we
16: return (t(i,p1 ) , t(i,p2 ) , · · · t(i,pl ) ) analysis the experimental results by comparing with the state-
17: end function of-the-art semi-supervised methods.
© 2022 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission.See https://www.ieee.org/publications/rights/index.html for more information.
Authorized licensed use limited to: Harbin University of Science and Technology. Downloaded on October 29,2022 at 04:18:54 UTC from IEEE Xplore. Restrictions apply.
This article has been accepted for publication in IEEE Transactions on Industrial Informatics. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/TII.2022.3199374
AUTHOR et al.: ITIMES: INVESTIGATING SEMI-SUPERVISED TIME SERIES CLASSIFICATION VIA IRREGULAR TIME SAMPLING 5
© 2022 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission.See https://www.ieee.org/publications/rights/index.html for more information.
Authorized licensed use limited to: Harbin University of Science and Technology. Downloaded on October 29,2022 at 04:18:54 UTC from IEEE Xplore. Restrictions apply.
This article has been accepted for publication in IEEE Transactions on Industrial Informatics. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/TII.2022.3199374
6 IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, VOL. XX, NO. XX, XXXX
75 24
s/
s/
41
(Train)
(Train)
23 77 73 89
(Test)
(Test)
ACC
ACC
72.42 0.00 47.03 0.00 48 22.71 0.00 73 72 86 34
73
s\
s\
s\
22 69 71 83 27
72.70 73.35 0.00 72 49.56 44.11 0.00 46 23.37 20.48 0.00 65 70 80 20
0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9
s
s
71 21 Sampling Rate ( ) Sampling Rate ( )
72.69 70.36 74.67 0.00 50.67 46.86 47.76 0.00 44 23.30 23.77 23.13 0.00
s
70
s
s/ s\ s s s/ s\ s s s/ s\ s s 20
(a) SmallKitchenAppliances (b) CricketX
(a) SmallKitchenAppli- (b) CricketX (c) Phoneme
27
(Train)
(Train)
87
(Test)
65
(Test)
78
ACC
ACC
72 21 83 60
75 94 80 66 18 79 55
0.00 0.00 0.00 60 15 75 50
78
s/
s/
s/
93 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9
63.48 0.00 70 92.37 0.00 75.76 0.00 76 Sampling Rate ( ) Sampling Rate ( )
s\
s\
s\
92
72.80 62.80 0.00 65 92.44 92.91 0.00 77.36 74.31 0.00 74
(c) Phoneme (d) MFPT
s
91
s
62.33 61.11 67.87 0.00 92.35 91.90 92.60 0.00 76.46 74.46 76.42 0.00 72
s
60
s
90
s
s/ s\ s s s/ s\ s s s/ s\ s s 70 93
(Train)
(Train)
90
(Test)
(Test)
98
ACC
ACC
(d) MFPT (e) FordB (f) EpilepticSeizure 97 89 85 68
96 87 80 64
95 85 75 60
Fig. 4. Impact of different compositions of two time sampling functions 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9
Sampling Rate ( ) Sampling Rate ( )
with sampling rate (α) 80% on different datasets (10% labeled data).
(e) FordB (f) EpilepticSeizure
74
Class. ACC
Class. ACC
(Test)
72 50 labeled data).
71 47
70 44
s\, s , s s/, s , s s/, s\, s s/, s\, s s/, s\, s , s s\, s , s s/, s , s s/, s\, s s/, s\, s s/, s\, s , s 76
51
27
75
24
Class. ACC
47
Class. ACC
Class. ACC
74
73 21
(a) SmallKitchenAppliances (b) CricketX 72
43
39 18
71
70 35 15
8 64 128 256 512 8 64 128 256 512 8 64 128 256 512
28 68
Class. ACC
Class. ACC
26 66
(Test)
(Test)
Class. ACC
Class. ACC
68 92 76
(c) Phoneme (d) MFPT 64 91 75
60 90 74
8 64 128 256 512 8 64 128 256 512 8 64 128 256 512
94 77
Class. ACC
Class. ACC
93 76
(Test)
(Test)
Fig. 5. Impact of different combinations of time sampling functions with and s/ that focuses on more information in the front may
sampling rate (α) 80% on different datasets (10% labeled data). not capture enough of the original information compared to a
uniform distribution of information in the time series.
In addition to examining the combination of two time
embedding size. As shown in Fig. 4, MFPT dataset and Epilep- sampling functions, we also examined the effect of the combi-
ticSeizure dataset obtains the best classification accuracy under nation of three time sampling functions and all time sampling
the combination of (s/ , s∨ ), CricketX, FordB, and SmallK- functions on time series classification. As shown in the Fig.
itchenAppliances dataset obtains the best classification under 5, compared with the combination of two time sampling func-
the combination of (s/ , s∧ ), (s\ , s∨ ), (s∨ , s∧ ) respectively, tions, the experimental results of three and all time sampling
and Phoneme dataset obtains the best classification accuracy function combinations do not have much variation, which
under the combination of (s\ , s∧ ) time sampling function. It is demonstrates the stability of iTimes.
evident that stable classification accuracies are achieved on the We next evaluate the effectiveness of the proposed irregular
MFPT and SmallKitchenAppliances datasets for all combina- time sampling module by investigating the effect of different
tions of two time sampling functions, more stable classification sampling rates α on time series classification with 10% labeled
accuracies are achieved on the two datasets containing the data, 64 embedding size under the optimal combination of time
largest number of samples for FordB and EpilepticSeizure, and sampling functions. As shown in Fig. 6, where the blue bar
some fluctuations in classification accuracies are observed on indicates transformation prediction accuracy on training data
the CricketX and Phoneme datasets. One possible reason for (St. Pred. ACC), and the line indicates classification accuracy
the fluctuations in classification accuracy in the two datasets on test data (ACC). As the sampling rate α increases, the
is that the information distribution of the time series is not classification accuracy keeps increasing, and after α increasing
suitable for the time sampling function. For example, the time to a certain value, the classification accuracy of all datasets
series may carry less information in the front than in the tail, except the MFPT dataset starts to decrease. We speculate that
© 2022 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission.See https://www.ieee.org/publications/rights/index.html for more information.
Authorized licensed use limited to: Harbin University of Science and Technology. Downloaded on October 29,2022 at 04:18:54 UTC from IEEE Xplore. Restrictions apply.
This article has been accepted for publication in IEEE Transactions on Industrial Informatics. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/TII.2022.3199374
AUTHOR et al.: ITIMES: INVESTIGATING SEMI-SUPERVISED TIME SERIES CLASSIFICATION VIA IRREGULAR TIME SAMPLING 7
TABLE III the MFPT and EpilepticSeizure, respectively, and 13.71% and
I MPACT OF DIFFERENT ENCODERS ON I T IMES . 11.29% over Fixmatch, respectively.
- SimConv4 ResNet18 ResNet34 3) Time series Classification: In this section, we evaluate
CricketX 49.78±3.42 52.81±2.91 55.21±3.05 iTimes by comparing it with other semi-supervised based state-
of-the-arts models on time series classification task. Following
the previous research [8], [14], we randomly select 10%, 20%,
40%, and 100% of the training set as labeled data and use the
as the sampling rate increases, it will make each sampled time
entire training set as unlabeled data for model training. As
series carry more original information, which is beneficial for
shown in the Table IV, our proposed method consistently out-
the model to learn the temporal structure. As the sampling
performs all benchmark models on all datasets. For example,
rate continues to increase, the distinguishability of time series
given 10% labeled data, the accuracy is 4.90% higher than
sampled by different time sampling functions becomes the less
that of SemiTime and 6.68% higher than that of MixMatch
noticeable to the extent that self-supervised learning cannot
on CricketX, 7.86% higher than that of FixMatch and 8.78%
learn the temporal structure.
higher than that of MTL on EpilepticSeizure, 7.89% higher
We then evaluate the impact of different encoders, a sim-
than that of Π-Model on EpilepticSeizure. Given 20% labeled
ple four-layer CNN and two residual network, on iTimes
data, our method improves the accuracy over SemiTime by
via withn 10% labeled CricketX dataset. The experimental
4.26% on Phoneme, over Π-Model by 14.10% on Small-
results are shown Table III, ResNet18, ResNet34 improved
KitchenAppliances, and 2.39% on FordB, respectively. The
the classification accuracy by 6.09% and 10.90% respectively
results show that predicting unlabeled data by using self-
compared with SimConv4. Although ResNet obtained better
ensembling or generating pseudo-label alone does not capture
experimental results, ResNet took up more GPU memory than
the potential temporal structure of time series effectively,
SimConv4.
which is of great significance to the semi-supervised repre-
We finally evaluate the stability of iTimes by examining sentation learning of time series, coarse-grained predicting
different backbone encoder embedding vector sizes |Zi,j | with whether past, future time series segments belong to the same
10% labeled data under 80% sampling rate. As shown in time series, cannot fully explore the temporal relationship
Fig. 7, where the line indicates classification accuracy on test within time series. It can also be seen that the classification
data (Class. ACC). As the embedding size |Zi,j | increases, accuracy decreases slightly as the amount of labeled data
the classification accuracy first increases and then stabilizes decreases. But all exceed the supervised and other baselines,
within a small fluctuation. The time series classification ac- demonstrating the stability of the model. Moreover, we also
curacy is poor at low-dimensional embedding because the use 100% training set as labeled and unlabeled data to evaluate
embedding vector size is so tiny that the representation learned the performance of our proposed method for supervised and
by iTimes is insufficient to express the semantic informa- self-supervised training, and the experimental results demon-
tion of the original samples. The classification accuracy of strate that our method is consistently better than supervised
iTimes achieves a steady state when the embedding vector learning and self-supervised learning based MTL. This shows
size exceeds 64, reflecting the stability of iTimes. We also that the forecasting pretext task of MTL can not capture the
discovered that the EpilepticSeizure dataset with more than internal temporal structure of unlabeled time series effectively,
10,000 samples achieved the maximum classification accuracy while we designed transformation prediction is able to learn
at 64 dimensions of embedding vector size, while the MFPT the internal structure of time series.
data with around 2,000 samples achieved the maximum at 4) Visualization: To qualitatively analyze the learned tempo-
256 dimensions of embedding vector size, indicating that the ral structure, we extract the representation embedding vector of
embedding vector size is not correlated with the sample size. the time series using a trained backbone encoder and visualize
We conjecture that although the sample quantity of MFPT is it in 2D space using the t-SNE [30] in order to verify the
tiny, each sample carries more diverse information compared semantic consistency of the learned embeddings. Fig. 8 shows
to EpilepticSeizure, resulting in the 64-dimensional embedding the extracted representation embedding vectors by iTimes as
vector size being insufficient to express the semantic informa- well as the baseline models on EpilepticSeizure dataset. As
tion of the sample within MFPT. can be seen from the Fig. 8, class Eyes open bounded by
2) Applications: In this section, we examine the perfor- circle obtains a better clustering performance compared to the
mance of iTimes in practical applications on the MFPT and other baselines. Thus by fine-grained examining the temporal
EpilepticSeizure datasets. MFPT is a bearing fault dataset relationships within unlabeled time series, iTimes is able to
that comprises one data from a bearing test rig and three learn more useful embeddings and enables better clustering
real-world fault data, divided into 14 abnormal classes under ability of time series, maintaining more semantic consistency.
different loads and one normal class. And EpilepticSeizure is Interestingly, we also found that EEG recordings from healthy
a commonly used EEG data for seizure detection, it includes brain regions and tumor brain regions were not clustered very
11,500 instances with 5 classes(Eyes open, Eyes closed, Tu- well, despite the better performance of our proposed model,
mor, Healthy, Seizure). The experimental results are shown in but this finding provides additional insights for data inspection
the Table IV. The classification accuracy of iTimes exceeds and model optimization.
that of all baseline methods. Given 10% labeled data, iTimes
improves 36.43% and 13.28% over Supervised method on
© 2022 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission.See https://www.ieee.org/publications/rights/index.html for more information.
Authorized licensed use limited to: Harbin University of Science and Technology. Downloaded on October 29,2022 at 04:18:54 UTC from IEEE Xplore. Restrictions apply.
This article has been accepted for publication in IEEE Transactions on Industrial Informatics. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/TII.2022.3199374
8 IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, VOL. XX, NO. XX, XXXX
TABLE IV
T EST CLASSIFICATION ACCURACY (%, AVERAGES OF 5 SEEDS AND 5 RUNS ) FOR SUPERVISED BASELINE AND SEMI - SUPERVISED LEARNING ON
DIFFERENT DATASETS . A LL METHODS USE THE SAME 4- LAYER CONVOLUTIONAL BACKBONE .
Label Ratio 10% 20% 40% 100% 10% 20% 40% 100% 10% 20% 40% 100%
Dataset SmallKitchenAppliances CricketX Phoneme
Supervised 63.11±3.50 63.11±2.30 68.23±3.51 73.03±2.65 33.62±0.95 38.97±2.08 52.64±2.53 62.98±2.01 19.54±0.69 22.45±1.14 27.82±1.62 35.90±1.4
Pseudo-Label [13] 52.46±4.09 62.29±2.96 73.04±3.75 - 38.87±2.26 44.44±2.91 53.39±2.18 - 18.92±0.72 21.40±0.99 26.34±0.68 -
Π-Model [14] 57.04±1.86 60.93±2.82 65.44±1.85 - 38.61±2.29 48.18±2.07 54.73±1.04 - 20.71±0.65 22.21±1.71 25.74±1.78 -
MTL [8] 65.03±2.71 66.69±3.03 71.23±2.19 75.30±3.24 40.94±1.97 50.12±1.22 55.10±1.12 63.58±1.72 19.98±1.67 20.00±0.66 25.58±1.17 32.34±0.78
MixMatch [15] 55.57±2.58 55.55±1.59 59.76±1.35 - 43.10±2.34 56.80±3.16 64.46±1.66 - 23.09±0.63 26.89±1.02 31.53±0.40 -
FixMatch [16] 65.14±3.36 65.60±4.23 74.10±2.48 - 38.33±2.07 50.89±2.04 60.66±1.27 - 18.89±1.80 21.58±1.57 27.72±1.0 -
SemiTime [9] 69.35±2.08 70.55±2.06 71.57±2.10 74.58±2.67 44.88±3.13 51.61±0.66 58.71±2.78 65.66±1.58 20.64±1.34 25.28±1.42 30.11±1.19 36.96±1.63
iTimes 74.13±4.91 75.03±4.94 77.38±2.41 81.23±1.28 49.78±3.42 58.64±2.59 68.37±2.75 77.71±1.40 24.23±1.53 29.54±2.12 36.54±0.42 45.12±1.2
Dataset MFPT EpilepticSeizure FordB
Label Ratio 10% 20% 40% 100% 10% 20% 40% 100% 10% 20% 40% 100%
Supervised 50.88±0.32 57.14±0.54 69.76±0.48 81.63±0.15 68.40±0.43 70.77±0.70 73.49±0.60 77.77±1.13 89.69±0.65 90.48±1.08 91.80±0.47 92.60±0.90
Pseudo-Label [13] 63.90±2.62 65.39±1.70 69.60±2.27 - 68.57±0.50 72.92±0.48 74.60±0.65 - 89.02±0.77 90.58±0.34 91.71±0.37 -
Π-Model [14] 55.41±0.65 59.68±0.43 70.15±1.01 - 69.60±0.34 71.58±0.64 74.54±0.55 - 88.94±0.74 90.64±0.37 91.02±0.14 -
MTL [8] 40.84±1.97 50.12±1.22 55.10±1.12 63.58±1.72 68.71±0.94 73.17±0.81 74.77±0.75 78.53±0.62 90.43±0.88 91.21±1.10 91.84±0.77 92.55±0.96
MixMatch [15] 52.56±1.01 68.62±0.80 81.87±1.60 - 67.99±0.09 69.67±0.30 72.22±0.63 - 90.75±0.44 91.34±0.39 91.72±0.58 -
FixMatch [16] 61.05±1.65 71.72±1.76 81.04±0.8 - 69.63±0.67 71.37±0.24 72.30±0.76 - 90.31±0.48 90.54±0.19 91.76±0.20 -
SemiTime [9] 64.16±0.85 69.84±0.94 76.49±0.54 84.33±0.50 74.86±0.42 75.54±0.63 77.01±0.79 79.26±1.02 90.97±0.88 91.46±0.68 91.80±0.48 92.40±0.74
iTimes 69.42±1.23 78.84±1.87 85.78±1.45 92.03±0.45 77.49 ± 0.49 78.93 ± 0.54 82.23 ± 0.56 82.87 ± 0.56 92.08±1.06 93.03±0.58 93.35±0.57 94.29±0.47
© 2022 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission.See https://www.ieee.org/publications/rights/index.html for more information.
Authorized licensed use limited to: Harbin University of Science and Technology. Downloaded on October 29,2022 at 04:18:54 UTC from IEEE Xplore. Restrictions apply.
This article has been accepted for publication in IEEE Transactions on Industrial Informatics. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/TII.2022.3199374
AUTHOR et al.: ITIMES: INVESTIGATING SEMI-SUPERVISED TIME SERIES CLASSIFICATION VIA IRREGULAR TIME SAMPLING 9
of natural images with convolutional neural network,” IEEE Transactions Xuxin Liu received the B.E. degree in Packag-
on Industrial Informatics, vol. 15, DOI 10.1109/TII.2018.2803041, no. 1, ing Engineeering from the Harbin University of
pp. 563–571, 2019. Commerce, Harbin, China, in 2020. He is cur-
[18] S. Gidaris, P. Singh, and N. Komodakis, “Unsupervised representation rently pursuing the M.S. degree in Computer Sci-
learning by predicting image rotations,” in 6th International Conference ence and Technology with the Harbin University
on Learning Representations, ICLR, 2018. of Science and Technology, Harbin, China. His
[19] T. Chen, S. Kornblith, M. Norouzi, and G. Hinton, “A simple framework current research interests include time series
for contrastive learning of visual representations,” in Proceedings of the analysis, network security, and machine learn-
37th international conference on machine learning (ICML), 2020. ing.
[20] D. Wei, J. J. Lim, A. Zisserman, and W. T. Freeman, “Learning and using
the arrow of time,” in Proceedings of the IEEE Conference on Computer
Vision and Pattern Recognition, CVPR, pp. 8052–8060, 2018.
[21] Y. Yao, C. Liu, D. Luo, Y. Zhou, and Q. Ye, “Video playback rate
perception for self-supervised spatio-temporal representation learning,”
in Proceedings of the IEEE/CVF Conference on Computer Vision and
Pattern Recognition (CVPR), Jun. 2020. Fengbin Zhang received the Ph.D. degree in
[22] J. Wang, J. Jiao, and Y.-H. Liu, “Self-supervised video representation Computer Application from Harbin Engineering
learning by pace prediction,” in European conference on computer University, China in 2005. He is currently a su-
vision, ECCV, 2020. pervisor and Professor at the Harbin University
[23] S. Pascual, M. Ravanelli, J. Serrà, A. Bonafonte, and Y. Ben- of Science and Technology. His current research
gio, “Learning problem-agnostic speech representations from multi- focuses on network and information security,
ple self-supervised tasks,” in Proc. of the Conf. of the Int. Speech firewall technology and intrusion detection tech-
Communication Association (INTERSPEECH), pp. 161–165, 2019. nology.
[24] A. v. d. Oord, Y. Li, and O. Vinyals, “Representation learning with
contrastive predictive coding,” arXiv preprint arXiv:1807.03748, 2018.
[25] H. Fan, F. Zhang, and Y. Gao, “Self-supervised time series repre-
sentation learning by inter-intra relational reasoning,” arXiv preprint Han Liu Han Liu received the M.S. degree
arXiv:2011.13548, 2020. in Computer Science and Technology from the
[26] A. Saeed, F. D. Salim, T. Ozcelebi, and J. Lukkien, “Federated Harbin University of Science and Technology,
self-supervised learning of multisensor representations for embed- Harbin, China, in 2021. He is currently studying
ded intelligence,” IEEE Internet of Things Journal, vol. 8, DOI for a Ph.D. degree at Harbin University of Sci-
10.1109/JIOT.2020.3009358, no. 2, pp. 1030–1040, 2021. ence and Technology, Harbin, China. His current
[27] D. Spathis, I. Perez-Pozuelo, S. Brage, N. J. Wareham, and C. Mascolo, research interests include artificial intelligence,
“Self-supervised transfer learning of physiological representations from network security, and machine learning.
free-living wearable data,” in Proceedings of the Conference on Health,
Inference, and Learning, pp. 69–78, 2021.
[28] P. Sarkar and A. Etemad, “Self-supervised ecg representation learning
for emotion recognition,” IEEE Transactions on Affective Computing,
DOI 10.1109/TAFFC.2020.3014842, pp. 1–1, 2020.
[29] T. T. Um, F. M. Pfister, D. Pichler, S. Endo, M. Lang, S. Hirche,
U. Fietzek, and D. Kulić, “Data augmentation of wearable sensor data for
parkinsonâs disease monitoring using convolutional neural networks,” in Haoyi Fan is an associate research fellow with
Proceedings of the 19th ACM International Conference on Multimodal the School of Computer and Artificial Intelli-
Interaction, pp. 216–220, 2017. gence, Zhengzhou University, Zhengzhou. He
[30] P. G. Poličar, M. Stražar, and B. Zupan, “opentsne: a modular python received the Ph.D. degree in the School of Com-
library for t-sne dimensionality reduction and embedding,” BioRxiv, p. puter Science and Technology, Harbin University
731877, 2019. of Science and Technology in 2021. His current
research interests include time series analysis,
data mining, and deep learning.
© 2022 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission.See https://www.ieee.org/publications/rights/index.html for more information.
Authorized licensed use limited to: Harbin University of Science and Technology. Downloaded on October 29,2022 at 04:18:54 UTC from IEEE Xplore. Restrictions apply.