A Dynamic Users' Interest Discovery Model With Distributed Inference Algorithm
A Dynamic Users' Interest Discovery Model With Distributed Inference Algorithm
Research Article
A Dynamic Users’ Interest Discovery Model with
Distributed Inference Algorithm
Shuo Xu,1 Qingwei Shi,1,2 Xiaodong Qiao,3 Lijun Zhu,1 Han Zhang,1 Hanmin Jung,4
Seungwoo Lee,4 and Sung-Pil Choi4
1
Information Technology Support Center, Institute of Scientific and Technical Information of China, No. 15 Fuxing Road,
Haidian District, Beijing 100038, China
2
School of Software, Liaoning Technical University, No. 188 Longwan Street South, Huludao, Liaoning 125105, China
3
College of Software, Northeast Normal University, 5268 Renmin Street, Changchun, Jilin 130024, China
4
Department of Computer Intelligence Research, Korea Institute of Science and Technology Information, 245 Daehak-ro, Yuseong-gu,
Daejeon 305-806, Republic of Korea
Copyright © 2014 Shuo Xu et al. This is an open access article distributed under the Creative Commons Attribution License, which
permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
One of the key issues for providing users user-customized or context-aware services is to automatically detect latent topics, users’
interests, and their changing patterns from large-scale social network information. Most of the current methods are devoted either
to discovering static latent topics and users’ interests or to analyzing topic evolution only from intrafeatures of documents, namely,
text content, without considering directly extrafeatures of documents such as authors. Moreover, they are applicable only to the
case of single processor. To resolve these problems, we propose a dynamic users’ interest discovery model with distributed inference
algorithm, named as Distributed Author-Topic over Time (D-AToT) model. The collapsed Gibbs sampling method following the
main idea of MapReduce is also utilized for inferring model parameters. The proposed model can discover latent topics and users’
interests, and mine their changing patterns over time. Extensive experimental results on NIPS (Neural Information Processing
Systems) dataset show that our D-AToT model is feasible and efficient.
am
am
xm,n
xm,n
𝛼 𝜗a zm,n
a ∈ [1, A]
wm,n
Figure 4: The graphical model representation of the AToT model.
𝜑k
𝛽
k ∈ [1, K] n ∈ [1, Nm ]
(1) For each topic 𝑘 ∈ [1, 𝐾],
m ∈ [1, M]
Figure 2: The graphical model representation of the AT model. (i) draw a multinomial 𝜑𝑘 from Dirichlet(𝛽);
zm,n
(3) for each word 𝑛 ∈ [1, 𝑁𝑚 ] in document 𝑚 ∈ [1, 𝑀],
From the above generative process, one can see that AToT
(1) For each topic 𝑘 ∈ [1, 𝐾], model is parameterized as follows:
3. Inference Algorithm 𝐾 × 𝑉 count matrix 𝑛𝑘(V) . From these data structures, one can
easily estimate the Φ and Θ as follows:
For inference, the task is to estimate the sets of the following
unknown parameters in the AToT model: (1) Φ = {𝜑𝑘 }𝐾 𝑘=1 , 𝑛𝑘(V) + 𝛽V
Θ = {𝜗𝑎 }𝐴 , and Ψ = {𝜓 }𝐾
and (2) the corresponding 𝜑𝑘,V = , (4)
𝑎=1 𝑘 𝑘=1
topic and author assignments 𝑧𝑚,𝑛 , 𝑥𝑚,𝑛 for each word token ∑𝑉 (V)
V=1 (𝑛𝑘 + 𝛽V )
𝑤𝑚,𝑛 . In fact, inference cannot be done exactly in this model.
A variety of algorithms have been used to estimate the param- 𝑛𝑎(𝑘) + 𝛼𝑘
𝜗𝑎,𝑘 = . (5)
eters of topics models, such as variational EM (expectation ∑𝐾
𝑘=1 (𝑛𝑎(𝑘) + 𝛼𝑘 )
maximization) [21, 22], expectation propagation [23, 24],
belief propagation [25], and Gibbs sampling [19, 20, 26, 27]. As for Ψ, similar to [14], for simplicity and speed, we
In this work, collapsed Gibbs sampling algorithm [26] is used, update it after each Gibbs sample by the method of moments
since it provides a simple method for obtaining parameter [28]:
estimates under Dirichlet priors and allows combination
of estimates from several local maxima of the posterior 𝑡𝑘 (1 − 𝑡𝑘 )
distribution. 𝜓𝑘,1 = 𝑡𝑘 ( − 1) ,
𝑠𝑘2
In the Gibbs sampling procedure, we need to calcu- (6)
late the conditional distribution 𝑃(𝑧𝑚,𝑛 , 𝑥𝑚,𝑛 | w, z¬(𝑚,𝑛) , 𝑡 (1 − 𝑡 )
x¬(𝑚,𝑛) , t, a, 𝛼, 𝛽, Ψ), where z¬(𝑚,𝑛) , x¬(𝑚,𝑛) represents the topic 𝜓𝑘,2 = (1 − 𝑡𝑘 ) ( 𝑘 2 𝑘 − 1) ,
and author assignments for all tokens except 𝑤𝑚,𝑛 , respec- 𝑠𝑘
tively. We begin with the joint distribution 𝑃(w, z, x, t |
a, 𝛼, 𝛽, Ψ) of a dataset, and, using the chain rule, we can get where 𝑡𝑘 and 𝑠𝑘2 indicate the sample mean and biased sample
the conditional probability conveniently as variance of the timestamps belonging to topic 𝑘, respectively.
The readers are invited to consult [28] for details. In fact,
similar to [14], since the Beta distribution with the support
[0, 1] can behave many more shapes including the bell
𝑃 (𝑧𝑚,𝑛 , 𝑥𝑚,𝑛 | w, z¬(𝑚,𝑛) , x¬(𝑚,𝑛) , t, a, 𝛼, 𝛽, Ψ)
curve than Gaussian distribution, it is utilized to model the
(𝑤 ) (𝑧 ) timestamps. But Wang and McCallum [14] did not provide
𝑛𝑧𝑚,𝑛𝑚,𝑛 + 𝛽𝑤𝑚,𝑛 − 1 𝑛𝑎 𝑚,𝑛 + 𝛼𝑧𝑚,𝑛 − 1
∝ × (1) much detail on how to handle documents with 0 and 1
∑𝑉 (V)
V=1 (𝑛𝑘 + 𝛽V ) − 1 ∑𝐾 (𝑘)
𝑘=1 (𝑛𝑥𝑚,𝑛 + 𝛼𝑘 ) − 1
timestamps so that they have some probability, so the time
range of the data is normalized to [0.01, 0.99] in the paper.
× Beta (𝜓𝑧𝑚,𝑛 ) , With (2)–(6), Gibbs sampling algorithm for AToT model
is summarized in Algorithm 1. The procedure itself uses only
seven larger data structures, the count variables 𝑛𝑎(𝑘) and 𝑛𝑘(V) ,
which have dimension 𝐴×𝐾 and 𝐾×𝑉, respectively, their row
where 𝑛𝑘(V) is the number of times tokens of word V are sums 𝑛𝑎 and 𝑛𝑘 with dimensions 𝐴 and 𝐾, Beta parameters Ψ
assigned to topic 𝑘 and 𝑛𝑎(𝑘) represents the number of times with dimension 𝐾 × 2, and the state variable 𝑧𝑚,𝑛 , 𝑥𝑚,𝑛 with
author 𝑎 is assigned to topic 𝑘. Detailed derivation of Gibbs dimension 𝑊 = ∑𝑀 𝑚=1 𝑁𝑚 .
sampling for AToT is provided in the appendix.
If one further manipulates the above (1), one can turn it
into separated update equations for the topic and author of 4. Distributed Inference Algorithm
each token, suitable for random or systematic scan updates:
Our distributed inference algorithm, named as D-AToT, is
inspired by AD-LDA algorithm [29, 30], following the main
(𝑧 )
idea of the well-known distributed programming model,
𝑛𝑥𝑚,𝑛
𝑚,𝑛
+ 𝛼𝑧𝑚,𝑛 − 1 MapReduce [18]. The overall distributed architecture for
𝑃 (𝑥𝑚,𝑛 | x¬(𝑚,𝑛) , z, a, 𝛼) ∝ , (2) AToT model is shown in Figure 5.
∑𝐾 (𝑘)
𝑘=1 (𝑛𝑥𝑚,𝑛 + 𝛼𝑘 ) − 1 As stated in Figure 5, the master firstly distributes 𝑀
𝑃 (𝑧𝑚,𝑛 | w, z¬(𝑚,𝑛) , x, t, 𝛼, 𝛽, Ψ) training documents over 𝑃 mappers, with nearly equal
number 𝑀/𝑃 of documents on each mapper. Specifically, D-
(𝑤 )
𝑛𝑧𝑚,𝑛𝑚,𝑛 + 𝛽𝑤𝑚,𝑛 − 1
(𝑧 )
𝑛𝑎 𝑚,𝑛 + 𝛼𝑧𝑚,𝑛 − 1 AToT partitions document {w}, {a}, and {t} into {{w|𝑝 }}𝑃𝑝=1 ,
∝ × (3) {{a|𝑝 }}𝑃𝑝=1 , and {{t|𝑝 }}𝑃𝑝=1 and corresponding topic and author
∑𝑉 (V)
V=1 (𝑛𝑘 + 𝛽V ) − 1 ∑𝐾 (𝑘)
𝑘=1 (𝑛𝑥𝑚,𝑛 + 𝛼𝑘 ) − 1
assignments {z} and {x} into {{z|𝑝 }}𝑃𝑝=1 and {{x|𝑝 }}𝑃𝑝=1 , where
× Beta (𝜓𝑧𝑚,𝑛 ) . {w|𝑝 }, {a|𝑝 }, {t|𝑝 }, {z|𝑝 }, and {x|𝑝 } exist only on mapper 𝑝.
The Author-Topic count {𝑛𝑎(𝑘) } and topic-word count {𝑛𝑘(V) } are
(𝑘) (V)
likewise distributed, denoted as {𝑛𝑎|𝑝 } and {𝑛𝑘|𝑝 } on mapper
During parameter estimation, the algorithm keeps track 𝑝, which are used to temporarily store local Author-Topic and
of two large data structures: an 𝐴 × 𝐾 count matrix 𝑛𝑎(𝑘) and a topic-word counts.
International Journal of Distributed Sensor Networks 5
Master
(k)
(k) ( )
} {na(k) } {n(k ) }
(k)
{na|p } {n(k|p) } {n(k) ( ) {na|p } {n(k|p) } {n(k) ( )
a } {n a }
{na|1 } {nk|1 a } {n k }
Reducer
{n(k)
a } {nk( ) }
Update Ψ
Calculate Φ and Θ
Table 2: Distribution of number of papers over year in NIPS dataset. the symmetric Dirichlet priors 𝛼 and 𝛽 are set at 0.5 and 0.1,
respectively. Gibbs sampling is run for 2000 iterations.
Year Number of papers
1987 90 (5.2%)
1988 95 (5.5%) 5.1. Examples of Topic, Author Distributions, and Topic Evo-
1989 101 (5.8%) lution. Table 3 illustrates examples of 8 topics learned by
1990 143 (8.2%) AToT model. The topics are extracted from a single sample
1991 144 (8.3%) at the 2000th iteration of the Gibbs sampler. Each topic
1992 127 (7.3%) is illustrated with (1) the top 10 words most likely to be
1993 144 (8.3%) generated conditioned on the topic, (b) the top 10 authors
1994 140 (8.0%) which have the highest probability conditioned on the topic,
1995 152 (8.7%)
and (c) histograms and fitted beta PDFs which show topics
evolution patterns over time.
1996 152 (8.7%)
1997 151 (8.7%)
1998 151 (8.7%) 5.2. Author Interest Evolution Analysis. In order to analyze
1999 150 (8.6%) further author interest evolution, it is interesting to calculate
𝑃 (𝑥𝑚,𝑛|𝑝 | x¬(𝑚,𝑛)|𝑝 , z|𝑝 , a|𝑝 , 𝛼) In this subsection, we take Sejnowski T as an example, who
published 43 papers in total from 1987 to 1999 in the NIPS
(𝑧 ) conferences, as shown in Figure 6(a). The research interest
𝑛𝑥𝑚,𝑛
𝑚,𝑛
+ 𝛼𝑧𝑚,𝑛 − 1
∝ , evolution for Sejnowski T is reported in Figure 6(b), in which
∑𝐾 (𝑘)
𝑘=1 (𝑛𝑥𝑚,𝑛 + 𝛼𝑘 ) − 1 the area occupied by a square is proportional to the strength
of his research interest.
𝑃 (𝑧𝑚,𝑛|𝑝 | w|𝑝 , z¬(𝑚,𝑛)|𝑝 , x|𝑝 , t|𝑝 , 𝛼, 𝛽, Ψ) (7) From Figure 6(b), one can see that Sejnowski T’s research
interest focused mainly on Topic 51 (Eye Recognition and
(𝑤 ) (𝑧 )
𝑛𝑧𝑚,𝑛𝑚,𝑛 + 𝛽𝑤𝑚,𝑛 − 1 𝑛𝑎 𝑚,𝑛 + 𝛼𝑧𝑚,𝑛 − 1 Factor Analysis), Topic 37 (Neural Networks), and Topic 58
∝ × (Data Model and Learning Algorithm) but with different
∑𝑉 (V)
V=1 (𝑛𝑘 + 𝛽V ) − 1 ∑𝐾 (𝑘)
𝑘=1 (𝑛𝑥𝑚,𝑛 + 𝛼𝑘 ) − 1 emphasis from 1987 to 1999. In the early phase (1989–1993),
Sejnowski T’s research interest is only limited to Topic 51
× Beta (𝜓𝑧𝑚,𝑛 ) and then extended to Topic 37 in 1994 and Topic 58 in
1996 with great research interest strength and finally back to
(𝑘) (V)
and updates local 𝑛𝑎|𝑝 and 𝑛𝑘|𝑝 according to the new topic and Topic 51 after 1997. Anyway, Sejnowski T did not change his
author assignments. After each iteration, each mapper sends main research direction, Topic 51, which is verified from his
the local counts to the reducer and then the reducer updates homepage again.
Ψ and broadcasts the global 𝑛𝑎(𝑘) , 𝑛𝑘(V) , and Ψ to all mappers.
After all sampling iterations, the reducer calculates the Φ and
Θ according to (4)-(5). 5.3. Predictive Power Analysis. Similar to [5], we further
divide the NIPS papers into a training set Dtrain of 1,557
papers and a test set Dtest of 183 papers of which 102 are single-
5. Experimental Results and Discussions authored papers. Each author in Dtest must have authored
at least one of the training papers. The perplexity, originally
NIPS proceeding dataset is utilized to evaluate the perfor- used in language modeling [31], is a standard measure for
mance of our model, which consists of the full text of the 13 estimating the performance of a probabilistic model. The
years of proceedings from 1987 to 1999 Neural Information perplexity of a test document 𝑚 ̃ ∈ Dtest is defined as the
Processing Systems (NIPS) Conferences. The dataset contains exponential of the negative normalized predictive likelihood
1,740 research papers and 2,037 unique authors. The distribu- under the model:
tion of the number of papers over year is shown in Table 2.
In addition to downcasing and removing stop words and
numbers, we also remove the words appearing less than five
times in the corpus. After the preprocessing, the dataset ̃ , t𝑚,⋅
perplexity (w𝑚,⋅ ̃ | a𝑚
̃ , 𝛼, 𝛽, Ψ)
contains 13,649 unique words and 2,301,375 word tokens in ln 𝑃 (w𝑚,⋅ (9)
̃ , t𝑚,⋅
̃ | a𝑚
̃ , 𝛼, 𝛽, Ψ)
total. Each document’s timestamp is determined by the year = exp [− ]
of the proceedings. In our experiments, 𝐾 is fixed at 100 and 𝑁𝑚̃
International Journal of Distributed Sensor Networks 7
Table 3: An illustration of 8 topics from a 100-topic solution for the NIPS collection. The titles are our own interpretation of the topics. Each
topic is shown with the 10 words and authors that have the highest probability conditioned on that topic. Histograms show how the topics are
distributed over time; the fitted beta PDFs is shown also.
1987
1988
1989
1990
1991
1992
1993
1994
1995
1996
1997
1998
1999
1987
1988
1989
1990
1991
1992
1993
1994
1995
1996
1997
1998
1999
1987
1988
1989
1990
1991
1992
1993
1994
1995
1996
1997
1998
1999
Table 3: Continued.
1987
1988
1989
1990
1991
1992
1993
1994
1995
1996
1997
1998
1999
1987
1988
1989
1990
1991
1992
1993
1994
1995
1996
1997
1998
1999
1987
1988
1989
1990
1991
1992
1993
1994
1995
1996
1997
1998
1999
7 87
6 37
Number of publications
5 11
88
4
Topic
47
3
78
2
51
1
58
0
1987
1988
1989
1990
1991
1992
1993
1994
1995
1996
1997
1998
1999
1987
1988
1989
1990
1991
1992
1993
1994
1995
1996
1997
1998
(a) Distribution of number of publications over time (b) Research interest evolution 1999
Figure 6: The distribution of number of publications and research interest evolution for Sejnowski T.
(10)
× ∫ 𝑝 (Φ | 𝛽, Dtrain ) ∑𝜑𝑧𝑚,𝑛 𝑑Φ 6. Conclusions
̃ ,𝑤𝑚,𝑛
̃
z𝑚,⋅
̃
With a dynamic users’ interest discovery model, one can
train answer many important questions about the content of
× ∫ 𝑝 (Θ | 𝛼, D ) ∑𝜗𝑥𝑚,𝑛 , 𝑧𝑚,𝑛
̃ 𝑑Θ.
x𝑚,⋅
̃
̃ information uploaded or shared to SNS. Based on our
previous work, Author-Topic over Time (AToT) model [19],
We approximate the integrals over Φ and Θ using the for documents using authors and topics with timestamps, this
point estimates obtained in (4)-(5) for each sample 𝑠 ∈ paper proposes a dynamic users’ interest discovery model
{1, 2, . . . , 10} of assignments x, z and then average over with distributed inference algorithm following the main idea
samples. Figure 7 shows the results for the AToT model and of MapReduce, named as Distributed AToT (D-AToT) model.
International Journal of Distributed Sensor Networks 9
𝑀 𝑁𝑚 𝐾
4000
= ∫ ∏ ∏𝑃 (𝑤𝑚,𝑛 | 𝜑𝑧𝑚,𝑛 ) ∏𝑝 (𝜑𝑘 | 𝛽) 𝑑Φ
𝑚=1 𝑛=1 𝑘=1
3500 𝑀 𝑁𝑚 𝐴
× ∫ ∏ ∏𝑃 (𝑧𝑚,𝑛 | 𝜗𝑥𝑚,𝑛 ) ∏𝑝 (𝜗𝑎 | 𝛼) 𝑑Θ
Perplexity
AT 𝐴 𝐾
𝑛𝑎(𝑘)
𝐴 Γ (∑𝐾
𝑘=1 𝛼𝑘 )
𝐾
𝛼 −1
AToT × ∫ ∏∏𝜗𝑎,𝑘 ∏ ( 𝐾 ∏𝜗𝑎,𝑘𝑘 ) 𝑑Θ
𝑎=1 𝑘=1 𝑎=1 ∏𝑘=1 Γ (𝛼𝑘 ) 𝑘=1
Figure 7: Perplexity of the 102 single-authored test documents.
𝑀 𝑁𝑚
× ∏ ∏𝑝 (𝑡𝑚,𝑛 | 𝜓𝑧𝑚,𝑛 )
𝑚=1 𝑛=1
𝐾 𝐴
The D-AToT model combines the merits of AT and ToT 1 Γ (∑𝑉
V=1 𝛽V ) Γ (∑𝐾
𝑘=1 𝛼𝑘 )
= 𝑁
( ) ( )
models. Specifically, it can automatically detect latent topics, ∏𝑀
𝑚=1 𝐴 𝑚
𝑚
∏𝑉
V=1 Γ (𝛽V ) ∏𝐾
𝑘=1 Γ (𝛼𝑘 )
users’ interests, and their changing patterns from large-scale
social network information. The results on NIPS dataset show 𝑀 𝑁𝑚 𝐾 ∏𝑉 (V)
V=1 Γ (𝑛𝑘 + 𝛽V )
the increase of salient topics and more reasonable users’ × ∏ ∏𝑝 (𝑡𝑚,𝑛 | 𝜓𝑧𝑚,𝑛 ) × ∏ 𝑉
𝑚=1 𝑛=1 𝑘=1 Γ (∑V=1 (𝑛𝑘(V) + 𝛽V ))
interest changing patterns.
One can generalize the approach in the work to construct
alternative dynamic models from other static users’ interest
𝐴 ∏𝐾 (𝑘)
𝑘=1 Γ (𝑛𝑎 + 𝛼𝑘 )
×∏ 𝐾
.
discovery models and ToT model with distributed inference 𝑎=1 Γ (∑𝑘=1 (𝑛𝑎(𝑘) + 𝛼𝑘 ))
algorithm. As a matter of fact, our work currently is limited
(A.1)
to deal with the users and latent topics with timestamps in
SNS. Though NIPS proceeding dataset is a benchmark data
for academic social network, the D-AToT model ignores the
links in SNS. In ongoing work, novel topic model, considering Using the chain rule, we can obtain the conditional
the links in SNS, will be constructed to identify the users with probability conveniently as follows:
similar interests from social networks.
𝑃 (w, t, z, x | a, 𝛼, 𝛽, Ψ)
𝑃 (w, z, x, t | a, 𝛼, 𝛽, Ψ) ∝
𝑃 (w¬(𝑚,𝑛) , t¬(𝑚,𝑛) , z¬(𝑚,𝑛) , x¬(𝑚,𝑛) | a, 𝛼, 𝛽, Ψ)
= 𝑃 (w | z, 𝛽) 𝑝 (t | Ψ, z) 𝑃 (z | x, 𝛼) 𝑃 (x | a) (𝑤 ) (𝑧 )
𝑛𝑧𝑚,𝑛𝑚,𝑛 + 𝛽𝑤𝑚,𝑛 − 1 𝑛𝑥𝑚,𝑛
𝑚,𝑛
+ 𝛼𝑧𝑚,𝑛 − 1
∝ ×
= ∫ 𝑃 (w | Φ, z) 𝑝 (Φ | 𝛽) 𝑑Φ × 𝑝 (t | Ψ, z) ∑𝑉 (V)
V=1 (𝑛𝑧𝑚,𝑛 + 𝛽V ) − 1 ∑𝐾 (𝑘)
𝑘=1 (𝑛𝑥𝑚,𝑛 + 𝛼𝑘 ) − 1
1
× ∫ 𝑃 (z | x, Θ) 𝑝 (Θ | 𝛼) 𝑑Θ × 𝑃 (x | a) × × 𝑝 (𝑡𝑚,𝑛 | 𝜓𝑧𝑚,𝑛 )
𝐴𝑚
10 International Journal of Distributed Sensor Networks
(𝑤 ) (𝑧 )
𝑛𝑧𝑚,𝑛𝑚,𝑛 + 𝛽𝑤𝑚,𝑛 − 1 𝑛𝑥𝑚,𝑛
𝑚,𝑛
+ 𝛼𝑧𝑚,𝑛 − 1 Mining (KDD ’07), pp. 500–509, San Jose, Calif, USA, August
∝ × 2007.
∑𝑉
V=1 (𝑛𝑧(V)𝑚,𝑛 + 𝛽V ) − 1 ∑𝐾
𝑘=1 (𝑛𝑥(𝑘)𝑚,𝑛 + 𝛼𝑘 ) − 1
[10] N. Kawamae, “Author interest topic model,” in Proceedings of the
33rd Annual International ACM SIGIR Conference on Research
× Beta (𝜓𝑧𝑚,𝑛 ) . and Development in Information Retrieval (SIGIR ’10), pp. 887–
(A.2) 888, ACM, Geneva, Switzerland, July 2010.
[11] N. Kawamae, “Latent interest-topic model: finding the causal
relationships behind dyadic data,” in Proceedings of the 19th
Conflict of Interests International Conference on Information and Knowledge Man-
agement and Co-located Workshops (CIKM ’10), pp. 649–658,
The authors declare that there is no conflict of interests ACM, Toronto, Canada, October 2010.
regarding the publication of this paper.
[12] J. Tang, J. Zhang, R. Jin et al., “Topic level expertise search over
heterogeneous networks,” Machine Learning, vol. 82, no. 2, pp.
Acknowledgments 211–237, 2011.
[13] X. Wang, N. Mohanty, and A. McCallum, “Group and topic
This work was funded partially by the Key Technolo- discovery from relations and their attributes,” in Advances in
gies R&D Program of Chinese 12th Five-Year Plan (2011– Neural Information Processing Systems 18, Y. Weiss, B. Schölkopf,
2015), Key Technologies Research on Large-Scale Seman- and J. Platt, Eds., pp. 1449–1456, MIT Press, Cambridge, Mass,
tic Calculation for Foreign STKOS, and Key Technologies USA, 2006.
Research on Data Mining from the Multiple Electric Vehicle [14] X. Wang and A. McCallum, “Topics over time: a non-markov
Information Sources under Grant nos. 2011BAH10B04 and continuous-time model of topical trends,” in Proceedings of
2013BAG06B01, respectively. the 12th ACM SIGKDD International Conference on Knowledge
Discovery and Data Mining (KDD ’06), pp. 424–433, August
2006.
References
[15] S. Xu, L. Zhu, Q. Xiaodong, S. Qingwei, and G. Jie, “Topic
[1] F. Qiu and J. Cho, “Automatic identification of user interest for linkages between papers and patents,” in Proceedings of the 4th
personalized search,” in Proceedings of the 15th International International Conference on Advanced Science and Technology,
Conference on World Wide Web (WWW ’06), pp. 727–736, pp. 176–183, Science & Engineering Research Support soCiety,
ACM, Edinburgh, UK, May 2006. Daejeon, Republic of Korea, 2012.
[2] J. Kim, D.-H. Jeong, D. Lee, and H. Jung, “User-centered [16] D. M. Blei and J. D. Lafferty, “Dynamic topic models,” in
innovative technology analysis and prediction application in Proceedings of the 23rd International Conference on Machine
mobile environment,” Multimedia Tools and Applications, 2013. Learning (ICML ’06), pp. 113–120, ACM, June 2006.
[3] M. Rosen-Zvi, T. Griffiths, M. Steyvers, and P. Smyth, “The [17] C. Wang, D. Blei, and D. Heckerman, “Continuous time
author-topic model for authors and documents,” in Proceedings dynamic topic models,” in Proceedings of the 24th Conference on
of the 20th Conference on Uncertainty in Artificial Intelligence Uncertainty in Artificial Intelligence (UAI ’08), pp. 579–586, July
(UAI ’04), pp. 487–494, AUAI Press, Arlington, Va, USA, 2004. 2008.
[4] M. Steyvers, P. Smyth, M. Rosen-Zvi, and T. Griffiths, “Prob- [18] J. Dean and S. Ghemawat, “MapReduce: simplified data process-
abilistic author-topic models for information discovery,” in ing on large clusters,” Communications of the ACM, vol. 51, no.
Proceedings of the 10th ACM SIGKDD International Conference 1, pp. 107–113, 2008.
on Knowledge Discovery and Data Mining (KDD ’04), pp. 306–
[19] S. Xu, Q. Shi, X. Qiao et al., “Author-topic over time (AToT):
315, ACM, Seattle, Wash, USA, August 2004.
a dynamic users’ interest model,” in Mobile, Ubiquitous, and
[5] M. Rosen-Zvi, C. Chemudugunta, T. Griffiths, P. Smyth, and Intelligent Computing: The 2nd International Conference on
M. Steyvers, “Learning author-topic models from text corpora,” Ubiquitous Context-Awareness and Wireless Sensor Network, vol.
ACM Transactions on Information Systems, vol. 28, no. 1, article 274, pp. 227–233, Springer, Berlin, Germany, 2014.
4, pp. 1–38, 2010.
[20] Q. Shi, X. Qiao, S. Xu, and G. Nong, “Author-topic evolution
[6] A. McCallum, A. Corrada-Emmanuel, and X. Wang, “The
model and its application in analysis of research interests evo-
author-recipient-topic model for topic and role discovery in
lution,” Journal of the China Society for Scientific and Technical
social networks: experiments with enron and academic email,”
Information, vol. 32, no. 9, pp. 912–919, 2013.
Tech. Rep. um-cs-2004-096, Department of Computer Science,
University of Massachusetts Amherst, 2004. [21] J. M. Winn, Variational message passing and its applications
[Ph.D. thesis], University of Cambridge, 2004.
[7] A. McCallum, A. Corrada-Emmanuel, and X. Wang, “Topic and
role discovery in social networks,” in Proceedings of the 19th [22] D. M. Blei, A. Y. Ng, and M. I. Jordan, “Latent Dirichlet
International Joint Conference on Artificial Intelligence, pp. 786– allocation,” Journal of Machine Learning Research, vol. 3, no. 4-5,
791, Morgan Kaufmann, San Francisco, Calif, USA, 2005. pp. 993–1022, 2003.
[8] A. McCallum, X. Wang, and A. Corrada-Emmanuel, “Topic and [23] T. P. Minka, “Expectation propagation for approximate
role discovery in social networks with experiments on enron Bayesian inference,” in Proceedings of the 17th Conference on
and academic email,” Journal of Artificial Intelligence Research, Uncertainty in Artificial Intelligence, pp. 362–369, Morgan
vol. 30, no. 1, pp. 249–272, 2007. Kaufmann, San Francisco, Calif, USA, 2001.
[9] D. Mimno and A. McCallum, “Expertise modeling for matching [24] T. Minka and J. Lafferty, “Expectation-propagation for the
papers with reviewers,” in Proceedings of the 13th ACM SIGKDD generative aspect model,” in Proceedings of the 18th Conference
International Conference on Knowledge Discovery and Data on Uncertainty in Artificial Intelligence, pp. 352–359, 2002.
International Journal of Distributed Sensor Networks 11