Enhancing Collaborative Filtering by User Interest Expansion Via Personalized Ranking
Enhancing Collaborative Filtering by User Interest Expansion Via Personalized Ranking
1, FEBRUARY 2012
Abstract—Recommender systems suggest a few items from automatically recommend the few optimal items, which users
many possible choices to the users by understanding their past might like or have interests to buy by learning the user profiles,
behaviors. In these systems, the user behaviors are influenced users’ previous transactions, the content of items, etc. [2]. In the
by the hidden interests of the users. Learning to leverage the
information about user interests is often critical for making better recent 20 years, many different types of recommender systems,
recommendations. However, existing collaborative-filtering-based such as collaborative-filtering-based methods [36], content-
recommender systems are usually focused on exploiting the in- based approaches [12], and hybrid approaches [46], have been
formation about the user’s interaction with the systems; the in- developed.
formation about latent user interests is largely underexplored.
To that end, inspired by the topic models, in this paper, we pro-
pose a novel collaborative-filtering-based recommender system by A. Collaborative Filtering
user interest expansion via personalized ranking, named iExpand.
The goal is to build an item-oriented model-based collaborative- Since collaborative-filtering methods only require the infor-
filtering framework. The iExpand method introduces a three- mation about user interactions and do not rely on the content
layer, user–interests–item, representation scheme, which leads to information of items or user profiles, they have more broad ap-
more accurate ranking recommendation results with less compu-
tation cost and helps the understanding of the interactions among plications [14], [16], [20], and more and more research studies
users, items, and user interests. Moreover, iExpand strategically on collaborative filtering have been reported [15], [26], [27].
deals with many issues that exist in traditional collaborative- These methods filter or evaluate items through the opinions
filtering approaches, such as the overspecialization problem and of other users [41]. They are usually based on the assumption
the cold-start problem. Finally, we evaluate iExpand on three that the given user will prefer the items which other users with
benchmark data sets, and experimental results show that iExpand
can lead to better ranking performance than state-of-the-art meth- similar preferences liked in the past [2].
ods with a significant margin. In the literature, there are model-based and memory-based
methods for collaborative filtering. Model-based approaches
Index Terms—Collaborative filtering, latent Dirichlet allocation
(LDA), personalized ranking, recommender systems, topic model. learn a model to make recommendation. Algorithms of this
category include the matrix factorization [38], the graph-based
approaches [14], etc. The common procedure of memory-based
I. I NTRODUCTION approaches is first to select a set of neighbor users for a
given user based on the entire collection of previously rated
T HE DEVELOPMENT of recommender systems has been
stimulated by the rapid growth of information on the
Internet. For information filtering, recommender systems can
items by the users. Then, the recommendations are made based
on the items that neighbor users like. Indeed, these methods
are referred to as user-oriented memory-based approaches. In
addition, an analogous procedure, which builds item similar-
Manuscript received November 8, 2010; revised March 24, 2011 and
June 24, 2011; accepted July 12, 2011. Date of current version December 7,
ity groups using corating history, is known as item-oriented
2011. This research was supported in part by the Natural Science Foundation memory-based collaborative filtering [40].
of China under Grants 60775037 and 71028002, by the Key Program of However, existing collaborative-filtering methods often di-
National Natural Science Foundation of China under Grant 60933013, and by
the Research Fund for the Doctoral Program of High Education of China under
rectly exploit the information about the users’ interaction with
Grant 20093402110017. A preliminary version of this work has been published the systems. In other words, they make recommendations by
in the Association for Computing Machinery Conference on Information and learning a “user–item” dualistic relationship. Therefore, exist-
Knowledge Management 2010. This paper was recommended by Associate
Editor J. Liu.
ing methods often neglect an important fact that there are many
Q. Liu and E. Chen are with the School of Computer Science and latent user interests which influence user behaviors. To that end,
Technology, University of Science and Technology of China, Hefei 230026, in this paper, we propose a three-layer, user–interests–item,
China (e-mail: [email protected]; [email protected]).
H. Xiong is with the Management Science and Information Systems representation scheme. Specifically, we interpret an interest as a
Department, Rutgers Business School, Rutgers University, Newark, NJ 07102 requirement from the user to items, while for the corresponding
USA (e-mail: [email protected]). item, the interest can be considered as one of its characteristics.
C. H. Q. Ding is with the Department of Computer Science and Engineering,
University of Texas at Arlington, Arlington, TX 76019 USA (e-mail: Indeed, it is necessary to leverage this three-layer representation
[email protected]). for enhancing collaborative filtering, since this representation
J. Chen is with the Department of Management Science and Engineering, leads to better explanation of why recommended items are
School of Economics and Management, Tsinghua University, Beijing 100084,
China (e-mail: [email protected]). chosen and helps the understanding of the interactions among
Digital Object Identifier 10.1109/TSMCB.2011.2163711 users, items, and user interests.
1083-4419/$26.00 © 2011 IEEE
LIU et al.: ENHANCING COLLABORATIVE FILTERING BY USER INTEREST EXPANSION 219
C. Contributions
To address the aforementioned challenges, in our prelim-
inary work [31], we proposed an item-oriented model-based
collaborative-filtering method named iExpand. In iExpand,
Fig. 1. Simple example of a movie recommender system (where the photos we assume that each user’s rating behavior depends on an
are downloaded from IMDB [http://www.imdb.com/)]. (a) When users decide underlying set of hidden interests and we use a three-layer,
to watch a movie, there are some latent interests that affect their choices.
(b) Users’ interests may change after they watch a movie. user–interests–item, representation scheme to generate recom-
mendations. Specifically, each user interest is first captured
by a latent factor which corresponds to a “topic” in topic
B. Motivating Example
models. Then, we learn the transition probabilities between
Fig. 1(a) shows an example of a movie recommender sys- different latent interests. Moreover, to deal with the cold-
tem. In the figure, user a is interested in kung fu movies, start and “overspecialization” problems, we model the possible
while user b likes Oscar movies. While both of them have expansion process of user interests by personalized ranking.
watched the movie Crouching Tiger, Hidden Dragon, which In other words, we exploit a personalized ranking strategy on
was recommended by the system, they have different reasons a latent interest correlation graph to predict the next possible
for watching this movie. Thus, if we can identify user latent interest for each user. At last, iExpand generates the recom-
interests, we will have a better understanding about the users’ mendation list by ranking the candidate items according to the
requirements, since user interests can better connect users and expanded user interests. We should note that, compared with
items. Also, when leveraging the information of user inter- previous topic-model-based collaborative-filtering approaches,
ests for developing recommender systems, we must be aware discovering the correlation between latent interests and using
that user interests can change from time to time under the personalized ranking to expand user current interests are the
influence of many internal and external factors. For instance, main advantages of iExpand.
as shown in Fig. 1(b), after watching the movie Crouching In addition, in many previous model-based recommender
Tiger, Hidden Dragon, user interests may be affected by it. systems, there are many parameters which are assigned default
For user a, while he is a fan of kung fu movies, he may start values. However, the best values for them should be determined
watching other movies directed by Ang Lee. Also, user b may in each particular scenario. In iExpand, we develop a model
become a fan of kung fu movies after her first-time exposure to select parameter values by combining Minka’s fixed-point
to this kung fu movie. If recommender systems cannot cap- iterations and an evaluation method for topic models.
ture these changes and only make recommendations according In this paper, we further explain why topic models can be
to the user’s past interests rather than exploring his/her new used to simulate the user latent interests and we demonstrate
preferences, then they are prone to the “overspecialization” the way of extracting these interests from the latent Dirichlet
problem [2]. allocation (LDA) model by the Gibbs sampling method. In
In addition, in real scenarios, the training data are far less addition, we illustrate how to use iExpand for making online
than plentiful and most of the items/users only have a few recommendations in the real-world applications. Finally, we
rating/buying records. At this time, typical measures fail to provide systematic experiments on three data sets selected from
capture actual similarities between items/users and the system a wide and diverse range of domains, and we use multiple
is unable to make meaningful recommendations. This situation evaluation metrics to evaluate the performance of iExpand.
is summarized as the cold-start problem [41]. Let us take Since iExpand views collaborative filtering as a ranking prob-
user b in Fig. 1(a) as an example. If she has only watched lem and aims to make recommendations by directly ranking the
one movie Crouching Tiger, Hidden Dragon and it has been candidate items, we report the ranking prediction accuracy. As
watched by few people before the rating of user b, for traditional shown in the experimental results, iExpand outperforms four
collaborative-filtering systems, it is difficult to find out the benchmark methods: two graph-based algorithms and two algo-
similar items or users for both Crouching Tiger, Hidden Dragon rithms based on dimension reduction. As many other algorithms
and user b. However, if we have identified that Crouching Tiger, formulate collaborative filtering as a regression problem (i.e.,
Hidden Dragon belongs to kung fu movies and Oscar movies, rating prediction) [30], we also report the comparison results of
then the system could recommend user b the movies that belong the rating predictions. In addition to this, these new experiments
to these two interests or some related interests. Thus, the cold- provide more insights into the iExpand model, such as the effect
start problem can be alleviated. of the parameters and the low computational cost.
220 IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS—PART B: CYBERNETICS, VOL. 42, NO. 1, FEBRUARY 2012
D. Outline
The rest of this paper is organized as follows. Section II gives
the detail of the iExpand method for effective recommendation.
In Section III, we show the experimental results and many
discussions. In Section IV, we introduce the related work.
Finally, Section V concludes this paper.
just deal with user current interests, the systems will suf-
fer from the overspecialization problem and the cold-start
problem.
To address this issue, we use PageRank [34], a personalized
ranking strategy on the user interest correlation graph. We
choose this strategy not only because it can create personalized
views of interest importance but also because it can predict user
interest expansion by exploiting the structure of the interest
Fig. 5. Example of interest–item bipartite graph. For simplicity, not all the correlation graph. Given a user interest (a vector), we do repeat
edges between each pair of item and interest are shown. PageRank iterations (i.e., guided random walks) until conver-
gence. The final converged PageRank score vector contains the
by the specific topic, i.e., they are class (topic)-conditional
expanded user interests. One can also view this as predicting the
distributions. Thus, the model representation of iExpand and
next possible interest for each user. Thus, we can make diverse
PLSA/LDA are significantly different from each other.
recommendations in a systematic way.
The algorithmic approach here is the personalized ranking
C. Correlation Graph of User Interests [25]. First, for user Ui , we represent his/her current interest
(0) (0)
In this section, we describe how to compute the transition model through vector θi in which the jth entry θi (j)
probabilities between latent interests by the correlation graph. corresponds to a latent interest Tj and is initialized as θij . From
(0)
In order to construct the correlation graph of latent interests, we Section II-C, we know θi is normalized and it represents
use the items as intermediary entities. ϕ is created to estimate the probability distribution on each latent interest when random
each item’s probability distribution over interests and ϕij can walk starts.
be estimated by Next, let θi perform Random Walk with Restart (RWR)
[18] (a specific implementation of the personalized ranking)
P (Tj , Ii ) φij ϑj
ϕij = P (Tj |Ii ) = = K . (3) on the correlation graph. Let us consider a random walk that
P (Ii ) (0)
ϑk φik starts from θi ; when arriving at Tj , it randomly chooses
k=1 Tj ’s neighbors and keeps walking. For each step, in ad-
Although the interests may be correlated to each other in dition to making such decisions, the random walker goes
reality, in LDA, when α is given, the distributions of interests back to the starting point with a certain probability c, so
are independent. Unlike the Correlated Topic Model [7], in as to counteract the dependence on far-away parts of the
iExpand, we model those correlations in the form of probabil- graph.
ities. Specifically, we first use a bipartite graph G = X, E to For example, the process of one step random walk of the user
represent the relationships between items and interests, with the Ui from step s to step (s + 1) can be formalized as
vertex set X = I ∪ T , as shown in Fig. 5. In the bipartite graph, (s+1) (s) (0)
the weight of the edge from interest Tj to item Ii is φij and the θi = (1 − c)θi ψ + cθi (5)
weight of the edge from Ii to Tj is ϕij .
while, for all the users, their one-step updates can be formal-
Then, by projecting G, we get the relationships between
ized as
interests, and we use ψ to represent them. Also, ψij indicates
the recommending strength of interest Ti for Tj , and it can be θ(s) = θ, s=0
computed by (6)
θ(s+1) = (1 − c)θ(s) ψ + cθ, s 0
N
N
(s)
ψij = P (Tj |Ti ) = P (Tj |In )P (In |Ti ) = ϕnj φni . (4) where θi serves as the interest vector for Ui after s steps of
n=1 n=1 random walk have completed. All users’ interest vectors form
(s)
At last, the bipartite graph is transformed into a correlation a matrix θ(s) where θij means the steady-state probability that
graph which describes the relations between interests, and ψ a random walk starts from user Ui and stops at interest Tj after
becomes its correlation matrix. It can be proven easily that s steps, meanwhile it implies the affinity of Tj with respect
(s)
each entry in ψ is equal to or greater than zero and ψ is row to Ui . The bigger θij , the closer Ui and Tj .
normalized. In terms of a correlation matrix, ψij means the The personalized ranking is run for all users simultaneously,
coefficient of correlation between Ti and Tj , from Ti ’s view. and it only takes several steps on average before θ(s) converges.
However, in terms of random walk, ψij is the probability that The parameter c indicates the restart probability, and (1 − c) is
current state jumps from Ti to Tj . the decay factor used to represent how much relationship is lost
in each step.
D. User Interest Expansion
E. From Expanded User Interests to the Item Recommendation
In this section, we describe the solution for the expan-
sion of user interests. As discussed previously, user interests In this section, we describe the last process of iExpand, the
often change from one to another. If recommender systems ranking of the items and the generation of recommendation
LIU et al.: ENHANCING COLLABORATIVE FILTERING BY USER INTEREST EXPANSION 223
lists. In iExpand, the items are ranked by their relevance with F. Estimating the Parameters
any given user. The user’s possible distribution on latent inter-
In this section, we present a method of selecting val-
ests serves as intermediary entities
ues for the parameters of iExpand. There are three para-
K
K meters: the hyperparameters α and β and the number of
(s)
P (Ij |Ui ) = P (Ij |t = k)Ps (t = k|Ui ) = φjk θik . (7) interests K.
k=1 k=1 First of all, we select the values for α and β. Previous
research works have found that α = 50/K and β = 0.01 work
It is easy to obtain the top K recommendations by well with different text collections, and they are often used
ranking the candidate items. Thus, iExpand directly gener- as the default values [10], [49]. However, Steyvers et al. [45]
ates recommendations without the step of predicting rating pointed out that good choices for these values depend on the
scores. number of interests and the item size. Furthermore, Asuncion
In addition, if the user rating has been taken into consid- et al. [5] suggested that hyperparameters play an important role
eration, iExpand can be used as a rating prediction method, in learning accurate topic models. Therefore, finding the best
such as the traditional memory-based collaborative-filtering α, β settings for each scenario is important. There are many
methods. Here, Pearson Correlation on expanded user interest ways for learning them [48], among which Minka’s fixed-point
vectors can be used to compute user similarities Sim(Ui , Uh ). iteration is widely used. It was proposed by Minka in [33] and
Therefore, the neighborhood N eighbor(Ui ) can be formed was carefully studied by Wallach [48]. In iExpand, each step of
for user Ui . Then, the rating from user Ui to item Ij can be fixed-point iteration is formalized as
predicted by
M K
α MK
Ψ Cmk + α − Ψ(α)
Sim(Ui , Uh ) ∗ (rh,j − rh )
α∗ ←−
m=1 k=1
Uh ∈Neighbor(Ui )
K
r̂i,j = ri +
|Sim(Ui , Uh )|
(8)
M MK
K Ψ Cmk + Kα − Ψ(Kα)
Uh ∈Neighbor(Ui ) m=1 k=1
K NK
N
where r̄i and r̄h are the average rating values for user Ui and β Ψ Cnk + β − Ψ(β)
Uh , respectively. rh,j refers to the rating value for item Ij from β ∗ ←− k=1 n=1
N (11)
user Uh .
K NK
N Ψ Cnk + N β − Ψ(N β) .
What we discussed earlier is about how to make recommen- k=1 n=1
dations in a general iExpand process. However, in real-world
applications, we face the challenge of online recommendations. Next, in addition to α and β, we choose the right value for
Since users’ interest distributions may change quickly from the interest number K. In previous works, if categories of the
time to time, while the correlation of interests evolves slowly, data sets are known, then K will be set equal to that number [9].
we can update both the inference process and the correlation However, in most scenarios, the category is unknown and how
graph periodically offline while renewing the user interests to set K becomes a problem. In most cases, K is randomly
whenever he/she rates. For example, when user Uu rates a new chosen or given a default value [5], [10], [51]. Until now,
item, Uu ’s interests can be resampled by Gibbs sampling. In one possible approach for setting this value is to compute the
each iteration, the interest assignment for every item in Uu ’s likelihood of the test data under different K values, then the
rating record is sampled by best one is chosen by a grid search. Exact computation of the
posterior probability is intractable, since it requires summing
P (tui = j|tu¬i , Uu , . . .) over all possible assignments. However, we can approximate it
by an estimator. In this paper, we refer to an approach proposed
CINuK u u
Cj¬i + α
j + Cj¬i + β by Wallach et al. [47] named Chib-style estimation which was
∝ i
(9)
N u
K u initially proposed as one evaluation method for topic models.
NK
Cnj + Cj + N β − 1 Ck + Kα − 1 The main idea of this approach is first to choose a special set
n=1 k=1
of latent topic assignments and then use Bayes’ rule to estimate
where tui = j means the interest assignment of item Iiu to the posterior probability.
u is a vector, and Cj u denotes the number of times Finally, as the posterior probability depends on α, β, and
interest Tj , C
K, we combine these factors together and propose a parameter
that interest Tj is assigned to the items in Uu ’s rating record.
learning algorithm, as shown in Algorithm 1. In Algorithm 1,
Also, ¬i refers to the interest assignments of all other items,
inputting the initial values of α and β, we first use Gibbs
not including the current instance. After performing interest
sampling and Minka’s fixed-point iteration to learn optimal
resampling, each interest distribution component of Uu can be
values for α and β specific to each number of interest K.
computed by
Then, Chib-style estimation is used to compute the posterior
Cju + α probability of the test data, under current parameter setting.
θuj = P (Tj |Uu ) = . (10) Lastly, the parameters with the best posterior probability are
K
chosen for the model, and they are used as the default settings
Cku + Kα
k=1 for performance comparison in the experimental part.
224 IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS—PART B: CYBERNETICS, VOL. 42, NO. 1, FEBRUARY 2012
TABLE II
D ESCRIPTION OF T HREE DATA S ETS
Algorithm 1: Estimating Parameters (a, b)
the training and test sets for Ui , where LUi and EUi mean the
item set that Ui rated in the training and test sets, respectively.
Furthermore, we define check_order as
1, if P RIj ≥ P RIk
check_orderUi (Ij , Ik ) =
0, otherwise
where P RIj denotes the predicted rank of item Ij in the
recommendation list. Then, the individual DOA for user Ui is
defined as
j∈EUi ,k∈N WUi check_orderUi (Ij , Ik )
DOAUi = .
|EUi | × |N WUi |
An ideal ranking corresponds to a 100% DOA, and we use
DOA to stand for the average of each individual DOA.
Top-K indicates the precision of the selected top K items,
and Recall measures the ratio of the number of hits to the size of
each user’s test data [39]. For each user Ui , these two measures
are defined as follows:
#hits #hits
T op − KUi = , RecallUi = .
K |EUi |
Fig. 6. Results of parameter selection for MovieLens. (a) Best α for different
For the purpose of evaluating the rating effectiveness, we number of interests. (b) Best β for different number of interests. (c) Log-
likelihood of posterior for different number of interests.
also choose the Mean Absolute Error (MAE) and the Root
Mean Squared Error (RMSE) as the evaluation metrics. Both TABLE III
of them are commonly used in traditional collaborative-filtering PARAMETER S ETTINGS
systems [2], [15], [20], [27].
C. Parameters in LDA
In this section, we investigate the learning performances of
two parameters, namely, hyperparameters and the number of mark approaches: ItemRank [18], L+̇ [14], UCF [36], SVD
interests, by Algorithm 1. Here, the first 893 users in Movie- [39],1 LDA, and RSVD [15]. For the purpose of comparison,
Lens are used as training data and the remaining 50 users form we record the best performance of each algorithm by tuning
the test set. Similarly, for Book-Crossing, the first 900 users their parameters. The training models of all these algorithms
are treated as training samples and the remaining users as test are learned only once, and ratings in the test set have never
data. Also, for Jester data set, the first 1800 users are treated as been used in the training process. Therefore, in order to make
training data and the remaining 200 users for testing. For each a clearer and fairer comparison, we do not take the online
run of Algorithm 1, we initialize the parameters as a = 0.5 and recommendation into consideration.
b = 0.5 and turn on Minka’s updates after 15 iterations, and First of all, we show a comparison of the effectiveness
these settings are similar to the ones in [5]. of all the algorithms. Tables IV and V and Fig. 7 show the
The estimation of posterior for the test set is computed for K performances of their recommendations with respect to dif-
sizes from 50 to 800 for both MovieLens and Book-Crossing ferent splits and different evaluation metrics. Table IV(a)–(c)
and K from 20 to 100 for Jester. The Gibbs sampling algorithm illustrates the evaluation results of the DOA/Recall measures.
runs 1000 iterations each time. Let us take the MovieLens Fig. 7 demonstrates the top K results on the three data sets, and
data set as an example. The results of parameter selection are Table V shows the evaluation results of the rating prediction
shown in Fig. 6. The results suggest that the test set are best accuracy on the MovieLens data set. Note that we did not report
accounted for by an LDA model incorporating 300 interests the rating prediction results on the Book-Crossing and Jester
and the corresponding best hyperparameter settings are α = data sets because the rating scale is too big for Jester and most
0.001 and β = 0.08. In Fig. 6, we can observe that the best of the ratings are 0 in Book-Crossing.
hyperparameters for collaborative filtering are different from DOA/Recall. In terms of DOA/Recall measures, from Ta-
those of text applications based on topic models. Finally, the ble IV, we can see that iExpand outperforms the other four
results of parameter selection are summarized in Table III. algorithms in each split. Also, the sparser the data, the more
significant improvement can be made. Indeed, both Item-
D. Performance Comparison
1 In our implementation, we rank the items by computing their Pearson
In this section, we present a performance comparison of both correlation with each user. This is slightly different from the implementation
effectiveness and efficiency between iExpand and the bench- in [39]; however, this way can yield better results for our situation.
226 IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS—PART B: CYBERNETICS, VOL. 42, NO. 1, FEBRUARY 2012
TABLE IV
P ERFORMANCE C OMPARISON OF D IFFERENT A LGORITHMS BASED ON DOA/R ECALL R ESULTS. (a) P ERFORMANCE C OMPARISON ON THE M OVIE L ENS
DATA S ET [(L EFT ) DOA IN P ERCENT. (R IGHT ) R ECALL IN P ERCENT.]. (b) P ERFORMANCE C OMPARISON ON THE B OOK -C ROSSING
DATA S ET [(L EFT ) DOA IN P ERCENT. (R IGHT ) R ECALL IN P ERCENT.]. (c) P ERFORMANCE C OMPARISON
ON THE J ESTER DATA S ET [(L EFT ) DOA IN P ERCENT. (R IGHT ) R ECALL IN P ERCENT.]
TABLE V
P ERFORMANCE C OMPARISON OF D IFFERENT A LGORITHMS BASED ON R ATING R ESULTS [(L EFT ) MAE. (R IGHT ) RMSE]
Rank and iExpand aim at alleviating the sparsity problem iExpand and LDA is interest expansion or not and because
and the cold-start problem, and they perform better than iExpand can expand user interests and increase the diversity
L+̇ , SVD, and LDA (except for Jester) when the training in a properly controlled manner, it performs much better than
sets are sparse, such as the 10–90 and 20–80 splits. How- LDA in all the cases. This means interest expansion can lead
ever, iExpand performs much better than ItemRank. For ex- to a better performance than only exploiting the current user
ample, in the three 10–90 splits, iExpand achieves nearly interests. Another interesting observation is that the smaller
two points of improvement on DOA values with respect to and sparser the training set, the more significant improve-
ItemRank. ment is made by iExpand compared with LDA, and when
In addition, both LDA and iExpand reduce data dimensions, the training set becomes larger and denser, the improvement
so they perform better when the data are dense, while SVD, becomes less obvious. The reason is that, when there are
another algorithm based on dimension reduction, does not enough interactions between a user and the system, the user
perform well. This may because of the use of different de- has experienced various types of items and his/her preference
composing techniques. Finally, as the main difference between has been decided. Hence, there will be not much difference
LIU et al.: ENHANCING COLLABORATIVE FILTERING BY USER INTEREST EXPANSION 227
Fig. 7. Performance comparison based on top K results. (a) 10-90, MovieLens. (b) 30-70, MovieLens. (c) 50-50, MovieLens. (d) 70-30, MovieLens. (e) 90-10,
MovieLens. (f) 10-90, Book-Crossing. (g) 30-70, Book-Crossing. (h) 50-50, Book-Crossing. (i) 70-30, Book-Crossing. (j) 90-10, Book-Crossing. (k) 10-90, Jester.
(l) 30-70, Jester. (m) 50-50, Jester. (n) 70-30, Jester. (o) 90-10, Jester.
from the current interest distribution to the next possible interest MAE/RMSE. From Table V, we can see that iExpand
distribution.2 performs the best on the two sparsest splits, while in gen-
Top-K. For better illustration, we select five splits from each eral, RSVD outperforms the other methods in terms of the
data set, and we only show the results of the three algorithms MAE/RMSE. On the sparse splits, the methods that can dis-
with the best top K performances. Fig. 7 shows the comparative cover the indirect correlations and deal with the cold-start
results of ItemRank, LDA, and iExpand, where the performance problem (i.e., iExpand and ItemRank) get better results than
of ItemRank is chosen as the baseline and the comparative other algorithms (i.e., RSVD, LDA, and UCF). However, on
results of LDA and iExpand against ItemRank on each k (k the remaining splits, the rating-oriented methods (i.e., RSVD
ranges from 5 to 25) are demonstrated. In Fig. 7, we can see that and UCF) generally perform better than the ranking-oriented
iExpand performs better than the baseline on almost every split, methods (i.e.,ItemRank, LDA, and iExpand). Another inter-
while there are more than five splits where LDA performs worse esting observation is that these two types of evaluation met-
than the baseline. Also, iExpand outperforms LDA, only except rics DOA/Recall/top K and MAE/RMSE lead to inconsistent
for the last two splits of Book-Crossing. In all, in terms of the judgements on the algorithms. The same observation has been
top K measure, in most cases, iExpand performs better than reported in many previous works [20], [30].
other methods. Finally, the sparser the data, the more significant Note that we chose SVD instead of RSVD for the ranking
improvement can be seen. This is similar to the results of comparison. The reason is that RSVD led to very bad results
DOA/Recall. which are not comparable with other methods in our ranking ex-
periments. In addition, the question about whether the ranking
prediction accuracy or the rating prediction accuracy is more
important is beyond the scope of this paper.
2 Only one exception is for the Jester data, where LDA performs nearly as
Runtime. Next, we compare the computational efficiency of
well as iExpand on the first two splits. This is because Jester data are a very
dense data, which can be seen from the data description in Table II, and this many algorithms. Fig. 8 shows the execution time of these algo-
alleviates the advantages of interest expansion. rithms on each data set. Without a surprise, on both MovieLens
228 IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS—PART B: CYBERNETICS, VOL. 42, NO. 1, FEBRUARY 2012
Fig. 8. Comparison of the execution time on each data set. (a) MovieLens data set. (b) Book-Crossing data set. (c) Jester data set.
sets is often bigger than that for smaller data sets. On the one
hand, when the training set is large, the correlation graph is
dense and there are plenty of direct contacts between vertices.
In this scenario, few indirect similarities needs to be considered,
and the random walk regresses to one step random walk or
there is no need for random walk. On the other hand, when
the correlation graph is sparse, random walk does not need to
restart frequently for lack of direct contacts. In this scenario, the
indirect contacts should be considered, and multistep random
walk will perform better than one-step random walk.
As an example, Fig. 9(d) shows the effect of the step of
random walk s on the performances of iExpand for two splits.
We can see that both curves converge after a few (no more than
ten) steps. The results show that random walk does enhance
the performance of iExpand and the best performance can be
achieved by just a few steps of random walk.
TABLE VI
T OP M OVIES IN THE F IRST T HREE L ATENT U SER I NTERESTS
TABLE VII
E XAMPLE OF U SER U140 AND THE C ORRESPONDING R ECOMMENDATION R ESULTS
movie Star Wars is given high probability in both latent interests Tables IV and V and Fig. 7, while this does not mean it will
1 and 3. This verifies that topic models can capture the multiple work for every single user and there may exist users whose
characteristics of each movie, and each characteristic can be interest expansion are different from the majority.
resolved by other movies in the corresponding latent interest.
The aforementioned analysis helps to map each latent interest
into explicit interests. This means that, even for collaborative G. Discussion
filtering, every latent factor still has a real meaning, although In this section, we analyze the advantages and limitations of
the interpretation may not be as easy and precise as that the iExpand method. From the experimental results, we can see
in text applications. Furthermore, this indicates that, in real that there are many key advantages of iExpand. First, iExpand
applications, if we only get several interest information input models the implicit relations between users and items through a
by the new user, we can still find out the possible items that a set of latent user interests. This three-layer representation leads
given user may like by the item–interest relationship described to more accurate ranking recommendation results. Second,
in iExpand model and thus mitigate the cold-start problem. iExpand can save the computational cost by reducing the num-
In the previous sections, we have showed that interest ex- ber of item dimensions. This dimensionality reduction can also
pansion can lead to a better performance than the method of help to alleviate the sparseness problem which is inherent to
only exploiting the current user interests. In the following, we many traditional collaborative-filtering systems. Third, iExpand
will illustrate the difference between these two recommending enables diverse recommendations by the interest expansion.
strategies by a user case. Let us consider the user U140 in This can help to avoid the overspecialization problem. Finally,
the MovieLens data set. The ratings of this user can be well iExpand can deal with the cold-start recommendations. This
divided into two types, thriller and nonthriller. According to means we only need several items or interests input by the new
this classification, we select the thriller movies to be the training user, and then, the corresponding items this user may like can
data and eight of the nonthriller movies to be the test set. Then, be predicted and recommended.
we run the two recommending strategies one by one, and we The main limitation of iExpand lies in its “bag of items”
get two types of recommendations in the end. The results are assumption, where in each user’s rating record, the rating
shown in Table VII. contextual information (e.g., rating time) is totally ignored.
In Table VII, we can see that the top eight recommendations However, Ding et al. [13] demonstrated that the ratings pro-
from the algorithm with interest expansion and the method duced at different times have different impacts on the prediction
without interest expansion are different from each other. First, of future user behaviors. Furthermore, Adomavicius et al. [3]
the method with interest expansion achieves a better result with presented a systematic discussion on the importance of contex-
more correctly predicted movies. Second, the recommendation tual information when providing recommendations. Thus, it is
results from the method with interest expansion are more possible for iExpand to further improve the recommendations
diversified.4 In other words, the interest expansion is more by considering the contextual information, such as time stamp
proper to capture the diversified interests and find potential and the rating orders.
interests for the users. Finally, we would like to point out that
this advantage is meaningful to most of the users which can be
seen from the results of the performance comparisons shown in IV. R ELATED W ORK
In general, related work can be grouped into four cat-
4 We should note that, in this case, we choose the movie genres as the criterion egories. The first category has a focus on the graph-
for diversity, and there may be other appropriate criterions. based collaborative-filtering methods. Here, the graph-based
230 IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS—PART B: CYBERNETICS, VOL. 42, NO. 1, FEBRUARY 2012
collaborative-filtering methods refer to those approaches which The third category of related work has a focus on solving
use the similarity of graph vertices to make recommendations the overspecialization problem in recommender systems. This
[14], [18], [44], [50], [52]. In these methods, users and items happens when the user is limited to being recommended the
are treated as vertices of a correlation graph and graph theory is items that are “similar” (with respect to content) to those
exploited for characterizing the relationship of user–item pairs. already rated [2]. In other words, at this time, users’ new or
The recommendation list is generated by considering how close latent interests will never be explored. This problem bothers
the candidate items are to a given user. The correlation graph most of the existing recommender systems, particularly for the
may consist of all users [44], all items [18], [50], [52], or both content-based approaches, where many studies have attempted
user and item vertices [14]. to find the solutions, for instance, filtering out the items which
While these graph-based collaborative-filtering methods are too similar to something the user has seen before [6] or
have elegant design ideas, they typically require more memory introducing some kind of serendipity [24].
and have high computational costs due to a large number of Since the overspecialization problem can be somewhat alle-
vertices. Moreover, most of these methods cannot explain why viated by the use of similar user interests, this problem has been
the items are chosen, and they provide limited understanding of largely ignored by most of the collaborative-filtering works.
the interactions among users, items, and user interests. However, some efforts have been dedicated to the solutions of
The second category includes the research work related to this issue. Among them, one possible approach is to consider
topic models, which are based upon the idea that documents are the transitive similarities in item-based collaborative filterings
mixtures of topics, where a topic is a probability distribution [18], [52]. However, directly computing the transitive simi-
over words. Many kinds of topic models have been proposed, larities between items will increase both the space and time
among which PLSA [21] and LDA [8] are most widely used costs. Another approach is to introduce diverse recommenda-
and studied. tions. For instance, Ziegler et al. [55] determined the overall
Before we describe topic models, we first introduce Latent diversity of the collaborative recommendations by introducing
Semantic Index (LSI), which was first proposed as a method the content information. Zhang et al. [54] modeled the goals of
for automatic indexing and retrieval [11]. LSI uses a technique maximizing the diversity of the recommendations while main-
called Singular Value Decomposition (SVD) to find the “latent taining adequate similarity to the user query as an optimization
semantic space” by decomposing the original matrix. LSI/SVD problem, and they applied this technique to an item-based
have been used for making recommendations probably since recommendation algorithm. Furthermore, a survey about some
2000 [28], [39]. Also, many SVD-based rating prediction meth- diversity enhancement algorithms was made in [53]. While the
ods are actually one of the successful competitors for the performances of these systems can be improved by introducing
Netflix prize [15], [27], [28]. These low rank recommenders diversity, most of them suffer from a tradeoff between diversity
usually treat collaborative filtering as a regression problem of and the recommending accuracy. A key reason is that they
user ratings. Although they perform well in rating predictions, neglect the fact that diversity should be made by exploiting
their effectiveness in generating recommendation lists should users’ possible interest expansion instead of randomly choosing
be further explored, since the rating prediction accuracy is not some explicit interests.
always consistent with the ranking accuracy [20], [30]. The fourth category of related work is focused on solving
The following PLSA topic model can be viewed as an the cold-start problem. Cold-start problem will happen when
enhancement of LSI. PLSA has a sound statistical foundation the recommender systems try to give recommendations to the
and has defined a proper generative model of the data [21]. users whose preference are underexplored or try to recommend
Also, PLSA is based on the observation that user preferences the new items whose characteristics are also unclear [2]. Thus,
and item characteristics are governed by a few latent semantics. it can be further classified as the item-side cold-start problem
As a statistical model, PLSA is able to capture the complex [42] and the user-side cold-start problem [29].
dependences among different factors using well-defined proba- For the content-based or the hybrid recommender systems,
bilistic semantics [30]. PLSA has been used both for automatic where there are profile descriptions, this problem can be al-
question recommendation [51] and collaborative filtering [22]. leviated by understanding items or users with such content
While PLSA has been successfully developed, it suffers from information. For instance, to deal with the item-side cold-
the overspecialization problem. start problem, Schein et al. proposed a probabilistic model
Compared with PLSA, the LDA model possesses fully gen- that combines item content and the collaborative information
erative semantics and has also been widely researched [5], [9], for recommendation [42]. To address the user-side cold-start
[47]. LDA is heavily cited in many text-related tasks, such as problem, Lam et al. proposed a User-Info Aspect Model by
finding scientific topics [19] and the information retrieval tasks using information of users, such as age and gender [29].
[49], but its feasibility and effectiveness in collaborative filter- However, for collaborative filtering, where there are no
ing is largely underexplored. Sometimes, topic models were content information, the only way to address the cold-start
only used to reduce the dimensionality of the data [10], [22], problem is to understand both users and items better from
like the function of principal component analysis [17]. In pre- the limited and sparse rating records. For instance, in order
vious topic-model-based collaborative-filtering algorithms, the to improve the recommendation performance under cold-start
correlation between latent factors has never been considered; conditions, Ahn [4] designed a heuristic similarity measure
thus, they easily suffer from the overspecialization problem and based on the minute meanings (i.e., proximity, impact, and
the cold-start problem. popularity) of coratings. Aside from exploring information
LIU et al.: ENHANCING COLLABORATIVE FILTERING BY USER INTEREST EXPANSION 231
from the direct relations among items (i.e., coratings), other sions,” IEEE Trans. Knowl. Data Eng., vol. 17, no. 6, pp. 734–749,
methods consider the indirect similarities. For instance, Huang Jun. 2005.
[3] G. Adomavicius and A. Tuzhilin, “Context-aware recommender systems,”
et al. [23] applied associative retrieval techniques to generate in Recommender Systems Handbook. New York: Springer-Verlag, 2011,
transitive associations in the user–item bipartite graph. In [35], pp. 217–253.
for alleviating the sparsity and the cold-start problems, the [4] H. J. Ahn, “A new similarity measure for collaborative filtering to alleviate
the new user cold-starting problem,” Inf. Sci., vol. 178, no. 1, pp. 37–51,
authors proposed a method using the trust inferences, which are Jan. 2008.
also transitive associations between users. Meantime, similar to [5] A. Asuncion, M. Welling, P. Smyth, and Y. W. Teh, “On
this paper, many random-walk-based similarity methods have smoothing and inference for topic models,” in Proc. Int. Conf. UAI,
2009, pp. 27–34.
been used in [14], [18], and [52]. However, these methods [6] D. Billsus and M. J. Pazzani, “User modeling for adaptive news access,”
consider the relationship between items or user–item pairs, User Model. User-Adapted Interaction, vol. 10, no. 2, pp. 147–180,
rather than the correlation between latent interests. Meanwhile, 2000.
[7] D. M. Blei and J. D. Lafferty, “A correlated topic model of science,” Ann.
as mentioned previously, with the increase of new items, users, Appl. Statist., vol. 1, no. 1, pp. 17–35, 2007.
or rating records, both their space and time costs will rise [8] D. M. Blei, Y. N. Andrew, and I. J. Michael, “Latent Dirichlet allocation,”
rapidly. J. Mach. Learn. Res., vol. 3, pp. 993–1022, 2003.
[9] K. R. Canini, L. Shi, and T. L. Griffiths, “Online inference of topics
with Latent Dirichlet allocation,” in Proc. 12th Int. Conf. AISTATS, 2009,
V. C ONCLUDING R EMARKS vol. 5, pp. 65–72.
[10] W. Chen, J. C. Chu, J. Luan, H. Bai, Y. Wang, and E. Y. Chang, “Collabo-
In this paper, we exploited user latent interests for devel- rative filtering for orkut communities: Discovery of user latent behavior,”
in Proc. 18th Int. Conf. WWW, 2009, pp. 681–690.
oping an item-oriented model-based collaborative framework, [11] S. Deerwester, S. T. Dumais, G. W. Furnas, T. K. Landauer, and
named iExpand. Specifically, in iExpand, a topic-model-based R. Harshman, “Indexing by latent semantic analysis,” J. Amer. Soc. Inf.
method is first used to capture each user’s interests. Then, Sci., vol. 41, no. 6, pp. 391–407, 1990.
[12] S. Debnath, N. Ganguly, and P. Mitra, “Feature weighting in content based
a personalized ranking strategy is developed for predicting a recommendation system using social network analysis,” in Proc. 17th Int.
user’s possible interest expansion. Moreover, a diverse recom- Conf. WWW, 2008, pp. 1041–1042.
mendation list is generated by using user latent interests as an [13] Y. Ding and X. Li, “Time weight collaborative filtering,” in Proc. 14th
ACM Int. CIKM, 2005, pp. 485–492.
intermediate layer between the user layer and the item layer. [14] F. Fouss, A. Pirotte, J.-M. Renders, and M. Saerens, “Random-walk
There are two key benefits of iExpand. First, the three-layer computation of similarities between nodes of a graph with application to
representation enables a better understanding of the interactions collaborative recommendation,” IEEE Trans. Knowl. Data Eng., vol. 19,
no. 3, pp. 355–369, Mar. 2007.
among users, items, and user interests and leads to more ac- [15] S. Funk, Netflix Update: Try This at Home, 2006. [Online]. Available:
curate ranking recommendation results. Second, since the user http://sifter.org/~simon/journal/20061211.html
interests and the change of the interests have been taken into [16] Y. Ge, H. Xiong, A. Tuzhilin, K. Xiao, M. Gruteser, and M. J. Pazzani,
“An energy-efficient mobile recommender system,” in Proc. 16th ACM
the consideration, iExpand can keep track of these changes and SIGKDD Int. Conf. KDD, 2010, pp. 899–908.
significantly mitigate the overspecialization problem and the [17] K. Goldberg, T. Roeder, D. Gupta, and C. Perkins, “Eigentaste: A con-
cold-start problem. stant time collaborative filtering algorithm,” Inf. Retrieval, vol. 4, no. 2,
pp. 133–151, Jul. 2001.
Finally, an empirical study has been conducted on three [18] M. Gori and A. Pucci, “A random-walk based scoring algorithm applied
benchmark data sets, namely, MovieLens, Book-Crossing, and to recommender engines,” in Proc. 8th Int. Workshop Knowl. Discov. Web
Jester. The corresponding experimental results demonstrate (WebKDD)—Advances in Web Mining and Web Usage Analysis, 2006,
pp. 127–146.
that iExpand can lead to better ranking performances than state- [19] T. L. Griffiths and M. Steyvers, “Finding scientific topics,” Proc. Nat.
of-the-art methods including two graph-based collaborative- Acad. Sci. U.S.A. (PNAS), vol. 101, pp. 5228–5235, 2004.
filtering algorithms and two dimension-reduction-based al- [20] J. L. Herlocker, J. A. Konstan, L. G. Terveen, and J. T. Riedl, “Evaluat-
ing collaborative filtering recommender systems,” ACM Trans. Inf. Syst.
gorithms. Due to an intellectual use of dimension-reduction (TOIS), vol. 22, no. 1, pp. 5–53, Jan. 2004.
techniques, iExpand also has low computational cost and is [21] T. Hofmann, “Probabilistic latent semantic analysis,” in Proc. 15th Conf.
highly scalable for a large number of users, items, and rating UAI, 1999, pp. 289–296.
[22] T. Hofmann, “Latent semantic models for collaborative filtering,” ACM
records. In the future, we plan to overcome the limitations of Trans. Inf. Syst. (TOIS), vol. 22, no. 1, pp. 89–115, Jan. 2004.
the current model and extend it to go beyond the usual recom- [23] Z. Huang, H. Chen, and D. Zeng, “Applying associative retrieval
mendations. In particular, we want to refine the iExpand model techniques to alleviate the sparsity problem in collaborative filter-
ing,” ACM Trans. Inf. Syst. (TOIS), vol. 22, no. 1, pp. 116–142,
so as to deal with the context-aware user–interests mining Jan. 2004.
problem. [24] L. Iaquinta, M. de Gemmis, P. Lops, and G. Semeraro, “Introducing
serendipity in a content-based recommender system,” in Proc. HIS, 2008,
pp. 168–173.
ACKNOWLEDGMENT [25] G. Jeh and J. Widom, “Scaling personalized web search,” in Proc. 12th
Int. Conf. WWW, 2003, pp. 271–279.
The authors are would like to thank the anonymous reviewers [26] Y. Koren, “Factorization meets the neighborhood: A multifaceted collab-
for their constructive comments. Q. Liu would like to thank the orative filtering model,” in Proc. 14th ACM SIGKDD Int. Conf. KDD,
2008, pp. 426–434.
China Scholarship Council for their support. [27] Y. Koren, “Collaborative filtering with temporal dynamics,” in Proc. 15th
ACM SIGKDD Int. Conf. KDD, 2009, pp. 447–456.
[28] M. Kurucz, A. A. Benczur, and K. Csalogany, “Methods
R EFERENCES for large scale SVD with missing values,” in Proc. KDDCup, 2007,
[1] Movielens Datasets, 2007. [Online]. Available: http://www.grouplens.org/ pp. 31–38.
node/73#attachments [29] X. N. Lam, T. Vu, T. D. Le, and A. D. Duong, “Addressing cold-start
[2] G. Adomavicius and A. Tuzhilin, “Toward the next generation of rec- problem in recommendation systems,” in Proc. 2nd Int. Conf. Ubiquitous
ommender systems: A survey of the state-of-the-art and possible exten- Inf. Manage. Commun., 2008, pp. 208–211.
232 IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS—PART B: CYBERNETICS, VOL. 42, NO. 1, FEBRUARY 2012
[30] N. N. Liu, M. Zhao, and Q. Yang, “Probabilistic latent preference Qi Liu received the B.E. degree in computer science
analysis for collaborative filtering,” in Proc. 18th ACM CIKM, 2009, from Qufu Normal University, Shandong, China, in
pp. 759–766. 2007. He is currently working toward the Ph.D.
[31] Q. Liu, E. Chen, H. Xiong, and C. H. Q. Ding, “Exploiting user interests degree from the School of Computer and Technol-
for collaborative filtering: Interests expansion via personalized ranking,” ogy, University of Science and Technology of China,
in Proc. 19th ACM CIKM, 2010, pp. 1697–1700. Hefei, China.
[32] Q. Mei, X. Shen, and C. Zhai, “Automatic labeling of multinomial He is currently supported by the China Scholar-
topic models,” in Proc. 13th ACM SIGKDD Int. Conf. KDD, 2007, ship Council and will stay for a year in Rutgers, The
pp. 490–499. State University of New Jersey, as a Visiting Re-
[33] T. Minka, Estimating a Dirichlet Distribution, 2000. [Online]. search Student in the Data Mining Group. His main
Available: http://research.microsoft.com/en-us/um/people/minka/papers/ research interests include intelligent data analysis,
dirichlet/minka-dirichlet.pdf recommender systems, and Web data mining. During his Ph.D. study, he has
[34] L. Page, S. Brin, R. Motwani, and T. Winograd, “The pagerank citation published several papers in refereed conference proceedings and journals.
ranking: Bringing order to the web,” Comput. Sci. Dept., Stanford Univ.,
Stanford, CA, Tech. Rep. 1999-0120, 1998.
[35] M. Papagelis, D. Plexousakis, and T. Kutsuras, “Alleviating the sparsity
problem of collaborative filtering using trust inferences,” in Proc. Trust Enhong Chen (SM’07) received the Ph.D. degree
Manage., 2005, pp. 224–239. from the University of Science and Technology of
[36] R. Paul, I. Neophytos, S. Mitesh, B. Peter, and R. John, “GroupLens: An China (USTC), Hefei, China.
open architecture for collaborative filtering of netnews,” in Proc. ACM He is a Professor and the Vice Dean of the School
Conf. CSCW, 1994, pp. 175–186. of Computer Science and Technology, USTC. His
[37] X. H. Phan, L. M. Nguyen, and S. Horiguchi, “Learning to classify short general areas of research are data mining, personal-
and sparse text & web with hidden topics from large-scale data collec- ized recommendation systems, and Web information
tions,” in Proc. 17th Int. Conf. WWW, 2008, pp. 91–100. processing. He has published more than 100 papers
[38] R. Salakhutdinov and A. Mnih, “Probabilistic matrix factorization,” in in refereed conferences and journals. His research is
Proc. Adv. Neural Inf. Process. Syst., 2000, pp. 1257–1264. supported by the National Natural Science Founda-
[39] B. Sarwar, G. Karypis, J. Konstan, and J. Riedl, “Application of dimen- tion of China, National High Technology Research
sionality reduction in recommender systems—A case study,” in Proc. and Development Program 863 of China, etc. He is the program committee
ACM WebKDD Workshop, 2000, pp. 82–90. member of more than 20 international conferences and workshops.
[40] B. Sarwar, G. Karypis, J. Konstan, and J. Riedl, “Item-based collaborative
filtering recommendation algorithms,” in Proc. 10th Int. Conf. WWW,
2001, pp. 285–295.
[41] J. B. Schafer, D. Frankowski, J. Herlocker, and S. Sen, “Collaborative
filtering recommender systems,” in Proc. Adapt. Web, Lecture Notes in Hui Xiong (SM’07) received the B.E. degree
Computer Science, 2007, pp. 291–324. from the University of Science and Technology
[42] A. I. Schein, A. Popescul, L. H. Ungar, and D. M. Pennock, “Methods of China, Hefei, China, the M.S. degree from the
and metrics for cold-start recommendations,” in Proc. 25th Annu. National University of Singapore, Singapore, and
Int. ACM SIGIR Conf. Res. Develop. Inf. Retrieval (SIGIR), 2002, the Ph.D. degree from the University of Minnesota,
pp. 253–260. Minneapolis, MN.
[43] S. Sen, J. Vig, and J. Riedl, “Tagommenders: Connecting He is currently an Associate Professor and the
users to items through tags,” in Proc. 18th Int. Conf. WWW, 2009, Vice Department Chair of the Management Science
pp. 671–680. and Information Systems Department, Rutgers Uni-
[44] X. Song, B. L. Tseng, C. Y. Lin, and M. T. Sun, “Personalized versity, NJ. His general area of research is data and
recommendation driven by information flow,” in Proc. 29th Annu. knowledge engineering, with a focus on developing
Int. ACM SIGIR Conf. Res. Develop. Inf. Retrieval (SIGIR), 2006, effective and efficient data analysis techniques for emerging data-intensive
pp. 509–516. applications. He has published over 90 technical papers in peer-reviewed
[45] M. Steyvers and T. Griffiths, “Probabilistic topic models,” in Handbook journals and conference proceedings. He is a Coeditor of Clustering and
of Latent Semantic Analysis, vol. 427. Mahwah, NJ: Lawrence Erlbaum Information Retrieval (Kluwer Academic Publishers, 2003) and a Co-Editor-
Associates, 2007, pp. 1–15. in-Chief of Encyclopedia of GIS (Springer, 2008). He is an Associate Editor of
[46] P. Symeonidis, A. Nanopoulos, and Y. Manolopoulos, “Providing justi- the Knowledge and Information Systems Journal and has served regularly in the
fications in recommender systems,” IEEE Trans. Syst., Man, Cybern. A, organization and program committees of a number of international conferences
Syst., Humans, vol. 38, no. 6, pp. 1262–1272, Nov. 2008. and workshops.
[47] H. M. Wallach, I. Murray, R. Salakhutdinov, and D. M. Mimno, “ Dr. Xiong is a senior member of the Association for Computing Machinery
Evaluation methods for topic models,” in Proc. 26th Annu. ICML, 2009, (ACM).
pp. 1105–1112.
[48] H. M. Wallach, “Structured topic models for language,”
Ph.D. dissertation, Univ. Cambridge, Cambridge, U.K., 2008.
[49] X. Wei and W. B. Croft, “LDA-based document models for ad-hoc Chris H. Q. Ding (M’09) received the Ph.D. degree
retrieval,” in Proc. 29th Annu. Int. ACM SIGIR Conf. Res. Develop. Inf. from Columbia University, New York, NY.
Retrieval (SIGIR), 2006, pp. 178–185. He is currently a Professor with the Department
[50] D. T. Wijaya and S. Bressan, “A random walk on the red carpet: Rating of Computer Science and Engineering, University
movies with user reviews and pagerank,” in Proc. 17th ACM CIKM, 2008, of Texas, Arlington (UTA). Prior to joining UTA,
pp. 951–960. he was in the Lawrence Berkeley National Labora-
[51] H. Wu, Y. Wang, and X. Cheng, “Incremental probabilistic latent semantic tory, University of California, Berkeley, and, prior
analysis for automatic question recommendation,” in Proc. ACM Conf. to that, with the California Institute of Technology,
RecSys, 2008, pp. 99–106. Pasadena. His general research areas are machine
[52] H. Yildirim and M. S. Krishnamoorthy, “A random walk method for learning/data mining and bioinformatics. He also
alleviating the sparsity problem in collaborative filtering,” in Proc. ACM works on information retrieval, Web link analysis,
Conf. RecSys, 2008, pp. 131–138. and high-performance computing. His research is supported by National Sci-
[53] M. Zhang, “Enhancing diversity in top-N recommendation,” in Proc. ence Foundation grants and by the University of Texas Regents STARS Award.
ACM Conf. RecSys, 2009, pp. 397–400. He has published over 150 research papers in peer-reviewed journals and
[54] M. Zhang and N. Hurley, “Avoiding monotony: Improving the conference proceedings, and these papers have been cited more than 5000
diversity of recommendation lists,” in Proc. ACM Conf. RecSys, 2008, times. He serves on many program committees of international conferences
pp. 123–130. and gave tutorials on spectral clustering and matrix models. He is an Associate
[55] C. Ziegler, S. M. McNee, J. A. Konstan, and G. Lausen, “Improving Editor of the journal Data Mining and Bioinformatics and is writing a book on
recommendation lists through topic diversification,” in Proc. 14th Int. spectral clustering to be published by Springer.
Conf. WWW, 2005, pp. 22–32. Dr. Ding is a member of the IEEE Computer Society since 2000.
LIU et al.: ENHANCING COLLABORATIVE FILTERING BY USER INTEREST EXPANSION 233