Opinion Based Mining
Opinion Based Mining
Opinion Based Mining
2 School
I. I NTRODUCTION
With the dramatic growth of opinionated user generated
content on the Web, to automatically understand, extract and
summarize the public opinions expressed in different online
media platforms has therefore become an important research
topic and gained much attention in recent years [13], [24].
Aspect-based opinion mining, a technique proposed originally
for nding elaborate opinions towards a perspective of a
product [19], has become a promising means for mining
aspect-level opinions for online public opinion analysis, where
the concept of an aspect here has been extended to be an
underlying theme, perspective or viewpoint as to a public
event. For instance, for the annual key event Two Sessions
(of the NPC and the CPPCC) 2015 in China, we would like
to know the elaborate public opinions towards a plenty of
relatively focused themes that have generated heated discussions, e.g., the downward pressure on GDP, the opportunities
in Jing-Jin-Ji integration, the Hukou reform, anti-corruption,
1550-4786/15 $31.00 2015 IEEE
DOI 10.1109/ICDM.2015.97
669
A. Model Description
As illustrated in Sect. II, our main task is to design a
model that can leverage complementary information from
diverse collections to jointly model aspects and opinions for
public opinion analysis. Along this line, two key problems
should be well addressed in the model. The rst one is how
to model aspects and opinions hidden in a collection in a
simultaneous and automatical way, which will be discussed
in detail in Sect. III-B. The second problem is how to capture
the complementarity across multiple collections with possible
severely asymmetric information, which is the main focus of
this subsection.
Let us again consider the case of aspect-based opinion mining from two asymmetric collections, i.e., the news collection
with clear aspects and the tweet collection with sharp opinions.
In the aspect level, to help extract clear aspects from tweets, it
is intuitive to share with the tweets side some similar aspects
found from the news side. This can be done by mining aspects
from the two collections separately, and then linking together
similar aspects from different sides. However, it would be
very difcult, if not impossible, to dene a proper similarity
measure and set a good threshold to it. Therefore, it would be
better to design a cross-collection model that could directly
mine some common aspects shared by different collections. In
the opinion level, however, it would be more interesting to read
public opinions from the tweet side, and compare them with
the opinions from the news side, which are often regarded as
the mainstream opinions from authoritative media. As a result,
our cross-collection model should be able to mine the opinions
separately from different sides for the purpose of comparison.
Based on the above reasoning, we now can describe our
model CAMEL, and give the generative process under it.
CAMEL is essentially a cross-collection LDA model with a
maximum entropy model embedded to determine the priors
for aspect and opinion words switching. CAMEL assumes
that different collections not only share some common aspects
but also have aspects of their own. Hereinafter, we call
aspects shared across collections as common aspects, and
call the aspects only contained in one collection as specic
aspects. CAMEL also assumes that each specic aspect has a
corresponding opinion, and each common aspect has multiple
corresponding opinions, one for each collection. We now describe how to generate a document under CAMEL as follows.
Suppose there are several multinomial word distributions
from a symmetric Dirichlet prior with parameter , including:
KI
K I common aspects {A
z }z=1 shared by all collections, with
O C
C opinions {z,c }c=1 for each A
z , where C is the number
S
S
S
corof collections; K specic aspects {zA }K
z=1 and K
670
KI
Notation
KI
KS
C
D
S
N
A
O
A
O
I
S
O
KS
{O, A}
C
w
z
x
y
r
Bern()
Beta()
M ulti()
Dir()
Description
the # of common aspects in total
the # of specic aspects for each collection
the # of collections
the # of documents in a collection
the # of sentences in a document
the # of words in a sentence
common aspect-word distribution
common opinion-word distribution
specic aspect-word distribution
specic opinion-word distribution
common aspect mixture for a document
collection specic aspect mixture for a document
parameters for common and specic aspect
switching for a sentence
parameters for aspect and opinion switching
for a word
an observed word
aspect index for a sentence
feature vector for the maximum entropy
(MaxEnt) model
weights learned by the MaxEnt model
aspect and opinion switcher for a word
common and specic aspect switcher
for a sentence
Dirichlet prior parameter for
Dirichlet prior parameter for all word distributions
symmetric Beta prior parameters for
Bernoulli distribution with parameter()
Beta distribution with parameter()
Multinomial distribution with parameter()
Dirichlet distribution with parameter()
b) Choose d Dir()
c) Choose d Beta()
d) For each sentence s:
i) Choose rd,s Bern(d )
ii) if rd,s = 0 choose zd,s M ulti(dI )
if rd,s = 1 choose zd,s M ulti(dS )
iii) For each word n:
A) Choose yd,s,n Bern(d,s,n )
B) if rd,s = 0 and yd,s,n = 0 choose
wd,s,n M ulti(A
zd,s )
if rd,s = 0 and yd,s,n = 1 choose
wd,s,n M ulti(O
zd,s ,c )
if rd,s = 1 and yd,s,n = 0 choose
wd,s,n M ulti(zAd,s ,c )
if rd,s = 1 and yd,s,n = 1 choose
wd,s,n M ulti(zOd,s ,c )
exp(l xd,s,n )
,
1l =0 exp(l xd,s,n )
671
A,j,k
is the total number
opinion word to common aspect k. C()
of times any word is assigned as an aspect word to common
O,j,k
is the total number of times any word is
aspect k, and C()
A,j,k
assigned as an opinion word to common aspect k. N(v)
is
the number of times word v is assigned as an aspect word to
O,j,k
aspect k in sentence s of document d, and similarly, N(v)
is the number of times word v is assigned as an opinion word
to aspect k in sentence s of document d. When j = 1, all
d
is the
counts mentioned above refer to specic aspects. C()
number of sentences in document d. Note that all these counts
represented by symbol C exclude sentence s of document d.
With assignments of z and r, we can sample y(d,s,n) for
y(d,s,n) = 0:
d
+
C(j)
,z
d,s d,s
C(wd,s,n
+
exp(1 xd,s,n )
)
.
1
O,r
,z
d,s
d,s
l =0 exp(l xd,s,n ) C
+V
()
,z
d,s d,s
C(wd,s,n
+
exp(0 xd,s,n )
)
,
1
A,r
,z
d,s
d,s
l =0 exp(l xd,s,n ) C
+V
()
d,j
+
C(k)
d + 2
d + Kj
C()
C(j)
A,j,k
V C A,j,k +N A,j,k +
C()
+V
(v)
(v)
A,j,k
A,j,k
A,j,k
C()
+N()
+V v=1
C(v)
+
O,j,k
V C O,j,k +N O,j,k +
+V
C()
(v)
(v)
.
O,j,k
O,j,k
O,j,k
C() +N() +V v=1
C(v) +
1 http://www.amazon.com
672
#Documents
3,535
3,659
7,194
9,662
82,366
1,015
13,004
#Sentences
56,053
60,935
116,988
251,023
144,149
42,651
25,425
#Words
315,471
339,192
654,663
3,561,412
1,136,422
262,613
74,353
B. Aspect Evaluation
We design a sentence classication experiment on reviews
to evaluate the quality of aspects learned by our method.
Note that all baseline methods as well as our method assigns
one aspect to each sentence, and each of induced aspect
often corresponds to one category, thus we can use sentence
classication as the evaluation method for learned aspects.
Specically, better sentence classication results indicate better quality of aspects. In order to illustrate asymmetric aspectbased opinion mining, we created a dataset with common
aspects across collections and specic ones only reside in each
collection as described in Sect.IV-A1.
Each review in our data set has one category in the set
L = {coffee machine, canister vacuum, MP3 player}. We use
the category of the review to label its sentences. Sentences of
coffee machine reviews are injected into reviews in collection
C0 or collection C1 randomly. Therefore, coffee machine sentences exist across collections. Sentences of canister vacuum
or MP3 player are only contained in C0 or C1 . In other words,
we expect aspects about coffee machine as common aspects,
those related to remain two categories as specic aspects.
We run BL-0, BL-1, BL-2 and our method over the data to
get inferred aspect set S A and aspect assignment k of each
sentence. In order to perform sentence classication evaluation, we have to manually map each aspect to one of the three
categories, i.e., we created a mapping functionf (k) : S A L.
Aspects can not be mapped to any category are labeled as
other. Given the mapping, we can get predicted labels of
sentences for each method. Then we evaluate all methods
according to metrics, namely precision, recall and f-measure.
For all methods, we set = 0.1, = 0.01. For our method,
we set = 0.1. In order to keep all methods comparable,
we have to set proper number of aspects for each method.
Specically, we set aspect number K2 of BL-2 equals to the
sum of K0 and K1 , where K0 is the aspect number for BL-0
2 http://nlp.stanford.edu/software/tagger.shtml
3 http://www.ltp-cloud.com/
673
precision
0.819
0.791
0.866
0.804
0.791
0.891
0.836
0.791
0.825
0.909
0.826
0.889
0.896
0.875
0.935
0.921
0.823
0.869
0.846
0.847
0.843
0.883
0.886
0.856
K=5
recall
0.783
0.833
0.646
0.780
0.819
0.727
0.787
0.847
0.836
0.876
0.943
0.841
0.871
0.954
0.832
0.881
0.819
0.862
0.794
0.811
0.845
0.841
0.809
0.864
f-measure
0.799
0.820
0.734
0.789
0.801
0.800
0.809
0.813
0.827
0.891
0.880
0.862
0.882
0.912
0.880
0.899
0.818
0.863
0.807
0.826
0.842
0.856
0.844
0.856
precision
0.745
0.861
0.856
0.695
0.856
0.858
0.826
0.865
0.797
0.919
0.861
0.882
0.904
0.894
0.935
0.934
0.780
0.869
0.859
0.789
0.880
0.876
0.881
0.899
K = 10
recall
f-measure
0.791
0.765
0.783
0.820
0.736
0.789
0.794
0.733
0.768
0.810
0.780
0.815
0.787
0.806
0.798
0.829
0.761
0.767
0.861
0.889
0.927
0.892
0.689
0.753
0.865
0.884
0.933
0.913
0.833
0.880
0.857
0.894
0.771
0.766
0.860
0.863
0.832
0.841
0.742
0.743
0.817
0.847
0.856
0.864
0.810
0.843
0.828
0.862
precision
0.795
0.829
0.864
0.771
0.820
0.879
0.827
0.838
0.810
0.923
0.857
0.886
0.910
0.895
0.940
0.934
0.805
0.868
0.860
0.828
0.865
0.887
0.883
0.886
K = 15
recall
f-measure
0.777
0.785
0.812
0.820
0.725
0.787
0.776
0.770
0.801
0.810
0.780
0.826
0.779
0.802
0.822
0.830
0.814
0.809
0.856
0.888
0.934
0.893
0.798
0.836
0.852
0.880
0.946
0.920
0.830
0.880
0.859
0.895
0.802
0.801
0.859
0.862
0.829
0.840
0.787
0.803
0.827
0.845
0.863
0.873
0.804
0.841
0.841
0.862
precision
0.833
0.819
0.846
0.824
0.816
0.877
0.843
0.821
0.810
0.922
0.855
0.892
0.907
0.890
0.930
0.937
0.818
0.869
0.851
0.858
0.843
0.884
0.886
0.879
K = 20
recall
f-measure
0.747
0.787
0.820
0.818
0.726
0.775
0.743
0.781
0.801
0.808
0.767
0.818
0.751
0.793
0.833
0.827
0.852
0.829
0.858
0.889
0.913
0.880
0.846
0.868
0.860
0.883
0.945
0.917
0.857
0.891
0.856
0.894
0.817
0.815
0.860
0.863
0.819
0.828
0.795
0.824
0.831
0.846
0.856
0.867
0.804
0.842
0.844
0.860
674
Method
BL-0
Ours-C0
BL-1
Ours-C1
BL-2
Ours
Ours
Ours-C0
Ours-C1
(k)
log
(k)
D(vl )
t=2 l=1
(k)
(k)
D(vt , vl ) +
(k)
T
T
t=1 l=1
A,(k)
log
D(vt
O,(k)
, vl
A,(k)
D(vt
)
)+
aspect-opinion coherence
T = 10
T = 15
-229.35.3
-572.17.1
-226.26.2
-571.315.7
-246.75.5
-621.111.6
-245.13.2
-613.04.8
-255.23.5
-640.712.1
-240.63.4
-604.06.4
-192.33.0
-483.54.8
-246.66.5
-598.820.1
-245.96.0
-612.110.2
C. Opinion Evaluation
CO (k; V (k) ) =
opinion coherence
T = 10
T = 15
-121.13.6
-307.67.2
-119.44.3
-307.79.0
-129.83.7
-334.911.9
-127.71.7
-326.75.4
-134.82.7
-345.67.9
-129.82.0
-329.74.1
-103.72.6
-260.93.9
-127.54.6
-309.66.5
-126.04.5
-320.29.6
675
(a) C0 : recall.
(b) C0 : f-measure.
Fig. 2: Common category classication results with different imbalance ratios on C0 or C0 combined with C1 .
simulate the imbalance circumstances, we reform the online
reviews data used in Sect.IV-B by removing sentences labeled
as coffee machine(the common category) from collection C0 ,
while collection C1 remains the same.
To creat different levels of imbalance, we remove coffee
machine sentences according to ratios from 10% to 80%,
which is the proportion of coffee machine sentences we
remove from collection C0 . With reformed data set, we can
perform coffee machine sentences classication on C0 and
C0 combined with C1 to discuss the complementary aspect
mining. We set = 0.1, = 0.01 for all methods, = 0.1
for our method. K = K0 = K1 = 15, K2 = 30, K I = 6 and
K S = 12. The resulting recall and f-measure are illustrated
in Fig. 2.
In Fig.2a and Fig. 2b we can see without complementary
information, the recall and f-measure of BL-0 both drop
rapidly and become extremely unstable when ratio reaches
80%. Since BL-2 and our method both utilize complementary
common aspects in C1 , therefore is relative stable than BL-0.
As we could see in Fig 2, the performance of BL-2 on C0 as
well as on C0 and C1 , also become rather poor and unstable
as ratio reaches 80%. This promising result strongly suggested
the explicit separation of common aspects and specic aspects
is necessary for complementary aspect mining.
= 10
= 20
= 30
= 40
= 50
P @5
MME
AME
0.90
0.64
0.80
0.64
0.86
0.84
0.80
0.84
0.82
0.90
P @10
MME
AME
0.82
0.58
0.74
0.56
0.82
0.81
0.75
0.85
0.81
0.83
P @20
MME
AME
0.71
0.51
0.67
0.50
0.70
0.76
0.71
0.78
0.71
0.76
News:
!
Thanks to you and those who also shouted fallback, you
are good man!
!
How calm and wise the fallback shouters are, great!
To justify the validity of our auto-labeled MaxEnt component, we give comparisons of manual labeled MaxEnt(MME
for short) with auto-labeled MaxEnt(AME for short). We compare the P recision@n(P @n for short) of MME with AME
on reviews data with varying number of training sentences
S. Here P @n means how many words are precisely opinion
words other than aspect words given the top n probability
words of an opinion, which is judged by human. For MME,
we select 50 sentences with opinions words and manually label
them. As to AME, we use our automatic procedure to acquire
the same number of training sentences. We increase the size
of training data inputted into the MaxEnt, and compare the
P @5, P @10 and P @20 of MaxEnt-LDA [27] with different
sources of training data. The results are reported in Table VI.
From the results, we can nd that AME is inaccuracy when
training size is small. However, when training size is larger
than 30, AME is even more accuracy than MME. Note that the
training data of AME is obtained fully automatically, therefore,
676
V. R ELATED W ORK
Opinion
Opinion
677
TABLE X: A sample of common aspect-opinions induced from APEC and Stampede data.
Common Aspect 1
crowd scene bund policeman
fallback shouters new year event
tourist order voice
opinion 0
opinion 1
(recede)
(above)
(great)
(tumble)
(touched)
(suspected)
(calm)
(at all)
(hope)
(down)
(tender)
(around)
(reduce)
(crowded)
(important)
(stable)
(dangerous)
(noisy)
(thank)
(nearby)
Common Aspect 2
victims namelist event bund
Shanghai identity stampede
square forenoon age
opinion 0
opinion 1
(grieve)
(announce)
(sympathize)
(verify)
(blessing)
(injure)
(regret)
(check)
(perfunctory)
(crowed)
(ignorant)
(happen)
(distressed)
(sympathise)
(appease)
(grieve)
(simple)
(heartfelt)
(fortunately)
(mourn)
Common Aspect 3
China society diplomacy
reformation world economy advance
international Jinping Xi legislation
opinion 0
opinion 1
(cheer)
(reform)
(progress)
(develop)
(thorough)
(overall)
(great)
(deepen)
(overall)
(innovate)
(happy)
(peaceful)
(civil)
(signicant)
(braw)
(great)
(strong)
(revival)
(fair)
(fair)
Common Aspect 4
APEC
China tariff APEC products
agreement free-trade price
commodity enterprise Australia
opinion 0
opinion 1
(import)
(invest)
(cheap)
(reduce)
(expect)
(cancel)
(bring)
(open)
(develop)
(signicant)
(convenient)
(investigate)
(recover)
(serve)
(share)
(export)
(free)
(promote)
(inuence)
(improve)
[4] Yi Fang, Luo Si, Naveen Somasundaram, and Zhengtao Yu. Mining
contrastive opinions on political texts using cross-perspective topic
model. In WSDM 12, pages 6372, 2012.
[5] Wei Gao, Peng Li, and Kareem Darwish. Joint topic modeling for event
summarization across news and social media streams. In CIKM 12.
[6] T. L. Grifths and M. Steyvers. Finding scientic topics. PNAS04.
[7] Honglei Guo, Huijia Zhu, Zhili Guo, XiaoXun Zhang, and Zhong
Su. Product feature categorization with multilevel latent semantic
association. In CIKM 09, pages 10871096, 2009.
[8] Thomas Hofmann. Probabilistic latent semantic indexing. In SIGIR 99.
[9] Minqing Hu and Bing Liu. Mining and summarizing customer reviews.
In KDD 04, pages 168177, 2004.
[10] Wei Jin and Hung Hay Ho. A novel lexicalized hmm-based learning
framework for web opinion mining. In ICML 09, pages 465472, 2009.
[11] Wei Jin, Hung Hay Ho, and Rohini K. Srihari. Opinionminer: A novel
machine learning system for web opinion mining and extraction. In
KDD 09, pages 11951204, 2009.
[12] Yohan Jo and Alice H. Oh. Aspect and sentiment unication model for
online review analysis. In WSDM 11, pages 815824, 2011.
[13] Kar Wai Lim and Wray Buntine. Twitter opinion topic model: Extracting
product opinions from tweets by leveraging hashtags and sentiment
lexicon. In CIKM 14, pages 13191328. ACM, 2014.
[14] Chenghua Lin and Yulan He. Joint sentiment/topic model for sentiment
analysis. In CIKM 09, pages 375384, 2009.
[15] Bing Liu, Minqing Hu, and Junsheng Cheng. Opinion observer:
Analyzing and comparing opinions on the web. In WWW 05.
[16] Qiaozhu Mei, Xu Ling, Matthew Wondra, Hang Su, and ChengXiang
Zhai. Topic sentiment mixture: Modeling facets and opinions in weblogs.
In WWW 07, pages 171180, 2007.
[17] David Mimno, Hanna M. Wallach, Edmund Talley, Miriam Leenders,
and Andrew McCallum. Optimizing semantic coherence in topic models.
In EMNLP 11, pages 262272, 2011.
[18] Samaneh Moghaddam and Martin Ester. Opinion digger: An unsupervised opinion miner from unstructured product reviews. In CIKM 10.
[19] Samaneh Moghaddam and Martin Ester. Aspect-based opinion mining
from product reviews. In SIGIR 12, pages 11841184, 2012.
[20] Samaneh Moghaddam and Martin Ester. On the design of lda models
for aspect-based opinion mining. In CIKM 12, pages 803812, 2012.
[21] Michael Paul and Roxana Girju. Cross-cultural analysis of blogs and
forums with mixed-collection topic models. In EMNLP 09.
[22] Keith Stevens, Philip Kegelmeyer, David Andrzejewski, and David
Buttler. Exploring topic coherence over many models and many topics.
In EMNLP-CoNLL 12, pages 952961, 2012.
[23] Ivan Titov and Ryan McDonald. Modeling online reviews with multigrain topic models. In WWW 08, pages 111120, 2008.
[24] Yao Wu and Martin Ester. Flame: A probabilistic model combining
aspect based opinion mining and collaborative ltering. In WSDM 15.
[25] Yuanbin Wu, Qi Zhang, Xuanjing Huang, and Lide Wu. Phrase
dependency parsing for opinion mining. In EMNLP 09.
[26] ChengXiang Zhai, Atulya Velivelli, and Bei Yu. A cross-collection
mixture model for comparative text mining. In KDD 04.
[27] Wayne Xin Zhao, Jing Jiang, Hongfei Yan, and Xiaoming Li. Jointly
modeling aspects and opinions with a maxent-lda hybrid. In EMNLP10.
VI. C ONCLUSIONS
In this paper, we propose CAMEL, a novel topic model for
complementary aspect-based opinion mining across asymmetric collections. By modeling both common and specic aspects
and keeping contrastive opinions, CAMEL is capable of integrating complementary information from different collections
in both aspect and opinion levels. An automatically labeling
scheme is also introduced to further boost the applicability
of CAMEL. Extensive experiments and real-world case study
on public events demonstrate the effectiveness of CAMEL
in leveraging complementarity for high-quality aspect and
opinion mining.
ACKNOWLEDGMENT
Deqing Wang was supported by National Natural Science
Foundation of China (NSFC) (71501003) and China Postdoctoral Science Foundation funded project (2014M550591). Junjie Wu was supported by National High Technology Research
and Development Program of China (SS2014AA012303),
NSFC (71322104,71171007,71531001,71471009), National
Center for International Joint Research on E-Business Information Processing (2013B01035), Foundation for the Author
of National Excellent Doctoral Dissertation of China (201189),
and Fundamental Research Funds for Central Universities.
R EFERENCES
[1] Yang Bao, Nigel Collier, and Anindya Datta. A partially supervised
cross-collection topic model for cross-domain text classication. In
CIKM 13, pages 239248. ACM, 2013.
[2] David M. Blei, Andrew Y. Ng, and Michael I. Jordan. Latent dirichlet
allocation. J. Mach. Learn. Res., 3:9931022, 2003.
[3] Samuel Brody and Noemie Elhadad. An unsupervised aspect-sentiment
model for online reviews. In HLT 10, pages 804812, 2010.
678