Understanding Wines
Understanding Wines
Article
Understanding 21st Century Bordeaux Wines from
Wine Reviews Using Naïve Bayes Classifier
Zeqing Dong, Xiaowan Guo, Syamala Rajana and Bernard Chen *
Department of Computer Science, University of Central Arkansas, Conway, AR 72034, USA;
[email protected] (Z.D.); [email protected] (X.G.); [email protected] (S.R.)
* Correspondence: [email protected]
Received: 31 October 2019; Accepted: 23 December 2019; Published: 14 January 2020
Abstract: Wine has been popular with the public for centuries; in the market, there are a variety of
wines to choose from. Among all, Bordeaux, France, is considered as the most famous wine region in
the world. In this paper, we try to understand Bordeaux wines made in the 21st century through
Wineinformatics study. We developed and studied two datasets: the first dataset is all the Bordeaux
wine from 2000 to 2016; and the second one is all wines listed in a famous collection of Bordeaux
wines, 1855 Bordeaux Wine Official Classification, from 2000 to 2016. A total of 14,349 wine reviews
are collected in the first dataset, and 1359 wine reviews in the second dataset. In order to understand
the relation between wine quality and characteristics, Naïve Bayes classifier is applied to predict the
qualities (90+/89−) of wines. Support Vector Machine (SVM) classifier is also applied as a comparison.
In the first dataset, SVM classifier achieves the best accuracy of 86.97%; in the second dataset, Naïve
Bayes classifier achieves the best accuracy of 84.62%. Precision, recall, and f-score are also used as our
measures to describe the performance of our models. Meaningful features associate with high quality
21 century Bordeaux wines are able to be presented through this research paper.
1. Introduction
The ancient beverage, wine, has remained popular in modern times. While the ancients had
mostly wine available from neighboring vineyards, the number and variety of wines available for
purchase have exploded in modern times. Consumers are assaulted with an endless number of
varieties and flavors. Some examples include red wine, white wine, rose wine, starch-based wine,
etc., which are then also based on a variety of grapes, fruits like apples, and berries. For a non-expert,
unfamiliar with the various nuances that make each brand distinct, the complexity of decision making
has vastly increased. In such a competitive market, wine reviews and rankings matter a lot since
they become part of the heuristics that drive consumers decision making. Producers of wine gain
a competitive advantage by knowing what factors contribute the most to quality as determined by
rankings. What has also changed is the amount of data available. Moore’s law and other advances in
computing have allowed for the collection and analysis of vast amounts of data. Data mining is the
utilization of various statistics, algorithms, and other tools of analysis to uncover useful insights into
all this data. The goal of data mining is to gain predictive or descriptive information in the domain
of interest. To help producers better understand the determinants of wine quality we decided to
harness the power of these data mining techniques on two datasets on wine produced in the Bordeaux
region. This region is the biggest wine delivering district in France and one of the most influential
wine districts in the world.
There is a lot of research that focuses on the price and vintage of Bordeaux wines [1–3] from a
historical and economic data. Shanmuganathan et al. applied decision tree and statistical methods
for modeling seasonal climate effects on grapevine yield and wine quality [4]. Noy et al. developed
the ontology on Bordeaux wine [5,6]. Most of these Bordeaux or wine related data mining researches
applied their works on small to medium sized wine datasets [7–10]. However, to the best of our
knowledge, there is almost no research utilizing data mining to determine the quality and character of
various vintages of wines in Bordeaux comparable to the size of our dataset.
The performance of data mining researches relies on the quality of the data. In this research
work, we focus on the wine reviews in human language format with a score as a verdict to wine.
Many researches point out the inconsistency between wine judges as well as the bias in taste [11,12].
The small group of wine experts many not agree with each other while they taste research designated
wines. Every wine expert might have their own tasting palate, wine preference, choice of word,
etc. [13–18]. However, this type of research falls into small or medium at best size of dataset, which is
not suitable for true data mining research works. Therefore, this paper focus on a single reputable wine
review organization: Wine Spectator to gather thousands of wine reviews as the research input dataset.
Although there are some challenges on Wine Spectator’s rating, ranking, and comments [19–21],
based on our previous researches [22–25], the correlation between wine reviews and their grading
are strong. To predict a wine receives a 90+ or 90− score based on Wine Spectators’ wine reviews,
the data mining models built on a dataset with more than 100,000 wine reviews achieved 87.21%
and 84.71% accuracy via Support Vector Machine (SVM) and Naïve Bayes model, accordingly. The
regression model built on the same dataset to predict a wine’s actual score based on Wine Spectator’s
wine reviews was only 1.6 points away on Mean Absolute Error (MAE) evaluation [25]. These findings
support that the large amount of Wine Spectators’ reviews are suitable for our data mining research.
To study 21 century Bordeaux wines based on our previous works, we developed two new
datasets related to Bordeaux. For the first dataset, we collected all the available Bordeaux wine reviews
on the latest vintage (2000–2016) from the Wine Spectator [26]. These reviews are then converted from
human language format into computer encoded codes via the computational wine wheel proposed
in the research of Wineinformatics [22–25]. For the second dataset, we are interested in a famous
collection of Bordeaux wines, 1855 Bordeaux Wine Official Classification [27]. The quality of the wine
in both data sets was determined by experts in a blind taste test. This was based upon an interval scale
from 50–100 in which 100 was determined to be the highest quality while 50 being the wine that was
not recommended due to quality issues. We will train algorithms on both datasets and see which
one is most effective at classifying it in 90+ category or 89− category via Naïve Bayes and SVM. If the
algorithms are effective, we can potentially uncover the words most predictive of wine quality and
enlighten producers on how to maintain and/or improve the quality of their wine allowing them to
succeed in such a competitive environment.
Beverages 2020, 6, 5 3 of 16
75–79 Mediocre: a drinkable wine that may have minor flaws
50–74 Not recommended
50–74 Not recommended
Following is a wine review example of the famous Château Latour Pauillac 2009, which cost
$1600 in 2012:
Following is a wine review example of the famous Château Latour Pauillac 2009, which cost $1600
in 2012:
Château Latour Pauillac 2009 99pts $1600
Château Latour Pauillac 2009 99pts $1600
This seems to come full circle, with a blazing iron note and mouthwatering acidity up front leading to intense,
This seems
vibrant to come
cassis, full circle,
blackberry and with
cherry a blazing iron note
skin flavors and mouthwatering
that course along, followed acidity
by theup front
same leading to
vivacious intense,
minerality
vibrant cassis, blackberry and cherry skin flavors that course along, followed by the
that started things off. The tobacco, ganache and espresso notes seem almost superfluous right now, but same vivacious minerality
that started
they'll join things
the frayoff.inThe
duetobacco,
time. The ganache andis,espresso
question can younotes
waitseem
longalmost
enough?superfluous
Best fromright
2020now, but they’ll
through 2040.
join
9580the fraymade—JM.
cases in due time. The question is, can you wait long enough? Best from 2020 through 2040. 9580 cases
made—JM.
Country: France • Region: Bordeaux • Issue Date: 31 March 2012
Country: France • Region: Bordeaux • Issue Date: 31 March 2012
2.2. The Computational Wine Wheel
2.2. The Computational Wine Wheel
Since the wine reviews are stored in human language format, we have to convert reviews into
Since the wine reviews
machine understandable viaarethestored in human language
computational wine wheel format,
[23]. we
Thehave to convert reviews
computational into
wine wheel
machine understandable via the computational wine wheel [23]. The computational
works as a dictionary to one-hot encoding to convert words into vectors. For example, in the wine wine wheel works
as a dictionary
review, there are to one-hot encoding
some words that to convert
contain words
fruits suchintoasvectors. For example,
apple, blueberry, in the
plum, etc.wine review,
If the word
there are some words that contain fruits such as apple, blueberry, plum,
matches the attribute in the computation wine wheel, it will be 1, otherwise, it will be 0. More etc. If the word matches the
attribute
examplesincan the computation
be found inwine Figurewheel,
1. it will be
Many 1, otherwise,
other it will be 0. More
wine characteristics are examples
included canin thebe
found in Figure 1. Many other wine characteristics are included
computational wine wheel other than fruit flavors, such as descriptive adjectives (balance,in the computational wine wheel
other than fruit
beautifully, etc.)flavors,
and body suchofasthe
descriptive adjectives
wine (acidity, level (balance,
of tannin,beautifully, etc.) and body wine
etc.). The computational of thewheel
wine
(acidity, level of tannin,
is also equipped etc.). The computational
with generalization wine wheel
function to generalize is also
similar equipped
words with
into the generalization
same coding. For
function to generalize similar words into the same coding. For example,
example, fresh apple, apple, and ripe apple are generalized into “Apple” since they represent fresh apple, apple, and ripe
the
apple are generalized into “Apple” since they represent the same flavor;
same flavor; yet, green apple belongs to “Green Apple” since the flavor of green apple is different yet, green apple belongs to
“Green Apple” since the flavor of green apple is different from apple.
from apple.
Figure 1. The
Figure Theflowchart
flowchartofof
coveting reviews
coveting intointo
reviews machine understandable
machine via the
understandable via computation wine
the computation
wheel.
wine wheel.
In
In this
thisresearch,
research,ininorder
ordertotounderstand
understand the
thecharacteristics of of
characteristics classic (95+)
classic and
(95+) outstanding
and outstanding(90–94)
(90–
wine, we use
94) wine, we 90usepoints as a as
90 points cutting point.
a cutting If a wine
point. receives
If a wine a score
receives equal/above
a score 90 points
equal/above out of
90 points out100,
of
we mark the label as a positive (+) class to the wine. Otherwise, the label would be a negative
100, we mark the label as a positive (+) class to the wine. Otherwise, the label would be a negative (−) class.
There areThere
(−) class. some are
wines that
some scored
wines a ranged
that scored ascore,
rangedsuch as 85–88.
score, such asWe use the
85–88. Weaverage of the ranged
use the average of the
score
ranged to score
decidetoand assign
decide andthe label.the label.
assign
2.3.
2.3. Datasets
Datasets
We
Wedeveloped
developedtwo
twodatasets
datasetsin in
thisthis
research. TheThe
research. firstfirst
one one
is theisreviews for allfor
the reviews theall
Bordeaux wines
the Bordeaux
made in the 21st century (2000–2016). The second one is the reviews for all available
wines made in the 21st century (2000–2016). The second one is the reviews for all available wine wine listed in
Bordeaux Wine Official Classification, made in the 21st century (2000–2016) as well. The second dataset
Beverages2020,
Beverages 2020,6,6,5x FOR PEER REVIEW 44of
of16
16
listed in Bordeaux Wine Official Classification, made in the 21st century (2000–2016) as well. The
issecond
a subset of theisfirst
dataset dataset.
a subset Allfirst
of the the available wine
dataset. All thereviews were
available collected
wine reviewsfrom
wereWine Spectator.
collected from
Details of each dataset will be discussed as follows.
Wine Spectator. Details of each dataset will be discussed as follows.
2.3.1.
2.3.1. ALL
ALL Bordeaux
Bordeaux Wine
WineDataset
Dataset
A
A total
total of
of 14,349 wines has
14,349 wines has been
beencollected.
collected.There
There areare 4263
4263 90+90+ wines
wines andand 10,086
10,086 89− 89− wines.
wines. The
The
number of 89− wines is more than 90+ wines. The score distribution is given in Figure 2a. Most
number of 89− wines is more than 90+ wines. The score distribution is given in Figure 2a. Most
wines
winesscore
scorebetween
between 86 86
andand
90. Therefore, they fall
90. Therefore, theyinto
fallthe category
into of “Veryof
the category Good” wine.
“Very Good”In Figure
wine.2b,
In
the line chart is used to represent the trend of number of wines has been
Figure 2b, the line chart is used to represent the trend of number of wines has been reviewedreviewed in each year.
in
The
eachchart
year.also
Thereflects
chart the
alsoquality
reflectsofthe
vintages.
qualityMore than 1200
of vintages. wines
More were
than reviewed
1200 in 2009
wines were and 2010,
reviewed in
which indicates
2009 and that 2009
2010, which and 2010
indicates thatare good
2009 vintages
and 2010 areingood
Bordeaux. Wine
vintages makers areWine
in Bordeaux. moremakers
willingare
to
send their wines to be reviewed if their wines are good.
more willing to send their wines to be reviewed if their wines are good.
2500 1400
1200
Number of reviews
2000
Number of reviews
1000
1500 800
1000 600
400
500
200
0 0
<80 81 83 85 87 89 91 93 95 97 99 2000 2005 2010 2015
Score Year
(a) (b)
Figure2.2. (a)
Figure (a)The
Thescore
scoredistribution
distributionof
ofALL
ALLBordeaux
BordeauxWines;
Wines;(b)
(b)The
Thenumber
numberof
ofwines
winesthat
thathave
havebeen
been
reviewed annually.
reviewed annually.
2.3.2. 1855 Bordeaux Wine Official Classification Dataset
2.3.2. 1855 Bordeaux Wine Official Classification Dataset
A total of 1359 wines has been collected. In this dataset, we have 882 90+ wines and 477 89− wines.
A total of 1359 wines has been collected. In this dataset, we have 882 90+ wines and 477 89−
The score distribution is given in Figure 2. Unlike the data distribution of the first dataset, which
wines. The score distribution is given in Figure 2. Unlike the data distribution of the first dataset,
has much more 89− wines than 90+ wines, in Wine Spectator, the wines selected in this research are
which has much more 89− wines than 90+ wines, in Wine Spectator, the wines selected in this
elite choices based on Bordeaux Wine Official Classification in 1855 (a complete list of Bordeaux Wine
research are elite choices based on Bordeaux Wine Official Classification in 1855 (a complete list of
Official Classification in 1855 is given in Appendix A). Therefore, classic (95+ points) and outstanding
Bordeaux Wine Official Classification in 1855 is given in Appendix A). Therefore, classic (95+
(90–94 points) wines are the majority of this dataset. The number of wines has been reviewed annually
points) and outstanding (90–94 points) wines are the majority of this dataset. The number of wines
is given in Figure 3b. Since Bordeaux Wine Official Classification in 1855 is a famous collection of
has been reviewed annually is given in Figure 3b. Since Bordeaux Wine Official Classification in
Bordeaux wines, wine makers send their wine for review almost every year. Therefore, the line chart
1855 is a famous collection of Bordeaux wines, wine makers send their wine for review almost
remains stable, which is very different from Figure 2b. Regardless, some wines listed in Bordeaux Wine
every year. Therefore, the line chart remains stable, which is very different from Figure 2b.
Official Classification in 1855 may still missing their wine reviews in Wine Spectator. A complete list of
Regardless, some wines listed in Bordeaux Wine Official Classification in 1855 may still missing
wines and vintages we cannot find within this dataset’s scope is listed in Appendix B.
their wine reviews in Wine Spectator. A complete list of wines and vintages we cannot find within
this dataset’s scope is listed in Appendix B.
Beverages 2020, 6, 5 5 of 16
Beverages 2020, 6, x FOR PEER REVIEW 5 of 16
250 90
80
Number of Reviews
Number of reviews
200 70
60
150 50
40
100
30
50 20
10
0 0
<80 81 83 85 87 89 91 93 95 97 99 2000 2005 2010 2015
Score Year
(a) (b)
Figure3.3.(a)
Figure (a)The
Thescore
scoredistribution
distributionofofBordeaux
Bordeaux Wine
Wine Official
Official Classification
Classification in 1855;
in 1855; (b) (b)
TheThe number
number of
of wines reviewed annually.
wines reviewed annually.
3.3.Methods
Methods
3.1.1.Naïve
NaïveBayes
Bayesis a commonly used machine learning classification algorithm. A Naïve Bayes
classifier is a simple probabilistic classifier by applying Bayes’ theorem with ignoring the dependency
Naïve Bayes is a commonly used machine learning classification algorithm. A Naïve Bayes
between features.
classifier is a simple probabilistic classifier by applying Bayes’ theorem with ignoring the
Formula of Naïve Bayes classifier algorithm [30]:
dependency between features.
Bayes’ Theorem:
Formula of Naïve Bayes classifier algorithm [30]:
P(X Y )P(Y )
Bayes’ Theorem: P(Y|X) = (1)
P(X )
( | ) ( )
P(Y|X): The posterior probability of Y 𝑃(𝑌|𝑋)
belongs =to a particular
( ) class when X happens; (1)
P(X|Y): The prior probability of certain feature value X when Y belongs to a certain class;
P(Y|X): The posterior probability of Y belongs to a particular class when X happens;
P(Y): prior
P(X|Y): Theprobability of Y; of certain feature value X when Y belongs to a certain class;
prior probability
P(X):
P(Y): prior
prior probability of X.
probability of Y;
P(X): prior probability of X.
Naïve Bayes Classifier:
Naïve Bayes Classifier:
P(X1, X2 , . . . Xn Y)P(Y) P(X1 |Y)P(X2 |Y) . . . P(Xn |Y)P(Y)
P(Y X1, X2 , . . . Xn ) = 𝑋 , 𝑋 , … 𝑋 𝑌 (=) 𝑋 𝑌 𝑋 𝑌… 𝑋 𝑌 ( ) (2)
𝑃 𝑌 𝑋 , 𝑋 , … 𝑋 P= (X1, X2 , . . . Xn ) = P(X1, X2 , . . . Xn ) (2)
, ,… , ,…
P𝑃(Y𝑌X𝑋1, X
, 𝑋2 ,, .…
. .𝑋Xn )::Compute
Computeall
allposterior
posteriorprobability
probabilityof
ofall
allvalues
valuesin
inXXfor
forall
allvalues
valuesin
inY.
Y.
Naïve Bayes classifier makes the prediction based on the maximum of posterior
Naïve Bayes classifier makes the prediction based on the maximum of posterior probability. probability.
Laplace
LaplaceSmoothing:
Smoothing:
N +1
P(Xi |Y) = ic (3)
𝑃(𝑋 |𝑌) N=C + c (3)
c: number of values in Y
Beverages
Beverages 2020,2020,
6, 5 6, x FOR PEER REVIEW 6 of 16 6 of 16
When a value of X never appears in the training set, the prior probability of that value of X will
be 0.
c: numberIf we do not useinany
of values Y techniques, 𝑃 𝑌 𝑋 , 𝑋 , … 𝑋 will be 0, even when some of other prior
probability
Beverages 2020,of6, X are
x FOR
When a value of X never very
PEER high.
REVIEW This case
appears does
in the not seem
training set,fair
thetoprior
otherprobability
X. Therefore, of we
thatuse Laplace
value 6 of 16
of X will
smoothing to handle zero prior probability.
be 0. If we Whendo not useofany
a value techniques,
X never appears in P(the 1, X2 , . . .set,
Y Xtraining Xn )thewill
priorbeprobability
0, even when of thatsome
valueofofother
X willprior
probability
be 0. SMV
3.1.2. of X are very high. This case does not seem fair to other
If we do not use any techniques, 𝑃 𝑌 𝑋 , 𝑋 , … 𝑋 will be 0, even when some of other prior X. Therefore, we use Laplace
smoothing to handle zero prior probability.
probability of X are very high. This case does not seem fair to other X. Therefore, we use Laplace
“SVM are supervised learning models with associated learning algorithms that analyze data
smoothing to handle zero prior probability.
3.1.2.and
SMV recognize patterns, used for classification and regression analysis” [31]. SVM for classification
will
3.1.2.based
SMV on the training data, building a model by constructing “a hyperplane or set of
“SVM are supervised
hyperplanes in a high-learning models with associated
or infinite-dimensional space, which learning canalgorithms
be used for thatclassification,
analyze data and
recognize “SVM
regression, are
patterns, supervised
or other used learning
for Intuitively,
tasks. classificationmodels with associated
andseparation
a good regression learning
analysis”by[31].
is achieved algorithms that analyze
that has data
SVM for classification
the hyperplane the will
based and
largest recognize
on the distance
trainingpatterns,
to the
data, used
nearest
buildingfor classification
training data point
a model and regression
of any class “a
by constructing analysis”
(so-called
hyperplane [31]. SVM
functional
or set for
margin),classification
since in in a
of hyperplanes
will
high-general based on the
the larger
or infinite-dimensional thetraining
margin data,
the
space, building
lower
which thecan a be
model
usedby
generalization for constructing
error “a regression,
hyperplane
of the classifier”.
classification, After havingor other
or settheoftasks.
hyperplanes
model, aa good in
test data a high-
is used to or infinite-dimensional
predict thebyaccuracy. space,
SVM light which[32] can
is the be used
the largest for
version distance classification,
of SVM that was
Intuitively, separation is achieved the hyperplane that has to the nearest
regression,
used or other
to perform tasks. Intuitively,
the classification a good for
of attributes separation
this project.is achieved by the hyperplane that has the
training data point of any class (so-called functional margin), since in
largest distance to the nearest training data point of any class (so-called functional margin), since in
general the larger the margin
the lower
general
3.2. thethegeneralization
Evaluations larger the margin errortheof the the
lower classifier”.
generalization Aftererrorhaving theclassifier”.
of the model, aAfter test data
having is the
used to
predict the accuracy.
model, a test dataSVM is usedlightto [32] is the
predict theversion
accuracy. of SVM
SVMlight that [32]
wasisused to perform
the version of SVM the that
classification
was
5-fold cross-validation, illustrated in Figures 4 and 5, is used to evaluate the predictive
of attributes for this the
used to perform project.
classification of attributes for this project.
performance of our models, especially the performance of the model for new data, which can
reduce overfitting to some extent. First, we shuffle the dataset randomly. Second, we group 90+/89−
3.2. Evaluations
3.2. Evaluations
wines. Third, split 90+ wine group and 89− wine group into 5 subsets separately. Fourth, combine
5-fold
first 5-fold
subset cross-validation,
cross-validation,
from 90+ wine illustrated
illustrated
group and in subset
infirst
FiguresFigures
4 and
from 45,and
is used
89− 5, istogroup
wine used
evaluateto evaluate
intothe the repeat
predictive
a new set, predictive
performance
the
of our performance
samemodels,
process of the
for our rest.
especially models, especially
this way, we the
theInperformance of the
split performance
ourmodel
dataset of
forinto
new the model
data,
5 subsets for the
which
with new
cansamedata,distribution
reduce which can to
overfitting
asreduce
the overfitting
original to
dataset. some extent. First, we shuffle the dataset randomly.
some extent. First, we shuffle the dataset randomly. Second, we group 90+/89− wines. Third, split 90+ Second, we group 90+/89−
wines. Third, split 90+ wine group and 89− wine group into 5 subsets separately. Fourth, combine
wine group and 89− wine group into 5 subsets separately. Fourth, combine first subset from 90+ wine
first subset from 90+ wine group and first subset from 89− wine group into a new set, repeat the
group and first subset from 89− wine group into a new set, repeat the same process for the rest. In this
same process for the rest. In this way, we split our dataset into 5 subsets with the same distribution
way, aswethesplit our dataset
original dataset.into 5 subsets with the same distribution as the original dataset.
After data splitting, we use the subset 1 as testing set, the rest of the subsets as training set as
fold 1; we use subset 2 as testing set, the rest of the subsets as training set as fold 2; Repeat the same
process for the rest.
Figure
Figure 4. This
4. This figuredemonstrates
figure demonstrates data
data splitting
splittinginin5-fold cross-validation.
5-fold cross-validation.
After data splitting, we use the subset 1 as testing set, the rest of the subsets as training set as
fold 1; we use subset 2 as testing set, the rest of the subsets as training set as fold 2; Repeat the same
process for the rest.
Figure
Figure 5. This
5. This figure
figure demonstratestraining
demonstrates training and
andtesting
testingsets assigning
sets in 5-fold
assigning cross-validation.
in 5-fold cross-validation.
After data splitting, we use the subset 1 as testing set, the rest of the subsets as training set as fold
1; we use subset 2 as testing set, the rest of the subsets as training set as fold 2; Repeat the same process
for the rest. Figure 5. This figure demonstrates training and testing sets assigning in 5-fold cross-validation.
Beverages 2020, 6, 5 7 of 16
To evaluate the effectiveness of the classification model, several standard statistical evaluation
metrics are used in this paper. First of all, we need to define True Positive (TP), True Negative (TN),
False Positive (FP), and False Negative (FN) as:
TP: The real condition is true (1) and predicted as true (1); 90+ wine correctly classified as 90+ wine;
TN: The real condition is false (−1) and predicted as false (−1); 89− wine correctly classified as
89− wine;
FP: The real condition is false (−1) but predicted as true (1); 89− wine incorrectly classified as 90+ wine;
FN: The real condition is true (1) but predicted as false (−1); 90+ wine incorrectly classified as 89− wine.
If we use 90 points as a cutting point, to describe TP is this research’s perspective would be “if a
wine scores equal/above 90 and the classification model also predicts it as equal/above 90”. In this
research, we include the following evaluation metrics:
Accuracy: The proportion of wines that has been correctly classified among all wines. Accuracy is
a very intuitive metric.
TP + TN
Accuracy = (4)
TP + TN + FP + FN
Recall: The proportion of 90+ wines was identified correctly. Recall explains the sensitivity of the
model to 90+ wine.
TP
Recall = (5)
TP + FN
Precision: The proportion of predicted 90+ wines was actually correct.
TP
Precision = (6)
TP + FP
F-score: The harmonic mean of recall and precision. F-score takes both recall and precision into
account, combining them into a single metric.
precision ∗ recall
F − score = 2 × (7)
(precision + recall)
4. Results
highest precision of 86.84%. In terms of recall, Naive Bayes Laplace classifier achieves the recall as high
as 90.02%. In the combination of precision and recall, Naive Bayes Laplace classifier has the highest
F-score of 88.38%. Overall, Naïve Bayes Laplace has better performance than SMV classifier in this
specific Bordeaux
Beverages 2020, 6, wine dataset.
x FOR PEER Details can be found in Table 2.
REVIEW 8 of 16
(a)
Figure 6. Cont.
Beverages 2020, 6, 5 9 of 16
Beverages 2020, 6, x FOR PEER REVIEW 9 of 16
(b)
Figure 6.
Figure 6. (a)
(a) The
The visualization
visualization ofof whole
wholedataset
datasetbybyNaïve
NaïveBayes.
Bayes.TheThe
horizontal
horizontalaxisaxis
indicates the the
indicates
probability that the sample is 90+, and the vertical axis indicates the probability that the sample is
probability that the sample is 90+, and the vertical axis indicates the probability that the sample is type type
90−; (b) Magnification of Figure 6a.
90−; (b) Magnification of Figure 6a.
4.5.20
4.5. Top Top 20 Keywords
Keywords
SVMSVM is considered
is considered as aasblack-box
a black-boxclassifier,
classifier,since
since the
the classification
classification processes are are
processes unexplainable.
unexplainable.
Naïve Bayes, on the other hand, is a white-box classification algorithm, since each attribute has its
Naïve Bayes, on the other hand, is a white-box classification algorithm, since each attribute has its own
own probability contribute to positive case and negative case. We extract keywords with 20 highest
probability contribute to positive case and negative case. We extract keywords with 20 highest positive
positive probabilities toward 90+ and 89− classes from both datasets.
probabilities toward
In ALL 90+ and
Bordeaux 89−dataset,
Wine classesthere
fromareboth datasets.keywords that appear in both 90+ and
11 common
In
89−ALL Bordeaux
wines. Wine
Details can dataset,
be found there3.are
in Table 11 common
These keywords
common keywords that appear
represent in both wine
the important 90+ and
89− wines. Details can be found
characteristics/attributes towardin 21st
Table 3. These
century common
general keywords
Bordeaux wines. represent theour
Furthermore, important
goal is towine
characteristics/attributes
understand the importanttoward wine
21st century general Bordeaux
characteristics/attributes wines.21st
toward Furthermore, our goal
century classic and is to
outstanding Bordeaux wines. Therefore, finding out the distinct keywords between
understand the important wine characteristics/attributes toward 21st century classic and outstanding 90+ and 89− is
our final goal. Details about the distinct keywords between 90+ and 89− from ALL
Bordeaux wines. Therefore, finding out the distinct keywords between 90+ and 89− is our final goal. Bordeaux Wine
Detailsdataset
aboutcan
thebedistinct
found in Tables 4 and
keywords 5. According
between 90+ and to 89−
Tablefrom
4, fruity
ALLcharacters
Bordeauxincluding “BLACK
Wine dataset can be
CURRANT”, “APPLE”, “RASPERBERRY”, and “FIG” are favorable flavors to 21st century
found in Tables 4 and 5. According to Table 4, fruity characters including “BLACK CURRANT”,
Bordeaux. Since Bordeaux is also famous for red wines that can age for many years, “SOLID”
“APPLE”, “RASPERBERRY”, and “FIG” are favorable flavors to 21st century Bordeaux. Since Bordeaux
(showed in Table 4 Body category) is preferred over “MEDIUM-BODIED” and “LIGHT-BODIED”
is also(showed
famousinfor red 5wines
Table Body that can age for many years, “SOLID” (showed in Table 4 Body category)
category).
is preferred over “MEDIUM-BODIED” and “LIGHT-BODIED” (showed in Table 5 Body category).
Table 3. Common keywords between 90+ and 89− wines from ALL Bordeaux Wine dataset.
Table 3. Common keywords between 90+ and 89− wines from ALL Bordeaux Wine dataset.
CATEGORY 90+ WINES AND 89− WINES
FLAVOR/DESCRIPTORS
CATEGORY GREAT 90+ WINES
FLAVORS AND 89− WINES
FRUITY
FLAVOR/DESCRIPTORS GREAT FRUIT PLUM
FLAVORS BLACKBERRY CURRENT
FRUITY BODY FULL-BODIED
FRUIT CORE
PLUM BLACKBERRY CURRENT
BODY FINISH FINISH
FULL-BODIED CORE
FINISH HERBS TOBACCO
FINISH
HERBSTANNINS TANNINS_LOW
TOBACCO
TANNINS TANNINS_LOW
Table 4. Distinct keywords between 90+ and 89− wines from ALL Bordeaux Wine dataset in 90+ wines.
Table 5. Distinct keywords between 90+ and 89− wines from ALL Bordeaux Wine dataset in 89− wines.
In 1855 Bordeaux Wine Official Classification dataset, there are 11 common keywords that appear
in both 90+ and 90− wines. Details can be found in Table 6. Comparing the common keywords with
ALL Boredeaux Wine dataset, 10 out of 11 are the same keywords. “TANNINES_LOW” only appears in
ALL Bordeaux Wine dataset, and “SWEET” only appears in 1855 Bordeaux Wine Official Classification
dataset. Details about the distinct keywords between 90+ and 89− from 1855 Bordeaux Wine Offical
Classification dataset can be found in Tables 7 and 8.
Table 6. Common keywords between 90+ and 89− wines from 1855 Bordeaux Wine Official
classification dataset.
Table 7. Distinct keywords between 90+ and 89− wines from 1855 Bordeaux Wine Official classification
dataset in 90+ wines.
Table 8. Distinct keywords between 90+ and 89− wines from 1855 Bordeaux Wine Official classification
dataset in 89− wines.
Comparing the dictinct keywords between 90+ and 89− wines from both datasets in 90+
wines, “LONG”, “BLACK CURRANT”, “APPLE”, and “FIG” appear in both datasets; “RANGE”,
“RIPE”, “RASPERBERRY”, “SOLID”, and “LICORICE” only appear in ALL Bordeaux Wine dataset;
“STYLE”, “LOVELY”, “IRON”, “TANNINS_LOW”, and “SPICE” only appear in 1855 Bordeaux Wine
Offical Classification.
5. Conclusions
In this research, we developed and studied two datasets: the first dataset is all the Bordeaux wine
from 2000 to 2016; and the second one is all wines listed in a famous collection of Bordeaux wines, 1855
Bordeaux Wine Official Classification, from 2000 to 2016. We used Naïve Bayes classifier and SMV
Beverages 2020, 6, 5 11 of 16
classifier to make wine quality prediction based on wine reviews. Overall, Naïve Bayes classifier works
better than SMV in the 1855 Bordeaux Wine Official Classification dataset, slightly worse than SMV in
the ALL Bordeaux Wine. Also, with the benefit of using Naïve Bayes classifier, we were able to find
the important wine characteristics/attributes toward 21st century classic and outstanding Bordeaux
wines. The list of common attributes in Tables 3 and 6 identifies the general wine characteristics in
Bordeaux; while the list of dominate attributes in Tables 4 and 7 (Tables 5 and 8) shows the preferable
characteristics for 90+ (90–) wines. Those characteristics/attributes can help producers improve the
quality of their wines allowing them to concentrate of dilute the wanted or unwanted characteristics
during the wine making process. To the best of our knowledge, this is the first paper that gives a
detailed analysis in all prestigious Bordeaux wines in the 21st century.
To go further in this research as future works, two follow up questions can be raised; 1. Instead
of dichotomous (90+ and 90−) analysis, can the research use finer label (classic, outstanding, very
good, and good) to categorize these Bordeaux wine to perform the analysis or even regression
analysis [32]? 2. What characteristics/attributes make the Bordeaux wines become a classic (95+)
instead of outstanding (90–94)? The first question can be studied as a multi-class problem in data
science since the computational model will be built into four different classes and produce important
characteristics for each class. The second question is a typical highly unbalanced problem in data
science. The number of wines scores 95+ is much less than 95− wines. The regular computational
model such as SVM and Naïve Bayes will not be able to identify the boundary between the two
classes and predict all testing wines into majority class. How to amplify the minority class and obtain
meaningful information is a big challenge in this type of question. Finally, we would like to address
the limitation of our current research. Since the computational wine wheel was developed from Wine
Spectators’ Top 100 lists, the proposed research might have optimal results in the dataset collected
from Wine Spectators’ review. While several other wine experts in the filed such as Robert Parker Wine
Advocate [33], Wine Enthusiast [34], and Decanter [35] may not agree with each other’s comments,
they can still agree in the overall score of the wine. The legendary Chateau Latour 2009 gives a great
example [36]; every reviewer scores the wine either 100 or 99 and their testing notes are very different
with each other. This would be our ultimate challenge in Wineinformatics research that involves the
true human language processing topic.
Author Contributions: B.C. and Z.D. literature review, conceived, designed the experiment and original writing.
X.G. and S.R. revised and improved the experiment. B.C. Critical reading and financial support. All authors have
read and agree to the published version of the manuscript.
Funding: We would like to thank the Department of Computer Science at UCA for the support of the new research
application domain development.
Conflicts of Interest: The authors declare no conflicts of interest.
DEUXIÈMES CRUS
TROISIÈMES CRUS
QUATRIÈMES CRUS
CINQUIÈMES CRUS
PREMIERS CRUS
DEUXIÈMES CRUS
References
1. Combris, P.; Lecocq, S.; Visser, M. Estimation of a hedonic price equation for Bordeaux wine: Does quality
matter? Econ. J. 1997, 107, 389–402. [CrossRef]
2. Cardebat, J.-M.; Figuet, J. What explains Bordeaux wine prices? Appl. Econ. Lett. 2004, 11, 293–296. [CrossRef]
3. Ashenfelter, O. Predicting the quality and prices of Bordeaux wine. Econ. J. 2008, 118, F174–F184. [CrossRef]
4. Shanmuganathan, S.; Sallis, P.; Narayanan, A. Data mining techniques for modelling seasonal climate
effects on grapevine yield and wine quality. In Proceedings of the 2010 2nd International Conference
on Computational Intelligence, Communication Systems and Networks, Liverpool, UK, 28–30 July 2010;
pp. 84–89.
5. Noy, F.N.; Sintek, M.; Decker, S.; Crubézy, M.; Fergerson, R.W.; Musen, M.A. Creating semantic web contents
with protege-2000. IEEE Intell. Syst. 2001, 16, 60–71. [CrossRef]
Beverages 2020, 6, 5 15 of 16
6. Noy, F.N.; McGuinness, D.L. Ontology Development 101: A Guide to Creating Your First Ontology. Stanford
Knowledge Systems Laboratory Technical Report KSL-01-05 and Stanford Medical Informatics Technical
Report SMI-2001-0880. March 2001. Available online: http://www.corais.org/sites/default/files/ontology_
development_101_aguide_to_creating_your_first_ontology.pdf (accessed on 1 January 2020).
7. Quandt, R.E. A note on a test for the sum of ranksums. J. Wine Econ. 2007, 2, 98–102. [CrossRef]
8. Ashton, R.H. Improving experts’ wine quality judgments: Two heads are better than one. J. Wine Econ. 2011,
6, 135–159. [CrossRef]
9. Ashton, R.H. Reliability and consensus of experienced wine judges: Expertise within and between? J. Wine
Econ. 2012, 7, 70–87. [CrossRef]
10. Bodington, J.C. Evaluating wine-tasting results and randomness with a mixture of rank preference models.
J. Wine Econ. 2015, 10, 31–46. [CrossRef]
11. Cardebat, J.M.; Livat, F. Wine experts’ rating: A matter of taste? Int. J. Wine Bus. Res. 2016, 28, 43–58.
[CrossRef]
12. Cardebat, J.M.; Figuet, J.M.; Paroissien, E. Expert opinion and Bordeaux wine prices: An attempt to correct
biases in subjective judgments. J. Wine Econ. 2014, 9, 282–303. [CrossRef]
13. Cao, J.; Stokes, L. Evaluation of wine judge performance through three characteristics: Bias, discrimination,
and variation. J. Wine Econ. 2010, 5, 132–142. [CrossRef]
14. Cardebat, J.M.; Paroissien, E. Standardizing expert wine scores: An application for Bordeaux en primeur.
J. Wine Econ. 2015, 10, 329–348. [CrossRef]
15. Hodgson, R.T. An examination of judge reliability at a major US wine competition. J. Wine Econ. 2008, 3,
105–113. [CrossRef]
16. Hodgson, R.T. An analysis of the concordance among 13 US wine competitions. J. Wine Econ. 2009, 4, 1–9.
[CrossRef]
17. Hodgson, R.; Cao, J. Criteria for accrediting expert wine judges. J. Wine Econ. 2014, 9, 62–74. [CrossRef]
18. Hopfer, H.; Heymann, H. Judging wine quality: Do we need experts, consumers or trained panelists?
Food Qual. Prefer. 2014, 32, 221–233. [CrossRef]
19. Ashenfelter, O.; Goldstein, R.; Riddell, C. Do expert ratings measure quality? The case of restaurant wine
lists. In Proceedings of the 4th Annual AAWE Conference at the University of California at Davis, Davis, CA,
USA, 20 June 2010.
20. Cardebat, J.M.; Corsinovi, P.; Gaeta, D. Do Top 100 wine lists provide consumers with better information?
Econ. Bull. 2018, 38, 983–994.
21. Reuter, J. Does advertising bias product reviews? An analysis of wine ratings. J. Wine Econ. 2009, 4, 125–151.
[CrossRef]
22. Chen, B.; Rhodes, C.; Crawford, A.; Hambuchen, L. Wineinformatics: Applying data mining on wine
sensory reviews processed by the computational wine wheel. In Proceedings of the 2014 IEEE International
Conference on Data Mining Workshop, Shenzhen, China, 14–14 December 2014; pp. 142–149.
23. Chen, B.; Rhodes, C.; Yu, A.; Velchev, V. The Computational Wine Wheel 2.0 and the TriMax Triclustering in
Wineinformatics. In Industrial Conference on Data Mining; Springer: Cham, Switzerland, 2016; pp. 223–238.
24. Chen, B.; Velchev, V.; Palmer, J.; Atkison, T. Wineinformatics: A Quantitative Analysis of Wine Reviewers.
Fermentation 2018, 4, 82. [CrossRef]
25. Palmer, J.; Chen, B. Wineinformatics: Regression on the Grade and Price of Wines through Their Sensory
Attributes. Fermentation 2018, 4, 84. [CrossRef]
26. Wine Spectator. Available online: https://www.winespectator.com (accessed on 1 January 2020).
27. Bordeaux Wine Official Classification of 1855. Available online: https://www.bordeaux.com/us/Our-Terroir/
Classifications/Grand-Cru-Classes-en-1855 (accessed on 1 January 2020).
28. Wine Spectator’s 100-Point Scale | Wine Spectator, Winespectator.com. 2019. Available online: https:
//www.winespectator.com/articles/scoring-scale (accessed on 1 January 2020).
29. Chen, B.; Le, H.; Rhodes, C.; Che, D. Understanding the Wine Judges and Evaluating the Consistency
Through White-Box Classification Algorithms. In Advances in Data Mining. Applications and Theoretical
Aspects. ICDM 2016; Perner, P., Ed.; Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2016;
Volume 9728.
Beverages 2020, 6, 5 16 of 16
30. Rish, I. An empirical study of the naive Bayes classifier. In Proceedings of the IJCAI 2001 Workshop
on Empirical Methods in Artificial Intelligence; 2001; Volume 3, pp. 41–46. Available online: https:
//www.cc.gatech.edu/~{}isbell/reading/papers/Rish.pdf (accessed on 1 January 2020).
31. Suykens, K.J.A.; Vandewalle, J. Least squares support vector machine classifiers. Neural Process. Lett. 1999, 9,
293–300. [CrossRef]
32. Thorsten, J. Svmlight: Support Vector Machine. Available online: https://www.researchgate.
net/profile/Thorsten_Joachims/publication/243763293_SVMLight_Support_Vector_Machine/links/
5b0eb5c2a6fdcc80995ac3d5/SVMLight-Support-Vector-Machine.pdf (accessed on 1 January 2020).
33. Robert Parker Wine Advocate. Available online: https://www.robertparker.com/ (accessed on 1 January 2020).
34. Wine Enthusiast. Available online: https://www.wineenthusiast.com/ (accessed on 1 January 2020).
35. Decanter. Available online: https://www.decanter.com/ (accessed on 1 January 2020).
36. Chateau Latour 2009 Wine Reviews. Available online: https://www.wine.com/product/chateau-latour-2009/
119875 (accessed on 1 January 2020).
© 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access
article distributed under the terms and conditions of the Creative Commons Attribution
(CC BY) license (http://creativecommons.org/licenses/by/4.0/).