Direct Marketing Decision Support Through Predictive Customer Response Modeling
Direct Marketing Decision Support Through Predictive Customer Response Modeling
a r t i c l e i n f o a b s t r a c t
Article history: Decision support techniques and models for marketing decisions are critical to retail success. Among different
Received 8 June 2011 marketing domains, customer segmentation or profiling is recognized as an important area in research and
Received in revised form 12 May 2012 industry practice. Various data mining techniques can be useful for efficient customer segmentation and
Accepted 19 June 2012
targeted marketing. One such technique is the RFM method. Recency, frequency, and monetary methods pro-
Available online 3 July 2012
vide a simple means to categorize retail customers. We identify two sets of data involving catalog sales and
Keywords:
donor contributions. Variants of RFM-based predictive models are constructed and compared to classical data
Customer response predictive model mining techniques of logistic regression, decision trees, and neural networks. The spectrum of tradeoffs is an-
Knowledge-based marketing alyzed. RFM methods are simpler, but less accurate. The effect of balancing cells, of the value function, and
RFM classical data mining algorithms (decision tree, logistic regression, neural networks) are also applied to the
Neural networks data. Both balancing expected cell densities and compressing RFM variables into a value function were
Decision tree models found to provide models similar in accuracy to the basic RFM model, with slight improvement obtained by
Logistic regression increasing the cutoff rate for classification. Classical data mining algorithms were found to yield better pre-
diction, as expected, in terms of both prediction accuracy and cumulative gains. Relative tradeoffs among
these data mining algorithms in the context of customer segmentation are presented. Finally we discuss prac-
tical implications based on the empirical results.
© 2012 Elsevier B.V. All rights reserved.
1. Introduction for marketing decisions for a long time and is recognized as a useful
data mining technique for customer segmentation and response models
The role of decision support techniques and models for marketing [3,30]. A survey [43] also shows that RFM is among the most popular
decisions has been important since the inception of decision support segmentation and predictive modeling techniques used by marketers.
systems (DSSs) [25]. Diverse techniques and models (e.g., optimization, RFM relies on three customer behavioral variables (how long since the
knowledge-based systems, simulation) have emerged over the last five last purchase by customer, how often the customer purchases, how much
decades. Many marketing domains, including pricing, new product de- the customer has bought) to find valuable customers or donors and devel-
velopment, and advertising, have benefited from these techniques and op future direct marketing campaigns. Having a reliable and accurate cus-
models [16]. Among these marketing domains, customer segmentation tomer response model is critical for marketing success since an increase
or profiling is recognized as an important area [18,19,26,43]. There are or decrease in accuracy of 1% could have a significant impact on their
at least two reasons for this. First, the marketing paradigm is becoming profits [1]. While there could be many other customer-related factors
customer-centric [41] and targeted marketing and service are suitable. [e.g.,42], previous studies have shown that RFM alone can offer a pow-
Second, unsolicited marketing is costly and ineffective (e.g., low re- erful way of predicting the future customer purchase [1,3,17].
sponse rate) [15,30]. Along with these reasons, there are increasing ef- Our research builds customer response models using RFM variables
forts on collecting and analyzing customer data for better marketing and compares them in terms of customer gains and prediction accuracy.
decisions [9,26,30]. The advancement of online shopping technologies The paper aims to increase understanding of how to find knowledge
and database systems has accelerated this trend. hidden in customer and transactional databases using data mining tech-
Data mining has been a valuable tool in this regard. Various data niques. This area is called knowledge-based marketing [26]. The next
mining techniques, including statistical analysis and machine learn- section briefly reviews various data mining techniques for building cus-
ing algorithms, can be useful for efficient customer segmentation tomer response or predictive models. Section 3 describes methodology.
and targeted marketing [4,26,38]. One such technique is RFM, stand- All the response models will be built upon the three RFM variables,
ing for recency, frequency, and monetary. RFM analysis has been used while different data mining techniques are used. Then, we present a re-
search design, including two direct marketing data sets with over
⁎ Corresponding author.
100,000 observations, a process of predictive modeling building, and
E-mail addresses: [email protected] (D.L. Olson), [email protected] methods to measure the performance of models. Section 4 includes
(B.(K.) Chae). analysis and results. There could be different methods to increase the
0167-9236/$ – see front matter © 2012 Elsevier B.V. All rights reserved.
doi:10.1016/j.dss.2012.06.005
444 D.L. Olson, B.(K.) Chae / Decision Support Systems 54 (2012) 443–451
prediction performance of an RFM-based predictive model and sophis- the original RFM model have been recognized in the literature [31,45].
ticated data mining techniques (decision tree, logistic regression, and Some previous studies have extended the original RFM model either
neural networks) appear to outperform more traditional RFM. These by considering additional variables (e.g., socio-demographics) [1] or
findings are further discussed in Section 5, comparing results with pre- by combining with other response techniques [6,7]. Because of the
vious studies of customer response models and in the broad contexts of high correlation between F and M, Yang [45] offered a version of RFM
knowledge-based marketing. We also discuss practical implications from model collapsing the data to a single variable “Value”= M / R. To over-
the findings and offer conclusions. come the problem of data skewed in RFM cells, Olson et al. [31] pro-
The contribution of this study is to demonstrate how RFM model posed an approach to balance observations in each of the 125 RFM cells.
variants can work, and supports general conclusions consistently Other variables that may be important include customer income,
reported by others that RFM models are inferior to traditional data customer lifestyle, customer age, product variation, and so on [14].
mining models. This study shows that RFM variables are very useful That would make traditional data mining tools such as logistic regres-
inputs for designing various customer response models with different sion more attractive. However, RFM is the basis for a continuing
strengths and weaknesses and the ones relying on classical data min- stream of techniques to improve customer segmentation marketing
ing (or predictive modeling) techniques can significantly improve the [12]. RFM has been found to work relatively well if expected response
prediction capability in direct marketing decisions. These predictive rate is high [24]. Other approaches to improve RFM results have in-
models using RFM variables are simple and easy to use in practice cluded Bayesian networks [1,8] and association rules [46].
than those with a complex set of variables. Besides descriptive model-
ing techniques popular in practice [43], thus, marketers should adopt
2.2.2. Classical data mining tools
those advanced predictive models in their direct marketing decisions.
Common data mining practice in classification is to gather a great
number of variables and apply different standard algorithms. Given
2. Customer response models using data mining techniques
the set of predefined classes and a number of attributes, these classi-
fication methods can provide a model to predict the class of other un-
2.1. Marketing DSS and customer response models
classified data. Mathematical techniques that are often used to
construct classification methods are binary decision trees, neural net-
The use of DSS in marketing goes back to the 1960s and 1970s
works, and logistic regression. By using binary decision trees, a tree
[22,44] and has been applied in various areas, including marketing
induction model with “Yes–No” format can be built to split data into
strategy, pricing, new product development, and product analysis
different classes according to its attributes. Such a model is very
and management [16]. There has been an increase of DSS use in
easy to apply to new cases, although the algorithms often produce
customer-side marketing activities, such as customer segmentation
an excessive number of rules. Neural networks often fit nonlinear re-
(or profiling), direct marketing, database marketing, and targeted ad-
lationships very well, but are difficult to apply to new data. Logistic
vertising. This reflects advances in database management and com-
regression models are easy to apply to new data, although the prob-
plex model building [11,16,35]. More convenient methods are
lem of a cutoff between classes can be an issue [32].
available for the acquisition and storage of large amounts of customer
Relative performance of data mining algorithms has long been un-
and transactional data. In addition, knowledge-based systems or intel-
derstood to depend upon the specific data. Since data mining soft-
ligent systems using data mining techniques (e.g., neural networks)
ware is widespread, common practice in classification is to try the
[37] have emerged in the marketing domain.
three basic algorithms (decision trees, neural networks, logistic re-
This trend is broadly termed knowledge-based marketing.
gression), and use the one that works best for the given data set.
Knowledge-based marketing is both data-driven and model-driven:
Studies have compared these algorithms with RFM. Levin and Zahavi
that is the use of sophisticated data mining tools and methods to
[20] compared RFM with decision trees (specifically CHAID), pointing
find knowledge discovery from customer and transactional databases
out that decision trees are more automatic (RFM requires extensive
[26]. Overall, this leads to more efficient and effective communication
data manipulation), but involve modeling issues such as controlling
with potential buyers and an increase in profits. An important ap-
tree size and determining the best split for branches and leaves.
proach to knowledge-based marketing is to understand customers
Kim and Street [19] proposed a neural network model and applied
and their behavioral patterns. This requires such transactional charac-
feature selection mechanisms to reduce input variables, enabling
teristics as recency of purchases, frequency of purchases, size of pur-
focus upon the most important variables. Baesens et al. [1] also ap-
chases, identifying customer groups, and predicting purchases [35].
plied neural networks to customer response models (adding custom-
The RFM model and other data mining-based customer response
er profile indicators to RFM), obtaining better prediction accuracy.
models have proven useful to marketers.
That is a consistent finding — data mining algorithms will be expected
to better predict customer response than RFM. However, RFM re-
2.2. Data mining techniques for customer response models
mains interesting because it relies upon the three fundamentally
basic inputs that are readily available.
2.2.1. RFM
R represents the period since the last purchase. F is the number of
purchases made by a customer during a certain period. M is the total 3. Methodology
purchase amount by a customer over that period. It is common practice
for each R, F, and M to have five groups or levels and thus there are 125 3.1. Problem description and data set
(=5 * 5 * 5) customer segmentation groups. Each customer is segment-
ed into one cell or group. This model allows markets to differentiate This research design includes two studies (Study 1 and Study 2
their customers in terms of three factors and to target the customer hereafter) using two datasets obtained from the Direct Marketing
groups that are likely to purchase products or services. This technique Educational Foundation. Study 1 uses a dataset including 101,532 in-
is known as the benchmark model in the area of database marketing [3]. dividual purchases from 1982 to 1992 in catalog sales. Study 2 is
Since its introduction in a major marketing journal [5], RFM has based on the data of 1,099,009 individual donors' contributions to a
received a great deal of interest from both academic and industry non-profit organization collected between 1991 and 2006. The pur-
communities [3,17]. Many studies [1,13,17] have recognized these chase orders (or donations) included ordering (or donation) date
three variables as important to predict the future responses by cus- and ordering amount. The last four months (Aug–Dec) of the data
tomers to potential direct marketing efforts. Certain limitations in were used as the target period: Aug–Dec 1992 for Study 1 and
D.L. Olson, B.(K.) Chae / Decision Support Systems 54 (2012) 443–451 445
Aug–Dec 2006 for Study 2. The average response rates in Studies 1 into customer behavior, as well as lead to ways to profitably act on
and 2 are 0.096 and 0.062 respectively. results. One of a number of algorithms automatically determines
Data preparation and manipulation are an important stage of which variables are most important, based on their ability to sort
knowledge discovery and learning in knowledge-base marketing the data into the correct output category. The method has relative
[35]. Fig. 1 describes our approach. The raw data contained customer advantage over neural network and genetic algorithms in that a
behavior represented by account, order (or donation) date, order (do- reusable set of rules are provided, thus explaining model
nation) dollars, and many other variables. We followed the general conclusions.
coding scheme to compute R, F, and M [17]. Various data preparation
techniques (e.g., filtering, transforming) were used during this pro-
3.2.6. Neural networks (NN)
cess. The order date of last purchase (or the date of last donation)
Neural networks are the third classical data mining tool found in
was used to compute R (R1, R2, R3, R4, R5). The data set contained
most commercial data mining software products, and have been ap-
order (or donation) history and order dollars (or donation amounts)
plied to direct marketing applications [4,8,19,36]. NN are known for
per each customer (or donor), which were used for F (F1, F2, F3, F4,
their ability to train quickly on sparse data sets. NN separates data
F5) and M (M1, M2, M3, M4, M5). We also included one response var-
into a specified number of output categories. NN are three layer net-
iable (Yes or No) to the direct marketing promotion or campaign.
works wherein the training patterns are presented to the input layer
and the output layer has one neuron for each possible category.
3.2. Predictive models
All three variables were significant at the 0.01 level. The relation- Table 2
ship between R and customer response is negative, as expected. In RFM boundaries.
contrast, F and M are positively associated with customer response. Factor Min Max Group 1 Group 2 Group 3 Group 4 Group 5
R and F are stronger predictors for customer response.
R 12 3810 1944+ 1291–1943 688–1290 306–687 12–305
RFM was initially applied, dividing the scales for each of the three Count 16,297 16,323 16,290 16,351 16,271
components into five groups based upon the scales for R, F, and M. F 1 39 1 2 3 4–5 6+
This was accomplished by entering bin limits in SPSS. Table 2 shows Count 43,715 18,274 8206 6693 4644
M 0 4640 0–20 21–38 39–65 66–122 123+
boundaries. Group 5 was assigned the most attractive group, which
Count 16,623 16,984 15,361 16,497 16,067
for R was the minimum, and for F and M the maximum.
Note the skewness of the data for F, which is often encountered.
Here the smaller values dominate that metric. Table 3 displays the cannot obtain the desired counts for each of the 125 combined cells
counts obtained for these 125 cells. because we are dealing with three scales. But we can come closer,
The proportion of responses (future order placed) for the data is as in Table 6. Difficulties arose primarily due to F having integer
given in Table 4. values. Table 6 limits were generated sequentially, starting by divid-
In the training set, 10 of 125 possible cells were empty, even with ing R into 5 roughly equal groups. Within each group, F was then
over 100,000 data points. The cutoff for profitability would depend sorted into groups based on integer values, and then within those
upon cost of promotion compared to average revenue and rate of 25 groups, M divided into roughly equally sized groups.
profit. For example, if cost of promotion were $50, average revenue The unevenness of cell densities is due to uneven numbers in the
per order $2000, and average profit rate $0.25 per dollar of revenue, few integers available for the F category. The proportion of positive
the profitability cutoff would be 0.1. In Table 4, those cells with return responses in the training set is given in Table 7.
ratios greater than 0.1 are shown in bold. Those cells with ratios at 0.1 If M = 5, this model predicts above average response. There is a
or higher with support (number of observations) below 50 are indi- dominance relationship imposed, so that cells 542 and better, 532
cated in italics. They are of interest because their high ratio may be spu- and better, 522 and better, 512 and better, 452 and better, 442 and
rious. The implication is fairly self-evident — seek to apply promotion to better, and 433 and better are predicting above average response.
those cases in bold without italics. The idea of dominance can also be Cells 422, 414, and 353 have above average training response, but
applied. The combinations of predicted success for different training cells with superior R or F ratings have below average response, so
cell proportions are given in Table 5. these three cells were dropped from the above average response
The RFM model from the Excel spreadsheet model yields predictive model. The prediction accuracy ((13,897 + 734) / 20,000) for this
model performance shown in the Appendix A for the line Basic on 0.1 model was 0.732 (see the balance on 0.1 row in the Appendix A). In
(because the cutoff used was a proportion of 0.1) along with results this case, balancing cells did not provide added accuracy over the
from the other models. This model was correct (13,961+ 1337 = basic RFM model with unbalanced cells. Using the cutoff rate of 0.5,
15,298) times out of 20,000, for a correct classification rate of 0.765. the model is equivalent to predict the combination of R = 5, F =4 or 5,
The error was highly skewed, dominated by the model predicting and M= 4 or 5 as responding and all others not. This model had a cor-
4113 observations to be 0 that turned out to respond. An alternative rect classification rate of 0.894, which was inferior to the degenerate
model would be degenerate — simply predict all observations to be 0. case. For this set of data, balancing cells accomplished better statistical
This would have yielded better performance, with 18,074 correct re- properties per cell, but was not a better predictor.
sponses out of 20,000, for a correct classification rate of 0.904. This Since F is highly correlated with M (0.631 in Table 1), the analysis
value could be considered a par predictive performance. This data is in- is simplified to one dimension. Dividing the training set into groups of
cluded in the Appendix A, where we will report results of all further 5%, sorted on V, generates Table 8.
models in terms of correct classification.
Increasing the test cutoff rate leads to improved models. We used
Table 3
increasing cutoffs of 0.2, 0.3, 0.4, and 0.5, yielding the results indicat-
Count by RFM cell – training set.
ed in the Appendix A. Only the model with a cutoff rate of 0.5 resulted
in a better classification rate than the degenerate model. In practice, RF R F M1 M2 M3 M4 M5
the best cutoff rate would be determined by financial impact analysis, 55 R 12–305 F 6+ 0 0 16 151 1761
reflecting the costs of both types of errors. Here we simply use classi- 54 F 4–5 2 18 118 577 1157
fication accuracy overall, as we have no dollar values to use. 53 F 3 9 94 363 756 671
52 F 2 142 616 1012 1135 559
The correlation across F and M (0.631 in Table 1) can be seen in 51 F 1 2425 1978 1386 938 387
Table 3, looking at the R = 5 categories. In the M = 1 column of 45 R306–687 F 6+ 0 1 11 101 1018
Table 3, F entries are 0 for every F5 category, usually increasing 44 F 4–5 0 16 87 510 927
through M = 2 through M = 5 columns. When F = 5, the heaviest 43 F 3 6 88 316 699 636
42 F 2 150 707 1046 1140 616
density tends to be in the column where M = 5. This skewness is
41 F 1 2755 2339 1699 1067 416
often recognized as one of the problems with RFM [13,27,31]. Our ap- 35 R688–1290 F 6+ 0 1 5 70 799
proach to this issue was through more equal density (size-coding) to 34 F 4–5 1 16 122 420 832
obtain data entries for all RFM cells. We accomplished this by setting 33 F 3 9 88 319 706 589
cell limits by count within the training set for each variable. We 32 F 2 163 697 1002 1128 645
31 F 1 2951 2567 1645 1078 437
25 R1291–1943 F 6+ 0 0 9 56 459
24 F 4–5 0 22 72 372 688
Table 1 23 F 3 9 95 290 678 501
Variable correlations. 22 F 2 211 749 1096 1128 561
21 F 1 3377 2704 1660 1108 478
R F M Ordered
15 R 1944+ F 6+ 0 0 3 22 170
R 1 14 F 4–5 1 11 74 243 409
F −0.192⁎⁎ 1 13 F 3 9 122 261 511 380
M −0.136⁎⁎ 0.631⁎⁎ 1 12 F 2 268 878 1108 995 522
Ordered −0.235⁎⁎ 0.241⁎⁎ 0.150⁎⁎ 1 11 F 1 4145 3177 1641 908 449
⁎⁎ Totals 16,623 16,984 15,361 16,497 16,067
Correlation is significant at the 0.01 level (2-tailed).
D.L. Olson, B.(K.) Chae / Decision Support Systems 54 (2012) 443–451 447
Table 4 Table 6
Response ratios by cell. Balanced group cell densities—training set.
RF R F M1 M2 M3 M4 M5 RF M1 M2 M3 M4 M5
Lift is the marginal difference in a segment's proportion of re- A logistic regression model was run on RFM variables. The model
sponse to a promotion and the average rate of response. Target cus- results were as shown in Table 9. The beta values of R and F are
tomers are identified as the small subset of people with marginally found to be significant.
higher probability of purchasing. Lift itself does not consider profit- Note that F was not included at all. This is explainable by the high
ability. In practice, this needs to be considered. For our purposes, we correlation between M and F, and the dominance of R in obtaining a
demonstrate without dollar values (which are not available), noting better fit. This model did very well on the test data, with a correct
that the relative cost of marketing and expected profitability per seg- classification rate of 0.984.
ment will determine the optimal number of segments to market. The neural network model used a popular architecture called mul-
Fig. 2 shows lift by value ratio. tilayer perceptron (MLP) [10]. This model built a hidden layer. The
In Fig. 2, the most responsive segment has an expected return of rate of neural network was 0.911, as shown in the Appendix A.
slightly over 20%. The lift line is the cumulative average response as Another performance measure used in this study is gains, which is a
segments are added (in order of response rate). useful tool for evaluating the value of predictive models in direct mar-
Using the value ratio as a predictive classifier, the training data keting [21]. We use gains to compare the performance of RFM score
was used to identify cells with better responses. Model fit is shown model and classical data mining-based predictive models. Table 11
in the Appendix A in the row value function. This model has a correct
classification rate of 0.721. This is inferior to a degenerate model that Table 7
would simply classify all cases as no response, indicating that the Training set proportion of responses by cell.
value function was non-productive in this case. While it is easier to
RF M1 M2 M3 M4 M5
manipulate than the RFM model, in this case the fit was inferior to
the basic RFM model. 55 0.129 0.178 0.101 0.673 0.818
54 0.059 0.118 0.189 0.541 0.629
Data mining is rich in classification models [2,23]. Three classical
53 0.064 0.130 0.287 0.392 0.647
data mining classification models were applied to the data: logistic 52 0.076 0.103 0.200 0.424 0.605
regression, decision trees, and neural networks. We next applied 51 0.054 0.102 0.274 0.406 0.527
these three basic data mining algorithms using SPSS. 45 0.037 0.109 0.141 0.211 0.378
44 0.041 0.108 0.116 0.281 0.417
43 0.033 0.052 0.125 0.072 0.483
42 0.049 0.118 0.098 0.073 0.544
41 0.045 0.038 0.092 0.116 0.531
35 0.045 0.067 0.138 0.060 0.458
Table 5 34 0.052 0.043 0.059 0.080 0.448
Basic RFM models by cutoff. 33 0.042 0.048 0.058 0.093 0.433
32 0.027 0.045 0.058 0.097 0.379
Cutoff R F M
31 0.050 0.040 0.062 0.080 0.414
0.1 R=5 Any Any 25 0.037 0.051 0.056 0.084 0.254
R=4 F=5 M = 3, 4, or 5 24 0.024 0.046 0.052 0.076 0.309
R=3 F=4 M = 4 or 5 23 0.051 0.047 0.055 0.080 0.273
F=3 M=5 22 0.027 0.040 0.055 0.068 0.246
F = 4 or 5 M = 3, 4, or 5 21 0.027 0.038 0.048 0.076 0.242
0.2 R=5 F = 2, 3, 4, or 5 Any 15 0.017 0.021 0.025 0.051 0.146
R=4 F=5 M = 3, 4, or 5 14 0.016 0.017 0.033 0.054 0.167
0.3 R=5 F = 3, 4, or 5 M = 2, 3, 4, or 5 13 0.010 0.019 0.034 0.052 0.156
0.4 R=5 F = 4 or 5 M = 2, 3, 4 or 5 12 0.018 0.021 0.036 0.043 0.137
0.5 R=5 F=5 M = 3, 4, or 5 11 0.016 0.022 0.014 0.044 0.154
448 D.L. Olson, B.(K.) Chae / Decision Support Systems 54 (2012) 443–451
Table 8 Table 9
V values by cell. Regression betas for logistic regression.
Table 11
Gains.
Table 13
Table 12 RFM boundaries.
Variable correlations.
Factor Min Max Group 1 Group 2 Group 3 Group 4 Group 5
R F M Response R 1 4950 2811+ 1932–2811 935–1932 257–935 1–257
R 1 Count 220,229 219,411 220,212 219,503 219,654
F −0.237⁎⁎ 1 F 1 1027 1 2 3 4 5+
M −0.125⁎⁎ 0.340⁎⁎ 1 Count 599,637 190,995 95,721 57,499 155,157
Response −0.266⁎⁎ 0.236⁎⁎ 0.090⁎⁎ 1 M 0 100,000 0–9 10–24 25–39 40–89 90+
⁎⁎
Count 248,639 343,811 77,465 209,837 219,257
Correlation is significant at the 0.01 level (2-tailed).
450 D.L. Olson, B.(K.) Chae / Decision Support Systems 54 (2012) 443–451
[11] E. Eisenstein, L. Lodish, Marketing decision support and intelligence systems: pre- [36] K.A. Smith, J. Gupta, Neural networks in business: techniques and applications for
cisely worthwhile or vaguely worthless? In: in: B. Weitz, R. Wensley (Eds.), Hand- the operations researcher, Computers and Operations Research 27 (11/12)
book of Marketing, Sage Publication, London, 2002. (2000) 1023–1044.
[12] R. Elsner, M. Krafft, A. Huchzemeier, Optimizing Rhenania's main-order business [37] E. Turban, J. Aronson, T. Liang, Decision Support Systems and Intelligent Systems,
through dynamic multilevel modeling, Interfaces 33 (1) (2003) 50–66. 7th ed. Prentice Hall, Upper Saddle River, NJ, 2004.
[13] P. Fader, B. Hardie, K. Lee, RFM and CLV: using iso-value curves for customer base [38] E. Turban, R. Sharda, D. Delen, D. King, Business Intelligence: A Managerial Approach,
analysis, Journal of Marketing Research 42 (4) (2005) 415–430. 2nd ed. Prentice Hall, New York, 2011.
[14] M. Fitzpatrick, Statistical analysis for direct marketers — in plain English, Direct [39] B. Van den Berg, T. Breur, Merits of interactive decision tree building: part 1,
Market 64 (4) (2001) 54–56. Journal of Targeting, Measurement and Analysis for Marketing 15 (3)
[15] R. Gopal, Ad mediation: new horizons in effective email advertising, Communica- (2007) 137–145.
tions of the ACM 19 (1) (2001) 17–30. [40] B. Van den Berg, T. Breur, Merits of interactive decision tree building: part 2: how
[16] M. Hart, Systems for supporting marketing decisions, In: in: F. Burstein, C. to do it journal of targeting, Measurement & Analysis for Marketing 15 (4) (2007)
Holsapple (Eds.), Handbook on Decision Support Systems, 2, Springer, 2008, 201–209.
pp. 395–418. [41] S. Vargo, R. Lusch, Evolving to a new dominant logic for marketing, Journal of
[17] A. Hughes, Strategic Database Marketing, Third ed. McGraw-Hill, New York, 2006. Marketing 68 (2004) 1–17.
[18] J. Jonker, N. Piersma, R. Potharst, A decision support system for direct mailing deci- [42] G. Verhaert, D. Poel, Empathy as added value in predicting donation behavior, Jour-
sions, Decision Support Systems 42 (2006) 915–925. nal of Business Research 64 (2011) 1288–1295.
[19] Y. Kim, W. Street, An intelligent system for customer targeting: a data mining ap- [43] P. Verhoef, P. Spring, J. Hoekstra, P. Leeflang, The commercial use of segmentation
proach, Decision Support Systems 37 (2004) 215–228. and predictive modeling techniques for database marketing in the Netherlands,
[20] N. Levin, J. Zahavi, Predictive modeling using segmentation, Journal of Interactive Decision Support Systems 34 (2002) 471–481.
Marketing 15 (2) (2001) 2–22. [44] B. Wierenga, The past, the present and the future of marketing decision models,
[21] G. Linoff, M. Berry, Data Mining Techniques, Wiley, Indianapolis, 2011. In: in: B. Wierenga (Ed.), Handbook of Marketing Decision Models, Springer,
[22] J. Little, Decision support systems for marketing managers, Journal of Marketing 2008, pp. 3–20.
43 (3) (1979) 9–26. [45] A. Yang, How to develop new approaches to RFM segmentation, Journal of
[23] N. Mastrogiannis, B. Boutsinas, I. Giannikos, A method for improving the accuracy Targeting, Measurement and Analysis for Marketing 13 (1) (2004) 50–60.
of data mining classification algorithms, Computers and Operations Research 36 [46] H. Yun, D. Ha, B. Hwang, K. Ryu, Mining association rules on significant rare data
(10) (2009) 2829–2839. using relative support, Journal of Systems and Software 67 (181–191) (2003).
[24] J. McCarthy, M. Hastak, Segmentation approaches in data-mining: a comparison [47] J. Zahavi, N. Levin, Applying neural computing to target marketing, Journal of Di-
of RFM, CHAID, and logistic regression, Journal of Business Research 60 (6) rect Marketing 11 (1) (1997) 5–22.
(2007) 656–662.
[25] R. McDaniel, Management strategies for complex adaptive systems, Performance David L. Olson is the James & H.K. Stuart Professor in MIS and Chancellor's Professor at
Improvement Quarterly 20 (2) (2007) 21–42. the University of Nebraska. He has published research in over 100 refereed journal ar-
[26] B. McKelvey, Avoiding complexity catastrophe in coevolutionary pockets: strategies ticles, primarily on the topic of multiple objective decision-making and information
for rugged landscape, Organization Science 10 (3) (1999) 294–321. technology. He has authored 17 books, is associate editor of Service Business and
[27] J. Miglautsch, Application of RFM principles: what to do with 1-1-1 customers? co-editor in chief of International Journal of Services Sciences. He has made over 100 pre-
Journal of Database Marketing 9 (4) (2002) 319–324. sentations at international and national conferences on research topics. He is a mem-
[28] P. Naik, M. Hagerty, C. Tsai, A new dimension reduction approach for data-rich ber of the Decision Sciences Institute, the Institute for Operations Research and
marketing environments, Journal of Marketing Research 37 (1) (2000) 88–101. Management Sciences, and the Multiple Criteria Decision Making Society. He was a
[29] E.W.T. Ngai, L. Xiu, D.C.K. Chau, Application of data mining techniques in custom- Lowry Mays endowed Professor at Texas A&M University from 1999 to 2001. He was
er relationship management: a literature review and classification, Expert Sys- named the Raymond E. Miles Distinguished Scholar award for 2002, and was a James
tems with Applications 36 (2) (2009) 2592–2602. C. and Rhonda Seacrest Fellow from 2005 to 2006. He was named Best Enterprise Infor-
[30] C. O'Reilly III, J. Harreld, M. Tushman, Organizational ambidexterity: IBM and emerg- mation Systems Educator by IFIP in 2006. He is a Fellow of the Decision Sciences Insti-
ing business opportunities, California Management Review 51 (4) (2009) 75–99. tute.
[31] D. Olson, Q. Cao, C. Gu, D. Lee, Comparison of customer response models, Service
Business 3 (2) (2009) 117–130.
[32] D.L. Olson, D. Delen, Advanced Data Mining Techniques, Springer, Heidelberg, 2008. Bongsug (Kevin) Chae is Associate Professor in Information & Operations Manage-
[33] P. Rossi, R. McCulloch, G. Allenby, The value of purchase history data in target ment at Kansas State University. He has published papers in such areas as business an-
marketing, Marketing Science 15 (4) (1996) 321–340. alytics, supply chain management, service innovation, and knowledge discovery. He
[34] G. Seni, J. Elder, Ensemble Methods in Data Mining: Improving Accuracy Through has made presentations in universities and global companies in several countries, pri-
Combining Predictions, Morgan & Claypool, 2010. marily on the topic of business analytics and intelligence, supply chain management,
[35] M. Shaw, C. Subramaniam, G. Tan, M. Welge, Knowledge management and data and service innovation. He is a recipient of the Ralph Reitz Teaching Award and a nom-
mining for marketing, Decision Support Systems 31 (2001) 127–137. inee of several other teaching awards at Kansas State University.