Ensemble Classifier For Stock Trading Recommendation
Ensemble Classifier For Stock Trading Recommendation
An International Journal
Chukiat Worasucheep
To cite this article: Chukiat Worasucheep (2022) Ensemble Classifier for Stock
Trading Recommendation, Applied Artificial Intelligence, 36:1, 2001178, DOI:
10.1080/08839514.2021.2001178
Introduction
Stock trend prediction is very valuable for investment management. Accurate
prediction makes it possible for investors to decide a proper moment to buy or
sell a stock, to achieve the goal to beat the market and make profits (Ding and
Qin 2020). However, stock trend prediction is really challenging due to the
high volatility in the stock market. The trend prediction in stock market has
drawn a lot of research attention for many decades using both statistical and
computing approaches including artificial intelligence. Recent approaches
focus more on machine learning with both regression and classification
techniques. Regression techniques aim to predict the future value of the
stock price (Wang et al. 2011), while classification techniques aim to predict
the trend of stock price movement (Deng et al. 2021; Kim and Han 2016;
Nobre and Neves 2019; Olson and Mossman 2003; Zhang et al. 2018, 2021).
In the second step, the predicted trend results are employed in a trading
recommendation system – that is to buy, to sell, or to do nothing if the
predicted trend is positive, negative, or very small, respectively. The invest
ment returns from trading simulation with the testing data are calculated.
Performance of the proposed model is evaluated using trading return and
Sharpe ratio. Sharpe ratio is a measure of risk-adjusted return, describing how
much excess return received for the volatility (or risk) of holding a riskier asset
(Sharpe 1994). The returns and Sharpe ratios are to be compared with the
trading using prediction results from individual classifier as well as with two
other cases – one is the buy and hold (B&H) and the other is if we know the
future (KF).
This research contributes to (1) construct a heterogeneous ensemble classi
fier that aggregates results from different base classifiers, each of which using
an extensive hyperparameter tuning. Inputs of each base classifier are various
commonly used technical indicators automatically chosen with a feature selec
tion method. In addition, we also (2) propose a trading recommendation
system based on the trend prediction result from the ensemble classifier.
The remaining of this article is organized as following. The next section
reviews related literature about technical analysis and financial trend predic
tion using classification algorithms. Then the proposed ensemble classifier will
be described, and its performance is evaluated and discussed. Finally, the last
section concludes this work with its limitations and future works.
Related Works
Technical Analysis and Indicators
indicator measures the range of price movement and can be used to identify
level of support and resistance. Common volatility indicator includes Bollinger
Band (BB) (Hu et al. 2015; Lorenzo 2013).
as sigmoid or tanh functions). Then the neuron sends its output to other
neurons or to be output of the network. The interconnection of all neurons
forms different types of architectures designed for various functions. The most
widely used architecture is called feed-forward multilayer perceptron (MLP)
which generally used back-propagation (BP) learning algorithm to adjust its
weights for supervised learning. BP works in an iteration of three steps. The
first step is propagating inputs forward through the hidden layers to the output
nodes. The second step is to propagate the errors backward through the
network starting from output layer. And the final step is to update the weight
and biases using approximate steepest descent rule. Such three steps repeat
until reaching the maximum iterations allowed with the aim to minimize the
errors of output during the training.
ANN has demonstrated promising results in predictions of stock price
trends during the past two decades. Olson and Mossman (2003) compared
the forecasts of one-year-ahead Canadian stock returns using ANN, logistic
regression and ordinary least squares. The results demonstrated that ANN
outperformed the other two algorithms. O’Connor and Madden (2006)
included some external indicators, such as currency exchange rates, in pre
dicting movements in the Dow Jones Industrial Average index using ANN
models. Kara, Boyacioglu, and Baykan (2011) predicted the direction of stock
price movement in the Istanbul Stock Exchange using ANN and SVM with ten
technical indicators as inputs. The empirical results demonstrated that their
ANN model performed better than SVM model. Chen, Leung, and Daouk
(2003) applied probabilistic neural network to predict the direction of return
on Taiwan Stock Exchange and found that their model demonstrated a more
predictive power than generalized methods of moments with Kalman filter
and random walk.
Random forest (RF) is an efficient learning algorithm for classification that
is constructed from many unpruned decision trees (DT) from random
subsets of features using bootstrapped training data (Breiman 2001). The
accuracy (or probability of correct prediction) of a single tree may not be
high, but the combination of many trees, forming the forest, results in
a higher accuracy. Construction of RF mainly includes two stages. The first
stage is the generation of forest, in which the training samples are divided
into many samples at random, construct CART (Classification and
Regression Tree) decision trees. When creating partitions to a feature, the
goodness of a partition is measured by purity. If a partition is pure, for each
sub-branch of this node, its instances belong to the same class (Zhang et al.
2018). The second stage is to determine the classification results from the
forest. For a classification task, the result combination can be as simple as
majority voting (Breiman 2001). This ensemble method enhances the accu
racy of RF over the single DT, while the problem of overfitting is controlled
simultaneously.
APPLIED ARTIFICIAL INTELLIGENCE e2001178-231
Feature Selection
First, one major question is how to choose a proper set of attributes, called
features, to be inputs of the model. Feature selection methods become an
important step of data preprocessing for the classification algorithm. Feature
selection approaches fall into three categories: filter based, wrapper based, and
embedded (Li et al. 2018; Xue et al. 2016).
Hyperparameter Tuning
Second, building an effective classification model is a complex and time-
consuming process that involves tuning its hyperparameters (Nguyen, Xue,
and Zhang 2020; Yang and Shami 2020). Tuning hyperparameters is consid
ered a key component of building an effective machine learning model,
especially for tree-based or neural-based models like RF and ANN, which
have many hyperparameters. Accurate results for ANN highly depend on
a careful selection of its hyperparameters such as number of hidden layers,
number of nodes in each layer the learning rate, input variables, etc. (Hussain
et al. 2008). Unfortunately, manual tuning of hyperparameters not only
requires expertise but also is prone to getting lower performance.
Grid search is one of the most commonly used methods to explore hyper
parameter configuration space. It involves exhaustively search that evaluates
all the hyperparameter combinations given to a grid of configurations
(Nguyen, Xue, and Zhang 2020; Yang and Shami 2020). In this work, we use
grid search for hyperparameter tuning of all selected classification algorithms
for an optimal performance. Although this approach is time consuming, this
work does not aim to run in a real-time environment.
which strategy clearly outperforms the others. This work employs an ensemble
classification model that works on the modified majority voting of results of
five classifiers.
Figure 1. Process.
e2001178-236 C. WORASUCHEEP
listed in Table 1. All these indicators, forming attribute, or feature set of the
classifiers, are commonly used by the technical practitioners in stock markets.
Their descriptions and formulas can be found in Hu et al. (2015) and Lorenzo
(2013). Some of them e.g. RSI, DISP, and ROC use varying periods to deal with
uncertainty. Then every feature is normalized to range [0, 1] before further
processing.
Feature Selection
Base Classification
The model aims to predict the trend of closing price at day d + 2 (closed + 2)
compared to the opening price on the next day d + 1 (opend + 1). Specifically,
output of the model, called classd ∈{-1, 0, 1}, represents direction of δd or
percentage change of day d as of the following equation:
APPLIED ARTIFICIAL INTELLIGENCE e2001178-237
closedþ2 opendþ1
δd ¼ 100 (1)
opendþ1
where closed and opend are the closing price and opening price of day d. The
value of classd is determined by comparing δd with a decision threshold θ as in
Figure 2. That is classd is 0 if the percentage change δd is between -θ and θ.
Otherwise classd will be −1 or 1 if δd is less than -θ or greater than θ
respectively. The sign of the class value indicates short-term direction, i.e.
positive is uptrend and negative is downtrend. Class 0 implies that the change
is so small that the potential profit is difficult to beat transaction fee.
Hyperparameter Tuning
As discussed in previous section, performance of any classifiers highly depends
on its hyperparameter setting. Grid search method is employed in this work to
explore hyperparameter configuration space. It exhaustively evaluates all the
hyperparameter combinations (Yang and Shami 2020) of all five base classi
fiers as following.
SVM
kernel [‘rbf,’ ‘poly’]
C [0.01, 0.1, 1, 10, 100]
gamma [1, 0.1, 0.01, 0.001]
ANN
solver [‘adam,’ ‘sgd’]
activation [‘tanh,’ ‘relu’]
learning_rate [‘invscaling,’ ‘adaptive’]
learning_rate_init [0.01, 0.001]
hidden_layer_sizes [(20,), (40,), (60,), (40, 40)]
RF
criterion [‘gini,’ ‘entropy’]
n_estimators [100, 400]
max_depth [5, 9, 13]
min_samples_split [2, 6]min_samples_leaf [1, 5]
XGBoost
max_depth [4, 8]
learning_rate [0.0005, 0.01]gamma [0.1, 1.0]
subsample [1.0, 0.7]
e2001178-238 C. WORASUCHEEP
tree. Parameter max_bin controls the maximum number of bins that features
will bucketed into. Large values of num_leaves and max_bin increase accuracy
of the training set but also are prone to overfitting.
Result Ensemble
The proposed model in this work employs five base classifiers, namely SVM,
ANN, RF, XGB, and LGB. Each of which runs independently using the same
input data of the selected features. However, the trend prediction results from
base classifiers might be different. Therefore, they are further aggregated using
a modified majority voting for a final trend prediction for the proposed
ensemble model. Since there are five base classifiers, we propose two following
ensemble methods for comparison:
(1) For each stock prediction, two base classifiers with the lowest accuracy
values are ignored. Then the prediction results of remaining three classi
fiers are considered with a majority voting scheme summarized in Table 2.
Last three rows are for the cases that results of all three base classifiers are
mutually different; then the final results will be from the one with highest
accuracy value. Let us call algorithm using this ensemble scheme as T3.
(2) A straightforward majority voting. If there is a tie (e.g. [1, 1, −1, −1,
and 0]), the result will come from the classifier pair with a greater sum of
accuracy values. Let us call algorithm using this ensemble scheme as E5.
Trading Recommendation
After the ensemble of prediction results, all predicted classd values are employed
in the recommendation of trading one stock as in Figure 3. That is if classd = 0,
then do nothing in response to the uncertainty in direction of price change. If
classd = 1, then take a long position (or buy) of the stock with a proportion ρ of
the current available cash. If classd = – 1, then take a short position (or sell) of
the stock for a proportion ρ of the current amount of in-hand stock.
Experimentation
Data in the experiment are historical daily prices of 20 stocks from Stock
Exchange of Thailand (SET). They are among top largest stocks and are ran
domly selected from leaders of different major industries including energy,
telecommunication, banking, foods, infrastructure, property, etc. To study the
performance in varying market situations, the datasets are taken from the
following three different periods:
Each dataset ranges 5 years or about 1220 days and is split into 80:20 proportion
for training and testing. The training set is for constructing the classification
models whose performance will be compared using the testing set. The character
istics of the stocks are summarized in Table 3. Column Sign Diff? indicates that the
direction of change in training set is different to the direction of change in testing
set. Such different patterns of training set and testing set makes it more difficult to
generalize the classification models. Figures 5-7 illustrate charts of the closing
APPLIED ARTIFICIAL INTELLIGENCE e2001178-241
prices of all stocks for the three datasets. It can be seen from both Table 3 and the
figures that the tested stocks have various characteristics: uptrend, downtrend,
fluctuation, and sideway, during the training and testing periods for evaluating the
proposed models under varying market situations. Figure 8 illustrates proportion
e2001178-242 C. WORASUCHEEP
Figure 5. Price charts of all stocks in dataset A (2014–2018). Training and testing datasets are
shown in green and blue colors.
of classd (or direction of price change) for testing set of each stock. Set {%Up, %
Notrend, %Down} maps to {1, 0, −1}. The Figure 8 shows that the classification
problem in this work is not imbalance.
Experimentation Setup
Figure 6. Price charts of all stocks in dataset B (2015–2019). Training and testing datasets are
shown in green and blue colors.
(3) Training with Hyperparameter Tuning: After feature selection, four base
classifiers are trained using the training set. We use Python scikit learn’s
gridsearchcv function to exhaustively search for the optimal hyperpara
meter set of SVM, ANN, RF, XGB, and LGB as reported in subsection 3.4.
(4) Prediction and Result Ensemble: After training, five base classifiers are
independently used for prediction of daily classd of the testing set. Their
results are then ensembled using algorithm in subsection 3.5 to obtain
the prediction results of T3 and E5.
(5) Trading Simulation to Observe Performance: The prediction results of all 7
classifiers are further used for simulation using trading recommendation
described in subsection 3.5 for comparison. In the simulation, the initial
fund is 1,000,000 baht. Note that the amount of initial fund does not
matter in this experiment since the performance is based on the invest
ment return R for whole testing period, and R is calculated as following.
valuef valuei
R¼ (2)
valuei
e2001178-244 C. WORASUCHEEP
Figure 7. Price charts of all stocks in dataset C (2016–2020). Training and testing datasets are
shown in green and blue colors.
Performance Comparison
RBH, as baseline. The return RBH in this case is computed from the closing
price of the last day (closef) and the opening price of first day (openi) of the
testing dataset, as following:
closef openi
RBH ¼ (3)
openi
e2001178-246 C. WORASUCHEEP
The main metrics for comparison are return surplus and Sharpe ratio. Return
surplus (Δ) is difference of the return from algorithm and the return from
B&H strategy, i.e. ΔT3 = RT3 – RBH. A positive surplus means that using the
algorithm obtains a higher return than using B&H strategy, and of course the
higher Δ achieved, the better the algorithm is.
To measure the volatility or risk of trading, Sharpe ratio (Sr) is the average
return earned in excess of the risk-free rate per unit of volatility, represented
by the standard deviation σp. For good return/risk ratio, Sr should be higher
than 1.0. Sr is calculated by
Rp Rf
Sr¼ (4)
σp
where Rp is return rate from investment and Rf or risk-free rate is the return of
an investment with risk-free asset, i.e. short-term government treasury bills. In
this experiment, Rf is set to 2%, 2% and 0.1% for datasets A, B, and
C respectively; these are average values obtained from web Interest Rate of
Bank of Thailand (https://www.bot.or.th/English/Statistics/FinancialMarkets/
InterestRate/Pages/InterestRate.aspx).
It’s also interesting to observe the case of If we know the future (KF). The
return RKF is computed based on the formulae which is similar to the case
of prediction. However, the trading decision (buying, selling or do-
nothing) is based on the actual class values, as if we know the future
closing prices.
valuef valuei
RKF ¼ (5)
valuei
Obviously, if we know the future, the return from trading (RKF) is very high.
This is confirmed from the column RKF in Table 4. It unquestionably beats
other trading strategies for every stock and thus is regarded as the ideal case.
Therefore, we here will focus on only the cases of seven classifiers compared
with B&H strategy.
In addition, we investigate the relationship of return of proposed ensem
ble model and accuracy metric. Accuracy (Ac) is the total number of
correct predictions divided by the total number of predictions made for
a dataset.
BJC.BK -13.065 120.869 0.426 0 0.418 -5.751 0.467 5.98 0.451 -1.276 0.467 9.684 16.116 6.797
CPALL.BK 4.965 84.797 0.332 4.597 0.348 6.551 0.377 -3.706 0.348 2.742 0.393 5.307 2.538 3.411
CPF.BK 10 140.335 0.353 3.15 0.348 0 0.426 25.88 0.418 28.6 0.443 24.707 30.647 26.371
CPN.BK -16.333 95.276 0.393 -8.162 0.393 -3.826 0.398 -0.451 0.398 -2.23 0.414 -3.963 2.797 4.822
EGCO.BK 34.146 143.653 0.312 -2.331 0.287 4.678 0.328 -6.366 0.385 4.295 0.361 9.745 17.463 -5.762
HANA.BK 10.236 276.257 0.406 -8.707 0.402 -0.133 0.443 -2.475 0.459 16.206 0.492 30.645 11.766 3.667
HMPRO.BK 7.285 103.741 0.451 8.531 0.488 24.299 0.508 9.656 0.434 4.445 0.455 16.051 14.13 21.002
e2001178-247
(Continued)
Table 4. (Continued).
Dataset B. 2015 – 2019
e2001178-248
KBANK.BK -19.571 105.524 0.426 -1.317 0.410 -4.843 0.422 7.716 0.443 7.361 0.410 -8.997 14.03 9.114
LH.BK -2 74.199 0.377 4.871 0.361 -1.017 0.381 -1.234 0.406 -2.764 0.426 6.46 0.328 0.881
PTT.BK -4.839 84.675 0.496 0 0.500 1.97 0.336 -8.191 0.320 -9.233 0.398 -7.332 -7.1 -3.322
PTTEP.BK 10.573 119.194 0.361 0 0.316 2.535 0.353 6.298 0.344 -4.037 0.373 14.512 11.273 -0.555
QH.BK -5.224 79.658 0.414 0 0.410 1.241 0.434 0.82 0.406 3.766 0.459 6.496 7.719 9.169
SCC.BK -10.502 75.854 0.385 -2.981 0.385 -1.989 0.381 -7.199 0.385 -11.208 0.353 -5.259 -5.059 -3.343
TASCO.BK 41.781 250.891 0.406 0 0.402 -3.573 0.406 13.475 0.422 29.601 0.406 8.582 22.683 11.077
TOP.BK 8.108 186.200 0.389 14.838 0.389 0 0.422 27.096 0.426 32.565 0.414 35.591 43.655 39.199
TVO.BK 3.774 99.555 0.430 0 0.439 0.189 0.443 14.187 0.430 3.762 0.459 11.829 13.779 12.382
C. WORASUCHEEP
Table 5. (Continued).
Dataset C. 2016 – 2020
KBANK.BK 23.43 -19.51 14.20 26.75 10.93 13.48 14.29
LH.BK 7.97 34.16 2.74 13.87 4.17 3.42 30.82
PTT.BK 2.26 4.64 0.13 14.51 12.51 0.94 -6.71
PTTEP.BK -0.62 -9.65 21.15 16.89 17.56 20.37 8.40
QH.BK -1.51 25.04 4.52 0.92 43.44 8.46 20.49
SCC.BK 24.38 -11.77 12.23 39.63 22.84 36.20 30.35
TASCO.BK 16.04 16.04 31.50 38.49 29.06 39.36 19.22
TOP.BK 5.42 60.02 14.28 10.35 -2.06 7.53 47.78
TVO.BK -27.52 -23.84 30.72 -1.89 22.81 32.40 10.78
accuracies (Ac), are displayed in Table 6. The higher Δ indicates more profit
given by such algorithm, whereas the higher Sr reflects a better excess return
per risk. As a baseline reference, Table 7 reports the averages of returns from
B&H (RBH). To identify the superior performance among 7 classifiers, their
returns are ranked from 1 (best) to 7 (worst) for 20 stocks. Next, those 20 ranks
are averaged for each classifier and listed in Table 8 as average rank. Table 9
reports the correlation coefficients (Cor) of return and accuracy of all classi
fiers. Figure 9 illustrates the comparative returns of all classifiers over testing
periods for some largest stocks in different sectors.
Here are the observations from Tables 6 to 9.
(1) In Table 6, all average Δs are greater than 0, indicating that on average
the proposed trading recommendation provides a more profitable return
than B&H.
(2) Among three datasets: A, B & C, T3 outperforms all base classifiers,
except in dataset A (ΔT3 = 12.375 < ΔRF = 12.720). In addition, E5 outper
forms all base classifiers, except in dataset C (ΔE5 = 15.466
< ΔLGB = 15.696).
(3) T3 and E5 defeat all individual base classifiers in terms of the average Δ �
of three datasets: Δ� T3 = 13.513 (the highest), Δ
� E5 = 11.993 (the second best),
with ranking as Δ � T3 > Δ
� E5 > Δ
� LGB > Δ
� XGB > Δ � RF > Δ
� ANN > Δ � SVM.
Table 6. Averages of return surplus (ΔX = RX – RBH), Sharpe ratios (Sr), and accuracy (Ac).
Metric Dataset SVM ANN RF XGB LGB T3 E5
Average Δ A. 2014 – 2018 12.113 11.506 12.720 12.074 10.340 12.375 12.878
B. 2015 – 2019 0.453 0.822 4.625 5.787 6.928 10.006 7.632
C. 2016 – 2020 7.966 8.366 13.356 14.115 15.696 18.150 15.466
Δ Grand average 6.844 6.898 10.234 10.659 10.988 13.510 11.992
Average Sr A. 2014 – 2018 0.394 0.040 0.476 0.265 -0.298 0.755 0.916
B. 2015 – 2019 -0.063 -0.587 0.710 1.134 1.006 1.261 1.160
C. 2016 – 2020 -0.063 -0.009 0.710 0.910 1.071 1.586 1.225
Sr Grand average 0.089 -0.185 0.632 0.770 0.593 1.201 1.100
Average Ac A. 2014 – 2018 0.412 0.414 0.427 0.415 0.419 0.423 0.445
B. 2015 – 2019 0.401 0.400 0.413 0.408 0.417 0.497 0.463
C. 2016 – 2020 0.427 0.430 0.443 0.440 0.443 0.481 0.476
Ac Grand average 0.414 0.414 0.428 0.421 0.426 0.467 0.461
APPLIED ARTIFICIAL INTELLIGENCE e2001178-251
Figure 9. Continued.
APPLIED ARTIFICIAL INTELLIGENCE e2001178-253
Disclosure statement
No potential conflict of interest was reported by the author(s).
ORCID
Chukiat Worasucheep http://orcid.org/0000-0001-7508-6150
References
Ampomah, E. K., Z. Qin, and G. Nyame. 2020. Evaluation of tree-based ensemble machine
learning models in predicting stock price direction of movement. Information 11 (6):332.
doi:10.3390/info11060332.
APPLIED ARTIFICIAL INTELLIGENCE e2001178-255
Bose, N. K., and P. Liang. 1996. Neural network fundamentals with graphs, algorithms, and
applications. In McGraw-Hill series in electrical and computer engineering, ed.
S. W. Director. 1-30. New York: McGraw-Hill.
Breiman, L. 1996. Bagging predictors. Machine Learning 24:123–40. doi:10.1007/BF00058655.
Breiman, L. 2001. Random forests. Machine Learning 45 (1):5–32. doi:10.1023/
A:1010933404324.
Cai, J., J. Luo, S. Wang, and S. Yang. 2018. Feature selection in machine learning: A new
perspective. Neurocomputing 300:70–79. doi:10.1016/j.neucom.2017.11.077.
Chen, A.-S., M.-T. Leung, and H. Daouk. 2003. Application of neural networks to an emerging
financial market: Forecasting and trading the Taiwan Stock Index. Computer and Operation
Research 30 (6):901–23. doi:10.1016/S0305-0548(02)00037-0.
Chen, T., and C. Guestrin. 2016. XGBoost: A Scalable Tree Boosting System. Proceedings of the
22nd ACM SIGKDD international conference on knowledge discovery and data mining
August 13 - 17, 2016 San Francisco, California, 785–94. doi:10.1145/2939672.2939785.
Chen, W., H. Zhang, M. K. Mehlawat, and L. Jia. 2021. Mean–variance portfolio optimization
using machine learning-based stock price prediction. Applied Soft Computing 100:106943.
doi:10.1016/j.asoc.2020.106943.
Cortes, C., and V. Vapnik. 1995. Support-vector networks. Machine Learning 20 (3):273–97.
doi:10.1007/BF00994018.
Deng, S., X. Huang, Z. Qin, Z. Fu, and T. Yang. 2021. A novel hybrid method for direction
forecasting and trading of Apple Futures. Applied Soft Computing 110:107734. doi:10.1016/j.
asoc.2021.107734.
Ding, G., and L. Qin. 2020. Study on the prediction of stock price based on the associated
network model of LSTM. International Journal of Machine Learning and Cybernetics
11:1307–17. doi:10.1007/s13042-019-01041-1.
Elshawi, R., M. Maher, and S. Sakr. 2019. Automated machine learning: State-of-the-art and
open challenges. arXiv Preprint arXiv:1906.02287, 1–26. http://arxiv.org/abs/1906.02287 .
Friedman, J. H. 2001. Greedy function approximation: A gradient boosting machine. The
Annals of Statistics 29 (5):1189–232. doi:10.1214/aos/1013203451.
Guang, L., and X. Wang. 2019. A new metric for individual stock trend prediction. Engineering
Applications of Artificial Intelligence 82:1–12. doi:10.1016/j.engappai.2019.03.019.
Hu, Y., K. Liu, X. Zhang, L. Su, E.-W.-T. Ngai, and M. Liu. 2015. Application of evolutionary
computation for rule discovery in stock algorithmic trading: A literature review. Applied Soft
Computing 36:534–51. doi:10.1016/j.asoc.2015.07.008.
Huang, W., Y. Nakamori, and S.-Y. Wang. 2005. Forecasting stock market movement direction
with support vector machine. Computers & Operations Research 32:2513–22. doi:10.1016/j.
cor.2004.03.016.
Hussain, A. J., A. Knowles, P. J. G. Lisboa, and W. El-Deredy. 2008. Financial time series
prediction using polynomial pipelined neural networks. Expert Systems and Applications
35 (3):1186–99. doi:10.1016/j.eswa.2007.08.038.
Kara, Y., M. A. Boyacioglu, and Ö. K. Baykan. 2011. Predicting direction of stock price index
movement using artificial neural networks and support vector machines: The sample of the
Istanbul stock exchange. Expert Systems with Applications 38 (5):5311–19. doi:10.1016/j.
eswa.2010.10.027.
Ke, G. et al, 2017. Lightgbm: A highly efficient gradient boosting decision tree. Advances in
Neural Information Processing Systems 30:3149–54.
Khairalla, M. A., X. Ning, N. T. AL-Jallad, and M. O. El-Faroug. 2018. Short-term forecasting
for energy consumption through stacking heterogeneous ensemble learning model. Energies
2018 (11):1–21. doi:10.3390/en11061605.
e2001178-256 C. WORASUCHEEP
Kim, H., and S.-T. Han. 2016. The enhanced classification for the stock index prediction.
Procedia Computer Science 91:284–86. doi:10.1016/j.procs.2016.07.077.
Li, J., K. Cheng, S. Wang, F. Morstatter, R. P. Trevino, J. Tang, and H. Liu. 2018. Feature
Selection: A data perspective. ACM Computing Survey (CSUR) 50 (6):94. doi:10.1145/
3136625.
Lorenzo, R. D. 2013. Basic Technical analysis of financial markets: A modern approach.
Heidelberg: Springer Milan.
Luo, L., S. You, Y. Xu, and H. Peng. 2017. Improving the integration of piece wise linear
representation and weighted support vector machine for stock trading signal prediction.
Applied Soft Computing 56:199–216. doi:10.1016/j.asoc.2017.03.007.
Nguyen, B. H., B. Xue, and M. Zhang. 2020. A survey on swarm intelligence approaches to
feature selection in data mining. Swarm and Evolutionary Computation 54:100633–49.
doi:10.1016/j.swevo.2020.100663.
Ni, L. P., Z. W. Ni, and Y. Z. Gao. 2011. Stock trend prediction based on fractal feature selection
and support vector machine. Expert Systems with Applications 38 (5):5569–76. doi:10.1016/j.
eswa.2010.10.079.
Nobre, J., and R. Neves. 2019. Combining principal component analysis, discrete wavelet
transform and XGBoost to trade in the financial markets. Expert Systems with Applications
125:181–94. doi:10.1016/j.eswa.2019.01.083.
Nti, I. K., A. F. Adekoya, and B. A. Weyori. 2020a. A comprehensive evaluation of ensemble
learning for stock-market prediction. Journal of Big Data 7 (20). doi:10.1186/s40537-020-
00299-5.
Nti, I. K., A. F. Adekoya, and B. A. Weyori. 2020b. Efficient stock-market prediction using
ensemble support vector machine. Open Computer Science 10 (1):153–63. doi:10.1515/
comp-2020-0199.
O’Connor, N., and M. G. Madden. 2006. A neural network approach to predicting stock
exchange movements using external factors. Knowledge Based Systems 19 (5):371–78.
doi:10.1016/j.knosys.2005.11.015.
Olson, D., and C. Mossman. 2003. Neural network forecasts of Canadian stock returns using
accounting ratios. International Journal of Forecasting 19 (3):453–65. doi:10.1016/S0169-
2070(02)00058-4.
Oram, E., P.-B. Dash, B. Naik, J. Nayak, S. Vimal, and S. K. Nataraj. 2021. Light gradient
boosting machine-based phishing webpage detection model using phisher website features
of mimic URLs. Pattern Recognition Letters 152:100–06. doi:10.1016/j.patrec.2021.09.018.
Polikar, R. 2006. Ensemble based systems in decision making. IEEE Circuits and Systems
Magazine 6:21–45. doi:10.1109/MCAS.2006.1688199.
Sharpe, W. F. 1994. The Sharpe Ratio. The Journal of Portfolio Management 21 (1):49–58.
doi:10.3905/jpm.1994.409501.
Slezak, J., L. Butler, and O. Akbilgic. 2021. The role of frailty index in predicting readmission
risk following total joint replacement using light gradient boosting machines. Informatics in
Medicine Unlocked 25:100657–62. doi:10.1016/j.imu.2021.100657.
Teixeira, L. A., and A. L. I. De Oliveira. 2010. A method for automatic stock trading combining
technical analysis and nearest neighbor classification. Expert Systems with Applications
37 (10):6885–90. doi:10.1016/j.eswa.2010.03.033.
Thakur, M., and D. Kumar. 2018. A hybrid financial trading support system using
multi-category classifiers and random forest. Applied Soft Computing 67:337–49.
doi:10.1016/j.asoc.2018.03.006.
Tsai, C.-F., Y.-C. Lin, D.-C. Yen, and Y.-M. Chen. 2011. Predicting stock returns by classifier
ensembles. Applied Soft Computing 11 (2):2452–59. doi:10.1016/j.asoc.2010.10.001.
APPLIED ARTIFICIAL INTELLIGENCE e2001178-257
Wang, J.-Z., -J.-J. Wang, Z.-G. Zhang, and S.-P. Guo. 2011. Forecasting stock indices with back
propagation neural network. Expert Systems with Applications 38 (11):14346–55.
doi:10.1016/j.eswa.2011.04.222.
Wolpert, D.-H., and W.-G. Macready. 1997. No free lunch theorems for search. IEEE
Transactions on Evolutionary Computation 1:67–82. doi:10.1109/4235.585893.
Xue, B., M. Zhang, W. Browne, and X. Yao. 2016. A survey on evolutionary computation
approaches to feature selection. IEEE Transactions on Evolutionary Computation 20:606–26.
doi:10.1109/TEVC.2015.2504420.
Yang, L., and A. Shami. 2020. On hyperparameter optimization of machine learning algo
rithms: Theory and practice. Neurocomputing 415:295–316. doi:10.1016/j.
neucom.2020.07.061.
Yun, K. K., S. W. Yoon, and D. Won. 2021. Prediction of stock price direction using a hybrid
GA-XGBoost algorithm with a three-stage feature engineering process. Expert Systems with
Applications 186:115716. doi:10.1016/j.eswa.2021.115716.
Zhang, J., S. Cui, Y. Xu, Q. Li, and T. Li. 2018. A novel data-driven stock price trend prediction
system. Expert Systems with Applications 97:60–69. doi:10.1016/j.eswa.2017.12.026.
Zhang, Y., W. Yu, Z. Li, S. Raza, and H. Cao. 2021. Detecting ethereum Ponzi schemes based on
improved LightGBM algorithm. IEEE Transactions on Computational Social Systems.
doi:10.1109/TCSS.2021.3088145.