Deep Learning Methods Application in Finance: A Review of State of Art
Deep Learning Methods Application in Finance: A Review of State of Art
Abstract
Artificial intelligence uses in financial markets or business units forms financial innovations. These innovations are the
key indicator for economic grow and intelligent finance system formation. Recants years scientist and most innovation
driving companies, such as Google, IBM, Microsoft and other, are focusing in deep learning methods. These methods have
achieved significant performances in diverse areas: image recognition, natural language processing, speech recognition,
video processing, etc. Therefore, it is necessary to understand the variety of deep learning methods and only then their
applicability in the financial field. Accordingly, in this paper firstly is presented differences in science community already
settled deep learning method’s architectures. Secondly, is shown a big picture of developing scientific articles of deep learning
uses in finance field, where the most used deep learning methods were identified. Finally, the conclusion, limitations and
future work have been shown.
Keywords
Artificial intelligence, Machine Learning, Deep Learning, Convolution Neural Network, Deep Belief Network,
Deep Boltzmann Machine, Deep neural network, Deep Q-Learning, Deep reinforcement learning,
The extreme learning machine, Generative adversarial network, Recurrent Neural Learning, Long short-term memory,
Gated Recurrent Unit, Finance, Financial innovations
60
2.5. Deep Q-Learning or Deep
Reinforcement Learning
Deep Q-Learning (DQL) or Deep reinforcement learn-
ing (DRL) concept is replaceable in scientific literature
(6). In DQL is always used reinforcement learning al-
gorithm, or in DRL is often used Q-learning function,
because of it is dealing with high-dimensional state
space inputs [30], [31]. A reinforcement learning (RL)
process involves an agent learning from interactions
Figure 1: The connection of AI, ML and DL with its environment in discrete time steps in order to
update its mapping between the perceived state and a
probability of selecting possible actions (policy) [32].
trained one-by-one until the hidden layer of the last In other words, RL is commonly used to solve an se-
RBM. During this phase, the parameter of each RBM quential decision making problem [30]. The RL prob-
can be obtained [23]. The back-propagation network lem is normally formalized using the Markov decision
(BP) is set in the last hidden layer of the DBN [25]. BP process (MDP) and includes a set of states S, set of ac-
is applied to fine tune the parameter using the output tions A, transition function T as action distributions,
labels of the sample data [23]. reward function R and discount factor 𝛾 [33]. The so-
lution to the MDP is a policy 𝜋 : S → A and the pol-
2.3. Deep Boltzmann Machine icy should maximize the expected discounted cumu-
lative reward [30]. Q-learning, as a typical reinforce-
Deep Boltzmann Machine (DBM) have only one undi- ment learning approach, mimics human behaviors to
rected network [24]. DBM as DBN is comprised of a take actions to the environment, in order to obtain the
Restricted Boltzmann Machines (RBM). The main dif- maximum long-term rewards [34]. The DQL process
ference is related to the interaction among layers of can be viewed as iteratively optimizing network pa-
RBMs [25]. For the computation of the conditional rameters process according to gradient direction of the
probability of the hidden units h1, both the lower visi- loss function at each stage [35]. Therefore, the inex-
ble layer v and the upper hidden layer h2 are incorpo- act approximate gradient estimation with a large vari-
rated, that makes DBM differentiated from DBN and ance can largely deteriorate the representation perfor-
also more robust to noisy observation [15]. There are mance of deep Q network by driving the network pa-
no direct connections between the units in the same rameter deviated from the optimal setting, causing the
layers. DBM parameters of all layers can be optimized large variability of DQL performance [35]. The advan-
jointly by following the approximate gradient of a vari- tages of deep Q-learning is good results and ease of use
ational lower-bound on the likelihood objective [26]. (code can be modified easy for different physical prob-
Different from the DBN, the DBM can incorporate lems) [36].
top-down feedback, which can better propagate un-
certainty and hence deal with ambiguous input more
robustly [27]. 2.6. The Extreme Learning Machine
The extreme learning machine (ELM) is a single-hidden
2.4. Deep Neural Network layer feedforward network, proposed by Huang in 2012.
In the traditional feed-forward ANN, the training of
Due to the novelty of the concept Deep neural network the network is iterative, while the process is trans-
(DNN) (Fig. 5)in the scientific literature can be identi- formed into an analytical equation in the ELM [37].
fied for all the algorithms analyzed in this paper. How- In ELM the weights between input and hidden layer
ever, in recent years the concept of DNN has become are assigned randomly following a normal distribution
known as Artificial Neural Network (ANN) with hid- and the weights between hidden and output layers are
den layers [9] [28]. DNN typically is feedforward net- learnt in single step by a pseudo-inverse technique.
work so it can be understood as the Multilayer Percep- During the training, the hidden layer is not learned
tron (MLP or MP). MLP consists of an input layer, sev- but the weight matrix of output layer is obtained by
eral hidden layers and one output layer ant it’s widely solving the optimization problem formulated by some
used for pattern classification, recognition and predic- learning criteria and regularizations [38], as showed
tion [29]. in the theory the output weights solved from regular-
61
Figure 2: Convolution neural network architecture.
2.7. Generative Adversarial Network same time, and finally the output is almost the same as
the real data [44].
The general idea of Generative adversarial network
(GAN) is that it aims to train a generator to recon-
2.8. Recurrent Neural Learning
struct high-resolution images for fooling a discrimi-
nator that is trained to distinguish generative images Recurrent Neural Learning (RNN) (Fig. 9) is different
from real ones [41] (Fig. 8). This idea involves two from the traditional feedforward neural networks, be-
competing neural network models: one of them takes cause have feedback connections, which can be be-
noise as input and produces some samples (genera- tween hidden units or from the output to the hidden
tor) and the other model (discriminator) accepts both units [44, 45]. This connections address the temporal
the data outputted by the generator and the real data, relationship of inputs by maintaining internal states
meanwhile, separates their sources [42]. The Discrim- that have memory . An RNN is able to process the
inator trains itself to discriminate real data and gen- sequential inputs by having a recurrent hidden state
erated data better while the Generator trains itself to whose activation at each step depends on that of the
fit the real data distribution so as to fool Discrimina- previous step [5, 46]. In other words, RNN not only
tor [43]. These two neural networks are trained at the processes the current element in the sequence, but also
62
Figure 5: Deep neural network architecture
63
Figure 9: Recurrent Neural Learning architecture
64
rent information should be treated as input in order to
generate the current state [51], whilst the forget gate
determine which information to be forgotten from the
memory state [52]. Finally, the output gate filters the
information that can be actually treated as significant
and produces the output [52]. The “gate” structure is
implemented using the sigmoid function, which de-
notes how much information can be allowed to pass.
For one hidden layer in LSTM, activation function is
used in forward propagation, and gradient is used in
backward propagation [38].
65
Table 1
Detailed topics from Finance perspective
4. Conclusions
learning machine, generative adversarial network, re-
current neural learning, long short-term memory, gated
recurrent unit; and it’s applicability in finance field.
This review reveals that financial article’s:
1. mainly focus for the forecasting task than clas-
sification;
2. starts using natural language processing tech-
niques, mostly sentiment analysis, for better re-
Figure 13: Use of deep learning methods in financial con- sults prediction;
text. 3. uses not ‘basic’ the deep learning methods, i.e.
they are often combined with several different
models or merged to voting classifier.
together [49, 59, 60] or make some model modifica-
tion’s [13, 50, 61, 62] for better prediction results. Some Furthermore, this analysis has shown the importance
authors [48, 63] analyze several different deep learning of balanced data set and normalization of the data, which
models results for the deeper future model develop- is submitted to deep learning networks.
ment, see Fig. 13. The main limitation of this work is representation
Most popular methods are CNN and LSTM. How- only a big picture of developing scientific articles in
ever, DBM and GAN method’s was not found any ad- Deep learning in Finance category. Therefore, in fu-
justment in finance field. ture research is needed to extend search keyword’s in
In some papers data is not normalized, i.e. cryp- electronic libraries, i.e. search by each method
tocurrency prices [51] or demand [18]. Therefore, pre-
dictive accuracy measurements, such as RMSE, MPE References
and others, can be comparable with different other au-
thors works or sometimes even in the same paper, i.e. [1] C. Huang, X. Wang, Financial innovation based
RMSE for Bitcoin is 2.75×103 or for Ripple 0.0499 [51]. on artificial intelligence technologies, in: Pro-
ceedings of the 2019 International Conference
on Artificial Intelligence and Computer Science,
2019, pp. 750–754.
66
[2] P. Yeoh, Artificial intelligence: accelerator or [15] Suk, Heung-Il, An introduction to neural net-
panacea for financial crime?, Journal of Finan- works and deep learning, in: Deep Learning for
cial Crime (2019). Medical Image Analysis, Elsevier, 2017, pp. 3–24.
[3] D. Mo, A survey on deep learning: one small step [16] MATHWORKS, Introducing Deep Learning with
toward ai, Dept. Computer Science, Univ. of New MATLAB, MATHWORKS, 2020.
Mexico, USA (2012). [17] Zheng, Chao, Wang, Shaorong, Liu, Yilu, Liuand,
[4] I. M. Cockburn, R. Henderson, S. Stern, The im- Chengxi, A novel rnn based load modelling
pact of artificial intelligence on innovation, Tech- method with measurement data in active distri-
nical Report, National bureau of economic re- bution system, Electric Power Systems Research
search, 2018. 166 (2019) 112–124.
[5] L. Mou, P. Ghamisi, X. X. Zhu, Deep recurrent [18] Kraus, Mathias, Feuerriegel, Stefan, Oztekin,
neural networks for hyperspectral image classi- Asil, Deep learning in business analytics and op-
fication, IEEE Transactions on Geoscience and erations research: Models, applications and man-
Remote Sensing 55 (2017) 3639–3655. agerial implications, European Journal of Oper-
[6] F. Beritelli, G. Capizzi, G. Lo Sciuto, C. Napoli, ational Research 281 (2020) 628–641.
M. Woźniak, A novel training method to preserve [19] Chollet, Francois, Deep Learning mit Python und
generalization of rbpnn classifiers applied to ecg Keras: Das Praxis-Handbuch vom Entwickler
signals diagnosis, Neural Networks 108 (2018) der Keras-Bibliothek, MITP-Verlags GmbH & Co.
331–338. KG, 2018.
[7] F. Beritelli, G. Capizzi, G. Lo Sciuto, C. Napoli, [20] keras, Guide to the Sequential model - Keras Doc-
F. Scaglione, Rainfall estimation based on the in- umentation, keras, 2020.
tensity of the received signal in a lte/4g mobile [21] G. Capizzi, G. L. Sciuto, P. Monforte, C. Napoli,
terminal by using a probabilistic neural network, Cascade feed forward neural network-based
IEEE Access 6 (2018) 30865–30873. model for air pollutants evaluation of single mon-
[8] J. Hearty, Advanced Machine Learning with itoring stations in urban areas, International
Python, Packt Publishing Ltd, 2016. Journal of Electronics and Telecommunications
[9] O. Lachiheb, M. S. Gouider, A hierarchical deep 61 (2015) 327–332.
neural network design for stock returns predic- [22] D. Saif, S. El-Gokhy, E. Sallam, Deep belief
tion, Procedia Computer Science 126 (2018) 264– networks-based framework for malware detec-
272. tion in android systems, Alexandria engineering
[10] D. Katayama, Y. Kino, K. Tsuda, A method of sen- journal 57 (2018) 4049–4057.
timent polarity identification in financial news [23] Balakrishnan, Nagaraj, Rajendran, Arunkumar,
using deep learning, Procedia Computer Science Pelusi, Danilo, Ponnusamy, Vijayakumar, Deep
159 (2019) 1287–1294. belief network enhanced intrusion detection sys-
[11] M.-Y. Day, C.-C. Lee, Deep learning for financial tem to prevent security breach in the internet of
sentiment analysis on finance news providers, in: things, Internet of Things (2019) 100112.
2016 IEEE/ACM International Conference on Ad- [24] J. Karhunen, T. Raiko, K. Cho, Unsupervised deep
vances in Social Networks Analysis and Mining learning: A short review, in: Advances in Inde-
(ASONAM), IEEE, 2016, pp. 1127–1134. pendent Component Analysis and Learning Ma-
[12] G. Capizzi, G. Lo Sciuto, C. Napoli, D. Polap, chines, Elsevier, 2015, pp. 125–142.
M. Wozniak, Small lung nodules detection based [25] Fan, Chaodong, Ding, Changkun, Zheng, Jin-
on fuzzy-logic and probabilistic neural network hua, Xiao, Leyi, Ai, Zhaoyang, Empirical mode
with bioinspired reinforcement learning, IEEE decomposition based multi-objective deep belief
Transactions on Fuzzy Systems 28 (2020). network for short-term power load forecasting,
[13] Wang, Wuyu, Li, Weizi, Zhang, Ning, Liu, Neurocomputing 388 (2020) 110–123.
Kecheng, Portfolio formation with preselec- [26] N. Srivastava, R. R. Salakhutdinov, Multimodal
tion using deep learning from long-term finan- learning with deep boltzmann machines, in: Ad-
cial data, Expert Systems with Applications 143 vances in neural information processing systems,
(2020) 11–42. 2012, pp. 2222–2230.
[14] Ongsulee, Pariwat, Artificial intelligence, ma- [27] S. Wang, M. He, Z. Gao, S. He, Q. Ji, Emotion
chine learning and deep learning, in: 2017 15th recognition from thermal infrared images using
International Conference on ICT and Knowledge deep boltzmann machine, Frontiers of Computer
Engineering (ICT&KE), IEEE, 2017, pp. 1–6. Science 8 (2014) 609–618.
67
[28] K. Akyol, Comparing of deep neural networks nary imbalanced learning, Neural Networks 119
and extreme learning machines based on grow- (2019) 235–248.
ing and pruning approach, Expert Systems with [41] R. Jiang, X. Li, A. Gao, L. Li, H. Meng, S. Yue,
Applications 140 (2020) 112875. L. Zhang, Learning spectral and spatial features
[29] C. Luo, D. Wu, D. Wu, A deep learning approach based on generative adversarial network for hy-
for credit scoring using credit default swaps, En- perspectral image super-resolution, in: IGARSS
gineering Applications of Artificial Intelligence 2019-2019 IEEE International Geoscience and Re-
65 (2017) 465–470. mote Sensing Symposium, IEEE, 2019, pp. 3161–
[30] W. Huang, Y. Wang, X. Yi, Deep q-learning to 3164.
preserve connectivity in multi-robot systems, in: [42] Y. Cui, W. Wang, Colorless video rendering sys-
Proceedings of the 9th International Conference tem via generative adversarial networks, in: 2019
on Signal Processing Systems, 2017, pp. 45–50. IEEE International Conference on Artificial In-
[31] Matta, Cardarilli, D. Nunzio, Fazzolari, Giardino, telligence and Computer Applications (ICAICA),
Nannarelli, Re, Spano, A reinforcement learning- IEEE, 2019, pp. 464–467.
based qam/psk symbol synchronizer, Ieee Access [43] Z. Zhai, J. Zhai, Identity-preserving conditional
7 (2019) 124147–124157. generative adversarial network, in: 2018 Inter-
[32] M. Ramicic, A. Bonarini, Attention-based expe- national Joint Conference on Neural Networks
rience replay in deep q-learning, in: Proceedings (IJCNN), IEEE, 2018, pp. 1–5.
of the 9th International Conference on Machine [44] I. Goodfellow, Y. Bengio, A. Courville, Y. Bengio,
Learning and Computing, 2017, pp. 476–481. Deep learning, volume 1, MIT press Cambridge,
[33] H. Shen, H. Hashimoto, A. Matsuda, Y. Taniguchi, 2016.
D. Terada, C. Guo, Automatic collision avoidance [45] F. Bonanno, G. Capizzi, C. Napoli, Some remarks
of multiple ships based on deep q-learning, Ap- on the application of rnn and prnn for the charge-
plied Ocean Research 86 (2019) 268–288. discharge simulation of advanced lithium-ions
[34] C. Qiu, F. R. Yu, F. Xu, H. Yao, C. Zhao, battery energy storage, in: International Sym-
Blockchain-based distributed software-defined posium on Power Electronics Power Electronics,
vehicular networks via deep q-learning, in: Pro- Electrical Drives, Automation and Motion, IEEE,
ceedings of the 8th ACM Symposium on Design 2012, pp. 941–945.
and Analysis of Intelligent Vehicular Networks [46] M. Miljanovic, Comparative analysis of recurrent
and Applications, 2018, pp. 8–14. and finite impulse response neural networks in
[35] W.-Y. Zhao, X.-Y. Guan, Y. Liu, X. Zhao, time series prediction, Indian Journal of Com-
J. Peng, Stochastic variance reduction for deep q- puter Science and Engineering 3 (2012) 180–191.
learning, arXiv preprint arXiv:1905.08152 (2019). [47] C. Yin, Y. Zhu, J. Fei, X. He, A deep learning
[36] I. Sajedian, H. Lee, J. Rho, Design of high trans- approach for intrusion detection using recurrent
mission color filters for solar cells directed by neural networks, Ieee Access 5 (2017) 21954–
deep q-learning, Solar Energy 195 (2020) 670– 21961.
676. [48] M. Kraus, S. Feuerriegel, Decision support from
[37] B. Çil, H. Ayyıldız, T. Tuncer, Discrimina- financial disclosures with deep neural networks
tion of 𝛽-thalassemia and iron deficiency ane- and transfer learning, Decision Support Systems
mia through extreme learning machine and regu- 104 (2017) 38–48.
larized extreme learning machine based decision [49] A. J. Balaji, D. H. Ram, B. B. Nair, Applicability
support system, Medical Hypotheses 138 (2020) of deep learning models for stock price forecast-
109611. ing an empirical study on bankex data, Procedia
[38] J. Chen, Y. Zeng, Y. Li, G.-B. Huang, Unsu- computer science 143 (2018) 947–953.
pervised feature selection based extreme learn- [50] Y. Chen, K. He, G. K. Tso, Forecasting crude oil
ing machine for clustering, Neurocomputing 386 prices: a deep learning based model, Procedia
(2020) 198–207. computer science 122 (2017) 300–307.
[39] H. Li, X. Yang, Y. Li, L.-Y. Hao, T.-L. Zhang, Evo- [51] S. Lahmiri, S. Bekiros, Cryptocurrency forecast-
lutionary extreme learning machine with sparse ing with deep learning chaotic neural networks,
cost matrix for imbalanced learning, ISA trans- Chaos, Solitons & Fractals 118 (2019) 35–40.
actions 100 (2020) 198–209. [52] C. Wang, D. Han, Q. Liu, S. Luo, A deep learning
[40] S. Shukla, B. S. Raghuwanshi, Online sequential approach for credit scoring of peer-to-peer lend-
class-specific extreme learning machine for bi- ing using attention mechanism lstm, IEEE Access
68
7 (2018) 2161–2168.
[53] Y. Santur, Sentiment analysis based on gated
recurrent unit, in: 2019 International Artifi-
cial Intelligence and Data Processing Symposium
(IDAP), IEEE, 2019, pp. 1–5.
[54] J. Kim, H. Kim, et al., Classification performance
using gated recurrent unit recurrent neural net-
work on energy disaggregation, in: 2016 in-
ternational conference on machine learning and
cybernetics (ICMLC), volume 1, IEEE, 2016, pp.
105–110.
[55] B. Zhu, W. Yang, H. Wang, Y. Yuan, A hybrid deep
learning model for consumer credit scoring, in:
2018 International Conference on Artificial Intel-
ligence and Big Data (ICAIBD), IEEE, 2018, pp.
205–208.
[56] A. Abdi, S. M. Shamsuddin, S. Hasan, J. Piran,
Deep learning-based sentiment classification of
evaluative text based on multi-feature fusion, In-
formation Processing & Management 56 (2019)
1245–1259.
[57] H. Maqsood, I. Mehmood, M. Maqsood, M. Yasir,
S. Afzal, F. Aadil, M. M. Selim, K. Muhammad, A
local and global event sentiment based efficient
stock exchange forecasting using deep learn-
ing, International Journal of Information Man-
agement 50 (2020) 432–451.
[58] F. Mai, S. Tian, C. Lee, L. Ma, Deep learning
models for bankruptcy prediction using textual
disclosures, European journal of operational re-
search 274 (2019) 743–758.
[59] L. Ni, Y. Li, X. Wang, J. Zhang, J. Yu, C. Qi, Fore-
casting of forex time series data based on deep
learning, Procedia computer science 147 (2019)
647–652.
[60] H. Yun, M. Lee, Y. S. Kang, J. Seok, Portfolio
management via two-stage deep learning with a
joint cost, Expert Systems with Applications 143
(2020) 113041.
[61] S. Huang, D. Wang, X. Wu, A. Tang, Dsanet:
Dual self-attention network for multivariate time
series forecasting, in: Proceedings of the 28th
ACM International Conference on Information
and Knowledge Management, 2019, pp. 2129–
2132.
[62] A. M. Aboussalah, C.-G. Lee, Continuous control
with stacked deep dynamic recurrent reinforce-
ment learning for portfolio optimization, Expert
Systems with Applications 140 (2020) 112891.
[63] C. Chen, P. Zhang, Y. Liu, J. Liu, Financial quan-
titative investment using convolutional neural
network and deep learning technology, Neuro-
computing 390 (2020) 384–390.
69