A Survey of Deep Learning Techniques Applied To Trading: Limit Order Book Modeling
A Survey of Deep Learning Techniques Applied To Trading: Limit Order Book Modeling
by Greg Harris
http://gregharris.info/a-survey-of-deep-learning-techniques-applied-t
o-trading/
Acronyms:
Dixon et al. (2016) use a deep neural network to predict the sign of
the price change over the next 5 minutes for 43 commodity and forex
futures.
Architecture – Their input layer has 9,896 neurons for input features
made up of lagged price differences and co-movements between
contracts. There are 5 learned fully-connected layers. The first of the
four hidden layers contains 1,000 neurons, and each subsequent
layer tapers by 100 neurons. The output layer has 135 neurons (3 for
each class {-1, 0, 1} times 43 contracts).
Zhu et al. (2016) make trade decisions using oscillation box theory
based on DBNs. Oscillation box theory says that a stock price will
oscillate within a certain range in a period of time. If the price moves
outside the range, then it enters into a new box. The authors try to
predict the boundaries of the box. Their trading strategy is to buy the
stock when it breaks through the top boundary or sell it when it
breaks through the bottom boundary.
Rönnqvist and Sarlin (2016) predict bank distress using news articles.
Specifically, they create a classifier to judge whether a given
sentence indicates distress or tranquility.
Architecture – They use two neural networks in this paper. The first
is for semantic pre-training to reduce dimensionality. For this, they
run a sliding window over text, taking a sequence of 5 words and
learning to predict the next word. They use a feed-forward topology
where a projection layer in the middle provides the semantic vectors
once the connection weights have been learned. They also include
the sentence ID as an input to the model, to provide context and
inform the prediction of the next word. They use binary Huffman
coding to map sentence IDs and word to activation patterns in the
input layer, which organizes the words roughly by frequency. They
say feed-forward topologies with fixed context sizes are more
efficient than recurrent neural networks for modeling text sequences.
The second neural network is for classification. Instead of a million
inputs (one for each word), they use 600 inputs from the learned
semantic model. The first layer has 600 nodes, the middle layer has
50 rectified linear hidden nodes, and the output layer has 2 nodes
(distress/tranquil).
Training – They train it with 243 distress events over 101 banks
observed during the financial crisis of 2007-2009. They use 716k
sentences mentioning the banks, taken from 6.6m Reuters news
articles published during and after the crisis.
Training – They initialize the weights with Gaussian noise, and then
update through back-propagation. They use an English ad-hoc news
announcement dataset (8,359 headlines) for the German market
covering 2004 to 2011.
Results – They find that structured events are better features than
words for stock market prediction. Their approach outperforms
baseline methods by 6%. They make predictions for the S&P 500
index and 15 individual stocks, and a table appears to show that they
can predict the S&P 500 with 65% accuracy.
Volatility Prediction
Xiong et al. (2015) predict the daily volatility of the S&P 500, as
estimated from open, high, low, close prices.
Portfolio Optimization
Training – They use weekly return data for the component stocks of
IBB from 2012 to 2016. They auto-encode all stocks in the index and
evaluate the difference between each stock and its auto-encoded
version. They keep the 10 most “communal” stocks that are most
similar to the auto-encoded version. They also keep a varying number
of other stocks, where the number is chosen with cross-validation.
Related Links
References
Ding, X., Zhang, Y., Liu, T., & Duan, J. (2015, June). Deep learning for
event-driven stock prediction. In Proceedings of the Twenty-Fourth
International Joint Conference on Artificial Intelligence (ICJAI) (pp.
2327-2333). abstract
Heaton, J. B., Polson, N. G., & Witte, J. H. (2016). Deep Portfolio Theory.
arXiv preprint arXiv:1605.07230. abstract
Sharang, A., & Rao, C. (2015). Using machine learning for medium
frequency derivative portfolio trading. arXiv preprint
arXiv:1512.06228. abstract
Sirignano, J. A. (2016). Deep Learning for Limit Order Books. arXiv
preprint arXiv:1601.01987. abstract
Xiong, R., Nicholas, E. P., & Shen, Y. (2015). Deep Learning Stock
Volatilities with Google Domestic Trends. arXiv preprint
arXiv:1512.04916. abstract
Zhu, C., Yin, J., & Li, Q. (2014). A stock decision support system based
on DBNs. Journal of Computational Information Systems, 10(2),
883-893. abstract·