Reinforcement Learning For Quantitative Trading - 2021
Reinforcement Learning For Quantitative Trading - 2021
complex sequential decision making problems. RL’s impact is pervasive, recently demonstrating its ability to conquer many challenging
QT tasks. It is a flourishing research direction to explore RL techniques’ potential on QT tasks. This paper aims at providing a
comprehensive survey of research efforts on RL-based methods for QT tasks. More concretely, we devise a taxonomy of RL-based
QT models, along with a comprehensive summary of the state of the art. Finally, we discuss current challenges and propose future
research directions in this exciting field.
Additional Key Words and Phrases: reinforcement learning, quantitative finance, survey
1 INTRODUCTION
Quantitative trading has been a lasting research area at the intersection of finance and computer science for many
decades. In general, QT research can be divided into two directions. In the finance community, designing theories
and models to understand and explain the financial market is the main focus. The famous capital asset pricing model
(CAPM) [103], Markowitz portfolio theory [82] and Fama & French factor model [35] are a few representative examples.
On the other hand, computer scientists apply data-driven ML techniques to analyze financial data [29, 94]. Recently,
deep learning becomes an appealing approach owing to not only its stellar performance but also to the attractive
property of learning meaningful representations from scratch.
RL is an emerging subfield of ML, which provides a mathematical formulation of learning-based control. With the
usage of RL, we can train agents with near-optimal behaviour policy through optimizing task-specific reward functions
[112]. In the last decade, we have witnessed many significant artificial intelligence (AI) milestones achieved by RL
approaches in domains such as Go [107], video games [83] and robotics [69]. RL-based methods also have achieved
Authors’ addresses: Shuo Sun, Nanyang Technological University, Singapore, [email protected]; Rundong Wang, Nanyang Technological University,
Singapore, [email protected]; Bo An, Nanyang Technological University, Singapore, [email protected].
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not
made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components
of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to
redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].
© 2021 Association for Computing Machinery.
Manuscript submitted to ACM
state-of-the-art performance on many QT tasks such as algorithmic trading (AT) [80], portfolio management [123], order
execution [37] and market making [109]. It is a promising research direction to address QT tasks with RL techniques.
Macro-level Micro-level
Both sides
One Multiple Single (buy/sell)
simultaneously
Considering the increasing popularity and potential of RL-based QT applications, a comprehensive survey will be of
high scientific and practical values. More than 100 high quality papers are shortlisted and categorized in this survey.
Furthermore, we analyze current situation of this area and point out future research directions.
Notation Description
ℎ length of a holding period
p𝑖 the time series vector of asset 𝑖’s price
𝑝𝑖,𝑡 the price of asset 𝑖 at time 𝑡
′
𝑝𝑖,𝑡 the price of asset 𝑖 after a holding period ℎ from time 𝑡
𝑝𝑡 the price of a single asset at time 𝑡
𝑠𝑡 position of an asset at time 𝑡
𝑢𝑡𝑖 trading volume of asset 𝑖 at time 𝑡
n the time series vector of net value
𝑛𝑡 net value at time 𝑡
𝑛𝑡′ net value after a holding period ℎ from time 𝑡
𝑤𝑡𝑖 portfolio weight of asset 𝑖 at time 𝑡
wt portfolio vector at time 𝑡
wt′ portfolio vector after a holding period ℎ from time 𝑡
𝑣𝑡 portfolio value at time 𝑡
𝑣𝑡′ portfolio value after a holding period ℎ from time 𝑡
𝑓𝑡𝑖 transaction fee for asset 𝑖 at time 𝑡
𝜉 transaction fee rate
𝑞 the quantity of a limit order
𝑄 total quantity required to be executed
r the time series vector of return rate
𝑟𝑡 return rate at time 𝑡
2.1 Overview
The financial market, an ecosystem involving transactions between businesses and investors, observed a market
capitalization exceeding $80 trillion globally as of the year 2019.1 For many countries, the financial industry has
become a paramount pillar, which spawns the birth of many financial centres. The International Monetary Fund (IMF)
categorizes financial centres as follows: international financial centres, such as New York, London and Tokyo; regional
financial centres, such as Shanghai, Shenzhen and Sydney; offshore financial centres, such as Hong Kong, Singapore
and Dublin. At the core of financial centres, trading exchanges,where trading activities involving trillions of dollars
take place everyday, are formed. Trading exchanges can be divided as stock exchanges such as NYSE, Nasdaq and
Euronext, derivatives exchanges such as CME and cryptocurrency exchanges such as Coinbase and Huobi. Participants
in the financial market can be generally categorized as financial intermediaries (e.g., banks and brokers), issuers (e.g.,
companies and governments), institutional investors (e.g., investment managers and hedge funds) and individual
investors. With the development of electronic trading platform, quantitative trading, which has been demonstrated quite
1 https://data.worldbank.org/indicator/CM.MKT.LCAP.CD/
profitable by many leading trading companies (e.g., Renaissance2 , Two Sigma3 , Cithadel4 , D.E. Shaw5 ), is becoming a
dominating trading style in the global financial markets. In 2020, quantitative trading accounts for over 70% and 40%
trading volume in developed market (e.g., US and Europe) and emerging market (e.g., China and India) respectively. 6
We introduce some basic QT concepts as follows:
• Financial Asset. A financial asset refers to a liquid asset, which can be converted into cash immediately during
trading time. Classic financial assets include stocks, futures, bonds, foreign exchanges and cryptocurrencies.
• Holding Period. Holding period ℎ refers to the time period where traders just hold the financial assets without
any buying or selling actions.
• Asset Price. The price of a financial asset 𝑖 is defined as a time series pi = {pi,1, pi,2 , pi,3 , ..., pi,t }, where 𝑝𝑖,𝑡
′ is the price of asset 𝑖 after a holding period ℎ from time 𝑡. 𝑝 is used to
denotes the price of asset 𝑖 at time 𝑡. 𝑝𝑖,𝑡 𝑡
denote the price at time 𝑡 when there is only one financial asset.
• OHLC. OHLC is the abbreviation of open price, high price, low price and close price. The candle stick, which is
consisted of OHLC, is widely used to analyze the financial market.
• Volume. Volume is the amount of a financial asset that changes hands. 𝑢𝑡𝑖 is the trading volume of asset 𝑖 at
time 𝑡.
• Technical Indicator. A technical indicator indicates a feature calculated by a formulaic combination of OHLC
and volume. Technical indicators are usually designed by finance experts to uncover the underlying pattern of
the financial market.
• Return Rate. Return rate is the percentage change of capital, where 𝑟𝑡 = (𝑝𝑡 +1 − 𝑝𝑡 )/𝑝𝑡 denotes the return
rate at time 𝑡. The time series of return rate is denoted as r = (𝑟 1, 𝑟 2, ..., 𝑟𝑡 ).
• Transaction Fee. Transaction fee is the expenses incurred during trading financial assets: 𝑓𝑡𝑖 = 𝑝𝑖,𝑡 × 𝑢𝑡𝑖 × 𝜉,
where 𝜉 is the transaction fee rate.
• Liquidity. Liquidity refers to the efficiency with which a financial asset can be converted into cash without
having an evident impact on its market price. Cash itself is the asset with the most liquidity.
discrete time steps. At the beginning of a trading period, traders are allocated some cash and set net value as 1. Then, at
each time step 𝑡, traders have the options to buy, hold or sell some amount of shares for changing positions. Net value
and position is used to represent traders’ status at each time step. The objective of AT is to maximize the final net value
at the end of the trading period. Based on trading styles, algorithmic trading is generally divide 5 categories: position
trading, swing trading, day trading, scalp trading and high-frequency trading. Specifically, position trading involves
holding the financial asset for a long period of time, which is unconcerned with short-term market fluctuations and only
focuses on the overarching market trend. Swing trading is a medium-term style that holds financial assets for several
days or weeks. The goal of swing trading is to spot a trend and then capitalise on dips and peaks that provide entry
points. Day trading tries to capture the fleeting intraday pattern in the financial market and all positions will be closed
at the end of the day to avoid overnight risk. Scalping trading aims at discovering micro-level trading opportunities
and makes profit by holding financial assets for only a few minutes. High-frequency trading is a type of trading style
characterized by high speeds, high turnover rates, and high order-to-trade. A summary of different trading styles is
illustrated in Table. 2.
Traditional AT methods discover trading signals based on technical indicators or mathematical models. Buy and
Hold (BAH) strategy, which invests all capital at the beginning and holds until the end of the trading period, is proposed
to reflect the average market condition. Momentum strategies, which assumes the trend of financial assets in the past
has the tendency to continue in the future, are another well-known AT strategies. Buying-Winner-Selling-Loser [57],
Times Series Momentum [87] and Cross Sectional Momentum [18] are three classic momentum strategies. In contrast,
mean reversion strategies such as Bollinger bands [12] assume the price of financial assets will finally revert to the
long-term mean. Although traditional methods somehow capture the underlying patterns of the financial market, these
simple rule-based methods exhibit limited generalization ability among different market conditions. We introduce some
basic AT concepts as follows:
• Position. Position 𝑠𝑡 is the amount of a financial asset owned by traders at time 𝑡. It represents a long (short)
position when 𝑠𝑡 is positive (negative).
• Long Position. Long position makes positive profit when the price of the asset increases. For long trading
actions, which buy a financial asset 𝑖 at time 𝑡 first and then sell it at 𝑡 + 1, the profit is 𝑢𝑡𝑖 (𝑝𝑖,𝑡 +1 − 𝑝𝑖,𝑡 ), where 𝑢𝑡𝑖
is the buying volume of asset 𝑖 at time 𝑡.
• Short Position. Short position makes positive profit when the price of the asset decreases. For short trading
actions, which buys a financial asset at time 𝑡 first and then sell it at 𝑡 + 1, the profit is 𝑢𝑡𝑖 (𝑝𝑖,𝑡 − 𝑝𝑖,𝑡 +1 ).
• Net Value. Net value represents a fund’s per share value. It is defined as a time series n = {𝑛 1, 𝑛 2, ..., 𝑛𝑡 }, where
𝑛𝑡 denotes the net value at time 𝑡. The initial net value is always set to 1.
market, the assets’ prices would change during the holding period. At the end of the holding period, the agent will get a
new portfolio value 𝑣𝑡′ and decide a new portfolio weight wt+1 of the next holding period. During the trading period,
the agent buys or sells some shares of assets to achieve the new portfolio weights. The lengths of the holding period
and trading period are based on specific settings and can change over time. In some previous works, the trading period
is set to 0, which means the change of portfolio weight is achieved immediately for convenience. The objective is to
maximize the final portfolio value given a long time horizon.
PM has been a fundamental problem for both finance and ML community for decades. Existing approaches can
be grouped into four major categories, which are benchmarks such as Constant Rebalanced Portfolio (CRP) and
Uniform Constant Rebalanced Portfolio (UCRP) [23], Follow-the-Winner approaches such as Exponential Gradient (EG)
[48] and Winner [41], Follow-the-Loser approaches such as Robust Mean Reversion (RMR) [53], Passive Aggressive
Online Learning (PAMR) [72] and Anti-Correlation [13], Pattern-Matching-based approaches such as correlation-driven
nonparametric learning (CORN) [71] and 𝐵𝐾 [47], and Meta-Learning algorithms such as Online Newton Step (ONS).
The readers can check this survey [70] for more details. We introduce some basic PM concepts as follows:
• Portfolio. A portfolio can be represented as:
𝑀
∑︁
𝑇
wt = [𝑤𝑡0, 𝑤𝑡1, ..., 𝑤𝑡𝑀 ] ∈ 𝑅 𝑀+1 𝑎𝑛𝑑 𝑤𝑡𝑖 = 1
𝑖=0
where M+1 is the number of portfolio’s constituents, including one risk-free asset, i.e., cash, and M risky assets.
𝑤𝑡𝑖 represents the ratio of the total portfolio value (money) invested at the beginning of the holding period 𝑡 on
asset i. Specifically, 𝑤𝑡0 represents the cash in hand.
• Portfolio Value. We define 𝑣𝑡 and 𝑣𝑡′ as portfolio value at the beginning and end of the holding period. So we
can get the change of portfolio value during the holding period and the change of portfolio weights:
′
𝑤𝑡𝑖 𝑝𝑖,𝑡
𝑀 𝑤𝑖 𝑝 ′
∑︁ 𝑝𝑖,𝑡
𝑣𝑡′ = 𝑣𝑡 𝑤𝑡′ =
𝑡 𝑖,𝑡
′
𝑓 𝑜𝑟 𝑖 ∈ [0, 𝑀]
𝑝𝑖,𝑡 Í𝑀 𝑤𝑡𝑖 𝑝𝑖,𝑡
𝑖=0
𝑖=0 𝑝𝑖,𝑡
Traditional OE solutions are usually designed based on some stringent assumptions of the market and then derive some
model-based methods with stochastic control theory. For instance, Time Weighted Average Price (TWAP) evenly splits
the whole order and execute at each time step with the assumption that the market price follows the Brownian motion [8].
The Almgren-Chriss model [2] incorporates temporary and permanent price impact
functions also with the Brownian motion assumption. Volume Weighted Average
Price (VWAP) distributes orders in proportion to the (empirically estimated) market
transaction volume. The goal of VWAP is to track the market average execution price
[61]. However, traditional solutions are not effective in the real market because of
the inconsistency between the assumptions and reality.
Formally, OE is to trade fixed amount of shares within a predetermined time
horizon (e.g., one hour or one day). At each time step 𝑡, traders can propose to trade a
Fig. 3. Limit Order Book
quantity of 𝑞𝑡 ≥ 0 shares at current market price 𝑝𝑡 , The matching system will then
return the execution results at time 𝑡 + 1. Taking the sell side as an example, assuming
a total of Q shares required to be executed during the whole time horizon, the OE task can be formulated as:
𝑇
∑︁ 𝑇
∑︁
arg max (𝑞𝑡 · 𝑝𝑡 ), s.t. 𝑞𝑡 = 𝑄
𝑞 1 ,𝑞 2 ,...,𝑞𝑇
𝑡 =1 𝑡 =1
OE not only completes the liquidation requirement but also the maximize/minimize average execution price for the
sell/buy side execution respectively. We introduce basic OE concepts as follows:
• Market Order. A market order refers submitting an order to buy or sell a financial asset at the current market
price, which expresses the desire to trade at the best available price immediately.
• Limit Order. A limit order is an order placed to buy or sell a number of shares at a specified price during a
specified time frame. It can be modeled as a tuple 𝑝𝑡𝑎𝑟𝑔𝑒𝑡 ± 𝑞𝑡𝑎𝑟𝑔𝑒𝑡 , where 𝑝𝑡𝑎𝑟𝑔𝑒𝑡 represents the submitted target
price, 𝑞𝑡𝑎𝑟𝑔𝑒𝑡 represents the submitted target quantity, and ± represents trading direction (buy/sell)
• Limit Order Book. A limit order book (LOB) is a list containing all the information about the current limit
orders in the market. An example of LOB is shown in Figure 3.
𝑞𝑡
• Average Execution Price Average execution price (AEP) is defined as 𝑝¯ =
Í𝑇
𝑡 =1 𝑄 · 𝑝𝑡 .
• Order Matching System. The electronic system that matches buy and sell orders for a financial market is
called the order matching system. The matching system is the core of all electronic exchanges, which decides the
execution results of orders in the market. The most common matching mechanism is first-in-first-out, which
means limit orders at the same price will be executed in the order in which the orders were submitted.
Traditional finance methods consider market making as a stochastic optimal control problem [16]. Agent-based method
[43] and RL [109] have also been applied to market making.
𝑃𝑅 = (𝑛𝑡 +ℎ − 𝑛𝑡 )/𝑛𝑡
• Win rate (WR). WR evaluates the proportion of trading days with positive profit among all trading days.
𝑉 𝑂𝐿 = 𝜎 [r]
• Maximum drawdown (MDD). MDD [81] measures the largest decline from the peak in the whole trading
period to show the worst case. The formal definition is:
𝑛𝑡 − 𝑛𝜏
𝑀𝐷𝐷 = max [ max ]
𝜏 ∈(0,𝑡 ) 𝑡 ∈(0,𝜏) 𝑛𝑡
• Downside deviation (DD). DD refers to the standard deviation of trade returns that are negative.
• Gain-loss ratio (GLR). GLR is a downside risk measure. It represents the relative relationship of trades with a
positive return and trades with a negative return. The formula is:
E[r|r > 0]
𝐺𝐿𝑅 =
E[−r|r < 0]
2.6.3 Risk-adjusted Metrics.
• Sharpe ratio (SR). SR [104] is a risk-adjusted profit measure, which refers to the return per unit of deviation:
E[r]
𝑆𝑅 =
𝜎 [r]
• Sortino ratio (SoR). SoR is a variant of risk-adjusted profit measure, which applies DD as risk measure:
E[r]
𝑆𝑜𝑅 =
𝐷𝐷
• Calmar ratio (CR). CR is another variant of risk-adjusted profit measure, which applies MDD as risk measure:
E[r]
𝐶𝑅 =
𝑀𝐷𝐷
and the environment is everything else except the agent. At each time step, the agent obtains some observations of the
environment, which is called state. Later on, the agent takes an action based on the current state. The environment will
then return a reward and a new state to the agent. Formally, an RL problem is typically formulated as a Markov decision
process (MDP) in the form of a tuple M = (𝑆, 𝐴, 𝑅, 𝑃, 𝛾), where 𝑆 is a set of states 𝑠 ∈ 𝑆, 𝐴 is a set of actions 𝑎 ∈ 𝐴, 𝑅 is
the reward function, 𝑃 is the transition probability, and 𝛾 is the discount factor. The goal of an RL agent is to find a
policy 𝜋 (𝑎 | 𝑠) that takes action 𝑎 ∈ 𝐴 in state 𝑠 ∈ 𝑆 in order to maximize the expected discounted cumulative reward:
𝜏
∑︁
max E[𝑅(𝜏)], 𝑤ℎ𝑒𝑟𝑒 𝑅(𝜏) = 𝛾 𝑡 𝑟 (𝑎𝑡 , 𝑠𝑡 ) 𝑎𝑛𝑑 0 ≤ 𝛾 ≤ 1
𝑡 =0
Sutton and Barto [112] summarise RL’s main components as: (i) policy, which refers to the probability of taking action
𝑎 when the agent is in state 𝑠. From policy perspective, RL algorithms are categorized into on-policy and off-policy
methods. The goal of on-policy RL methods is to evaluate or improve the policy, which they are now using to make
decisions. As for off-policy RL methods, they aim at improving or evaluating the policy that is different from the one
used to generate data. (ii) reward: after taking selected actions, the environment sends back a numerical signal reward
to inform the agent how good or bad are the actions selected. (iii) value function, which means the expected return if
the agent starts in that state 𝑠 or state-action pair (𝑠, 𝑎), and then acts according to a particular policy 𝜋 consistently.
Value function tells how good or bad your current position is in the long run. (iv) model, which is an inference about
the behaviour of the environment in different states.
Plenty of algorithms have been proposed to solve RL problems. Tabular methods and approximation methods are
two mainstream directions. For tabular algorithms, a table is used to represent the value function for every action
and state pair. The exact optimal policy can be found through checking the table. Due to the curse of dimensionality,
tabular methods only work well when the action and state space is small. Dynamic programming (DP), Monto Carlo
(MC) and temporal difference (TD) are a few widely studied tabular methods. Under perfect model of environment
assumption, DP uses a value function to search for good policies. Policy iteration and value iteration are two classic DP
algorithms. MC methods try to learn good policies through sample sequences of states, actions, and reward from the
environment. For MC methods, the assumption of perfect environment understanding is not required. TD methods are
a combination of DP and MC methods. While they do not need a model from the environment, they can bootstrap,
which is the ability to update estimates based on other estimates. From this family, Q-learning [125] and SARSA [96]
are popular algorithms, which belong to off-policy and on-policy methods respectively.
On the other hand, approximation methods try to find a great approximate function with limited computation.
Learning to generalize from previous experiences (already seen states) to unseen states is a reasonable direction.
Policy gradient methods are popular approximate solutions. REINFORCE [126] and actor-critic [65] are two important
examples. With the popularity of deep learning, RL researchers use neural networks as function approximator. DRL is
the combination of DL and RL, which lead to great success in many domains [83, 118]. Popular DRL algorithms for QT
community include deep Q-network (DQN) [83], deterministic policy gradient (DPG) [108], deep deterministic policy
gradient (DDPG) [75], proximal policy optimization (PPO) [100]. More details for RL can be found in [112]. Recurrent
reinforcement learning (RRL) is another widely used RL approach for QT. "Recurrent" means the previous output is fed
into the model as part of the input here. RRL achieves more stable performance when exposed to noisy data such as
financial data.
methods can tackle this obstacle through learning an end-to-end agent, which maps market information into trading
actions directly. In the next section, we will discuss notable RL methods for QT tasks and why they are superior to
traditional methods.
Foreign Commodity
Others Others
Exchange 4%
13% 14%
8%
Cryptocurrency
12%
Europe
Stock
15% The US
41%
42% Day Level
46%
Minute Level
26%
Artificial Data
14%
China
30% Stock Index Hour Level
21% 14%
management on the Korea stock market compared to supervised learning baselines. In [56], the authors firstly design
some local traders based on dynamic programming and heuristic rules. Later on, they apply Q-learning to learn a meta
policy of these local traders on Korea stock markets. de Oliveira et al. [25] implemented a SARSA-based RL method and
tested it on 10 stocks in the Brazil market.
DQN is used to enhance trading systems by considering trading frequencies, market confusion and transfer learning
[58]. The trading frequency is determined in 3 ways: (1) a heuristic function related to Q-value, (2) an action-dependent
NN regressor, and (3) an action-independent NN regressor. Another heuristic function is applied to add a filter as
the agent’s certainty on market condition. Moreover, the authors train the agent on selected component stocks and
apply the pre-train weights as the starting point for different stock indexes. Experiments on 4 different stock indexes
demonstrate the effectiveness of the proposed framework.
Other methods. iRDPG [80] is an adaptive DPG-based framework. Due to the noisy nature of financial data, the
authors formulate algorithmic trading as a Partially Observable Markov Decision Process (POMDP). GRU layers are
introduced in iRDPG to learn recurrent market embedding. In addition, the authors apply behavior cloning with expert
trading actions to guide iRDPG and achieve great performance on two Chinese stock index futures. There are also
some works focusing on evaluating the performance of different RL algorithms on their own data. Zhang et al. [144]
evaluated DQN, PG and A2C on the 50 most liquid futures contracts. Yuan et al. [138] tested PPO, DQN and SAC on
three selected stocks. Based on these two works, DQN achieves the best overall performance among different financial
assets.
Summary. Although existing works demonstrate the potential of RL for quantitative trading, there is seemingly no
consensus on a general ranking of different RL algorithms (notably, we acknowledge that no free lunch theorem exists).
The summary of algorithmic trading publications is in Table 4. In addition, most existing RL-based works only focus on
general AT, which tries to make profit through trading one asset. In finance, extensive trading strategies have been
designed based on trading frequency (e.g., high-frequency trading) and asset types (e.g., stock and cryptocurrency).
describe the interrelationships among stocks. Later on, the output of CAAN (winning score of each stock) is feed into
a heuristic portfolio generator to construct the final portfolio. Policy gradient is used to optimize the Sharpe Ratio.
Experiments on both U.S. and Chinese stock market show that Alphastock achieves robust performance over different
market states. 𝐸𝐼 3 [105] is another RRL-based method, which tries to build profitable cryptocurrency portfolios by
extracting multi-scale patterns in the financial market. Inspired by the success of Inception networks [113], the authors
design a multi-scale temporal feature aggregation convolution framework with two CNN branches to extract short-term
and mid-term market embedding and a max pooling branch to extract the highest price information. To bridge the gap
between the traditional Markowitz portfolio and RL-based methods, Benhamou et al. [6] applied PG with a delayed
reward function and showed better performance than the classic Markowitz efficient frontier.
Zhang et al. [143] proposed a cost-sensitive PM framework based on direct policy gradient. To learn more robust
market representation, a novel two-stream portfolio policy network is designed to extract both price series pattern
and the relationship between different financial assets. In addition, the authors design a new cost-sensitive reward
function to take the trading cost constrain into consideration with theoretically near-optimal guarantee. Finally, the
effectiveness of the cost-sensitive framework is demonstrated on real-world cryptocurrency datasets. Xu et al. [132]
proposed a novel relation-aware transformer (RAT) under the classic RRL paradigm. RAT is structurally innovated
to capture both sequential patterns and the inner corrections between financial assets. Specifically, RAT follows an
encoder-decoder structure, where the encoder is for sequential feature extraction and the decoder is for decision
making. Experiments on 2 cryptocurrency and 1 stock datasets not only show RAT’s superior performance over existing
baselines but also demonstrate that RAT can effectively learn better representation and benefit from leverage operation.
Bisi et al. [10] derived a PG theorem with a novel objective function, which exploited the mean-volatility relationship.
Manuscript submitted to ACM
16 Sun et al.
The new objective could be used in actor-only algorithms such as TRPO with monotonic improvement guarantees.
Wang et al. [124] proposed DeepTrader, a PG-based DRL method, to tackle the risk-return balancing problem in PM.
The model simultaneously uses negative maximum drawdown and price rising rate as reward functions to balance
between profit and risk. The authors propose an asset scoring unit with graph convolution layer to capture temporal
and spatial interrelations among stocks. Moreover, a market scoring unit is designed to evaluation the market condition.
DeepTrader achieves great performance across three different markets.
Actor-critic methods. Jiang et al. [60] proposed a DPG-based RL framework for portfolio management. The
framework consists of 3 novel components: 1) the Ensemble of Identical Independent Evaluators (EIIE) topology; 2) a
Portfolio Vector Memory (PVM); 3) an Online Stochastic Batch Learning (OSBL) scheme. Specifically, the idea of EIIE is
that the embedding concatenation of output from different NN layers can learn better market representation effectively.
In order to take transaction costs into consideration, PVM uses the output portfolio at the last time step as part of the
input of current time step. The OSBL training scheme makes sure that all data points in the same batch are trained in the
original time order. To demonstrate the effectiveness of proposed components, extensive experiments using different
NN architectures are conducted on cryptocurrency data. Later on, more comprehensive experiments are conducted in
an extended version [59]. To model the data heterogeneity and environment uncertainty in PM, Ye et al. [136] proposed
a State-Augmented RL (SARL) framework based on DPG. SARL learns the price movement prediction with financial
news as additional states. Extensive experiments on both cryptocurrency and U.S. stock market validation that SARL
outperforms previous approaches in terms of return rate and risk-adjusted criteria. Another popular actor-critic RL
method for portfolio management is DDPG. Xiong et al. [131] constructed a highly profitable portfolio with DDPG
on the Chinese stock market. PROFIT [99] is another DDPG-based approach that makes time-aware decisions on PM
with text data. The authors make use of a custom policy network that hierarchically and attentively learns time-aware
representations of news and tweets for PM, which is generalizable among various actor-critic RL methods. PROFIT
shows promising performance on both China and U.S. stock markets.
Other methods. Neuneier [88] made an attempt to formalize portfolio management as an MDP and trained an RL
agent with Q-learning. Experiments on German stock market demonstrate its superior performance over heuristic
benchmark policy. Later on, a shared value-function for different assets and model-free policy-iteration are applied to
further improve the performance of Q-learning in [89]. There are a few model-based RL methods that attempt to learn
some models of the financial market for portfolio management. [137] proposed the first model-based RL framework for
portfolio management, which supports both off-policy and on-policy settings. The authors design an Infused Prediction
Module (IPM) to predict future price, a Data Augmentation Module (DAM) with recurrent adversarial networks to
mitigate the data deficiency issue, and a Behavior Cloning Module (BCM) to reduce the portfolio volatility. Wang
et al. [123] focused on a more realistic PM setting where portfolio managers assign a new portfolio periodically for
a long-term profit, while traders care about the best execution at the favorable price to minimize the trading cost.
Motivated by this hierarchy scenario, a hierarchical RL system (HRPM) is proposed. The high level policy was trained
by REINFORCE with an entropy bonus term to encourage portfolio diversification. The low level framework utilizes
the branching dueling Q-Network to train agents with 2 dimensions (price and quantity) action space. Extensive
experiments are conducted on both US and China stock market to demonstrate the effectiveness of HRPM.
Portfolio management is also formulated as a multi-agent RL problem. MAPS [66] is a cooperative multi-agent RL
system in which each agent is an independent "investor" creating its own portfolio. The authors design a novel loss
function to guide each agent to act as diversely as possible while maximizing its long-term profit. MAPS outperforms
most of baselines with 12 years of U.S. stock market data. In addition, the authors find that adding more agents to MAPS
Manuscript submitted to ACM
Reinforcement Learning for Quantitative Trading 17
can lead to a more diversified portfolio with higher Sharpe Ratio. MSPM [55] is a multi-agent RL framework with a
modularized and scalable architecture for PM. MSPM consists of the Evolving Agent Module (EAM) to learn market
embedding with heterogeneous input and the Strategic Agent Module (SAM) to produce profitable portfolios based on
the output of EAM.
Some works compare the profitability of portfolios constructed by different RL algorithms on their own data. Liang
et al. [74] compared the performance of DDPG, PPO and PG on Chinese stock market. Yang et al. [135] firstly tested the
performance of PPO, A2C and DDPG on the U.S. stock market. Later on, the authors find that the ensemble strategy
of these three algorithms can integrate the best features and shows more robust performance adjusting to different
market situations.
Summary. Since a portfolio is a vector of weights for different financial assets, which naturally corresponds to a
policy, policy-based methods are the most widely-used RL methods for PM. There are also many successful examples
based on actor-critic algorithms. The summary of portfolio management publications is in Table 5. We point out two
issues of existing methods: (1) Most of them ignore the interrelationship between different financial assets, which is
valuable for human portfolio managers. (2) Existing works construct portfolios from a relative small pool of stocks (e.g.,
20 in total). However, the real market contains thousands of stocks and common RL methods are vulnerable when the
action space is very large [32].
In practice, professional traders usually finish the execution process in much shorter time window (e.g., 10 minute).
Third, existing works will fail when the trading volume is huge, because all of them assume there is no obvious market
impact, which is impossible for large volume settings. In the real-world, the requirement of institutional investors is
to execute large amount of shares in a relatively short time window. There is still a long way to go for researchers to
tackle these limitations.
First, data scarcity is a major challenge on applying RL for QT tasks. Model-based RL can speed up the training process
by learning a model of the financial market [137]. The worst-case (e.g., financial crisis) can be used as a regularizer
for maximizing the accumulated reward. Second, the key objective of QT is to balance between maximizing profit
and minimizing risk. Multi-objective RL techniques provide a weapon to balance the trade-off between profit and risk.
Training diversified trading policies with different risk tolerance is an interesting direction. Third, graph learning
[130] has shown promising results on modeling the ubiquitous relationship between stocks in supervised learning
[38, 98]. Combing graph learning with RL for modeling the internal relationship between different stocks or financial
market is an interesting future direction. Fourth, the severe distribution shift of financial market makes RL-based
methods exhibit poor generalization ability in new market condition. Meta-RL and transfer learning techniques can
help improve RL-based QT models’ generalization performance across different financial assets or market. Fifth, for
high risky decision-making tasks such as QT, we need to explain its actions to human traders as a condition for their
full acceptance of the algorithm. Hierarchical RL methods decompose the main goal into sub-goals for low-level agents.
By learning the optimal subgoals for the low-level agent, the high-level agent forms a representation of the financial
market that is interpretable by human traders. Sixth, for QT, learning through directly interacting with the real market
is risky and impractical. RL-based QT normally use historical data to learn a policy, which fits in offline RL settings.
Offline RL techniques can help to model the distribution shift and risk of financial market while training RL agents.
depth knowledge of RL. We believe that it is a promising research direction to facilitate the development of RL-based
QT models with auto-ML techniques.
7 CONCLUSION
In this article, we provided a comprehensive review of the most notable works on RL-based QT models. We proposed a
classification scheme for organizing and clustering existing works, and we highlighted a bunch of influential research
prototypes. We also discussed the pros/cons of utilizing RL techniques for QT tasks. In addition, we point out some of
the most pressing open problems and promising future directions. Both RL and QT are ongoing hot research topics in
the past few decades. There are many newly developing techniques and emerging models each year. We hope that
this survey can provide readers with a comprehensive understanding of the key aspects of this field, clarify the most
notable advances, and shed some lights on future research.
Manuscript submitted to ACM
22 Sun et al.
REFERENCES
[1] Saud Almahdi and Steve Y Yang. 2017. An adaptive portfolio trading system: A risk-return portfolio optimization using recurrent reinforcement
learning with expected maximum drawdown. Expert Systems with Applications 87 (2017), 267–279.
[2] Robert Almgren and Neil Chriss. 2001. Optimal execution of portfolio transactions. Journal of Risk 3 (2001), 5–40.
[3] Adebiyi A Ariyo, Adewumi O Adewumi, and Charles K Ayo. 2014. Stock price prediction using the ARIMA model. In Proceedings of the 6th
International Conference on Computer Modelling and Simulation (ICCMS). 106–112.
[4] Arash Bahrammirzaee. 2010. A comparative survey of artificial intelligence applications in finance: Artificial neural networks, Expert system and
hybrid intelligent systems. Neural Computing and Applications 19, 8 (2010), 1165–1195.
[5] Suryoday Basak, Saibal Kar, Snehanshu Saha, Luckyson Khaidem, and Sudeepa Roy Dey. 2019. Predicting the direction of stock market prices
using tree-based classifiers. The North American Journal of Economics and Finance 47 (2019), 552–567.
[6] Eric Benhamou, David Saltiel, Sandrine Ungari, and Abhishek Mukhopadhyay. 2020. Bridging the gap between Markowitz planning and deep
reinforcement learning. arXiv preprint arXiv:2010.09108 (2020).
[7] Francesco Bertoluzzo and Marco Corazza. 2012. Testing different reinforcement learning configurations for financial trading: Introduction and
applications. Procedia Economics and Finance 3 (2012), 68–77.
[8] Dimitris Bertsimas and Andrew W Lo. 1998. Optimal control of execution costs. Journal of Financial Markets 1, 1 (1998), 1–50.
[9] Dinesh Bhuriya, Girish Kaushal, Ashish Sharma, and Upendra Singh. 2017. Stock market predication using a linear regression. In Proceedings of 1st
International Conference of Electronics, Communication and Aerospace Technology (ICECA). 510–513.
[10] Lorenzo Bisi, Luca Sabbioni, Edoardo Vittori, Matteo Papini, and Marcello Restelli. 2019. Risk-averse trust region optimization for reward-volatility
reduction. arXiv preprint arXiv:1912.03193 (2019).
[11] Fischer Black and Myron Scholes. 1973. The Pricing of Options and Corporate Liabilities. The Journal of Political Economy 81, 3 (1973), 637–654.
[12] John Bollinger. 2002. Bollinger on Bollinger Bands. McGraw-Hill New York.
[13] Allan Borodin, Ran El-Yaniv, and Vincent Gogan. 2004. Can we learn to beat the best stock. Journal of Artificial Intelligence Research 21 (2004),
579–594.
[14] David Byrd, Maria Hybinette, and Tucker Hybinette Balch. 2019. Abides: Towards high-fidelity market simulation for AI research. arXiv preprint
arXiv:1904.12066 (2019).
[15] Álvaro Cartea, Sebastian Jaimungal, and José Penalva. 2015. Algorithmic and High-frequency Trading.
[16] Álvaro Cartea, Sebastian Jaimungal, and Jason Ricci. 2014. Buy Low, Sell High: A High Frequency Trading Perspective. SIAM Journal on Financial
Mathematics 5, 1 (2014), 415–444.
[17] Stephan K. Chalup and Andreas Mitschele. 2008. Kernel Methods in Finance. Chapter 27, 655–687.
[18] Louis KC Chan, Narasimhan Jegadeesh, and Josef Lakonishok. 1996. Momentum strategies. The Journal of Finance 51, 5 (1996), 1681–1713.
[19] Nicholas Tung Chan and Christian Shelton. 2001. An electronic market-maker. Technical report. (2001).
[20] Lakshay Chauhan, John Alberg, and Zachary Lipton. 2020. Uncertainty-aware lookahead factor models for quantitative investing. In Proceedings of
the 37th International Conference on Machine Learning (ICML). 1489–1499.
[21] Chi Chen, Li Zhao, Jiang Bian, Chunxiao Xing, and Tie-Yan Liu. 2019. Investment behaviors can tell what inside: Exploring stock intrinsic
properties for stock trend prediction. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining
(KDD). 2376–2384.
[22] Yingmei Chen, Zhongyu Wei, and Xuanjing Huang. 2018. Incorporating corporation relationship via graph convolutional neural networks for
stock price prediction. In Proceedings of the 27th ACM International Conference on Information and Knowledge Management (CIKM). 1655–1658.
[23] Thomas M. Cover. 1991. Universal Portfolios. Mathematical Finance 1, 1 (1991), 1–29.
[24] Kevin Dabérius, Elvin Granat, and Patrik Karlsson. 2019. Deep Execution-Value and Policy Based Reinforcement Learning for Trading and Beating
Market Benchmarks. Available at SSRN 3374766 (2019).
[25] Renato Arantes de Oliveira, Heitor S Ramos, Daniel Hasan Dalip, and Adriano César Machado Pereira. 2020. A tabular Sarsa-based stock market
agent. In Proceedings of the 1st ACM International Conference on AI in Finance (ICAIF).
[26] Michael AH Dempster and Vasco Leemans. 2006. An automated FX trading system using adaptive reinforcement learning. Expert Systems with
Applications 30, 3 (2006), 543–552.
[27] Yue Deng, Feng Bao, Youyong Kong, Zhiquan Ren, and Qionghai Dai. 2016. Deep direct reinforcement learning for financial signal representation
and trading. IEEE Transactions on Neural Networks and Learning Systems 28, 3 (2016), 653–664.
[28] A Victor Devadoss and T Antony Alphonnse Ligori. 2013. Forecasting of stock prices using multi layer perceptron. International Journal of
Computing Algorithm 2 (2013), 440–449.
[29] Xiao Ding, Yue Zhang, Ting Liu, and Junwen Duan. 2015. Deep learning for event-driven stock prediction. In Proceedings of the 24th International
Joint Conference on Artificial Intelligence (IJCAI). 2327–2333.
[30] Xiao Ding, Yue Zhang, Ting Liu, and Junwen Duan. 2016. Knowledge-driven event embedding for stock prediction. In Proceedings of the 26th
International Conference on Computational Linguistics. 2133–2142.
[31] Yi Ding, Weiqing Liu, Jiang Bian, Daoqiang Zhang, and Tie-Yan Liu. 2018. Investor-imitator: A framework for trading knowledge extraction. In
Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining (KDD). 1310–1319.
[32] Gabriel Dulac-Arnold, Richard Evans, Hado van Hasselt, Peter Sunehag, Timothy Lillicrap, Jonathan Hunt, Timothy Mann, Theophane Weber,
Thomas Degris, and Ben Coppin. 2015. Deep reinforcement learning in large discrete action spaces. arXiv preprint arXiv:1512.07679 (2015).
[33] Sophie Emerson, Ruairí Kennedy, Luke O’Shea, and John O’Brien. 2019. Trends and applications of machine learning in quantitative finance. In
Proceedigns of the 8th International Conference on Economics and Finance Research (ICEFR).
[34] Eugene F Fama. 2021. Efficient capital markets: A review of theory and empirical work. The Fama Portfolio (2021), 76–121.
[35] Eugene F Fama and Kenneth R French. 1993. Common risk factors in the returns on stocks and bonds. Journal of Financial Economics 33, 1 (1993),
3–56.
[36] Jie Fang, Shutao Xia, Jianwu Lin, Zhikang Xia, Xiang Liu, and Yong Jiang. 2019. Alpha discovery neural network based on prior knowledge. arXiv
preprint arXiv:1912.11761 (2019).
[37] Yuchen Fang, Kan Ren, Weiqing Liu, Dong Zhou, Weinan Zhang, Jiang Bian, Yong Yu, and Tie-Yan Liu. 2021. Universal trading for order execution
with oracle policy distillation. In Proceedings of the 35th AAAI Conference on Artificial Intelligence (AAAI).
[38] Fuli Feng, Xiangnan He, Xiang Wang, Cheng Luo, Yiqun Liu, and Tat-Seng Chua. 2019. Temporal relational ranking for stock prediction. ACM
Transactions on Information Systems (TOIS) 37, 2 (2019), 27.
[39] Thomas G Fischer. 2018. Reinforcement learning in financial markets: A survey. Technical Report. FAU Discussion Papers in Economics.
[40] Keke Gai, Meikang Qiu, and Xiaotong Sun. 2018. A survey on FinTech. Journal of Network and Computer Applications 103 (2018), 262–273.
[41] Alexei A Gaivoronski and Fabio Stella. 2000. Stochastic nonstationary optimization for finding universal portfolios. Annals of Operations Research
100, 1 (2000), 165–188.
[42] Xiu Gao and Laiwan Chan. 2000. An algorithm for trading and portfolio management using Q-learning and sharpe ratio maximization. In
Proceedings of the 14th International Conference on Neural Information Processing (NIPS). 832–837.
[43] Dhananjay K Gode and Shyam Sunder. 1993. Allocative Efficiency of Markets with Zero-intelligence Traders: Market as a Partial Substitute for
Individual Rationality. Journal of Political Economy 101, 1 (1993), 119–137.
[44] Shihao Gu, Bryan Kelly, and Dacheng Xiu. 2020. Empirical asset pricing via machine learning. The Review of Financial Studies 33, 5 (2020),
2223–2273.
[45] Shihao Gu, Bryan Kelly, and Dacheng Xiu. 2021. Autoencoder asset pricing models. Journal of Econometrics 222, 1 (2021), 429–450.
[46] Olivier Guéant and Iuliia Manziuk. 2019. Deep reinforcement learning for market making in corporate bonds: Beating the curse of dimensionality.
Applied Mathematical Finance 26, 5 (2019), 387–452.
[47] László Györfi, Gábor Lugosi, and Frederic Udina. 2006. Nonparametric kernel-based sequential investment strategies. Mathematical Finance: An
International Journal of Mathematics, Statistics and Financial Economics 16, 2 (2006), 337–357.
[48] David P Helmbold, Robert E Schapire, Yoram Singer, and Manfred K Warmuth. 1998. On-line portfolio selection using multiplicative updates.
Mathematical Finance 8, 4 (1998), 325–347.
[49] Dieter Hendricks and Diane Wilcox. 2014. A reinforcement learning extension to the Almgren-Chriss framework for optimal trade execution. In
Proceedings of the IEEE Conference on Computational Intelligence for Financial Engineering & Economics. 457–464.
[50] Ehsan Hoseinzade and Saman Haratizadeh. 2019. CNNpred: CNN-based stock market prediction using a diverse set of variables. Expert Systems
with Applications 129 (2019), 273–285.
[51] Ziniu Hu, Weiqing Liu, Jiang Bian, Xuanzhe Liu, and Tie-Yan Liu. 2018. Listening to chaotic whispers: A deep learning framework for news-oriented
stock trend prediction. In Proceedings of the 11th ACM International Conference on Web Search and Data Mining (WSDM). 261–269.
[52] Chien-Feng Huang. 2012. A hybrid stock selection model using genetic algorithms and support vector regression. Applied Soft Computing 12, 2
(2012), 807–818.
[53] Dingjiang Huang, Junlong Zhou, Bin Li, Steven HOI, and Shuigeng Zhou. 2013. Robust median reversion strategy for on-line portfolio selection. In
Proceedings of the 23rd International Joint Conference on Artificial Intelligence (IJCAI). 2006–2012.
[54] Wei Huang, Yoshiteru Nakamori, and Shou-Yang Wang. 2005. Forecasting stock market movement direction with support vector machine.
Computers & Operations Research 32, 10 (2005), 2513–2522.
[55] Zhenhan Huang and Fumihide Tanaka. 2021. A Modularized and Scalable Multi-Agent Reinforcement Learning-based System for Financial Portfolio
Management. arXiv preprint arXiv:2102.03502 (2021).
[56] O Jangmin, Jongwoo Lee, Jae Won Lee, and Byoung-Tak Zhang. 2006. Adaptive stock trading with dynamic asset allocation using reinforcement
learning. Information Sciences 176, 15 (2006), 2121–2147.
[57] Narasimhan Jegadeesh and Sheridan Titman. 1993. Returns to buying winners and selling losers: Implications for stock market efficiency. The
Journal of Finance 48, 1 (1993), 65–91.
[58] Gyeeun Jeong and Ha Young Kim. 2019. Improving financial trading decisions using deep Q-learning: Predicting the number of shares, action
strategies, and transfer learning. Expert Systems with Applications 117 (2019), 125–138.
[59] Zhengyao Jiang and Jinjun Liang. 2017. Cryptocurrency portfolio management with deep reinforcement learning. In 2017 Intelligent Systems
Conference (IntelliSys). 905–913.
[60] Zhengyao Jiang, Dixing Xu, and Jinjun Liang. 2017. A deep reinforcement learning framework for the financial portfolio management problem.
arXiv preprint arXiv:1706.10059 (2017).
[61] Sham M Kakade, Michael Kearns, Yishay Mansour, and Luis E Ortiz. 2004. Competitive algorithms for VWAP and limit order trading. In Proceedings
of the 5th ACM conference on Electronic Commerce (EC). 189–198.
Manuscript submitted to ACM
24 Sun et al.
[62] Zura Kakushadze. 2016. 101 formulaic alphas. Wilmott 2016, 84 (2016), 72–81.
[63] Guolin Ke, Qi Meng, Thomas Finley, Taifeng Wang, Wei Chen, Weidong Ma, Qiwei Ye, and Tie-Yan Liu. 2017. Lightgbm: A highly efficient gradient
boosting decision tree. In Proceedings of the 30th Neural Information Processing Systems. 3146–3154.
[64] Luckyson Khaidem, Snehanshu Saha, and Sudeepa Roy Dey. 2016. Predicting the direction of stock market prices using random forest. arXiv
preprint arXiv:1605.00003 (2016).
[65] Vijay R Konda and John N Tsitsiklis. 2000. Actor-critic algorithms. Proceedings of the 14th Neural Information Processing Systems (NIPS), 1008–1014.
[66] Jinho Lee, Raehyun Kim, Seok-Won Yi, and Jaewoo Kang. 2020. MAPS: Multi-agent reinforcement learning-based portfolio management system.
arXiv preprint arXiv:2007.05402 (2020).
[67] Jae Won Lee and O Jangmin. 2002. A multi-agent Q-learning framework for optimizing stock trading systems. In Proceedings of the 13rd International
Conference on Database and Expert Systems Applications (DESA). 153–162.
[68] Ming-Chi Lee. 2009. Using support vector machine with a hybrid feature selection method to the stock trend prediction. Expert Systems with
Applications 36, 8 (2009), 10896–10904.
[69] Sergey Levine, Chelsea Finn, Trevor Darrell, and Pieter Abbeel. 2016. End-to-end training of deep visuomotor policies. Journal of Machine Learning
Research 17, 1 (2016), 1334–1373.
[70] Bin Li and Steven CH Hoi. 2014. Online portfolio selection: A survey. Comput. Surveys 46, 3 (2014), 1–36.
[71] Bin Li, Steven CH Hoi, and Vivekanand Gopalkrishnan. 2011. Corn: Correlation-driven nonparametric learning approach for portfolio selection.
ACM Transactions on Intelligent Systems and Technology 2, 3 (2011), 1–29.
[72] Bin Li, Peilin Zhao, Steven CH Hoi, and Vivekanand Gopalkrishnan. 2012. PAMR: Passive aggressive mean reversion strategy for portfolio selection.
Machine Learning 87, 2 (2012), 221–258.
[73] Wei Li, Ruihan Bao, Keiko Harimoto, Deli Chen, Jingjing Xu, and Qi Su. 2020. Modeling the stock relation with graph network for overnight stock
movement prediction. In Proceedings of the 29th International Joint Conference on Artificial Intelligence (IJCAI). 4541–4547.
[74] Zhipeng Liang, Hao Chen, Junhao Zhu, Kangkang Jiang, and Yanran Li. 2018. Adversarial deep reinforcement learning in portfolio management.
arXiv preprint arXiv:1808.09940 (2018).
[75] Timothy P Lillicrap, Jonathan J Hunt, Alexander Pritzel, Nicolas Heess, Tom Erez, Yuval Tassa, David Silver, and Daan Wierstra. 2015. Continuous
control with deep reinforcement learning. arXiv preprint arXiv:1509.02971 (2015).
[76] Bryan Lim, Stefan Zohren, and Stephen Roberts. 2019. Enhancing time-series momentum strategies using deep neural networks. The Journal of
Financial Data Science 1, 4 (2019), 19–38.
[77] Ye-Sheen Lim and Denise Gorse. 2018. Reinforcement learning for high-frequency market making. In Proceedings of the 26th European Symposium
on Artificial Neural Networks, Computational Intelligence and Machine Learning (ESANN).
[78] Siyu Lin and Peter A Beling. 2020. An end-to-end optimal trade execution framework based on proximal policy optimization. In Proceedings of the
29th International Joint Conference on Artificial Intelligence (IJCAI). 4548–4554.
[79] Guang Liu, Yuzhao Mao, Qi Sun, Hailong Huang, Weiguo Gao, Xuan Li, JianPing Shen, Ruifan Li, and Xiaojie Wang. 2020. Multi-scale two-way
deep neural network for stock trend prediction. In Proceedings of the 29th International Joint Conference on Artificial Intelligence (IJCAI). 4555–4561.
[80] Yang Liu, Qi Liu, Hongke Zhao, Zhen Pan, and Chuanren Liu. 2020. Adaptive quantitative trading: An imitative deep reinforcement learning
approach. In Proceedings of the 32nd AAAI Conference on Artificial Intelligence (AAAI). 2128–2135.
[81] Malik Magdon-Ismail and Amir F Atiya. 2004. Maximum drawdown. Risk Magazine 17, 10 (2004), 99–102.
[82] Harry Markowitz. 1959. Portfolio Selection. Yale University Press New Haven.
[83] Volodymyr Mnih, Koray Kavukcuoglu, David Silver, Andrei A Rusu, Joel Veness, Marc G Bellemare, Alex Graves, Martin Riedmiller, Andreas K
Fidjeland, Georg Ostrovski, et al. 2015. Human-level control through deep reinforcement learning. Nature 518, 7540 (2015), 529–533.
[84] John Moody and Matthew Saffell. 2001. Learning to trade via direct reinforcement. IEEE Transactions on Neural Networks 12, 4 (2001), 875–889.
[85] John Moody and Lizhong Wu. 1997. Optimization of trading systems and portfolios. In Proceedings of the IEEE/IAFE 1997 Computational Intelligence
for Financial Engineering. 300–307.
[86] John Moody, Lizhong Wu, Yuansong Liao, and Matthew Saffell. 1998. Performance functions and reinforcement learning for trading systems and
portfolios. Journal of Forecasting 17, 5-6 (1998), 441–470.
[87] Tobias J Moskowitz, Yao Hua Ooi, and Lasse Heje Pedersen. 2012. Time series momentum. Journal of Financial Economics 104, 2 (2012), 228–250.
[88] Ralph Neuneier. 1996. Optimal asset allocation using adaptive dynamic programming. Proceedings of the 10th Neural Information Processing Systems
(NIPS), 952–958.
[89] Ralph Neuneier. 1998. Enhancing Q-learning for optimal asset allocation. In Proceedings of the 12nd Neural Information Processing Systems (NIPS).
936–942.
[90] Yuriy Nevmyvaka, Yi Feng, and Michael Kearns. 2006. Reinforcement learning for optimized trade execution. In Proceedings of the 23rd International
Conference on Machine Learning (ICML). 673–680.
[91] Brian Ning, Franco Ho Ting Lin, and Sebastian Jaimungal. 2018. Double deep Q-learning for optimal execution. arXiv preprint arXiv:1812.06600
(2018).
[92] Ahmet Murat Ozbayoglu, Mehmet Ugur Gudelek, and Omer Berat Sezer. 2020. Deep learning for financial applications: A survey. Applied Soft
Computing (2020), 106384.
[93] Theodore Panagiotidis, Thanasis Stengos, and Orestis Vravosinos. 2018. On the determinants of bitcoin returns: A LASSO approach. Finance
Research Letters 27 (2018), 235–240.
[94] Jigar Patel, Sahil Shah, Priyank Thakkar, and Ketan Kotecha. 2015. Predicting stock and stock price index movement using trend deterministic data
preparation and machine learning techniques. Expert Systems with Applications 42, 1 (2015), 259–268.
[95] James M Poterba and Lawrence H Summers. 1988. Mean reversion in stock prices: Evidence and implications. Journal of Financial Economics 22, 1
(1988), 27–59.
[96] Gavin A Rummery and Mahesan Niranjan. 1994. On-line Q-learning Using Connectionist Systems. University of Cambridge, Department of
Engineering Cambridge, UK.
[97] Francesco Rundo, Francesca Trenta, Agatino Luigi di Stallo, and Sebastiano Battiato. 2019. Machine learning for quantitative finance applications:
A survey. Applied Sciences 9, 24 (2019), 5574.
[98] Ramit Sawhney, Shivam Agarwal, Arnav Wadhwa, and Rajiv Shah. 2021. Exploring the Scale-Free Nature of Stock Markets: Hyperbolic Graph
Learning for Algorithmic Trading. In Proceedings of the Web Conference 2021. 11–22.
[99] Ramit Sawhney, Arnav Wadhwa, Shivam Agarwal, and Rajiv Shah. 2021. Quantitative Day Trading from Natural Language using Reinforcement
Learning. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics (NAACL). 4018–4030.
[100] John Schulman, Filip Wolski, Prafulla Dhariwal, Alec Radford, and Oleg Klimov. 2017. Proximal policy optimization algorithms. arXiv preprint
arXiv:1707.06347 (2017).
[101] Sreelekshmy Selvin, R Vinayakumar, EA Gopalakrishnan, Vijay Krishna Menon, and KP Soman. 2017. Stock price prediction using LSTM, RNN
and CNN-sliding window model. In Proceedings of the 6th International Conference on Advances in Computing, Communications and Informatics
(ICACCI). 1643–1647.
[102] Omer Berat Sezer, Mehmet Ugur Gudelek, and Ahmet Murat Ozbayoglu. 2020. Financial time series forecasting with deep learning: A systematic
literature review: 2005–2019. Applied Soft Computing 90 (2020), 106181.
[103] William F Sharpe. 1964. Capital asset prices: A theory of market equilibrium under conditions of risk. The Journal of Finance 19, 3 (1964), 425–442.
[104] William F Sharpe. 1994. The sharpe ratio. Journal of Portfolio Management 21, 1 (1994), 49–58.
[105] Si Shi, Jianjun Li, Guohui Li, and Peng Pan. 2019. A multi-scale temporal feature aggregation convolutional neural network for portfolio management.
In Proceedings of the 28th ACM International Conference on Information and Knowledge Management (CIKM). 1613–1622.
[106] Weiyu Si, Jinke Li, Peng Ding, and Ruonan Rao. 2017. A multi-objective deep reinforcement learning approach for stock index future’s intraday
trading. In Proceeding of the 10th International Symposium on Computational Intelligence and Design (ISCID). 431–436.
[107] David Silver, Aja Huang, Chris J Maddison, Arthur Guez, Laurent Sifre, George Van Den Driessche, Julian Schrittwieser, Ioannis Antonoglou, Veda
Panneershelvam, Marc Lanctot, et al. 2016. Mastering the game of Go with deep neural networks and tree search. Nature 529, 7587 (2016), 484–489.
[108] David Silver, Guy Lever, Nicolas Heess, Thomas Degris, Daan Wierstra, and Martin Riedmiller. 2014. Deterministic policy gradient algorithms. In
Proceedings of the 31st International Conference on Machine Learning (ICML). 387–395.
[109] Thomas Spooner, John Fearnley, Rahul Savani, and Andreas Koukorinis. 2018. Market making via reinforcement learning. arXiv preprint
arXiv:1804.04216 (2018).
[110] Thomas Spooner and Rahul Savani. 2020. Robust market making via adversarial reinforcement learning. arXiv preprint arXiv:2003.01820 (2020).
[111] Xiaolei Sun, Mingxi Liu, and Zeqian Sima. 2020. A novel cryptocurrency price trend forecasting model based on LightGBM. Finance Research
Letters 32 (2020), 101084.
[112] Richard S Sutton and Andrew G Barto. 2018. Reinforcement Learning: An Introduction. MIT press.
[113] Christian Szegedy, Wei Liu, Yangqing Jia, Pierre Sermanet, Scott Reed, Dragomir Anguelov, Dumitru Erhan, Vincent Vanhoucke, and Andrew
Rabinovich. 2015. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 1–9.
[114] Lawrence Takeuchi and Yu-Ying Albert Lee. 2013. Applying deep learning to enhance momentum trading strategies in stocks. Technical report.
[115] Leigh Tesfatsion and Kenneth L Judd. 2006. Handbook of Computational Economics: Agent-based Computational Economics.
[116] Alaa Tharwat, Tarek Gaber, Abdelhameed Ibrahim, and Aboul Ella Hassanien. 2017. Linear discriminant analysis: A detailed tutorial. AI
communications 30, 2 (2017), 169–190.
[117] Chih-Fong Tsai and Yu-Chieh Hsiao. 2010. Combining multiple feature selection methods for stock prediction: Union, intersection, and multi-
intersection approaches. Decision Support Systems 50, 1 (2010), 258–269.
[118] Oriol Vinyals, Igor Babuschkin, Wojciech M Czarnecki, Michaël Mathieu, Andrew Dudzik, Junyoung Chung, David H Choi, Richard Powell, Timo
Ewalds, Petko Georgiev, et al. 2019. Grandmaster level in StarCraft II using multi-agent reinforcement learning. Nature 575, 7782 (2019), 350–354.
[119] Edoardo Vittori, Martino Bernasconi de Luca, Francesco Trovò, and Marcello Restelli. 2020. Dealing with transaction costs in portfolio optimization:
Online gradient descent with momentum. In Proceedings of the 1st ACM International Conference on AI in Finance (ICAIF). 1–8.
[120] Edoardo Vittori, Michele Trapletti, and Marcello Restelli. 2020. Option hedging with risk averse reinforcement learning. arXiv preprint
arXiv:2010.12245 (2020).
[121] Svitlana Vyetrenko, David Byrd, Nick Petosa, Mahmoud Mahfouz, Danial Dervovic, Manuela Veloso, and Tucker Hybinette Balch. 2019. Get real:
Realism metrics for robust limit order book market simulations. arXiv preprint arXiv:1912.04941 (2019).
[122] Jingyuan Wang, Yang Zhang, Ke Tang, Junjie Wu, and Zhang Xiong. 2019. Alphastock: A buying-winners-and-selling-losers investment strategy
using interpretable deep reinforcement attention networks. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery
& Data Mining (KDD). 1900–1908.
Manuscript submitted to ACM
26 Sun et al.
[123] Rundong Wang, Hongxin Wei, Bo An, Zhouyan Feng, and Jun Yao. 2020. Commission fee is not enough: A hierarchical reinforced framework for
portfolio management. arXiv preprint arXiv:2012.12620 (2020).
[124] Zhicheng Wang, Biwei Huang, Shikui Tu, Kun Zhang, and Lei Xu. 2021. DeepTrader: A deep reinforcement learning approach to risk-return
balanced portfolio management with market conditions embedding. In Proceedings of the 35th AAAI Conference on Artificial Intelligence (AAAI).
[125] Christopher JCH Watkins and Peter Dayan. 1992. Q-learning. Machine Learning 8, 3-4 (1992), 279–292.
[126] Ronald J Williams. 1992. Simple statistical gradient-following algorithms for connectionist reinforcement learning. Machine learning 8, 3-4 (1992),
229–256.
[127] Svante Wold, Kim Esbensen, and Paul Geladi. 1987. Principal component analysis. Chemometrics and Intelligent Laboratory Systems 2, 1-3 (1987),
37–52.
[128] Bo K Wong and Yakup Selvi. 1998. Neural network applications in finance: A review and analysis of literature (1990–1996). Information &
Management 34, 3 (1998), 129–139.
[129] Lan Wu and Yuehan Yang. 2014. Nonnegative elastic net and application in index tracking. Appl. Math. Comput. 227 (2014), 541–552.
[130] Zonghan Wu, Shirui Pan, Fengwen Chen, Guodong Long, Chengqi Zhang, and S Yu Philip. 2020. A comprehensive survey on graph neural
networks. IEEE Transactions on Neural Networks and Learning Systems 32, 1 (2020), 4–24.
[131] Zhuoran Xiong, Xiao-Yang Liu, Shan Zhong, Hongyang Yang, and Anwar Walid. 2018. Practical deep reinforcement learning approach for stock
trading. arXiv preprint arXiv:1811.07522 (2018).
[132] Ke Xu, Yifan Zhang, Deheng Ye, Peilin Zhao, and Mingkui Tan. 2020. Relation-Aware Transformer for Portfolio Policy Learning.. In Proceedings of
the 29th International Joint Conference on Artificial Intelligence (IJCAI). 4647–4653.
[133] Wentao Xu, Weiqing Liu, Chang Xu, Jiang Bian, Jian Yin, and Tie-Yan Liu. 2021. REST: Relational Event-driven Stock Trend Forecasting. In
Proceedings of the Web Conference 2021. 1–10.
[134] Yumo Xu and Shay B Cohen. 2018. Stock movement prediction from tweets and historical prices. In Proceedings of the 56th Annual Meeting of the
Association for Computational Linguistics (ACL). 1970–1979.
[135] Hongyang Yang, Xiao-Yang Liu, Shan Zhong, and Anwar Walid. 2020. Deep reinforcement learning for automated stock trading: An ensemble
strategy. Available at SSRN (2020).
[136] Yunan Ye, Hengzhi Pei, Boxin Wang, Pin-Yu Chen, Yada Zhu, Ju Xiao, and Bo Li. 2020. Reinforcement-learning based portfolio management with
augmented asset movement prediction states. In Proceedings of the 34th AAAI Conference on Artificial Intelligence (AAAI). 1112–1119.
[137] Pengqian Yu, Joon Sern Lee, Ilya Kulyatin, Zekun Shi, and Sakyasingha Dasgupta. 2019. Model-based deep reinforcement learning for dynamic
portfolio optimization. arXiv preprint arXiv:1901.08740 (2019).
[138] Yuyu Yuan, Wen Wen, and Jincui Yang. 2020. Using data augmentation based reinforcement learning for daily stock trading. Electronics 9, 9 (2020),
1384.
[139] Chuheng Zhang, Yuanqi Li, Xi Chen, Yifei Jin, Pingzhong Tang, and Jian Li. 2020. DoubleEnsemble: A new ensemble method based on sample
reweighting and feature selection for financial data analysis. arXiv preprint arXiv:2010.01265 (2020).
[140] Dongsong Zhang and Lina Zhou. 2004. Discovering golden nuggets: Data mining in financial application. IEEE Transactions on Systems, Man, and
Cybernetics, Part C (Applications and Reviews) 34, 4 (2004), 513–522.
[141] Liheng Zhang, Charu Aggarwal, and Guo-Jun Qi. 2017. Stock price prediction via discovering multi-frequency trading patterns. In Proceedings of
the 23rd ACM SIGKDD International Conference on Knowledge Discovery & Data Mining (KDD). 2141–2149.
[142] Tianping Zhang, Yuanqi Li, Yifei Jin, and Jian Li. 2020. AutoAlpha: An efficient hierarchical evolutionary algorithm for mining alpha factors in
quantitative investment. arXiv preprint arXiv:2002.08245 (2020).
[143] Yifan Zhang, Peilin Zhao, Bin Li, Qingyao Wu, Junzhou Huang, and Mingkui Tan. 2020. Cost-sensitive portfolio selection via deep reinforcement
learning. IEEE Transactions on Knowledge and Data Engineering (2020).
[144] Zihao Zhang, Stefan Zohren, and Stephen Roberts. 2020. Deep reinforcement learning for trading. The Journal of Financial Data Science 2, 2 (2020),
25–40.
[145] Yueyang Zhong, YeeMan Bergstrom, and Amy Ward. 2020. Data-driven market-making via model-free learning. In Proceedings of the 29th
International Joint Conference on Artificial Intelligence (IJCAI). 2327–2333.
[146] Dawei Zhou, Lecheng Zheng, Yada Zhu, Jianbo Li, and Jingrui He. 2020. Domain adaptive multi-modality neural attention network for financial
forecasting. In Proceedings of the 29th Web Conference (WWW). 2230–2240.