Generative Modeling For Time Series Via SCHR Odinger Bridge: Mohamed HAMDOUCHE Pierre Henry-Labordere Huy en Pham
Generative Modeling For Time Series Via SCHR Odinger Bridge: Mohamed HAMDOUCHE Pierre Henry-Labordere Huy en Pham
Abstract
arXiv:2304.05093v1 [math.OC] 11 Apr 2023
We propose a novel generative model for time series based on Schrödinger bridge
(SB) approach. This consists in the entropic interpolation via optimal transport be-
tween a reference probability measure on path space and a target measure consistent
with the joint data distribution of the time series. The solution is characterized by a
stochastic differential equation on finite horizon with a path-dependent drift function,
hence respecting the temporal dynamics of the time series distribution. We can esti-
mate the drift function from data samples either by kernel regression methods or with
LSTM neural networks, and the simulation of the SB diffusion yields new synthetic
data samples of the time series.
The performance of our generative model is evaluated through a series of numerical
experiments. First, we test with a toy autoregressive model, a GARCH Model, and the
example of fractional Brownian motion, and measure the accuracy of our algorithm
with marginal and temporal dependencies metrics. Next, we use our SB generated
synthetic samples for the application to deep hedging on real-data sets. Finally, we
illustrate the SB approach for generating sequence of images.
Keywords: generative models, time series, Schrödinger bridge, kernel estimation, deep
hedging.
1 Introduction
Sequential data appear widely in our society like in video and audio data, and simulation of
time series models are used in various industrial applications including clinical predictions
[17], and weather forecasts [5]. In the financial industry, simulations of dynamical scenarios
are considered in market stress tests, risk measurement, and risk management, e.g. in
deep hedging [2]. The design of time series model is a delicate issue, requiring expensive
calibration task, and subject to error misspecification and model risk. Therefore, the
generation for synthetic samples of time series has gained an increasing attention over the
last years, and opens the door in the financial sector for a pure data driven approach in
risk management.
∗
LPSM, Université Paris Cité and Sorbonne Université, and Qube Research and Technologies, ham-
douche at lpsm.paris
†
Qube Research and Technologies, pierre.henrylabordere at qube-rt.com
‡
LPSM, Université Paris Cité and Sorbonne Université, pham at lpsm.paris; This author is supported by
the BNP-PAR Chair “Futures of Quantitative Finance”, and by FiME, Laboratoire de Finance des Marchés de
l’Energie, and the “Finance and Sustainable Development” EDF - CACIB Chair
1
Generative modeling (GM) has become over the last years a successful machine lear-
ning task for data synthesis notably in (static) image processing. Several competing
methods have been developed and state-of-the-art includes Likelihood-based models like
energy-based models (EBM) [15], variational auto-encoders (VAE) [14], Implicit genera-
tive models with the prominent works on generative adversarial network (GAN) [11] and
its extensions [1], and recently the new generation of score-based models using Langevin
dynamics, [21], [22], [12], and diffusion models via Schrödinger bridge, see [13] for the
application to a class of stochastic volatility models, and [23], [8]. Generative methods
for time series raises challenging issues for learning efficiently the temporal dependencies.
Indeed, in order to capture the potentially complex dynamics of variables across time, it is
not sufficient to learn the time marginals or even the joint distribution without exploiting
the sequential structure. An increasing attention has been paid to these methods in the
literature and state-of-the-art generative methods for time series are: Time series GAN
[27] which combines an unsupervised adversarial loss on real/synthetic data and super-
vised loss for generating sequential data, Quant GAN [25] with an adversarial generator
using temporal convolutional networks, Causal optimal transport COT-GAN [26] with ad-
versarial generator using the adapted Wasserstein distance for processes, Conditional loss
Euler generator [20] starting from a diffusion representation time series and minimizing the
conditional distance between transition probabilities of real/synthetic samples, Signature
embedding of time series [9], [18], [3], and Functional data analysis with neural SDEs [6].
In this paper, we develop a novel generative model based on Schrödinger bridge ap-
proach that captures the temporal dynamics of the time series. This consists in the entropic
interpolation via optimal transport between a reference probability measure on path space
and a target measure consistent with the joint data distribution of the time series. The
solution is characterized by a stochastic differential equation on finite horizon with a path-
dependent drift function, called Schrödinger bridge time series (SBTS) diffusion, and the
simulation of the SBTS diffusion yields new synthetic data samples of the time series.
Our SB approach differs from related works that have been recently designed for lear-
ning marginal (or static) distributions. In [23], the authors perform generative models by
solving two SB problems. The paper [8] formulates generative modeling by computing the
SB problem between the data and prior distribution. The very recent work [4] proposes
momentum SB by considering an additional velocity variable for learning multi marginal
distributions. Let us mention also the recent paper [19] that combines SB with h-transform
in order to respect aligned data. Instead, our SBTS diffusion interpolates the joint time
series distribution starting from an initial deterministic value. Moreover, we propose
an alternate method for the estimation of the drift function, which is path-dependent
in our case. While [23] uses a logistic regression for estimating the density ratio and
then the drift function, which requires additional samples from Gaussian noises, and [8]
performs an extension of the Sinkhorn algorithm, we propose a kernel regression method
relying solely on data samples, and this turns out to be quite simple, efficient and low-cost
computationally. Compared to GAN type methods, the simulation of synthetic samples
from SBTS is much faster as it does not require the training of neural networks.
We validate our methodology with several numerical experiments. We first test on
some time series models like autoregressive, GARCH models, and also for the fractional
Brownian motion with rough paths. The accuracy is measured by some metrics aiming to
2
capture the temporal dynamics and the correlation structure. We also provide operational
metrics of interest for the financial industry by implementing our results on real data-
sets, and applying to the deep hedging of call options. Finally, we show some numerical
illustrations of our SB method in high dimension for the generation of sequential images.
2 Problem formulation
Let µ be the distribution of a time series representing the evolution of some Rd -valued
process of interest (e.g. asset price, claim process, audio/video data, etc), and suppose
that one can observe samples of this process at given fixed times of a discrete time grid T
= {ti , i = 1, . . . , N } on (0, ∞). We set T = tN as the terminal observation horizon. Our
goal is to construct a model that generates time series samples according to the unknown
target distribution µ ∈ P((Rd )N ) the set of probability measures on (Rd )N . For that
objective, we propose a dynamic modification of the Schrödinger bridge as follows. Let
Ω = C([0, T ]; Rd ) be the space of Rd -valued continuous functions on [0, T ], X = (Xt )t
the canonical process with initial value X0 = 0, and F = (Ft )t the canonical filtration.
Denoting by P(Ω) as the space of probability measures on Ω, we search for P (representing
the theoretical generative model) in P(Ω), close to the Wiener measure W in the sense
of Kullback-Leibler (or relative entropy), and consistent with the observation samples. In
other words, we look for a probability measure P∗ ∈ P(Ω) solution to:
P∗ ∈ arg min
µ
H(P|W), (2.1)
P∈PT (Ω)
where PTµ (Ω) is the set of probability measures P on Ω with joint distribution µ at
(t1 , . . . , tN ), i.e., P ◦ (Xt1 , . . . , XtN )−1 = µ, and H(.|.) is the relative entropy between
two probability measures defined by
dP
R
ln dW dP, if P W
H(P|W) =
∞, otherwise .
dP
Denoting by EP and EW the expectation under P and W, we see that H(P|W) = EP [ln dW ]
dP dP
= EW [ dW ln dW ] when P W. Compared to the classical Schrödinger bridge (SB) (see
[16], and the application to generative modeling in [23]), which looks for a probability
measure P that interpolates between an initial probability measure and a target probability
distribution at terminal time T , here, we take into account via the constraint in PTµ the
temporal dependence of the process observed at sequential times t1 < . . . < tN , and look
for an entropic interpolation of the time series distribution. We call (2.1) the Schrödinger
bridge for time series (SBTS) problem.
Let us now formulate (SBTS) as a stochastic control problem following the well-known
connection established for classical (SB) in [7]. Given P ∈ P(Ω) with finite relative entropy
H(P|W) < ∞, it is known by Girsanov’s theorem that one can associate to P an F-adapted
RT
Rd -valued process α = (αt ) with finite energy: EP [ 0 |αt |2 dt] < ∞ such that
Z T Z T
dP 1
ln = αt .dXt − |αt |2 dt, (2.2)
dW 0 2 0
3
Rt
and Xt − 0 αs ds, 0 ≤ t ≤ T , is a Brownian motion under P. We then have
h1 Z T i
H(P|W) = EP |αt |2 dt .
2 0
where W is a Brownian motion under P, A is the set of Rd -valued F-adapted processes s.t.
RT
EP [ 0 |αt |2 dt] < ∞, and (Xt1 , . . . , XtN ) ∼ µ is the usual notation for P ◦ (Xt1 , . . . , XtN )−1
P
Rt
where AµT is the set of controls α in A satisfying (Xt1 , . . . , XtN ) ∼ µ with Xt = 0 αs ds+Wt .
Our goal is to prove the existence of an optimal control α∗ that can be explicitly derived,
and then used to generate samples of the time series distribution µ via the probability
measure P∗ on Ω, i.e., the simulation of the optimal diffusion process X controlled by the
drift α∗ .
for (x1 , . . . , xN ) ∈ (Rd )N (with the convention that t0 = 0, x0 = 0). The measure µ is
absolutely continuous with respect to µW T , and we shall assume that its relative entropy is
finite, i.e.,
Z
µ
H(µ|µWT ) = ln W dµ < ∞.
µT
4
Rt
Theorem 3.1. The diffusion process Xt = 0 αs∗ ds + Wt , 0 ≤ t ≤ T , with α∗ defined as
where we set X ti = (Xt1 , . . . , Xti ), induces a probability measure P∗ = µµW (Xt1 , . . . , XtN )W,
T
which solves the Schrödinger bridge time series problem. Moreover, we have
Proof. First, observe that EW [ µµW (Xt1 , . . . , XtN )] = 1, and thus one can define a probability
T
measure P∗ W with density process
h dP∗ i h µ i
Zt = EW Ft = EW (Xt1 , . . . , X t ) F t , 0 ≤ t ≤ T.
dW µW
T
N
Notice from the Markov and Gaussian properties of the Brownian motion that for t ∈
[ti , ti+1 ), i = 0, . . . , N − 1, we have Zt = hi (t, Xt ; X ti ), where for a path xi = (x1 , . . . , xi )
∈ (Rd )i , hi (; xi ) is defined on [ti , ti+1 ) × Rd by
h µ p √ i
hi (t, x; xi ) = EY ∼N (0,Id ) W (xi , x + ti+1 − tY, . . . , x + tN − tY )
µT
for t ∈ [ti , ti+1 ), x ∈ Rd , and EY ∼N (0,Id ) [.] is the expectation when Y is distributed ac-
cording to the Gaussian law N (0, Id ). Moreover, by the law of conditional expectations,
we have
h i
hi (t, x; xi ) = EW hi+1 (ti+1 , Xti+1 ; xi , Xti+1 ) Xt = x ,
µ
with the convention that hN (tN , x; xN −1 , x) = µW
(x1 , . . . , xN −1 , x). Therefore, for i
T
= 0, . . . , N − 1, and x ∈ (Rd )i , the function (t, x) 7→ hi (t, x; xi ) is a strictly positive
C 1,2 ([ti , ti+1 ) × Rd ) ∩ C 0 ([ti , ti+1 ] × Rd ) classical solution to the heat equation
∂hi (.; xi ) 1
+ ∆x hi (.; xi ) = 0, on [ti , ti+1 ) × Rd ,
∂t 2
with the terminal condition: hi (ti+1 , x; xi ) = hi+1 (ti+1 , x; xi , x) (here ∆x is the Laplacian
operator). By applying Itô’s formula to the martingale density process Z of P∗ under the
Wiener measure W, we derive
5
for i = 0, . . . , N − 1. Thus, by defining the process α∗ by αt∗ = ∇x ln hi (t, Xt ; X ti ), for t
∈ [ti , ti+1 ), i = 0, . . . , N − 1, we have
dP∗ Z T 1 T ∗2
Z
= exp αt∗ dXt − |αt | dt ,
dW 0 2 0
Rt
and by Girsanov’s theorem, Xt − 0 αs∗ ds is a Brownian motion under P∗ . On the other
hand, by definition of P∗ , and Bayes formula, we have for any bounded measurable function
ϕ on (Rd )N :
h µ i
EP∗ ϕ(Xt1 , . . . , XtN ) = EW W (Xt1 , . . . , XtN )ϕ(Xt1 , . . . , XtN )
µT
Z
µ
= (x1 , . . . , xN )ϕ(x1 , . . . , XN )µWT (x1 , . . . , xN )dx1 . . . dxN
µW
T
Z
= ϕ(x1 , . . . , XN )µ(x1 , . . . , xN )dx1 . . . dxN ,
P∗
which shows that (Xt1 , . . . , XtN ) ∼ µ. Moreover, by noting that
hZ T
1 ∗2 i h dP∗ i
∗
J(α ) = E P∗ |αt | dt = EP∗ ln = H(P∗ |W) = H(µ|µW
T ) < ∞,
0 2 dW
P∗
where we used in the last inequality the fact that (Xt1 , . . . , XtN ) ∼ µ, this shows in
particular that α∗ ∈ AµT .
It remains to show that for any α ∈ AµT associated to a probability measure P W
with density given by (2.2), i.e. J(α) = H(P|W), we have
J(α) ≥ H(µ|µW
T ). (3.2)
Rt
For this, we write from Bayes formula and since Wt = Xt − 0 αs ds is a Brownian motion
under P by Girsanov theorem:
h µ i
1 = EW W (Xt1 , . . . , XtN )
µT
Z T
1 T
Z
h µ i
= EP exp ln W (Xt1 , . . . , XtN ) − αt dWt − |αt |2 dt
µT 0 2 0
Z T
1 T
Z
h µ i
≥ exp EP ln W (Xt1 , . . . , XtN ) − αt dWt − |αt |2 dt
µT 0 2 0
= exp H(µ|µW T ) − J(α) ,
P
where we use Jensen’s inequality, and the fact that (Xt1 , . . . , XtN ) ∼ µ in the last equality.
This proves the required inequality (3.2), and ends the proof.
6
Remark 3.2. The optimal drift of the Schrödinger bridge time series diffusion is in general
path-dependent: it depends at given time t not only on its current state Xt , but also on
the past values X η(t) = (Xt1 , . . . , Xη(t) ), where η(t) = max{ti : ti ≤ t}, and we have:
Moreover, the proof of the above theorem shows that this drift function is explicitly given
by
∇x hi (t, x; xi )
a∗ (t, x; xi ) = , t ∈ [ti , ti+1 ), xi ∈ (Rd )i , x ∈ Rd , (3.3)
hi (t, x; xi )
for i = 0, . . . , N − 1, where
h p √ i
hi (t, x; xi ) = EY ∼N (0,Id ) ρ(xi , x + ti+1 − tY, . . . , x + tN − tY ) ,
µ
with ρ := µW
the density ratio.
T
The following result states an alternate representation of the drift function that will
be useful in the next section for estimation.
Proposition 3.3. For i = 0, . . . , N − 1, t ∈ [ti , ti+1 ), xi = (x1 , . . . , xi ) ∈ (Rd )i , x ∈ Rd , we
have
∗ 1 Eµ (Xti+1 − x)Fi (t, xi , x, Xti+1 ) X ti = xi
a (t, x; xi ) = , (3.4)
ti+1 − t Eµ Fi (t, xi , x, Xti+1 ) X ti = xi
where
|xi+1 − x|2 |xi+1 − xi |2
Fi (t, xi , x, xi+1 ) = exp − + ,
2(ti+1 − t) 2(ti+1 − ti )
and Eµ [·] denotes the expectation under µ.
Proof. Fix i ∈ J0, N − 1K, and t ∈ [ti , ti+1 ). From the expression of µW T in (3.1), we have
µ
EW W (Xt1 , · · · , XtN ) (Xt1 , · · · , Xti ) = (x1 , · · · , xi ), Xt = x (3.5)
µT
NY −1
|xi+1 − x|2 |xj+1 − xj |2
Z
µ
=C (x1 , · · · , xN ) exp − exp − dxi+1 · · · dxN
µW
T 2(ti+1 − t)
j=i+1
2∆tj
Z
µ(x1 , . . . , xN ) h i
= C Fi (t, xi , x, xi+1 ) dxi+1 · · · dxN = CEµ Fi (t, xi , x, Xti+1 ) X ti = xi ,
µi (x1 , . . . , xi )
where C is a constant varying from line to line and depending only on t and xi =
(x1 , · · · , xi ), but not on x, and µi is the density of (Xt1 , · · · , Xti ) under µ, i.e.,
Z
µi (x1 , · · · , xi ) = µ(x1 , · · · , xN )dxi+1 · · · dxN .
7
By plugging the new expression (3.5) into a∗ and differentiating with respect to x, we then
get
∗ 1 Eµ (Xti+1 − x)Fi (t, xi , x, Xti+1 ) X ti = xi
a (t, x; xi ) = .
ti+1 − t Eµ Fi (t, xi , x, Xti+1 ) X ti = xi
Remark 3.4. In the case where µ is the distribution arising from a Markov chain, i.e., in
Q −1
the form µ(dx1 , . . . , xN ) = N d
i=1 νi (xi , dxi+1 ), for some transition kernels νi on R , then
the conditional expectations in (3.4) will depend on the past values X ti = (Xt1 , . . . , Xti )
only via the last value Xti .
4 Generative learning
From Theorem 3.1, we can run an Euler scheme for simulating the Schrödinger bridge
diffusion, and then samples of the target distribution µ. For that purpose, we need an
accurate estimation of the drift terms, i.e., of the functions a∗i , for i = 0, . . . , N − 1. We
propose several estimation methods. In the sequel, for a probability measure ν on (Rd )N ,
we denote by Eν [·] the expectation under the distribution ν.
(m) (m
Therefore, given data samples X (m) = (Xt1 , . . . , XtN ), m = 1, . . . , M from µ, and using
samples Y (m) from µWT , we estimate the density ratio ρ by
ρ̂ = exp(rθ̂ )
where rθ̂ is the neural network that minimizes the empirical logistic loss function:
M
1 X
ln 1 + exp(−r(X (m) ) + ln 1 + exp(r(Y (m) ) .
θ 7→
M
m=1
By writing from (3.3) the Schrödinger drift a∗ as
h √ √ i
EY ∼N (0,Id ) ρ∇x ρ(xi , x + ti+1 − tY, . . . , x + tN − tY )
a∗ (t, x; xi ) = h √ √ i , (4.1)
EY ∼N (0,Id ) ρ(xi , x + ti+1 − tY, . . . , x + tN − tY )
for t ∈ [ti , ti+1 ), we obtain an estimator of a∗ by plugging into (4.1) the estimate ρ̂, and rθ̂
of ρ, and ln ρ, and then computing the expectation with Monte-Carlo approximations from
samples in N (0, Id ). Notice that this method is very costly as it requires in addition to
the training of the neural networks for estimating the density ratio, another Monte-Carlo
sampling for estimating finally the drift.
8
Kernel estimation of the drift. In order to overcome the computational issue of the above
estimation method, we propose an alternative approach that relies on the representation
of the drift term in Proposition 3.3. Indeed, the key feature of the formula (3.4) is that it
involves (conditional) expectations under the target distribution µ, which is amenable to
direct estimation using data samples. For the approximation of the conditional expecta-
tion, we can then use the classical kernel methods.
(m) (m)
From data samples X (m) = (Xt1 , . . . , XtN ), m = 1, . . . , M from µ, the Nadaraya-
Watson estimator of the drift function is given by
M i
(m) (m) (m) (m)
X Y
(Xti+1 − x)Fi (t, Xti , x, Xti+1 ) Kh (xj − Xtj )
1 m=1 j=1
â(t, x; xi ) = , (4.2)
ti+1 − t M i
(m) (m) (m)
X Y
Fi (t, Xti , x, Xti+1 ) Kh (xj − Xtj )
m=1 j=1
is approximated by Ψθf (t, pti , x), where Ψθf ∈ N Nd+k+1,d is a feed-forward neural network
with input dimension 1 + k + d, and output dimension d, and pti is an output vector of
9
dimension k from an LSTM network, i.e., pti = ψθr (xi ) with ψθr ∈ LSTMi,d,k at time ti .
This neural network is trained by minimizing over the parameter θ = (θf , θr ) the quadratic
loss function
N −1
X 2
L(θ) = Ê (Xti+1 − X)F (τ, Xti , X, Xti+1 ) − Ψθf (τ, ψθr (X ti ), X) .
i=0
Here Ê is the empirical loss expectation where (Xt1 , . . . , XtN ) are sampled from the data
distribution µ, τ is sampled according to an uniform law on [ti , ti+1 ), and X is sampled
e.g. from a Gaussian law with mean Xti for i = 0, . . . , N − 1. The conditional expectation
function in the denominator is similarly approximated.
The output of this neural network training yields an approximation:
of the drift term function, which is then used for generating samples (Xt1 , . . . , XtN ) of µ
from the simulation of the diffusion:
end
Set xi+1 = yN π .
end
Return: x1 , · · · , xN
10
5 Numerical experiments
In this section, we demonstrate the effectiveness of our SBTS algorithm on several examples
of time series models, as well as on real data sets for an application to deep hedging. The
algorithms are performed on a computer with the following characteristics: Intel(R) i7-
7500U CPU @ 2.7GHz, 2 Core(s).
• Marginal metrics for quantifying how well are the marginal distributions from the
generated samples compared to the data ones. These include
• Temporal dynamics metrics for quantifying the ability of the generator to capture
the time structure of the P
time series data. We compute the empirical distribution of
the quadratic variation: i |Xti+1 − Xti |2 .
• Correlation structure for evaluating the ability of the generator to capture the multi-
dimensional structure of the time series. We shall compare the empirical covariance
or correlation matrix induced by the generator SBTS and the ones from the data
samples.
where the noises εi ∼ N (0, σi2 ), i = 1, . . . , 3 are mutually independent. The model param-
eters are b = 0.7, σ1 = 0.1, σ2 = σ3 = 0.05 and β1 = β2 = −1.
We use samples of size M = 1000 for simulated data of the AR model. The drift of
the SBTS diffusion is estimated with a kernel of bandwith h = 0.05, and simulated from
euler scheme with N π = 100. The runtime for generating 500 paths of SBTS is 8 seconds.
In Figure 2, we plot the empirical distribution of each pair (Xti , Xtj ) from the AR
model, and from the generated SBTS. We also show the marginal empirical distributions.
Table 1 presents the marginal metrics for the AR model and generator (p-value and per-
centiles at level 5% and 95%). In Table 2, we give the difference between the empirical
correlation from generated samples and AR data samples.
11
Data Data Data
SBTS SBTS SBTS
Xt2
Xt3
Xt3
Xt1 Xt1 Xt2
Figure 2: Comparison between the true and generated distribution for each couple (Xti , Xtj ) with
i, j ∈ J1, 3K with i 6= j
Table 1: Marginal metrics for AR model and generator (q̃ for percentile)
Table 2: Difference between empirical correlation from generated samples and reference samples
with α0 = 5, α1 = 0.4, α2 = 0.1, and the noises εti ∼ N (0, 0.1), i = 1, . . . , N , are i.i.d.
The size of the time series is N = 60.
The hyperparameters for the training and generation of SBTS are M = 1000, N π =
100, and a bandwith for the kernel estimation h = 0.2 (larger than for the AR model since
by nature the GARCH process is more ”volatile”). The runtime for generating 1000 paths
is 120 seconds.
In Figure 3, we plot four sample paths of the GARCH process to be compared with
four sample paths of the SBTS. Figure 4 represents samples plot of the joint distribution
between (Xt1 and the terminal value XtN ) of the time series. Figure 5 provides some
metrics to evaluate the performance of SBTS. On the left, we represent the p-value for
12
each of the marginals of the generated SBTS. On the right, we compute for each marginal
index i = 1, . . . , N = 60, the difference between ρi and ρ̂i where ρi (resp. ρ̂i ) is the sum
over j of the empirical correlation between Xi and Xtj from GARCH (resp. generated
SBTS), and plot its mean and standard deviation.
0.8 0.8
0.4 0.4
Garch
SBTS
0.0 0.0
0.4 0.4
0.8 0.8
0 10 20 30 40 50 60 0 10 20 30 40 50 60
time time
Figure 3: Samples path of reference GARCH (left) and generator SBTS (right)
0.8 SBTS
Data
0.6
0.4
0.2
XtN
0.0
0.2
0.4
0.6
0.8
0.8 0.6 0.4 0.2 0.0 0.2 0.4 0.6
Xt1
5%
Correlation
p-value
50% 0%
5%
10%
10%
0 10 20 30 40 50 60 0 10 20 30 40 50 60
Marginal index Marginal index
Figure 5: Left: p-value for the marginals Xti . Right: Difference between the term-by-term empirical
correlation from generated samples and reference samples.
13
1000 sample paths, and the hyperparameters used for the simulation are N π = 100, with
bandwith h = 0.05 for the kernel estimation of the Schrödinger drift. The runtime for
1000 paths is 120 seconds.
6 % 7 6
I % P
W L P H W L P H
Figure 6: Samples path of reference FBM (left) and generator SBTS (right)
Figure 7 represents the covariance matrix of (Xt1 , . . . , XtN ) for N = 60 of the FBM
and of the generated
PSBTS, while we plot in Figure 8 the empirical distribution of the
N −1
quadratic variation i=0 |Xti+1 − Xti |2 for the FBM and the SBTS.
1.0
1.0
Theoretical Covariance Matrix Generative Model
1 1
0.8 0.8
0.6 0.6
0.4 0.4
0.2 0.2
60 60
1 60 1 60
0.0 0.0
14
6 % 7 6 6 % 7 6
I % P I % P
'