EE746 - Report
EE746 - Report
TOPIC:
Improving LSM performance by Deep
time delay reservoir computing:-
Machine learning approaches can generate models for forecasting the behaviour of
dynamical systems using only observed data, but reservoir computing is especially
well-suited for learning such models.
RCs are based on a recurrent artificial neural network with a pool of interconnected
neurons, an input layer feeding observed data to the network, and an output layer
weighing the network states. They perform as well as other ML methods on dynamical
systems tasks but have smaller data set requirements.
But in RC we use nonlinearity in the reservoir and linearity in the readout layer this
seems to be very complex when there is a small dataset so we move to next-generation
reservoir computing where we interchange the linearity between the reservoir and
readout. The next generation RC (NG-RC) is presented that can perform equally well as
an optimized RC on three challenging RC benchmark problems:
Working of TD-RC:-
● since no output signals (e.g., linear regression) are generated at each layer
separately. The input enters only the first layer, and each consecutive layer
receives only the dynamical state of the previous layer. As a result, sequential
training of the layers isn’t performed.
● We consider there are L nodes. They are coupled unidirectionally with
self-feedback after a delay τl. The nodes with their corresponding feedback loops
are referred to as layers in the following because of the unidirectional topology
and their high-dimensional dynamics.
● 1.) According to experiments in general, additional layers can improve the
performance of deep TRC as compared to single-layer TRC.
● The deep TRC enables faster computation by a constant separation of nodes θ,
i.e., a deep TRC with L = 5 layers is 5 times faster than the single-layer TRC with
the same amount of total nodes in line with a 5 times shorter clock cycle.
Part 1: Introduction
● Examines the impact of the mismatch between delay time (τ) and clock cycle
time (τ') on TDRC.
● Suggests that deviations from a perfect match between τ and τ' can enhance the
system's capabilities, particularly when using a linear activation function.
1. τ: Indicates the time duration over which past input signals are retained and
integrated into the current state of the system. It is essentially a measure of the
memory capacity of the system. A larger τ allows the system to retain
information from further back in time, whereas a smaller τ provides shorter
memory.
2. τ': the clock cycle time of the system. In other words, it denotes the time interval
at which the system updates or processes its state. τ' acts as a reference time scale
for the system's operation.
Example of a resonant case scenario depicting the same task with values given above
Discussion:
● Discusses a notable increase in computing error and a drop in linear memory
capacity for resonant cases of bτ' ≈ aτ, with a, b ∈ N, particularly when b is small.
● The resulting memory capacity will be small for cases where τ and τ ′ are
resonant because the information within the equivalent network will be
overwritten by new inputs very quickly.
● Suggests that these results are not limited to the linear case and can be expected
to apply in more general situations. We can go to higher orders using the same
methods given in the paper.
Conclusion:
As mentioned we went through the first RC models used, which are of least
optimized properties. Then came the NG-RC model with slight improvement
still not giving the best results. Then came a time-delayed model with optimized
performance which requires even less training data set. Even in the time-delayed
RC model(TDRC) synchronous model doesn’t perform very well, opposite to
what is generally expected.
So, we are done with the literature survey. The next step is implementation. ;)