PaperThe following article is Open access

Chaotic attractor reconstruction using small reservoirs—the influence of topology

Published 27 August 2024 © 2024 The Author(s). Published by IOP Publishing Ltd
, , Citation Lina Jaurigue 2024 Mach. Learn.: Sci. Technol. 5 035058DOI 10.1088/2632-2153/ad6ee8

2632-2153/5/3/035058

Abstract

Forecasting timeseries based upon measured data is needed in a wide range of applications and has been the subject of extensive research. A particularly challenging task is the forecasting of timeseries generated by chaotic dynamics. In recent years reservoir computing has been shown to be an effective method of forecasting chaotic dynamics and reconstructing chaotic attractors from data. In this work strides are made toward smaller and lower complexity reservoirs with the goal of improved hardware implementability and more reliable production of adequate surrogate models. We show that a reservoir of uncoupled nodes more reliably produces long term timeseries predictions than more complex reservoir topologies. We then link the improved attractor reconstruction of the uncoupled reservoir with smaller spectral radii of the resulting surrogate systems. These results indicate that, the node degree plays an important role in determining whether the desired dynamics will be stable in the autonomous surrogate system which is attained via closed-loop operation of the trained reservoir. In terms of hardware implementability, uncoupled nodes would allow for greater freedom in the hardware architecture because no complex coupling setups are needed and because, for uncoupled nodes, the system response is equivalent for space and time multiplexing.

Export citation and abstractBibTeXRIS

Original content from this work may be used under the terms of the Creative Commons Attribution 4.0 license. Any further distribution of this work must maintain attribution to the author(s) and the title of the work, journal citation and DOI.

1. Introduction

Data driven models are needed in a wide range of applications and there are many statistical and machine learning methods of developing them. Low training costs and energy efficient approaches are paramount for wide spread use, particularly in edge devices. For data driven timeseries forecasting, particularly chaotic timeseries forecasting, reservoir computing has been shown to perform well. Reservoir computing is a machine learning approach wherein a dynamical system is driven with an input timeseries and the responses of the system to the input are linearly combined to approximate the desired output timeseries [1, 2]. In contrast to other machine learning methods, the training is restricted to the linear output weights and can be performed via linear regression. Due to the low training cost and the fact that the dynamical system remains unchanged in the training procedure, reservoir computing is well suited for hardware implementation.

In this work strides are made toward smaller and lower complexity reservoirs which can lead to easier hardware implementation and interpretable surrogate models. We show that a small reservoir of uncoupled nodes more reliably produces better long term timeseries prediction performance than more complex reservoir topologies. These results have implications for the hardware implementability, as uncoupled nodes are easier to construct and can equally be mutliplexed in space and time, giving more freedom for possible hardware architectures. Furthermore, to gain understanding about the connections between the reservoir topology and both the short- and long-term prediction performance, error measures which quantify the training accuracy, short-term predictions and the accuracy of the attractor reconstruction are analysed, as well as properties of the trained system. It is shown that an uncoupled reservoir topology leads to a smaller spectral radius of the final trained system and that this correlates with more stable attractor reconstruction.

The low training cost of reservoir computing has also lead to extensive theoretical and numerical studies. These have typically used larger numbers of reservoir nodes than the 20 used in this study. In [3] the closed loop operation of a reservoir trained on one-step-ahead prediction for the Lorenz and Kuramoto–Sivashinsky systems was investigated. The focus of this study was the estimation of Lyaponov exponents of the Lorenz and Kuramoto–Sivashinsky systems and reservoirs of 300 nodes were used. In [4] conditions for good short and long term prediction were established. The reservoir used in this study were sparse Erdős–Renyi networks of 2000 nodes. In [5] the authors used a valid forecasting time measure, the correlation dimension and the largest Lyaponov exponent to quantify short and long term prediction quality. They investigate the influence of small world, Erdős-Renyi and scale free network topologies, as well as the influence of the spectral radius of the original system. Networks of 300 nodes were used. In [6] various reservoir topologies were also compared for networks of 100 nodes, and Bayesian optimization was applied for hyperparameter optimisation. Optimisation and validation strategies for chaotic attractor reconstruction are investigated in [7]. In [8] various block diagonal topologies were tested, with networks of 400–600 nodes. A block size of one is specifically not considered. The authors of [8] compare the block diagonal topology with other reservoir computing related approaches in [9] and find good performance compared with the SINDy [10] approach, nonlinear vector regression (referred to as 'next generation reservoir computing' [11]) and reservoir computing using a Erdős-Renyi network. Performance comparisons of reservoir computing with other machine learning and statistical methods are given in [12]. Applications of trained chaotic attractors using reservoir computing have been investigated. For example, in [13] chaos synchronisation and cryptography applications are investigated and in [14] it is shown that such systems can be used to infer unseen attractors.

Several hybrid reservoir computing approaches for chaotic timeseries forecasting have been developed. One approach is to couple the reservoir at the output layer with a so-called knowledge-based model [15]. This approach requires an approximate model for the underlying dynamics of the system to be reconstructed and has been shown to perform well, for example on atmospheric modelling [16]. Often an knowledge based model is not available, therefore in [17] a hybrid reservoir-SINDy approach was investigate. There the knowledge-based model was replaced with a model produced from data using the SINDy approach [10]. Another important topic is the influence of input noise on the quality of the reconstruction of chaotic attractors. The authors of [18] have shown that injection noise can lead to improved short- and long-term predictions. And, in [19] the influence of input noise on the stability of reconstructed attractors, as well as the frequency components whereof, are investigated.

For many applications it is not only the reconstruction or forecasting of the dynamics that are relevant, it is often also desirable to gain insights into the underlying systems which are producing the data. This is difficult or even impossible with most machine learning approaches as they do not produce interpretable models. One key step toward an interpretable model is to greatly restrict the size of the model. In terms of reservoir computing, this means restricting the dynamical degrees of freedom of the system used as the reservoir. A commonly used type of reservoir is a network of randomly connected nodes with nonlinear activation functions. For such a reservoir, the size of the reservoir is given by the number of nodes. In the context of interpretable models, the dynamics of the final trained system are of particular relevance, and these have so far not been thoroughly investigated.

In this work the influence of the reservoir topology on the ability to predict the short and long term behaviour of the Lorenz chaotic attractor is investigated, while restricting the number of reservoir nodes. The paper is structured as follows. In section 2 the reservoir computing appproach is explain and the reservoir model and network topologies used in this work are introduced. In section 3 the timeseries prediction task is presented and open- and closed-open predictions are explained. In section 4 various error measures are defined and simulation parameters are given. In section 5 the results are presented and discussed. Finally, the conclusions are given in section 6.

2. Reservoir computing

In reservoir computing the response of a dynamical system to an input sequence is sampled a number of times and then a linear weighted sum of these sampled responses is used to make predictions or assign labels. In contrast to other machine learning approaches, the training is restricted to the linear output weights, as illustrated in figure 1(a). Training the reservoir requires the following steps:

  • Feed a sequence of inputs of length KB + KT into the reservoir.
  • For each input step sample N responses of the reservoir.
  • Disregard the responses to the first KB inputs, where KB is long enough that the reservoir state no longer depends on the initial state.
  • Collect the N responses to the remaining KT inputs into a time ordered state matrix $\textbf{S}\in\mathbb{R}^{K_T}\times \mathbb{R}^{N+1}$, where a bias column of ones has been added. The entry $s_{k,j}$ of the state matrix then corresponds to the $j{\textrm{th}}$ sampled response of the reservoir to the $K_B+k{\textrm{th}}$ input.
  • Find the vector of weights $\textbf{w} \in \mathbb{R}^{N+1}$ that minimises the difference between the target output $\textbf{y}\in \mathbb{R}^{K_T}$ and the output $\hat{\mathbf{y}} = \textbf{ S}\textbf{w} \in \mathbb{R}^{K_T}$. The solution to this problem is given by
    Equation (1)
    using the Moore-Penrose pseudoinverse, where λ is the Tikhonov regularisation parameter and $\textbf{I} \in \mathbb{R}^{N+1}\times \mathbb{R}^{N+1}$ the identity matrix. For multiple targets $\mathbf{Y} = \left[\mathbf{y_1},..,\mathbf{y_q}\right]$ the matrix of output weights $\mathbf{W} = \left[\mathbf{w_1},..,\mathbf{w_q}\right] \in \mathbb{R}^{N+1}\times \mathbb{R}^{q} $ must be calculated.
Figure 1. Refer to the following caption and surrounding text.

Figure 1. (a) Sketch of the reservoir computing approach. An input is fed into the reservoir with fixed random weights $\textbf{W}_{in}$ (grey arrows) and the output is produced by linearly combining states of the reservoir with trained output weights W (white arrows). (b) In open-loop operation externally sourced input data is fed into the reservoir to produce the output sequence. (c) In closed-loop operation the trained system is run autonomously by feeding the output back in as the next input step.

Standard image High-resolution image

2.1. The reservoir

Many different dynamical systems are suitable reservoirs and have been investigated both numerically and experimentally. These include optical [2024], optoelectronic [25, 26], micromechanical [27, 28], quantum reservoirs [2932]. In this study networks of the following form are used as the reservoirs:

Equation (2)

where $\textbf{x}\left( k\right)\in \mathbb{R}^M$ is the vector describing the state of the M network nodes at discrete time k, $\textbf{W}_{\mathrm{int}}\in \mathbb{R}^{M}\times \mathbb{R}^{M}$ is the matrix of internal reservoir node couplings (adjacency matrix), $\textbf{i}_{\mathrm{in}}\left( k\right) \in \mathbb{R}^p$ is the vector of p input signals at time k and $\textbf{W}_{\mathrm{in}}\in \mathbb{R}^M\times \mathbb{R}^{p}$ is the matrix of input weights. The function $\tanh\left( \cdot\right)$ is the nonlinear function which is applied element wise and commonly referred to as the activation function. The $k{\textrm{th}}$ row of the state matrix is filled by sampling N of the reservoir states $\textbf{x}\left( k\right)$, with $N \unicode{x2A7D} M$ and $k\in \left[0,K_T\right)$. For N = M the elements of the state matrix are $s_{k,j} = x_j\left( k\right)$ with $x_j\left( k\right)$ ($j\in \left[0,..M\right)$) being the elements of the reservoir state vector $\textbf{x}\left( k\right)$.

In this study we consider the case that all nodes are sampled, N = M, and compare three reservoir coupling matrix types; a random coupling matrix $\tilde{\mathbf{W}}_{\mathrm{rand}}$ with entries chosen from a uniform distribution on the interval $[0,1]$, an identity matrix $\tilde{\mathbf{W}}_{\mathrm{id}}$ and a diagonal matrix $\tilde{\mathbf{W}}_{\mathrm{ring}}$ with the entries shifted one down from the main diagonal:

Equation (3)

In each case the matrices are rescaled such that $\textbf{W}_{\mathrm{int}} = \rho_R \tilde{\mathbf{W}}a_0$, where a0 is the scaling factor needed such that ρR is the spectral radius of $\textbf{W}_{\mathrm{int}}$. The spectral radius is defined as

Equation (4)

where $\lambda_1,{\ldots},\lambda_N$ are the eigenvalues of the N×N matrix W. For the ring and identity matrices a0 = 1 and for the random matrix a0 depends on the randomly drawn values.

For $\textbf{W}_{\mathrm{int}} = \textbf{W}_{\mathrm{id}}$ the nodes are uncoupled in the training phase, i.e. each node only couples to its own state in the previous time step and the node degree is zero. This coupling scheme is identical to delay-based time-multiplexed reservoir computing when the feedback delay time of the reservoir is equal to the input clocktime [33]. When $\textbf{W}_{\mathrm{int}} = \textbf{W}_{\mathrm{ring}}$, each node is coupled to one neighbour, thereby forming a unidirectional ring. Sketches of the three coupling topologies are shown in figure 2.

Figure 2. Refer to the following caption and surrounding text.

Figure 2. Reservoir network topologies corresponding to the coupling matrices $\tilde{\mathbf{W}}_{\mathrm{rand}}$ (Random), $\tilde{\mathbf{W}}_{\mathrm{ring}}$ (Ring) and $\tilde{\mathbf{W}}_{\mathrm{id}}$ (Uncoupled).

Standard image High-resolution image

If the coupling matrices are filled with zeros, then the reservoir reduces to an extreme learning machine (ELM), i.e. a single layer of uncoupled (also no self-coupling) nodes with random input weights [34]. An example of a hardware implemented ELM is presented in [35], where an array of microresonators is used as a photonic ELM. In this limit, the system has no memory of past inputs and the method can also be interpreted as nonlinear vector regression where the feature vector contains various nonlinear transforms of the current input step. In the results section this limit is included and corresponds to the $\rho_R = 0$ results (see figures 36 in section 5).

Figure 3. Refer to the following caption and surrounding text.

Figure 3. NRMSE of the X, Y and Z components for one-step-ahead prediction in open-loop operation as a function of spectral radius ρ of the coupling matrix. The mean over 100 realisations of the random weights and Lorenz trajectories is depicted for the uncoupled (teal), random (orange) and ring (grey) reservoirs. The error bars indicate the standard deviation.

Standard image High-resolution image

3. Timeseries prediction

The aim of this study is to investigate the influence of the topology on the ability of the reservoir to predict chaotic dynamics and act as a surrogate model. To perform multi-step-ahead predictions and to act as a surrogate model the reservoir is trained to predict a timeseries one step ahead and then in the prediction phase the predicted output is used as the next input step. This is the so-called closed-loop operation and is illustrated in figure 1(c) (open-loop operation is shown in figure 1(b)). By feeding the reservoir output back in as the next input, the system becomes autonomous. The desired result would then be that the dynamics of the trained autonomous system emulate those of the original system which produced the training data. In this study the Lorenz system [36] is used to test to reservoir performance. Results for the Rössler system are given in the supplemental material.

3.1. Lorenz 63 system

The Lorenz system [36] is given by

Equation (5)

With c1 = 10, c2 = 28 and $c_3 = 8/3$ this system exhibits chaotic dynamics and has a largest Lyaponov exponent of $\lambda_L\approx 0.91$. We generate the Lorenz timeseries using Runge-Kutta fourth order numerical integration with a time step of $h = 10^{-3}$. We use all three variables, sampled with a step size of $\Delta t = 0.1$ and each rescaled to the range $\left[0,1\right]$, as the reservoir input $\textbf{i}_{in}(k) = \left[X(k),Y(k),Z(k)\right]$ (X, Y, Z indicate the rescaled versions of x, y ,z). In the training phase the target is given by $\textbf{y}(k) = \left[X(k+1),Y(k+1),Z(k+1)\right]$. The step size of $\Delta t = 0.1$ is chosen, as it has been shown to be well suited for the reconstruction of the Lorenz attractor [37, 38].

3.2. The trained autonomous system

After training, the reservoir output can be written in terms of the reservoir state:

Equation (6)

where the reservoir state vector is extended by one term, $x_{N+1} = 1$, to include the bias. Labelling the entries of the output weight matrix W which correspond to the bias terms as $W_{N+1,i} = W{^\textrm{bias}_i}$, the reservoir outputs can be written as

Equation (7)

for $i\in \left[1,q\right]$, where q is the number of target sequences. Then the next input is given by $I_i(k+1) = \hat{y}_i(k)$, i.e. $\textbf{i}_{in}(k+1) = \left[\hat{y}_X(k),\hat{y}_Y(k),\hat{y}_Z(k)\right]$.

The reservoir then becomes an autonomous dynamical system described by

Equation (8)

where $\textbf{W}^a_{\mathrm{int}}$ is the modified coupling matrix and b is the vector of constant bias terms:

Equation (9)

with $\textbf{W}_{in} = [\textbf{w}_1^{\mathrm{in}},{\ldots},\textbf{w}_q^{\mathrm{in}}]$. The righthand side of the final system now only dynamically depends on the previous states of the nodes.

For the case of reconstructing a three dimensional dynamical system (q = 3) the new coupling matrix is given by

Equation (10)

the entries of which depend on the initial reservoir, the input weights and the trained output weights. We will relate the spectral radius of this new coupling matrix to the ability of the reservoir to operate as a surrogate model. The spectral radius of $\textbf{W}^a_{\mathrm{int}}$ is calculated using equation (4) and will be denoted ρa.

4. Methods

4.1. Performance measures

4.1.1. Normalised root mean squared error (NRMSE)

The performance of the open-loop (see figure 1(b)) one-step-ahead prediction is quantified using the NRMSE, defined as

Equation (11)

where yk are the target values, $\hat{y}_{k}$ are the outputs produced by the reservoir computer, KT is the number of testing steps (i.e. the length of the vector y) and $\textrm{var}\left( \textbf{y} \right)$ is the variance of the target sequence.

4.1.2. Valid prediction time (VPT)

We quantify the short-term closed-loop (see figure 1(c) performance using a VPT as in [17]. The VPT is a measure which quantifies how long the closed-loop trained reservoir can be used to accurately predict a particular trajectory. This measure has an upper bound which is related to the initial separation between the true trajectory and the first prediction step by the reservoir and the largest Lyaponov exponent of the chaotic target system. The VPT is given by

Equation (12)

where t is the time determined by the discretisation step Δt of the training data set and number of iterations from the beginning of the prediction phase, and δVPT is a measure of the prediction error defined as:

Equation (13)

Here $\langle \cdot\rangle$ indicates the time average over the length of the target timeseries, $\textbf{y}(k) = [X(k+1),Y(k+1),$ $Z(k+1)]$ is the vector of target values at discrete time k and $\hat{\mathbf{y}}(k) = [\hat{X}(k+1),\hat{Y}(k+1),\hat{Z}(k+1)]$ is the vector of predicted values. The threshold of 0.4 used in equation (12) is chosen for the sake of comparability with other studies [15, 17]. For a real world problem this threshold would be determined by the error tolerances allowed for the particular task.

Various other definitions of prediction times are used in the literature. For example in [8] a forecast horizon is defined as the time at which the difference in the predicted and target trajectories exceeds the standard deviation of the target trajectory. Care must be taken when comparing these measures as they can lead to very different times.

4.1.3. Attractor deviation

To quantify how well the desired attractor is reconstructed we use an approach similar to that used in [18]. We divided the three-dimensional phase-space into cubes of volume $\mathrm dX\times \mathrm dY\times \mathrm dZ$. Each cube is assigned a value of one or zero; zero if the trajectory never passes through the cube and one if the trajectory passed through at least once. We do this for both the reconstructed trajectory and the target trajectory. We then calculate the sum of the absolute differences between the values assigned to each cube for the reconstructed and target trajectories:

Equation (14)

where $c^{\mathrm{true}(\mathrm{pred})}_{ijk}$ is one if the trajectory passed through the corresponding element of the phase space at least once and otherwise zero.

The purpose of the attractor deviation measure is quantify whether two trajectories reside within the same region of the three-dimensional phase-space. For this measure to be useful, the discretisation of the phase-space must be fine enough to distinguish dynamics on the desired attractor from different dynamical solutions, but it must not be so fine that different true trajectories on a chaotic attractor lead to very large ADev values. Additionally, the length of the test timeseries influences the ADev value, as for a longer timeseries more of the phase-space can be traversed. Therefore, the ADev value depends on the size of the cubes, on the length of the timeseries and on the target trajectory. We choose a cube dimension of $0.1\times 0.1\times 0.1$ and a timeseries length of 5000 time units ($5000\Delta t$), which is about 450 Lyaponov times ($1/\lambda_L\approx 1.1$) for the Lorenz 63 system with the standard parameters used here. In all subsequent results we also provide a reference ADev value which is obtained for two true Lorenz trajectories.

The ADev measure alone is not sufficient to determine whether a trajectory has remained on (or near) the desired attractor for the entire prediction phase. This is because a sudden collapse to the fixed point solution after a long oscillatory phase will not have a large influence on the ADev value. It is therefore also important to characterise the dynamics of the test trajectory near the end of the prediction phase, i.e. fixed point, periodic oscillations or irregular oscillation.

4.1.4. Power spectral density (PSD)

To compare the frequency components of the target and reconstructed trajectories we calculate the PSD of the Z component. This is done using the timeseries over the prediction horizon of $5000\Delta t$, applying a Hamming windowing function and then calculating the fast Fourier transform (FFT). The PSD is then given by

Equation (15)

where $\tilde{Z}(f)$ is the FFT of Z(t) with $t = k\Delta t$. Finally, the spectrum is smoothed by calculating a running average with a range of 20 points.

4.2. Simulation parameters

For all tasks we rescale the input and target sequences to the range between zero and one. Before beginning the training and testing phases, the reservoirs are initialised by feeding in a sequence of 10000 inputs $\textbf{i}_{\mathrm{in}}(k)$. In the training phase 10 000 inputs are used while in the testing phase 5000 inputs are used. The lengths of the training and testing phases were not optimised, as this was not the focus of this study. For the calculation of the output weights a regularisation parameter of $\lambda = 5\times 10^{-6}$ was used for all simulations.

In the next section results are presented in dependence of the spectral radius of the initial reservoir ρR. The testing error measures are averaged over 100 realisations of the random input weights $\textbf{W}_{\mathrm{in}}$ and coupling weights $\textbf{W}_{\mathrm{rand}}$. For each ρR value new random weights are drawn and the coupling matrix is rescaled such that the spectral radius is equal to the desired ρR. The input weights $\textbf{W}_{\mathrm{in}}$ are drawn from a uniform distribution on the interval $[0,1]$. In each realisation the initial conditions for the Lorenz target trajectory are chosen randomly for the prediction phase, meaning that the start of the prediction phase is randomly distributed over the Lorenz attractor. A reservoir of N = 20 nodes is used. Results for different node numbers are given in the supplemental material.

5. Results and discussion

To judge the forecasting capabilities of a reservoir, typically only the performance in closed-loop operation is considered. We take a different approach, and assess the quality of the one-step-ahead prediction by calculating the NRMSE on a test timeseries of 5000 steps in open-loop operation, i.e. the true Lorenz timeseries is used as the input throughout the testing phase. We do this to gain understanding about the influence of the training accuracy on the closed-loop prediction performance. Figure 3 shows the NRMSE for the X, Y and Z components of the Lorenz timeseries for the three reservoir topologies. As a function of the spectral radius of the coupling matrices, ρR, the NRMSEs are very similar for all three topologies, with the random network producing slightly lower errors. The lowest errors are achieved near $\rho_R = 0$, which is the ELM limit. For $\rho_R = 0$ the reservoir has no memory of past inputs, meaning that the predictions are based entirely on the current input step. This lack of memory is not detrimental to the task performance since the state of the Lorenz system at any point in time is fully defined by the current X, Y and Z values and these are all being provided as input. A study of the difference in reservoir performance for partial and full input information in given in [39].

In the closed-loop (autonomous) operation, the predicted X, Y and Z values are fed back into the reservoir as the input. In this case three types of resulting dynamics are found; diverging, oscillatory and fixed point. In the diverging case the dynamic variables become much larger than the target values and therefore the attractor is not successfully reconstructed. Since we are interested in accurate timeseries prediction and attractor reconstruction, it is not relevant in this study if the dynamics are truly diverging or if the resulting system produces large oscillations. We therefore set an upper limit of two and check if $|X|$, $|Y|$ and $|Z|$ remain below this value for the entire prediction phase. Trajectories that remain below this limit we refer to as bounded. In figure 4(a) the percentage of the realisations which were bounded is depicted as a function of the spectral radius. Here, a clear difference can be seen between the uncoupled reservoir and the random and ring topologies. The uncoupled reservoir produces near 90% bounded trajectories over a much large range of ρR. The bounded results can be separated into those that remain oscillatory for the entire prediction phase and those that collapse into a fixed point. The percentage of the results which remain oscillatory for the complete prediction phase is shown in figure 4(b). Here, a clear difference is also evident between uncoupled reservoir and the random and ring topologies. Figure 4 shows that the uncoupled reservoir produces bounded results more reliably and that fewer of the realisations collapse into a fixed point.

Figure 4. Refer to the following caption and surrounding text.

Figure 4. Percentage of the 100 realisations of the random weights and Lorenz trajectories for which the resulting closed-loop (autonomous) trajectory; (a) was bounded by $|X|,|Y|,|Z| \lt 2$ for the entire prediction phase of 5000 steps, (b) was bounded and remained oscillatory for the entire prediction phase of 5000 steps.

Standard image High-resolution image

To evaluate the short term closed-loop prediction performance the VPT is plotted in figures 5(a)–(c) for the uncoupled, random and ring topologies, respectively. For $\rho_R = 0$ all three systems are identical, and consequently show very similar performance as ρR tends to zero. The best performance in terms of the VPT is achieved near $\rho_R = 0$ for all three topologies. This correlates with the ρR values for which the lowest NRMSE errors are achieved in the open-loop operations, see figures 3. But, in the case of the uncoupled reservoir, this does not correspond to the ρR values for which bounded and oscillatory trajectories are produced most reliably, see figure 4. The finding that the highest VPT of to the Lorenz task are achieved at $\rho_R = 0$ are in agreement with the results presented in [40]. In [40] reservoir hyperparameters are optimised for various tasks and for the Lorenz task it is shown that the best prediction performance is achieved when the adjacency matrix density is zero, which is equivalent to $\rho_R = 0$ in this study.

Figure 5. Refer to the following caption and surrounding text.

Figure 5. Closed-loop prediction performance: (a)–(c) The VPT for the three reservoir topologies. (d)–(f) The attractor deviation ADev, with a reference ADev calculated for two true Lorenz trajectories indicated by the dashed line. (g)–(i) The time interval over which the trajectories were oscillatory. For the ADev and run time values only the bounded trajectories were considered. For all plots the median is plotted and the shaded regions show the 25th–75th percentiles.

Standard image High-resolution image

To judge the absolute performance it is useful to compare with other studies. However, when making such comparisons differences, in the task implementation and the error measures must be considered. In [17] the Lorenz task is implemented with the same time discretisation as in this study and the same definition for the VPT is used. There VPTs of approximately 2 are achieved using a network of 100 nodes. By coupling the reservoir with a SINDy model, this is improved to VPTs of about 3. In [15] a hybrid reservoir of 50 nodes coupled with a knowledge based model, resulted in VPTs of approximately 1.5. The VPTs in [17] and [15] are larger than in this study, however this was only achieved with larger node numbers and hybrid reservoir methods.

As well as information about the performance relative to other reservoir computing or machine learning approaches, we are also interested in the performance relative to the Lorenz system itself. The largest Lyaponov exponent of the Lorenz system indicates how quickly two trajectories on the true attractor will diverge given a certain difference in the initial conditions. Therefore, to obtain an objective benchmark with which the absolute performance in terms of the VPT can be compared, we calculate the VPT of the true Lorenz system given the error after the first prediction step (one prediction step is made using the trained reservoir, then this prediction is used as the initial conditions for the Lorenz system to generate a second timeseries with which to compare the target trajectory). This comparison is useful as it sets an upper limit on the VPT that can be expected given the accuracy in the one-step-ahead training stage. This comparison is depicted in figure 6 for the uncoupled reservoir. As is to be expected, the VPTs of the true Lorenz system are slightly larger, however the 75${\textrm{th}}$ percentile of the reservoir results coincide with the median of the true Lorenz VPTs, indicating considerable overlap in the performance.

Figure 6. Refer to the following caption and surrounding text.

Figure 6. VPT for the uncoupled reservoir (orange) and for the true Lorenz system (grey). For the true Lorenz VPTs the first prediction step of the trained reservoir was used as the initial conditions the Lorenz equations and these were then integrated to generate the predicted timeseries. The median is plotted and the shaded regions show the 25th–75th percentiles.

Standard image High-resolution image

Figure 6 shows that even for the true Lorenz system, given differences in the initial conditions of the order of the NRMSEs depicted in figure 3, trajectories diverge after one to three Lyaponov times ($1/\lambda_L\approx 1.1$). Therefore, to evaluate the quality of the attractor reconstruction over hundreds of Lyaponov times we use the attractor deviation measure introduced in section 4.1.3. The attractor deviation (ADev) is a measure of the difference in the three-dimensional phase space which is traversed by the trajectory of the target Lorenz timeseries and the predicted timeseries. Figures 5(d)–(f) shows the attractor deviation for the uncoupled, random and ring reservoirs. Only the realisations resulting in bounded trajectories were considered. The horizontal dashed line is a reference attractor deviation of two true Lorenz trajectories. The random and ring topologies show very similar performance, and their large attractor deviations can be explained by the low percentage of trajectories which do not collapse into a fixed point during the prediction phase (as shown in figure 4(b)). The uncoupled reservoir, on the other hand, shows quite low attractor deviation values corresponding to the ρR range for which a high percentage of the trajectories remain oscillatory for the entire prediction phase.

Since the attractor deviation measure does not capture any temporal information, it is also useful to compare the power spectral densities of the predicted and true trajectories as these are calculated from the dynamics over the entire prediction phase. Figure 7 shows phase-space plots, power spectral densities and the attractor deviation for three example trajectories generated using the uncoupled reservoir. Different random input weights and initial conditions of the Lorenz system were used to generate the trajectories. The trajectory in figures 7(c) and (d) collapses near the end of the prediction phase and corresponds to trajectory two in figure 8. Figures 7(e) and (f) corresponds to trajectory three in figure 8. As long as the trajectory remains on the attractor, the spectral features are reproduced quite well and are comparable with the performance achieved using a data-informed-reservoir hybrid model in [17], even though we have used a much smaller reservoir.

Figure 7. Refer to the following caption and surrounding text.

Figure 7. Phase spaces plots (a), (c), (e) and the corresponding PSD plots (b), (d), (f) for three example trajectories generated using the uncoupled reservoir in closed-loop operation. The predicted results are shown in orange and the target results in green. The trajectories in (c), (d) and (e), (f) correspond to trajectories two and three in figure 8, respectively.

Standard image High-resolution image
Figure 8. Refer to the following caption and surrounding text.

Figure 8. Timeseries of the X, Y and Z components for two predicted trajectories generated using the uncoupled reservoir in closed-loop operation. The predicted results are shown in orange and the target results in green. Trajectories two and three correspond to the trajectories in (c), (d) and (e), (f) of figure 7, respectively.

Standard image High-resolution image

The Lorenz-like dynamics achieved with the trained reservoir appear to be meta-stable, i.e. the solutions have small basins of attraction or are sensitive to perturbations in particular directions. This is indicated by the trajectories appearing Lorenz-like for hundreds of Lyaponov times and then suddenly collapsing to a stable fixed-point, as depicted in the timeseries shown in figure 8. Similar behaviour has also been reported in other works [5]. Comparing the results depicted in figures 35, a trade-off between the short-term prediction performance and the stable attractor reconstruction is evident.

For the reservoir to accurately reconstruct the Lorenz attractor, the Lorenz-like dynamics of the trained system must be stable. For the uncoupled reservoir a large percentage of the solutions appear to be stable or meta-stable over the prediction range. However, for the random and ring topologies most of the solutions diverge quickly, indicating that the desired solutions either do not exist in the final autonomous system or are unstable. To work towards understanding the difference in the results between the reservoir topologies, the spectral radius of the trained autonomous systems ρa is plotted in figure 9 as a function of the spectral radius of the original reservoir. For all three reservoirs ρa increases with ρR and for $\rho_R \gt 0.5$ the large ρa values correspond to short run times (time before the system diverges or collapses to a fixed point), as shown in figures 5(g)–(i). For $\rho_R \lt 0.35$, ρa of the uncoupled system is smaller than those of the random and ring reservoirs (see figure 9(b) for an enlargement of the $\rho_R \lt 0.5$ results). The region of low ρa for the uncoupled system corresponds to the regions of lower attractor deviation and longer run times in figures 5(d) and (g). These results indicate that the topology of the initial reservoir plays a deciding role in determining the stability of the desired solution in the final system. Specifically, less connectivity within the network leads to improved results. These results are in agreement with the findings of [8] where the authors showed improved performance for block diagonal coupling matrices with low node degree (zero node degree was not considered).

Figure 9. Refer to the following caption and surrounding text.

Figure 9. (a) Spectral radius of the trained autonomous system ρa as a function of the spectral radius of the original reservoir ρ. The solid lines indicate the mean and the error bars give the standard deviation. (b) Enlargement of the $\rho_R\in\left[0,0.5\right]$ range of (a).

Standard image High-resolution image

6. Conclusion

In this work the chaotic Lorenz attractor has been reconstructed using small reservoirs of only 20 nodes. We have shown that the short term performance depends on the error in the one-step-ahead prediction, but that the stable reconstruction of the attractor is related to dynamical properties of the trained autonomous system. By reducing the reservoir to a system of uncoupled nodes, smaller spectral radii of the trained system can be achieved compared with random or ring reservoirs. And, the parameter regions resulting in small spectral radii for the trained uncoupled reservoir correspond to more reliable reconstruction of the Lorenz attractor.

These results pave the way towards interpretable surrogate models. Due to the small size of the reservoir and the uncoupled reservoir nodes it becomes feasible to perform bifurcation analysis on the final system and additional optimisation can be performed after the training phase without effecting all reservoir nodes, for example pruning of selected nodes. Furthermore, hardware implementability is improved, as complex network topologies are not required and for a system of uncoupled nodes there is more freedom in the choice of multiplexing.

The insights gained in this study into the relationships between training accuracy, network topology and short- and long-term prediction performance can also be applied to related machine learning approaches. For example, nonlinear vector regression (referred to as 'next generation reservoir computing' [11]) has recently been shown to perform very well at chaotic timeseries forecasting. Nonlinear vector regression could be thought of a network of node degree zero (and no self-coupling) and would then be similar to the uncoupled reservoir in this study.

In this study all three dynamical variables of the Lorenz system were provided to the reservoir in the training stage. This meant that the reservoir did not need to have memory for the one-step-ahead training to perform well. In contrast, for tasks where only partial information is provided in the training stage the reservoir must perform a delay embedding and therefore requires memory [39, 41, 42]. Since the memory of the reservoir is determined by the coupling topology and the spectral radius, different dependencies on these properties can be expected for partial information tasks compared with the results for the full information Lorenz task presented in this study. In order to restrict the reservoir topology to uncoupled nodes for partial information tasks, but still incorporate the required memory, a possible solution could be to apply memory augmentation methods at the levels of the input or output layers. In [43] it was shown that adding task relevant timescales at the level of the input or output layers can improve the reservoir performance for partial information timeseries prediction tasks and decreases the sensitivity to other reservoir hyperparameters.

Data availability statement

The data is not yet in a form in which it can be made publicly available. The data that support the findings of this study are available upon reasonable request from the authors.

Please wait… references are loading.

Supplementary data (7.9 MB PDF)