Next Article in Journal
Analysis and Radiometric Calibration for Backscatter Intensity of Hyperspectral LiDAR Caused by Incident Angle Effect
Previous Article in Journal
Automated Non-Contact Respiratory Rate Monitoring of Neonates Based on Synchronous Evaluation of a 3D Time-of-Flight Camera and a Microwave Interferometric Radar Sensor
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhancing Performance of Reservoir Computing System Based on Coupled MEMS Resonators

1
The State Key Laboratory of Transducer Technology, Aerospace Information Research Institute, Chinese Academy of Sciences, Beijing 100010, China
2
School of Electronic, Electrical and Communication Engineering, University of Chinese Academy of Sciences, Beijing 100010, China
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(9), 2961; https://doi.org/10.3390/s21092961
Submission received: 7 March 2021 / Revised: 15 April 2021 / Accepted: 20 April 2021 / Published: 23 April 2021
(This article belongs to the Section Electronic Sensors)

Abstract

:
Reservoir computing (RC) is an attractive paradigm of a recurrent neural network (RNN) architecture, owning to the ease of training and existing neuromorphic implementation. Its simulated performance matches other digital algorithms on a series of benchmarking tasks, such as prediction tasks and classification tasks. In this article, we propose a novel RC structure based on the coupled MEMS resonators with the enhanced dynamic richness to optimize the performance of the RC system both on the system level and data set level. Moreover, we first put forward that the dynamic richness of RC comprises linear dynamic richness and nonlinear dynamic richness, which can be enhanced by adding delayed feedbacks and nonlinear nodes, respectively. In order to set forth this point, we compare three typical RC structures, a single-nonlinearity RC structure with single-feedback, a single-nonlinearity RC structure with double-feedbacks, and the couple-nonlinearity RC structure with double-feedbacks. Specifically, four different tasks are enumerated to verify the performance of the three RC structures, and the results show the enhanced dynamic richness by adding delayed feedbacks and nonlinear nodes. These results prove that coupled MEMS resonators offer an interesting platform to implement a complex computing paradigm leveraging their rich dynamical features.

1. Introduction

Artificial neural networks (ANN) have played an important role in the current boom in artificial intelligence (AI), especially with the invention of the Internet of Things (IoT) and ubiquitous sensing [1]. ANN has strong self-learning adaptability, parallel information processing capabilities, and nonlinear mapping capabilities, but they require complex and time-consuming algorithms to train the connection weights in the large-scale network. When interested in processing time-dependent tasks, standard feed-forward ANNs can no longer meet the requirements, so the recurrent neural network (RNN) has gradually become the mainstream [2,3,4]. RNN requires complex algorithms to train the connection weights, however, which causes slow convergence and consumes lots of calculations [5,6]. Therefore, reservoir computing (RC) gradually attracted attention because the weights of its recursive network are initialized randomly and untrained.
Reservoir computing (RC) is a brain-inspired computational framework suited for temporal data processing, owing to its derivation from the recurrent neural network (RNN) [5,7]. The main difference between RC and conventional RNNs is that the weights on the recurrent connections in the reservoir are not trained, but only the weights in the readout are trained, which avoids the well-known limitations of RNN, such as iterative parameter optimization and the condition of convergence [8]. More importantly, the fixed reservoir without updating is suitable for hardware implementation using various nonlinear dynamical systems. The hardware reservoir needs two crucial properties, nonlinearity and fading memory, which can be realized by using numerous randomly interacting nonlinear nodes or using a time-delayed nonlinear system. Actually, RC based on the time-delayed nonlinear system greatly reduces the implementation difficulty, which has been successfully demonstrated in a wide variety of systems, such as memristor [9,10], electronic [8,11,12], spintronic [13,14,15], optoelectronic [16,17,18], all-optical [19,20,21,22], quantum [23], and mechanical resonator [24,25,26,27]. In spite of these encouraging results, the RC implementation based on the single nonlinear node with single time-delay feedback adopted in these previous works has a limited memory capacity. The memory capacity is a key property of the RC that allows the processing of dynamical signals. The RC structures with the multiple delayed feedback or bidirectionally coupled nodes was proven to increase the processing speed of the system by boosting the memory length of the RC system, which are discussed numerically in detail in Appeltant’s doctoral dissertation [28]. Furthermore, this pioneering work inspires many studies on different configurations for the time-delayed RC. Particularly, double optical feedbacks RC structures are proposed to increase prediction performance under the condition that the long delay time and short delay time meet the specific relationship [29]. In addition, the mutually delay-coupled semiconductor lasers structure was studied for enhancing calculation speed and accuracy of the RC systems [30,31,32]. Despite these instructive works based on the optoelectronic devices promoting research on RC with the multiple delayed feedbacks and coupled nonlinear nodes, the potential of RC with other nonlinear devices and the reason why the coupled RC structure improves the accuracy has not yet been fully investigated.
In this work, we present a couple-nonlinearity RC structure with double-feedbacks (CRSD) based on the coupled MEMS resonators with the enhanced dynamic richness to promote the performance of the RC. Moreover, we first put forward that the dynamic richness of RC comprises linear dynamic richness and nonlinear dynamic richness, which can be enhanced by adding delayed feedbacks and nonlinear nodes, respectively. In order to elaborate this point in detail, we compare three typical RC structures, a single-nonlinearity RC structure with a single-feedback (SRSS), a single-nonlinearity RC structure with double-feedbacks (SRSD), and the couple-nonlinearity RC structure with double-feedbacks (CRSD). In addition, we describe the evolution for the promotion of the performance of the three structures. Furthermore, four different tasks—the parity benchmark task, nonlinear autoregressive moving average (NARMA) task, isolated spoken digits recognition, and human action recognition—are investigated to prove the feasibility of these RC structures. The results show the enhanced dynamic richness by adding delayed feedbacks and nonlinear nodes.
The rest of this paper is organized as follows. In Section 2, we describe the methods we used in this paper, including the three typical structures of the RC system, the nonlinear mapping node of the RC structure, and the dynamic richness analysis of three RC structures. In Section 3, we give a detailed discussion about the numerical results of our system on four typical tasks. In Section 4, the paper ends with a brief conclusion and an outlook for possible future work.

2. Methods

2.1. The Structures of the Reservoir Computing

A general single nonlinear node RC structure with a single-feedback (SRSS) can be conceptually decomposed into three layers, which are shown schematically in Figure 1a. The input layer (in blue) is the preprocessing stage of the RC system. After transformed by the mask m t , m t + T = m t , the original signal u t , n T t < n + 1 T , with the duration T is injected into the reservoir layer. The mask is random samples at the range of (−1, 1), with zero mean and unit variance, which breaks the symmetry of the input data and creates various transient trajectories for the nonlinear nodes in a transient state [17]. Input gain β determines the amplitude of the input signal of reservoir layer, which determines the position of the system operating point. The reservoir layer (in orange) in Figure 1a consists of a single nonlinear node and a delayed feedback loop. Particularly, a nonlinear node is used to nonlinearly map the time-dependent input to a high-dimensional feature space based on its own nonlinear output capability and inherent attenuation characteristics, which are represented by the state of the nodes forming the reservoir. This nonlinear mapping makes the initial complex input linearly separable in the new space based on the reservoir state, so a simple linear network layer (output layer) can be used for further software processing [33]. The delayed feedback loop helps the RC system build long-term memory [34]. This property assures the execution of computations that demand input retrieval for several time steps in the immediate past so that the system can handle prediction tasks that need more memory length. N together with the duration θ , θ = T / N , are denoted as virtual nodes because they exist in a time-multiplexed way rather than the nodes corresponding to real physical spatially distributed and interconnected nodes. The values of virtual nodes x i are injected into the output layer (in green) for the postprocessing; x i are linearly combined, y n = i = 0 N 1 W x i n , where W is a set of readout weights that are trained by the ridge regression algorithm.
The dynamic reservoir layer is the key to determine the performance of the RC because of its ability to nonlinearly transform input data into high-dimensional space. In addition, the reservoir is fixed and only the readout is trained with a simple method such as linear regression and classification. Therefore the reservoir is sensitive to its optimizable architecture properties, and attention should be paid to the optimization for the reservoir layer rather than the optimization of the input and output layers. The dynamic richness describes the response of the reservoir, which provides memory and nonlinear computing capabilities of the system. Based on our analysis, the dynamic richness consists of linear richness and nonlinear richness. The linear richness represents the linear combination of the current virtual node state and various previous nodes information, which are determined by the delay feedback loop. The general SRSS has only one delay feedback loop with a delay feedback gain α and a delay length τ . Here, we define τ = N + 1 T / N , where T is the period of the input data, making the system in an unsynchronized state where the virtual node is connected with several dependent nodes in a short period of time [17]. Thus, the reservoir of SRSS has only one input channel and one output channel. To enhance the linear richness of the RC system, we add another feedback loop, as shown in Figure 1b. Compared with SRSS, SRSD has an additional delay feedback loop. The short delay feedback loop (in orange) has the same delay length as SRSS’s, τ 1 = τ , and the delay length of the long delay feedback loop (in black) is τ 2 = 2 τ . The short feedback gain equals the long feedback gain, α 1 = α 2 . In the output layer, all responses from the virtual nodes of the short delay feedback loop are weighted and linearly summed. This structure is regarded as a transition structure for CRSD and has one input channel and two output channels. Since the nonlinear richness represents the nonlinear mapping ability of the nonlinear unit itself, we change the nonlinear mapping node from a single node to the coupled nonlinear node based on SRSD to optimize the nonlinear richness of the RC system. As illustrated in Figure 1c, CRSD is a system with a coupled nonlinear unit as the core, and two output channels and one input channel. The parameters of two delay feedback loops (in orange and black) are the same as those in SRSD. The details of the coupled nonlinear node are shown in the next section. Consequently, compared with the general structure SRSS, the CRSD increases the linear richness and nonlinear richness of the system.

2.2. The Nonlinear Mapping Node of the RC Structure

The nonlinear transforming of injected time-dependent data into a higher dimensional state is one of the most crucial properties of RC [35]. This transforming property determined by the nonlinear mapping node of the system is directly related to the accuracy of testing different data sets. The nonlinear mapping nodes of the SRSS and SRSD are based on the Duffing oscillator:
m x ¨ + c x ˙ + k 1 x + k 3 x 3 = F c o s ω t
where x is the displacement of silicon beam, m is the lumped effective mass, c is the damping coefficient of the system, k 1 is the linear spring stiffness, k 3 is the nonlinear spring stiffness that affects the nonlinear behavior of the beam, ω is drive frequency, and F is harmonic force using the electrostatic drive. This higher-order ordinary differential equation only has one parameter ( k 3 ) related to nonlinearity; that is, the nonlinear transforming property is not rich enough.
In this paper, we propose the coupled nonlinear mapping nodes to enhance the transforming property of CRSD, as illustrated in Figure 2. There are two resonators, Resonator 1 on the left and Resonator 2 on the right, enclosed by the red dotted line [36]. For subsequent silicon on glass (SOG) processing in our future work, each clamped-clamped beam is designed to be 40   μ m thick, 500   μ m long, and 6   μ m wide, with a gap of 2   μ m between two beams. Resonator 1 is driven and sensed using capacitive transduction while another vibrates simultaneously because of electrostatic coupling. The actuation is achieved using parallel plates of equal dimensions with 400   μ m in length and 40   μ m in thickness. The gaps of electrostatic drive and electrostatic detection are both 2   μ m . The dynamics of the system is approximated by these equations:
m x ¨ 1 + c x ˙ 1 + k 1 1 + k 3 x 1 2 x 1 + k c 1 + k c 3 x 1 x 2 2 x 1 x 2 = F c o s ω t ,
m x ¨ 2 + c x ˙ 2 + k 1 1 + k 3 x 2 2 x 2 + k c 1 + k c 3 x 2 x 1 2 x 2 x 1 = 0 ,
where x 1 and x 2 is the displacement of Resonator 1 and Resonators 2, respectively, m is lumped effective mass, c is the damping coefficient of system, k 1 is linear spring stiffness, k 3 is nonlinear spring stiffness, ω is drive frequency, F is harmonic force using the electrostatic drive, k c is the linear coupled stiffness, and k c 3 is the nonlinear coupled stiffness. These differential equations have two parameters ( k 3 , k c 3 ) to impact the nonlinearity of the system. Due to the coupled stiffness ( k c , k c 3 ), the nonlinear relationship between output and input is greatly strengthened. The simulated values are k 1 = 586.6   N / m , k 3 = 3.404 × 10 12   N / m 3 , and k c 3 = 5.21 × 10 12   N / m 3 . When two resonators separated by an electrically coupled gap are subjected to different DC polarization voltage, the displacement component of the electrostatic force generated between the resonators forms a negative electrostatic spring, whose magnitude is [37]:
k c = F x = ε 0 A g 3 Δ V b 2 ,
where A is the coupled electrode area, g is the gap between two coupled electrodes, and V b is the DC polarization voltage, which is also shown in Figure 2. When V b = 5   V , the linear coupled stiffness k c = 0.2878   N / m .
After processed by mask signal and input gain, and modulated by the drive voltage, the input signal is injected into the drive electrode of the coupled resonators. The coupled resonators produce two different current output signals. The current signals are converted to voltage signals by the custom transimpedance amplifier (TIA) and then through the band pass filter to remove the noise. The amplitude modulation signals restore the amplitude information through the envelope (ENV) and then enter the long delay feedback loop and short feedback loop. Finally, two feedback signals are injected into the input signal to form a closed loop. The values of the virtual node from the short feedback loop are used as the processing data in the postprocessing algorithm.

2.3. Dynamic Richness Analysis of Three RC Structure

2.3.1. The Linear Richness of Dynamic Richness

The memory of the previous input is very significant in the RC system. Insufficient memory can degrade performance, thereby depriving the reservoir of all prediction or processing ability [28]. Linear richness, which depends on the delay feedback loop and represents the memory ability of the system, is a part of dynamic richness. After going through the delay feedback loop with delay length τ , the value of the virtual node x i t at time t inject into the input end and linear combine with the virtual node x i t + τ at time t + τ . Compared with SRSS, the SRSD has an extra delay feedback loop to enhance the linear richness of the RC system.
Figure 3 shows the output signals of SRSS and SRSD with the same input spike. A voltage pulse of 0.6 ms duration is sent to the reservoir layer, as shown in the top panel. The middle panel shows the output amplitude signal of SRSS while the bottom presents the information from SRSD. Due to the intrinsic attenuation characteristics of the resonators, the output signal at each interval decays to zero after a few microseconds. Since SRSS has only one feedback loop, the output signal can maintain about five intervals. When the interval is greater than five, the information of the input pulse has already disappeared. Due to the addition of a long delay feedback loop, the output signal of SRSD can keep at least nine intervals, which indicates that SRSD can preserve the useful information of the input signal for a longer time.
In order to study the effect of the delay feedback loop on memory ability in more detail, we used memory capacity (MC) to make a quantitative analysis. Jaeger proposed a test for system memory capacity in 2002 [38]. The input u k of the reservoir is the point derived from the uniform distribution in the interval (−0.8, 0.8). The output y k is constructed as an infinite number of output sequences y i , each output sequence is a copy of the time sequence u , which is shifted by i steps, so y i k is the construction of u k i for i = 1 . In the actual test, the maximum value of i is selected high enough that it will not have a significant impact on the results. Therefore, we choose i = 40 for efficiency. The total memory capacity is defined as the normalized correlation between the approximate value of the target returned by the readout layer and its associated delay input, expressed by the following formula:
u c = i = 1 m i ,
with m i being the memory function, that is to say, the normalized correlation between y i ^ and y i , given by:
m i = c o r r y i k , y i ^ k = c o r r y i k , u k i .
Figure 4 shows the memory function as the function of the delay step of two RC structures. The blue line stands for the SRSS. The memory function decreases sharply with the increase of delay step. When the delay step increases to 10, the value of the memory function decays to near 0. That means this SRSS is not well suited for long-term prediction tasks. The red line represents the SRSD. When the delay step is less than seven, the memory function almost has no attenuation. When the range of the delay step is 7~28, the memory function decreases gradually. The MC of SRSD is 13.3412, which is more than twice that of SRSS, whose MC equals 6.3689.
Obviously, SRSD with two delay feedback loops has better memory capacity than SRSS with only one delay feedback loop. That is to say, due to the addition of a delayed feedback loop, the RC system enhances the linear richness. Since the coupled resonators have two output signals, we do not add more (three or more) feedback loops in the simulation.

2.3.2. The Nonlinear Richness of Dynamic Richness

As noted above, the reservoir has the ability to nonlinear transform the time-dependent input signal into a higher dimensional state space. The higher dimensional input signal greatly simplifies the difficulty of linear regression in postprocessing. Compared with SRSD, the CRSD uses the coupled nonlinear resonators to replace the single resonators, which improves the nonlinear richness of the RC system. The coupled resonators have two output signals from each resonator, and they connect a long delay feedback loop and a short delay feedback loop, which is the same as SRSD.
Figure 5 shows the output signal of two RC structures without the delay feedback loop. The top panel shows the input square wave signal with a period of 0.2 ms. One of the crucial properties of the RC system is consistency, which means the same response output can be observed by using a repeated input signal [39]. The outputs of the middle panel and the bottom panel remain consistent in each period, which indicated that these two systems all have the consistency property. In order to facilitate comparison, the two output signals have been processed by the same normalization method. Compared with the output of SRSS, the output of CRSD transforms the input signal into a relatively more complex data stream. That means the same input represents more data with different values.
Because the value of the virtual node directly determines the quality of the classification, analyzing the value of the virtual node can more intuitively show the nonlinear mapping ability of the RC system. Figure 6 shows the histogram of the virtual nodes in three RC structures with the same input data. We randomly intercepted 50,000 points in the values of a virtual node output for histogram statistics. In order to facilitate subsequent comparisons, the values of the virtual nodes are all normalized. The virtual nodes of SRSS are shown in the top panel. Most virtual nodes are concentrated near (0, 0.1), and there are around 8000 points in the range of (0.12, 0.13). The middle panel shows the virtual node of SRSD. After adding another delay feedback, the distribution of the virtual node is relatively scattered. Most virtual nodes are concentrated near (0.1, 0.3), and there are around 5000 points in the range of (0.17, 0.18). Because of the coupled nonlinear resonators, the distribution of virtual nodes is very scattered, as shown in the bottom panel. Most virtual nodes are concentrated near (0, 0.6) and the range with the maximum number of points is also limited to around 2500. The value distribution of the virtual nodes is more uniform, indicating that the virtual nodes have more abundant states, and the abundant node states can greatly improve the performance of the reservoir system.
In brief, CRSD has a richer node status than the other two systems. That is to say, the system enhances its nonlinear richness because of the coupled nonlinear resonators. Compared with the general structure SRSS, the CRSD increases the linear richness through another delay feedback loop and nonlinear richness through the coupled MEME resonators, which we call dynamic richness of the system.

3. Results and Discussion

The performance of the system lies in the following two aspects, prediction ability and classification ability, which are assessed by four typical data sets to prove that the coupled nonlinear resonators improve the dynamic richness of the RC system. The prediction ability is numerically investigated via the parity benchmark prediction task and NARMA task, while the classification ability is discussed via the TI-46 isolated word task and human action recognition task.

3.1. Parity Benchmark Prediction Task

The prediction ability of the system is first assessed by means of the parity benchmark task. The parity function is considered as a benchmark, as it requires both memory ability and nonlinear computational capability [40]. For this task, u t is a binary time-dependent signal that randomly switches between two states −1 and +1, where τ is its switching period. The n t h -order parity function can be defined as:
P n = i = 1 n u t i 1 τ ,
where n is the order of benchmark, τ is the period, and P n is the training target of this n t h -order parity function. As n increases, the difficulty of this prediction task will gradually increase. When n = 1 , the benchmark performs a linearly separable data set and easy to predict [41]. As n > 1 , the data of P n at this time are related to the previous n data. Namely, the system needs to have at least a n τ memory length to process this data set. One hundred fifty training data and 50 testing data are used for our test and we choose the parameters, θ = 0.1 ms , N = 400 , f = 352 , 500   H z, β = 1 , and V d c = 20   V , V a c = 2   V for the better success rate. The success rate of P 2 ~ 4 = 100 % , P 5 = 95 % , and P 6 = 88 % is reported in [24]. The method of parameter selection has been explained in our previously published articles [42] and will not be explained here.
Figure 7 shows the performance of SRSS and SRSD when the order n equals 7. Because the success rate of P 1 ~ P 6 has already achieved 100% for two RC systems, we do not show the result of P 1 ~ P 6 here. The parameters of these two RC systems are the same except for the delay feedback gain. The delay feedback gain α = 5 in SRSS while α 1 = 2.5 and α 2 = 2.5 in SRSD. The black dotted line represents the 50 target value of P 7 . Compared with the result of SRSS (red line), the result of SRSD (blue line) is closer to the target value. At the same time, the success rate is 78% for SRSS and 100% for SRSD. Since the success rate has already up to 100% for n = 1 , 2 , 3 , 4 , 5 , 6 , 7 with SRSD, there is no room for improvement, so we did not use CRSD to further optimize this data set.
The result of this prediction task shows SRSD with another delay feedback loop can improve the success rate of the RC system by increasing its linear richness.

3.2. NARMA Prediction Task

The nonlinear autoregressive moving average (NARMA) prediction task is one of the most popular tasks in the RC community and it is also a more difficult task than the parity benchmark. In this prediction task, the RC system is trained to predict the behaviors of the system, for instance, a nonlinear autoregressive moving average (NARMA) of order m driven by white noise. The m th order NARMA task is given by the following recursive formula:
y n + 1 = 0.3 y n + 0.05 y n i = 0 m 1 y n 1 + 1.5 u n m + 1 u n + 0.1 ,
where y n is the output target of the system, m is the order of this benchmark, and u t stands for the white noise, which is the random input derived from a uniform distribution over the interval (0, 0.5). In this task, the RC system is trained in a sequence of 1000 data length and tested in the following sequence of 1000 data length. The performance metric used to evaluate NARMA is normalized mean square error (NMSE), NMSE = 1 L n = 1 L y n y ^ n 2 / v a r y , where y n is the target and y ^ n is the prediction result of the system. We use three RC systems (SRSS, SRSD, and CRSD) to test the performance of this data set, and the same parameters were set as N = 50 , β = 3.6 , and V d c = 80 V , and V a c = 2   V . In particular, θ exhibits a crucial role in this task and we set θ = 0.01   ms to enhance the connection of current input to the previous virtual node state. The prediction difficulty of the system increases exponentially with the increase of order m . We test the prediction accuracy of the three RC systems under the conditions of m = 1 , 2 , 5 , and 10.
Figure 8 shows the performance of three RC systems for the NARMA prediction test. The lines with different colors represent the different orders of prediction. The driving frequency f = 352 , 000   H z in SRSS and SRSD, while f = 264 , 300   H z in CRSD. The delay feedback loop α = 0.6 in SRSS, while α 1 = 0.4 , α 2 = 0.4 in SRSD and CRSD with two delay feedback loops. In the same structure, the NMSE increases gradually with the increase of order because of increasing difficulty. Under the condition of the same order m , the NMSE decreases as the structure is optimized from SRSS to SRSD to CRSD at the same time. When m = 1 , CRSD reduces the NMSE of the system by around 70% from SRSS, and when m = 2 , 5 , CRSD reduces the NMSE of the system by around 50%. Concurrently, CRSD reduces the NMSE of the system by around 35%.
The results of this complicated prediction task show CRSD with two delay feedback loops and the coupled resonators can decrease NMSE by increasing its linear richness and nonlinear richness.

3.3. TI-46 Isolated Word Classification Task

Through the isolated spoken word task, the classification performance of reservoir computing when inputting complex data is evaluated. This task is the classification of isolated audio sequences from a subset of the National Institute of Standards and Technology Texas Instrument-46 Corpus (NIST TI-46 Corpus) [43]. The data set has 500 audio sequences, each audio sequence represents a number (0–9), recorded ten times by five different female speakers. Five hundred audio waveforms with different time lengths are sampled at a rate of 12.5 kHz. Every input representing a spoken digit is preprocessed using the Lyon cochlear ear model [44] before injecting into the RC system. In the training phase, ten weight vectors are calculated for the ten numbers in the vocabulary, and then the winner-takes-all method is applied to select the actual numbers. Each classifier is trained to output a value of 1 when presented with an utterance of the digit corresponding to its category, and 0 otherwise. The highest averaged classifier corresponds to the correct digit. The performance indicator used to evaluate the isolated word classification is the word error rate (WER), namely, the fraction of digits incorrectly classified. There are only 500 sequences in the subset corpus, so we divide them into 10 parts and use the tenfold cross-validation method to avoid the influence of the specific division of available data in some processes (such as training and testing), so as to make the results more convincing. Ten sections were randomly selected, including nine for training and one for testing.
In this task, we use three RC systems (SRSS, SRSD, and CRSD) to test the performance, and the same parameters were set as N = 100 , θ = 0.1   ms , β = 200 , and V d c = 20   V , V a c = 2   V . The driving frequency f = 349 , 200   H z is in SRSS and SRSD, while f = 244 , 300   H z in CRSD; the delay feedback loop is α = 0.4 in SRSS, while α 1 = 0.2 , α 2 = 0.2 in SRSD and CRSD with two delay feedback loops. The original result is WER = 0.2% in SRSS. After increasing the linear richness of the system, the result is improved to WER = 0.12% in SRSD. CRSD improves the dynamic richness of the system and the result is optimized to WER = 0.04%.
In order to visualize how the time-dependent data separation occurs and understand the recognition capabilities of the different RC structures, the t-distributed stochastic neighbor embedding (t-SNE) technique [45] is used to represent input data of the TI-46 data set in a 2D figure. The t-SNE technique is a nonlinear dimensionality reduction technique used to map high-dimensional data to a low-dimensional space with two dimensions or three dimensions. During data reduction, the probability of two vectors becoming neighbors is preserved, so that the structure in the data can be visualized. Figure 9 shows the t-SNE result of two structures. For all of the data points of the digit isolated word, each number is represented by a colored dot. Figure 9a shows the result of the TI-46 raw data only after processing by the Lyon cochlear ear model. Since all the color points seem to be randomly distributed, there is no data separation. In particular, numbers in the same category will not form separate clusters. For the data set after preprocessing by SRSS, data separation can clearly be seen in Figure 9b, correlating with the word error rate of 0.2%. A few clusters such as Digit 8 and Digit 9, however, are very close to some clusters, and some even overlap with others. After optimized by the CRSD, the t-SNE result is shown in Figure 9c with WER = 0.04%. The clusters of Digit 8 and Digit 9 can be relatively well separated from the clustering of other numbers and corroborate the high recognition rates exhibited by CRSD.
The results of this isolated word classification task show that CRSD can decrease the WER by increasing its dynamic richness.

3.4. Human Action Recognition Task

Recognizing human actions in video streams is a difficult task in computer vision. This task was first applied to RC systems in 2019 [46] as a standard data set for testing RC systems. This recognition task is a subset of the KTH video database [47], which contains four different scenarios. For brevity, we limited the data set in the first scenario, called “S1,” which included an outdoor video of six different motions (walking, jogging, running, boxing, hand waving, and hand clapping) performed by 25 subjects. All videos were recorded with a static camera and 25 fps against a uniform background and then sampled down to 160 × 120 pixels of spatial resolution. The length of each motion is different, the average is 4 s. Our dataset consists of 25 × 6 × 4 = 600 sequences, each combination consisting of 25 subjects, 6 actions, and 4 replicates. These 600 motion sequences are preprocessed by the histograms of oriented gradients (HOG) algorithm [47], which is used to extract spatial and shape information from a single video frame. HOG features are performed by the computer, with a cell size of 8 × 8 and a block size of 2 × 2, individually for each frame of every sequence. The size of each frame is 160 × 120 pixels, and the HOG algorithm returns 19 × 14 × 4 × 9 = 9576 features per frame. These features are fed into the RC system, which outputs the values of the virtual nodes, and then trains it to classify each frame. The classifiers are implemented by defining six binary output nodes. Training them to output 1 for a frame of the corresponding class and 0 for the other classes, which is the same as training method for TI-46 isolated word classification task. Classification by frame is obtained by selecting the node with the largest output, namely the winner-take-all method. The final determination of a video sequence is given by the classes that belong to the majority of the frames in the sequence.
In this recognition task, we also use three RC systems (SRSS, SRSD, and CRSD) to test the performance, and the same parameters were set as N = 3000 , θ = 0.1   ms , β = 0.01 , and V d c = 20   V , V a c = 2   V . The driving frequency was f = 350 , 000   H z in SRSS and SRSD, while f = 244 , 000   H z in CRSD. The delay feedback loop was α = 0.2 in SRSS, while α 1 = 0.1 , α 2 = 0.1 in SRSD and CRSD with two delay feedback loops. Figure 10 displays the confusion matrices [48] for three different RC systems. After processing by SRSS, the success rate is only 73.3%. Relatively good recognition results were obtained for the hand gestures (walking, jogging, and running), but for fast spatial movements (jogging and running) that are difficult to distinguish, the recognition results are very unsatisfactory. After optimizing the linear richness, the success rate rises to 89.33% in SRSD. The hand gestures can get perfect accuracy and the accuracy of fast spatial movements is greatly improved. The CRSD optimizes the nonlinear richness of the system, so the success rate up to 92%. Compared with SRSD, the accuracy of jogging and walking are relatively improved.
In addition, we test the effect of the number of virtual nodes N on the accuracy rate of the CRSD system. The number of virtual nodes N is a crucial parameter for RC because it determines the speed as well as the performance of the system. If N is relatively small, the performance will decrease, but the system operation speed will increase on the contrary. A higher N usually means better state diversity, which improves performance at the expense of lower computing speed. Because this recognition video data set has a much higher data complexity than the other three above, more virtual nodes are needed to process the data to increase the accuracy rate. The photonic computer gets the accuracy rate = 91.3% at the condition of N = 16 , 384 [46]. In our test, the accuracy rate as a function of the number of virtual nodes is shown in Figure 11. In the range of N = 400 to N = 3000 , as the number of virtual nodes increases, the accuracy rate increases significantly. When N > 3000 , the accuracy rate fluctuates in a small range and gradually stabilizes. The simulation time increases rapidly as the number of virtual nodes increases. If increasing the number of virtual nodes does not increase the accuracy rate, then too large a number of virtual nodes will seriously affect the calculation efficiency. As a compromise between the performance and the efficiency, we choose N = 3000 in this task.

4. Conclusions

We propose the coupled MEMS resonators can optimize the performance of the RC system based on the enhanced dynamic richness that is composed of linear richness and nonlinear richness. In order to set forth this point, we compare three typical RC structures with a single-nonlinearity RC structure with single-feedback, a single-nonlinearity RC structure with double-feedbacks that is treated as a transition structure, and the couple-nonlinearity RC structure with double-feedbacks. We also describe the evolution for the promotion of the performance of the three structures. At the system level, we use the delay characteristic diagram and memory capacity to illustrate that the delayed feedback loop increases the linear richness, and use the output signal timing diagram and the histogram of the virtual nodes to illustrate that the coupled resonator increases the nonlinear richness of the system. At the data set level, four different tasks, the parity benchmark task, nonlinear autoregressive moving average (NARMA) task, isolated spoken digits recognition, and human action recognition, are investigated to prove that CRSD can improve the accuracy of the system to a certain extent. These results have significant implications with respect to hardware implementation of the reservoir and open new future possibilities for exploring real-world applications of coupling resonators in the RC system.

Author Contributions

Conceptualization, X.Z. and W.Y.; methodology, T.Z.; software, T.Z. and W.Y.; validation, T.Z. and J.S.; formal analysis, X.X.; investigation, Z.L. and Z.W.; writing—original draft preparation, T.Z.; writing—review and editing, T.Z. and W.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was partially supported by the National Key Research and Development Program of China under Grant 2018YFB2002300, the National Natural Science Foundation of China (Grant No.61971399), and Key Research Program of Frontier Science (CAS, Grant No.ZDBS-LY-JSC028).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article. More detailed data and data presented in this study are available on request from the corresponding author. Part of the data could be included in the final reports to the corresponding funding organizations.

Acknowledgments

The authors would like to thank the members of The State Key Laboratory of Transducer Technology for helpful discussions. We thank Lilin Yi’s group from Shanghai Jiao Tong University for sharing the TI-46 dataset with us.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  2. Feldkamp, L.A.; Prokhorov, D.V.; Eagen, C.F.; Yuan, F. Enhanced multi-stream Kalman filter training for recurrent networks. In Nonlinear Modeling; Springer: Berlin, Germany, 1998; pp. 29–53. [Google Scholar]
  3. Graves, A.; Mohamed, A.-R.; Hinton, G. Speech recognition with deep recurrent neural networks. In Proceedings of the 2013 IEEE International Conference on Acoustics, Speech and Signal Processing, Vancouver, BC, Canada, 26–31 May 2013; pp. 6645–6649. [Google Scholar]
  4. Mo, W.; Gutterman, C.L.; Li, Y.; Zussman, G.; Kilper, D.C. Deep neural network based dynamic resource reallocation of BBU pools in 5G C-RAN ROADM networks. In Proceedings of the Optical Fiber Communication Conference, San Diego, CA, USA, 11–15 March 2018; p. Th1B. 4. [Google Scholar]
  5. Maass, W.; Natschläger, T.; Markram, H.J.N.C. Real-time computing without stable states: A new framework for neural computation based on perturbations. Neural Comput. 2002, 14, 2531–2560. [Google Scholar] [CrossRef] [PubMed]
  6. Pearlmutter, B.A. Gradient calculations for dynamic recurrent neural networks: A survey. IEEE Trans. Neural Netw. 1995, 6, 1212–1228. [Google Scholar] [CrossRef] [Green Version]
  7. Jaeger, H.; Haas, H. Harnessing Nonlinearity: Predicting Chaotic Systems and Saving Energy in Wireless Communication. Science 2004, 304, 78–80. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  8. Caluwaerts, K.; D’Haene, M.; Verstraeten, D.; Schrauwen, B. Locomotion without a brain: Physical reservoir computing in tensegrity structures. Artif. Life 2013, 19, 35–66. [Google Scholar] [CrossRef]
  9. Du, C.; Cai, F.; Zidan, M.A.; Ma, W.; Lee, S.H.; Lu, W.D. Reservoir computing using dynamic memristors for temporal information processing. Nat. Commun. 2017, 8, 1–10. [Google Scholar] [CrossRef] [PubMed]
  10. Moon, J.; Ma, W.; Shin, J.H.; Cai, F.; Du, C.; Lee, S.H.; Lu, W.D. Temporal data classification and forecasting using a memristor-based reservoir computing system. Nat. Electron. 2019, 2, 480–487. [Google Scholar] [CrossRef]
  11. Appeltant, L.; Soriano, M.C.; Van der Sande, G.; Danckaert, J.; Massar, S.; Dambre, J.; Schrauwen, B.; Mirasso, C.R.; Fischer, I.J. Information processing using a single dynamical node as complex system. Nat. Commun. 2011, 2, 1–6. [Google Scholar] [CrossRef] [Green Version]
  12. Degrave, J.; Caluwaerts, K.; Dambre, J.; Wyffels, F. Developing an embodied gait on a compliant quadrupedal robot. In Proceedings of the 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Hamburg, Germany, 28 September–2 October 2015; pp. 4486–4491. [Google Scholar]
  13. Riou, M.; Araujo, F.A.; Torrejon, J.; Tsunegi, S.; Khalsa, G.; Querlioz, D.; Bortolotti, P.; Cros, V.; Yakushiji, K.; Fukushima, A. Neuromorphic computing through time-multiplexing with a spin-torque nano-oscillator. In Proceedings of the 2017 IEEE International Electron Devices Meeting (IEDM), San Francisco, CA, USA, 2–6 December 2017; pp. 36.33.31–36.33.34. [Google Scholar]
  14. Riou, M.; Torrejon, J.; Garitaine, B.; Araujo, F.A.; Bortolotti, P.; Cros, V.; Tsunegi, S.; Yakushiji, K.; Fukushima, A.; Kubota, H.J. Temporal pattern recognition with delayed-feedback spin-torque nano-oscillators. Phys. Rev. Appl. 2019, 12, 024049. [Google Scholar] [CrossRef]
  15. Araujo, F.A.; Riou, M.; Torrejon, J.; Tsunegi, S.; Querlioz, D.; Yakushiji, K.; Fukushima, A.; Kubota, H.; Yuasa, S.; Stiles, M.D. Role of non-linear data processing on speech recognition task in the framework of reservoir computing. Sci. Rep. 2020, 10, 1–11. [Google Scholar]
  16. Larger, L.; Soriano, M.C.; Brunner, D.; Appeltant, L.; Gutiérrez, J.M.; Pesquera, L.; Mirasso, C.R.; Fischer, I.J. Photonic information processing beyond Turing: An optoelectronic implementation of reservoir computing. Opt. Express 2012, 20, 3241–3249. [Google Scholar] [CrossRef] [PubMed]
  17. Paquot, Y.; Duport, F.; Smerieri, A.; Dambre, J.; Schrauwen, B.; Haelterman, M.; Massar, S.J. Optoelectronic reservoir computing. Sci. Rep. 2012, 2, 287. [Google Scholar] [CrossRef]
  18. Antonik, P.; Hermans, M.; Haelterman, M.; Massar, S. Towards adjustable signal generation with photonic reservoir computers. In Proceedings of the International Conference on Artificial Neural Networks, Barcelona, Spain, 6–9 September 2016; pp. 374–381. [Google Scholar]
  19. Duport, F.; Smerieri, A.; Akrout, A.; Haelterman, M.; Massar, S.J. Fully analogue photonic reservoir computer. Sci. Rep. 2016, 6, 22381. [Google Scholar] [CrossRef] [Green Version]
  20. Brunner, D.; Soriano, M.C.; Mirasso, C.R.; Fischer, I.J. Parallel photonic information processing at gigabyte per second data rates using transient states. Nat. Commun. 2013, 4, 1–7. [Google Scholar] [CrossRef] [Green Version]
  21. Vinckier, Q.; Duport, F.; Smerieri, A.; Vandoorne, K.; Bienstman, P.; Haelterman, M.; Massar, S.J.O. High-performance photonic reservoir computer based on a coherently driven passive cavity. Optica 2015, 2, 438–446. [Google Scholar] [CrossRef]
  22. Denis-Le Coarer, F.; Sciamanna, M.; Katumba, A.; Freiberger, M.; Dambre, J.; Bienstman, P.; Rontani, D.J. All-optical reservoir computing on a photonic chip using silicon-based ring resonators. IEEE J. Sel. Top. Quantum Electron. 2018, 24, 1–8. [Google Scholar] [CrossRef] [Green Version]
  23. Govia, L.; Ribeill, G.; Rowlands, G.; Krovi, H.; Ohki, T.A. Quantum reservoir computing with a single nonlinear oscillator. Phys. Rev. Res. 2021, 3, 013077. [Google Scholar] [CrossRef]
  24. Dion, G.; Mejaouri, S.; Sylvestre, J. Reservoir computing with a single delay-coupled non-linear mechanical oscillator. J. Appl. Phys. 2018, 124, 152132. [Google Scholar] [CrossRef] [Green Version]
  25. Coulombe, J.C.; York, M.C.; Sylvestre, J. Computing with networks of nonlinear mechanical oscillators. PLoS ONE 2017, 12, e0178663. [Google Scholar] [CrossRef] [Green Version]
  26. H Hasan, M.; Al-Ramini, A.; Abdel-Rahman, E.; Jafari, R.; Alsaleem, F. Colocalized Sensing and Intelligent Computing in Micro-Sensors. Sensors 2020, 20, 6346. [Google Scholar] [CrossRef]
  27. Barazani, B.; Dion, G.; Morissette, J.-F.; Beaudoin, L.; Sylvestre, J. Microfabricated Neuroaccelerometer: Integrating Sensing and Reservoir Computing in MEMS. J. Microelectromechanical Syst. 2020, 29, 338–347. [Google Scholar] [CrossRef] [Green Version]
  28. Appeltant, L. Reservoir Computing Based on Delay-Dynamical Systems. Ph.D. Thesis, Vrije Universiteit Brussel/Universitat de les Illes Balears, Ixelles, Belgium, 2012. [Google Scholar]
  29. Hou, Y.; Xia, G.; Yang, W.; Wang, D.; Jayaprasath, E.; Jiang, Z.; Hu, C.; Wu, Z. Prediction performance of reservoir computing system based on a semiconductor laser subject to double optical feedback and optical injection. Opt. Express 2018, 26, 10211–10219. [Google Scholar] [CrossRef] [PubMed]
  30. Hou, Y.-S.; Xia, G.-Q.; Jayaprasath, E.; Yue, D.-Z.; Yang, W.-Y.; Wu, Z.-M. Prediction and classification performance of reservoir computing system using mutually delay-coupled semiconductor lasers. Opt. Commun. 2019, 433, 215–220. [Google Scholar] [CrossRef]
  31. Chen, Y.; Yi, L.; Ke, J.; Yang, Z.; Yang, Y.; Huang, L.; Zhuge, Q.; Hu, W. Reservoir computing system with double optoelectronic feedback loops. Opt. Express 2019, 27, 27431–27440. [Google Scholar] [CrossRef]
  32. Hou, Y.; Xia, G.; Jayaprasath, E.; Yue, D.; Wu, Z. Parallel information processing using a reservoir computing system based on mutually coupled semiconductor lasers. Appl. Phys. A 2020, 126, 40. [Google Scholar] [CrossRef]
  33. Lukoševičius, M.; Jaeger, H. Reservoir computing approaches to recurrent neural network training. Comput. Sci. Rev. 2009, 3, 127–149. [Google Scholar] [CrossRef]
  34. Kovac, A.D.; Koall, M.; Pipa, G.; Toutounji, H. Persistent memory in single node delay-coupled reservoir computing. PLoS ONE 2016, 11, e0165170. [Google Scholar] [CrossRef]
  35. Soriano, M.C.; Ortín, S.; Keuninckx, L.; Appeltant, L.; Danckaert, J.; Pesquera, L.; Van der Sande, G. Delay-based reservoir computing: Noise effects in a combined analog and digital implementation. IEEE Trans. Neural Netw. Learn. Syst. 2014, 26, 388–393. [Google Scholar] [CrossRef]
  36. Thiruvenkatanathan, P.; Yan, J.; Lee, J.-Y.; Seshia, A. Enhancing parametric sensitivity using mode localization in electrically coupled MEMS resonators. In Proceedings of the TRANSDUCERS 2009-2009 International Solid-State Sensors, Actuators and Microsystems Conference, Denver, CO, USA, 21–25 June 200; pp. 2350–2353.
  37. Pourkamali, S.; Ayazi, F.J.S.; Physical, A.A. Electrically coupled MEMS bandpass filters: Part I: With coupling element. Sens. Actuators A Phys. 2005, 122, 307–316. [Google Scholar] [CrossRef] [Green Version]
  38. Jaeger, H. Tutorial on Training Recurrent Neural Networks, Covering BPPT, RTRL, EKF and the" Echo State Network" Approach; GMD-Forschungszentrum Informationstechnik: Bonn, Germany, 2002; Volume 5. [Google Scholar]
  39. Nakayama, J.; Kanno, K.; Uchida, A. Laser dynamical reservoir computing with consistency: An approach of a chaos mask signal. Opt. Express 2016, 24, 8679–8692. [Google Scholar] [CrossRef]
  40. Büsing, L.; Schrauwen, B.; Legenstein, R. Connectivity, dynamics, and memory in reservoir computing with binary and analog neurons. Neural Comput. 2010, 22, 1272–1311. [Google Scholar] [CrossRef]
  41. Bertschinger, N.; Natschläger, T. Real-time computation at the edge of chaos in recurrent neural networks. Neural Comput. 2004, 16, 1413–1436. [Google Scholar] [CrossRef]
  42. Zheng, T.; Yang, W.; Sun, J.; Xiong, X.; Li, Z.; Zou, X. Parameters optimization method for the time-delayed reservoir computing with a nonlinear duffing mechanical oscillator. Sci. Rep. 2021, 11, 1–11. [Google Scholar] [CrossRef]
  43. Instruments-Developed, Texas. 46-Word Speaker-Dependent Isolated Word Corpus (TI46), NIST Speech Disc, September 1991. Available online: https://catalog.ldc.upenn.edu/LDC93S9 (accessed on 1 March 2021).
  44. Lyon, R. A computational model of filtering, detection, and compression in the cochlea. In Proceedings of the ICASSP’82. IEEE International Conference on Acoustics, Speech, and Signal Processing, Paris, France, 3–5 May 1982; pp. 1282–1285. [Google Scholar]
  45. Van der Maaten, L.; Hinton, G. Visualizing data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
  46. Antonik, P.; Marsal, N.; Brunner, D.; Rontani, D. Human action recognition with a large-scale brain-inspired photonic computer. Nat. Mach. Intell. 2019, 1, 530–537. [Google Scholar] [CrossRef] [Green Version]
  47. Schuldt, C.; Laptev, I.; Caputo, B. Recognizing human actions: A local SVM approach. In Proceedings of the 17th International Conference on Pattern Recognition, Cambridge, UK, 26 August 2004; pp. 32–36. [Google Scholar]
  48. Lakkam, S.; Balamurali, B.; Bouffanais, R. Hydrodynamic object identification with artificial neural models. Sci. Rep. 2019, 9, 1–12. [Google Scholar] [CrossRef]
Figure 1. Schematic diagrams of three RC structures: (a) single-nonlinearity RC structure with single-feedback (SRSS); (b) single-nonlinearity RC structure with double-feedbacks (SRSD); and (c) couple-nonlinearity RC structure with double-feedbacks (CRSD). The blue part stands for the input layer and the orange part shows the reservoir layer, while the green part performs the output layer. The reservoir layer of the three structures is different, and the input layer and output layer are the same.
Figure 1. Schematic diagrams of three RC structures: (a) single-nonlinearity RC structure with single-feedback (SRSS); (b) single-nonlinearity RC structure with double-feedbacks (SRSD); and (c) couple-nonlinearity RC structure with double-feedbacks (CRSD). The blue part stands for the input layer and the orange part shows the reservoir layer, while the green part performs the output layer. The reservoir layer of the three structures is different, and the input layer and output layer are the same.
Sensors 21 02961 g001
Figure 2. Signal chain of CRSD. The blue part stands for the input layer and the orange part shows the reservoir layer, while the green part performs the output data. The coupled MEMS resonators enclosed by the red dashed line has one signal input electrode and two signal output electrodes.
Figure 2. Signal chain of CRSD. The blue part stands for the input layer and the orange part shows the reservoir layer, while the green part performs the output data. The coupled MEMS resonators enclosed by the red dashed line has one signal input electrode and two signal output electrodes.
Sensors 21 02961 g002
Figure 3. Delay characteristic diagram of two structures. From top to bottom, respectively, waveforms show the input signal, the output signal of SRSS, and the output signal of SRSD.
Figure 3. Delay characteristic diagram of two structures. From top to bottom, respectively, waveforms show the input signal, the output signal of SRSS, and the output signal of SRSD.
Sensors 21 02961 g003
Figure 4. Memory curve for two RC structures. The blue line stands for the SRSS and the red line represents the SRSD.
Figure 4. Memory curve for two RC structures. The blue line stands for the SRSS and the red line represents the SRSD.
Sensors 21 02961 g004
Figure 5. Output signal timing diagram of two reservoir layer structures. From top to bottom, respectively, waveforms show the input signal, the output signal of SRSS without the feedback loop, and the output signal of CRSD without the feedback loop.
Figure 5. Output signal timing diagram of two reservoir layer structures. From top to bottom, respectively, waveforms show the input signal, the output signal of SRSS without the feedback loop, and the output signal of CRSD without the feedback loop.
Sensors 21 02961 g005
Figure 6. The histogram of the virtual nodes in three RC structures. From top to bottom, the histogram shows the virtual nodes of SRSS, SRSD, CRSD. For the convenience of comparison, the values of virtual nodes are all normalized.
Figure 6. The histogram of the virtual nodes in three RC structures. From top to bottom, the histogram shows the virtual nodes of SRSS, SRSD, CRSD. For the convenience of comparison, the values of virtual nodes are all normalized.
Sensors 21 02961 g006
Figure 7. Performance of two RC systems for the parity benchmark test with n = 7 . The black dotted line stands for the target of P 7 and the red line represents the result of SRSS, while the blue line shows the result of SRSD in the time span of ( 1.5 × 10 4 , 2 × 10 4 ) after the training phase. The success rate is 78% for SRSS and 100% for SRSD.
Figure 7. Performance of two RC systems for the parity benchmark test with n = 7 . The black dotted line stands for the target of P 7 and the red line represents the result of SRSS, while the blue line shows the result of SRSD in the time span of ( 1.5 × 10 4 , 2 × 10 4 ) after the training phase. The success rate is 78% for SRSS and 100% for SRSD.
Sensors 21 02961 g007
Figure 8. Performance of three RC systems for the NARMA prediction test with m = 1 , 2 , 5 , 10 . The blue line stands for the NMSE result of NARMA1, the orange line represents the result of NARMA2, and the yellow line represents the result of NARMA5, while the purple line shows the result of NARMA10 for three different RC systems. The numbers on the circle represent the results of the NMSE.
Figure 8. Performance of three RC systems for the NARMA prediction test with m = 1 , 2 , 5 , 10 . The blue line stands for the NMSE result of NARMA1, the orange line represents the result of NARMA2, and the yellow line represents the result of NARMA5, while the purple line shows the result of NARMA10 for three different RC systems. The numbers on the circle represent the results of the NMSE.
Sensors 21 02961 g008
Figure 9. Two-dimensional representation of the two t-SNE components for SRSS and CRSD: (a) t-SNE result of TI-46. raw data only after processing by the Lyon cochlear ear model; (b) t-SNE result of TI-46 data after preprocessing by SRSS; and (c) t-SNE result of TI-46 data after preprocessing by CRSD.
Figure 9. Two-dimensional representation of the two t-SNE components for SRSS and CRSD: (a) t-SNE result of TI-46. raw data only after processing by the Lyon cochlear ear model; (b) t-SNE result of TI-46 data after preprocessing by SRSS; and (c) t-SNE result of TI-46 data after preprocessing by CRSD.
Sensors 21 02961 g009
Figure 10. The confusion matrixes of the human action recognition benchmark show the probability (colors) that a motion presented to the system (columns) is classified to a certain motion (lines). (a) Confusion matrices for the SRSS with the success rate = 73.3% (b) Confusion matrices for the SRSD with the success rate = 89.33%. (c) Confusion matrices for the CRSD with the success rate = 92%.
Figure 10. The confusion matrixes of the human action recognition benchmark show the probability (colors) that a motion presented to the system (columns) is classified to a certain motion (lines). (a) Confusion matrices for the SRSS with the success rate = 73.3% (b) Confusion matrices for the SRSD with the success rate = 89.33%. (c) Confusion matrices for the CRSD with the success rate = 92%.
Sensors 21 02961 g010
Figure 11. Virtual node sweep results of the human action task.
Figure 11. Virtual node sweep results of the human action task.
Sensors 21 02961 g011
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zheng, T.; Yang, W.; Sun, J.; Xiong, X.; Wang, Z.; Li, Z.; Zou, X. Enhancing Performance of Reservoir Computing System Based on Coupled MEMS Resonators. Sensors 2021, 21, 2961. https://doi.org/10.3390/s21092961

AMA Style

Zheng T, Yang W, Sun J, Xiong X, Wang Z, Li Z, Zou X. Enhancing Performance of Reservoir Computing System Based on Coupled MEMS Resonators. Sensors. 2021; 21(9):2961. https://doi.org/10.3390/s21092961

Chicago/Turabian Style

Zheng, Tianyi, Wuhao Yang, Jie Sun, Xingyin Xiong, Zheng Wang, Zhitian Li, and Xudong Zou. 2021. "Enhancing Performance of Reservoir Computing System Based on Coupled MEMS Resonators" Sensors 21, no. 9: 2961. https://doi.org/10.3390/s21092961

APA Style

Zheng, T., Yang, W., Sun, J., Xiong, X., Wang, Z., Li, Z., & Zou, X. (2021). Enhancing Performance of Reservoir Computing System Based on Coupled MEMS Resonators. Sensors, 21(9), 2961. https://doi.org/10.3390/s21092961

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop