Transferring learning from external to internal weights in echo-state networks with sparse connectivity
- PMID: 22655041
- PMCID: PMC3360031
- DOI: 10.1371/journal.pone.0037372
Transferring learning from external to internal weights in echo-state networks with sparse connectivity
Abstract
Modifying weights within a recurrent network to improve performance on a task has proven to be difficult. Echo-state networks in which modification is restricted to the weights of connections onto network outputs provide an easier alternative, but at the expense of modifying the typically sparse architecture of the network by including feedback from the output back into the network. We derive methods for using the values of the output weights from a trained echo-state network to set recurrent weights within the network. The result of this "transfer of learning" is a recurrent network that performs the task without requiring the output feedback present in the original network. We also discuss a hybrid version in which online learning is applied to both output and recurrent weights. Both approaches provide efficient ways of training recurrent networks to perform complex tasks. Through an analysis of the conditions required to make transfer of learning work, we define the concept of a "self-sensing" network state, and we compare and contrast this with compressed sensing.
Conflict of interest statement
Figures
, contain units that interact through a sparse weight matrix
, and produce an output
, obtained by summing activity from the entire network weighted by the modifiable components of the vector
. (A) The output unit sends feedback to all of the network units through connections of fixed weight
. Learning affects only the output weights
. (B) The same network as in A, but without output feedback. Learning takes place both in the network through the modification
, to implement the effect of the feedback loop, and at the output weights
, to correctly learn
.
for different
values. The “error” here is either
(red points and curve) or
(blue points and curve). The input was
with
and
= 0, 0.4, 0.6 in the three panels, from left to right. The value of
was adjusted by changing
. Inserts show the PC eigenvalues (blue) and the exponential fits to them (red), using the value of
indicated. Logarithms are base 10.
in A and
in B for the five sample units. The similarity in these traces shows that the transfer was successful at getting the recurrent input in B to approximate well the feedback input in A for each unit.
for a correlation matrix from simulations similar to those in figure 2 used to demonstrate the approximately Gaussian distribution for the elements of
. The red distribution in the front is for
, and the black distribution in the back is for
, with intermediate layers corresponding to intermediate values. The
matrix was randomly initialized for each value of
.Similar articles
-
A Geometrical Analysis of Global Stability in Trained Feedback Networks.Neural Comput. 2019 Jun;31(6):1139-1182. doi: 10.1162/neco_a_01187. Epub 2019 Apr 12. Neural Comput. 2019. PMID: 30979353
-
Multi-source sequential knowledge regression by using transfer RNN units.Neural Netw. 2019 Nov;119:151-161. doi: 10.1016/j.neunet.2019.08.004. Epub 2019 Aug 17. Neural Netw. 2019. PMID: 31446234
-
A novel time series analysis approach for prediction of dialysis in critically ill patients using echo-state networks.BMC Med Inform Decis Mak. 2010 Jan 21;10:4. doi: 10.1186/1472-6947-10-4. BMC Med Inform Decis Mak. 2010. PMID: 20092639 Free PMC article.
-
Basic concepts of artificial neural network (ANN) modeling and its application in pharmaceutical research.J Pharm Biomed Anal. 2000 Jun;22(5):717-27. doi: 10.1016/s0731-7085(99)00272-1. J Pharm Biomed Anal. 2000. PMID: 10815714 Review.
-
Supervised learning in spiking neural networks: A review of algorithms and evaluations.Neural Netw. 2020 May;125:258-280. doi: 10.1016/j.neunet.2020.02.011. Epub 2020 Feb 25. Neural Netw. 2020. PMID: 32146356 Review.
Cited by
-
Building functional networks of spiking model neurons.Nat Neurosci. 2016 Mar;19(3):350-5. doi: 10.1038/nn.4241. Nat Neurosci. 2016. PMID: 26906501 Free PMC article. Review.
-
Working Memory Requires a Combination of Transient and Attractor-Dominated Dynamics to Process Unreliably Timed Inputs.Sci Rep. 2017 May 30;7(1):2473. doi: 10.1038/s41598-017-02471-z. Sci Rep. 2017. PMID: 28559576 Free PMC article.
-
Learning Universal Computations with Spikes.PLoS Comput Biol. 2016 Jun 16;12(6):e1004895. doi: 10.1371/journal.pcbi.1004895. eCollection 2016 Jun. PLoS Comput Biol. 2016. PMID: 27309381 Free PMC article.
-
From statistical inference to a differential learning rule for stochastic neural networks.Interface Focus. 2018 Dec 6;8(6):20180033. doi: 10.1098/rsfs.2018.0033. Epub 2018 Oct 19. Interface Focus. 2018. PMID: 30443331 Free PMC article.
-
Constraints of Metabolic Energy on the Number of Synaptic Connections of Neurons and the Density of Neuronal Networks.Front Comput Neurosci. 2018 Nov 20;12:91. doi: 10.3389/fncom.2018.00091. eCollection 2018. Front Comput Neurosci. 2018. PMID: 30524259 Free PMC article.
References
-
- Doya K. Bifurcations in the learning of recurrent neural networks. Proceedings of the IEEE International Symposium on Circuits and Systems, ISCAS ’92, vol. 1992;6:2777–2780.
-
- Bengio Y, Simard P, Frasconi P. Learning long-term dependencies with gradient descent is difficult. IEEE Transactions on Neural Networks. 1994;5:157–166. - PubMed
-
- Martens J, Sutskever I. Learning recurrent neural networks with hessian-free optimization. Proceedings of the 28th International Conference on Machine Learning. 2011;4 Available: http://www.cs.toronto.edu/~jmartens/docs/RNN_HF.pdf. Accessed 2012 May.
-
- Maass W, Natschläger T, Markram H. Real-time computing without stable states: a new framework for neural computation based on perturbations. Neural Computation. 2002;14:2531–2560. - PubMed
-
- Jaeger H. Adaptive nonlinear system identification with echo state networks. In: Becker S, Thrun S, Obermayer K, editors. Advances in Neural Information Processing Systems 15. Cambridge, MA: MIT Press. 1713 pp; 2003.
Publication types
MeSH terms
Grants and funding
LinkOut - more resources
Full Text Sources
Other Literature Sources
