Relating Bayesian learning to training in recurrent networks | IEEE Conference Publication | IEEE Xplore

Relating Bayesian learning to training in recurrent networks


Abstract:

It is demonstrated that a recurrent neural network relying on an error correcting learning algorithm and a localist coding scheme is able to converge to a solution that w...Show More

Abstract:

It is demonstrated that a recurrent neural network relying on an error correcting learning algorithm and a localist coding scheme is able to converge to a solution that would be expected from Bayesian learning. This is possible even without implementing Bayes theorem and without assigning prior probabilities to the model.
Date of Conference: 20-24 July 2003
Date Added to IEEE Xplore: 26 August 2003
Print ISBN:0-7803-7898-9
Print ISSN: 1098-7576
Conference Location: Portland, OR, USA
References is not available for this document.

I. Introduction

Machine Learning and Statistics have many aspects in common. It will be shown that Bayes theorem can contribute to understand and predict to what solution recurrent networks with truncated gradient descent learning algorithms and a localist coding scheme converge. I have chosen a simple recurrent network, SRN [1], for my demonstration.

Select All
1.
J.L. Elman, "Finding structure in time", Cognitive Science, vol. 14, pp. 179-211, 1990.
2.
J. A. Anderson and E. Rosenfeld, Talking Nets [Interview with Bernard Widrow], Cambridge, MA:MIT-Press, pp. 65, 1998.
3.
T.M. Mitchell, Machine Learning, New York:McGraw-Hill, pp. 154-200, 1997.
4.
E.A. Wan, "Neural Network Classification: A Bayesian Interpretation", IEEE Transactions on Neural Networks, vol. 1, pp. 303-305, 1990.
5.
C.M. Bishop, Neural Networks for Pattern Recognition, Oxford:Oxford University Press, pp. 385-439, 1995.
6.
D.J.C. Mackay, "A practical Bayesian framework for backpropagation networks", Neural Computation, vol. 4, no. 5, pp. 448-472, 1992.
7.
S. Haykin, Neural Networks. A Comprehensive Foundation, Upper Saddle River, N.J:Prentice-Hall, pp. 135-148, 1999.
8.
S. Santini and A. Del Bimbo, "Recurrent Neural Networks can be trained to be Maximum a Posteriori Probability Classifiers", Neural Networks, vol. 8, pp. 25-29, 1995.
9.
S. Santini, A. Del Bimbo and R. Jain, "Block-Structured Recurrent Neural Networks", Neural Networks, vol. 8, pp. 135-147, 1995.
10.
P. Wcrbos, Beyond regression: new tools for prediction and analysis in behavior sciences, 1974.
11.
Y. LeCun, "Une Procédure dApprentissagc pour Réseau à Scuil Asymétrique", Cognitiva 85: A la Frontiere de lIntelligence Artificielle des Sciences de la Conaissance des Neurosciences, pp. 599-604, 1985.
12.
D.B. Parker, Learning logic Technical Report TR-47, 1985.
13.
D.E. Rumelhart, G.E. Hinton and R.J. Williams, "Learning Internal Representations by Error Propagation" in Parallel distributed processing, Cambridge, MA:MIT-Press, vol. I, pp. 318-362, 1986.
14.
D. Servan-Schrciber, A. Cleeremans and J.L. McClclland, "Graded State Machines: The Representation of Temporal Contingencies in Simple Recurrent Networks", Machine Learning, vol. 7, pp. 161-193, 1991.
15.
A. Cleeremans, Mechanisms of Implicit Learning. Connectionist Models of Sequence Processing, Cambridge. MA:MIT-Press, pp. 65, 1993.
16.
Z. Dienes, G.T. Altmann and S.J. Gao, "Mapping across Domains Without Feedback: A Neural Network Model of Transfer of Implicit Knowlcdge", Cognitive Science, vol. 23, pp. 53-82, 1999.
17.
R. Spiegel, Human and Machine Learning of Spatio-Temporal Sequences: An Experimental and Computational Investigation, 2002.
18.
P. McLeod, K. Plunkett and E.T. Rolls, Introduction to Connectionist Modelling of Cognitive Processes, Oxford:Oxford University Press, pp. 236, 1998.
19.
R. Spiegel, M. Surct, M.E. Le Pellcy and I.P.L. McLaren, "Analyzing State Dynamics in a Recurrent Neural Network", Proceedings of the IEEE World Congress on Computational Intelligence. Conference: International Joint INNS/IEEE Conference on Neural Networks, pp. 834-839, 2002.
20.
R. Spiegel and I.P.L. McLaren, "Recurrent Neural Networks and Symbol Grounding", Proceedings of the International Joint INNS/IEEE Conference on Neural Networks, pp. 320-325, 2001.
21.
R. Spiegel and I.P.L. McLaren, "Abstract and associatively-based representations in human sequence learning", Phil. Trans. Roy. Soc. London, vol. B, July 2003.

Contact IEEE to Subscribe

References

References is not available for this document.