A two-step algorithm for learning from unspecific reinforcement

We study a simple learning model based on the Hebb rule to cope with `delayed', unspecific reinforcement. In spite of the unspecific nature of the information-feedback, convergence to asymptotically perfect generalization is observed, with a rate depending, however, in a non-universal way on le...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Kühn, Reimer (VerfasserIn) , Stamatescu, Ion-Olimpiu (VerfasserIn)
Dokumenttyp: Article (Journal)
Sprache:Englisch
Veröffentlicht: 1999
In: Journal of physics. A, Mathematical and theoretical
Year: 1999, Jahrgang: 32, Heft: 31
ISSN:1751-8121
DOI:10.1088/0305-4470/32/31/301
Online-Zugang:Verlag, Volltext: http://dx.doi.org/10.1088/0305-4470/32/31/301
Verlag, Volltext: http://stacks.iop.org/0305-4470/32/i=31/a=301
Volltext
Verfasserangaben:Reimer Kühn and Ion-Olimpiu Stamatescu
Beschreibung
Zusammenfassung:We study a simple learning model based on the Hebb rule to cope with `delayed', unspecific reinforcement. In spite of the unspecific nature of the information-feedback, convergence to asymptotically perfect generalization is observed, with a rate depending, however, in a non-universal way on learning parameters. Asymptotic convergence can be as fast as that of Hebbian learning, but may be slower. Morever, for a certain range of parameter settings, it depends on initial conditions whether the system can reach the regime of asymptotically perfect generalization, or rather approaches a stationary state of poor generalization.
Beschreibung:Gesehen am 06.03.2018
Beschreibung:Online Resource
ISSN:1751-8121
DOI:10.1088/0305-4470/32/31/301