A two-step algorithm for learning from unspecific reinforcement
We study a simple learning model based on the Hebb rule to cope with `delayed', unspecific reinforcement. In spite of the unspecific nature of the information-feedback, convergence to asymptotically perfect generalization is observed, with a rate depending, however, in a non-universal way on le...
Gespeichert in:
| Hauptverfasser: | , |
|---|---|
| Dokumenttyp: | Article (Journal) |
| Sprache: | Englisch |
| Veröffentlicht: |
1999
|
| In: |
Journal of physics. A, Mathematical and theoretical
Year: 1999, Jahrgang: 32, Heft: 31 |
| ISSN: | 1751-8121 |
| DOI: | 10.1088/0305-4470/32/31/301 |
| Online-Zugang: | Verlag, Volltext: http://dx.doi.org/10.1088/0305-4470/32/31/301 Verlag, Volltext: http://stacks.iop.org/0305-4470/32/i=31/a=301 |
| Verfasserangaben: | Reimer Kühn and Ion-Olimpiu Stamatescu |
| Zusammenfassung: | We study a simple learning model based on the Hebb rule to cope with `delayed', unspecific reinforcement. In spite of the unspecific nature of the information-feedback, convergence to asymptotically perfect generalization is observed, with a rate depending, however, in a non-universal way on learning parameters. Asymptotic convergence can be as fast as that of Hebbian learning, but may be slower. Morever, for a certain range of parameter settings, it depends on initial conditions whether the system can reach the regime of asymptotically perfect generalization, or rather approaches a stationary state of poor generalization. |
|---|---|
| Beschreibung: | Gesehen am 06.03.2018 |
| Beschreibung: | Online Resource |
| ISSN: | 1751-8121 |
| DOI: | 10.1088/0305-4470/32/31/301 |