Direct policy search: intrinsic vs. extrinsic perturbations
Publikation: Bidrag til bog/antologi/rapport › Konferencebidrag i proceedings › Forskning › fagfællebedømt
Reinforcement learning (RL) is a biological inspired learning
paradigm based on trial-and-error learning. A successful RL algorithm
has to balance exploration of new behavioral strategies and exploitation
of already obtained knowledge. In the initial learning phase exploration
is the dominant process. Exploration is realized by stochastic perturbations,
which can be applied at different levels. When considering direct
policy search in the space of neural network policies, exploration can be
applied on the synaptic level or on the level of neuronal activity. We
propose neuroevolution strategies (NeuroESs) for direct policy search in
RL. Learning using NeuroESs can be interpreted as modelling of extrinsic
perturbations on the level of synaptic weights. In contrast, policy
gradient methods (PGMs) can be regarded as intrinsic perturbation of
neuronal activity. We compare these two approaches conceptually and
experimentally.
paradigm based on trial-and-error learning. A successful RL algorithm
has to balance exploration of new behavioral strategies and exploitation
of already obtained knowledge. In the initial learning phase exploration
is the dominant process. Exploration is realized by stochastic perturbations,
which can be applied at different levels. When considering direct
policy search in the space of neural network policies, exploration can be
applied on the synaptic level or on the level of neuronal activity. We
propose neuroevolution strategies (NeuroESs) for direct policy search in
RL. Learning using NeuroESs can be interpreted as modelling of extrinsic
perturbations on the level of synaptic weights. In contrast, policy
gradient methods (PGMs) can be regarded as intrinsic perturbation of
neuronal activity. We compare these two approaches conceptually and
experimentally.
Originalsprog | Engelsk |
---|---|
Titel | Workshop New Challenges in Neural Computation |
Redaktører | B. Hammer, T. Villmann |
Antal sider | 7 |
Publikationsdato | 2010 |
Sider | 33-39 |
Status | Udgivet - 2010 |
Eksternt udgivet | Ja |
Begivenhed | Workshop New Challenges in Neural Computatation 2010 - Karlsruhe, Tyskland Varighed: 21 sep. 2010 → 21 sep. 2010 |
Konference
Konference | Workshop New Challenges in Neural Computatation 2010 |
---|---|
Land | Tyskland |
By | Karlsruhe |
Periode | 21/09/2010 → 21/09/2010 |
Navn | Machine Learning Reports |
---|---|
Vol/bind | 04/2010 |
ISSN | 1865-3960 |
Links
- https://www.techfak.uni-bielefeld.de/~fschleif/mlr/mlr_04_2010.pdf
Forlagets udgivne version
ID: 33863042