Anda belum login :: 27 Nov 2024 08:31 WIB
Home
|
Logon
Hidden
»
Administration
»
Collection Detail
Detail
Evolution and Learning: Evolving Sensors in a Simple MDP Environment
Oleh:
Jung, Tobias
;
Dauscher, Peter
;
Uthmann, Thomas
Jenis:
Article from Journal - e-Journal
Dalam koleksi:
Adaptive Behavior vol. 11 no. 3 (Sep. 2003)
,
page 159–177.
Topik:
sensor evolution
;
adaptive
;
state generalization
;
cognitive economy
;
relevant information
;
reinforcement learning
Fulltext:
159.pdf
(1,007.74KB)
Isi artikel
Natural intelligence and autonomous agents face difficulties when acting in information-dense environments. Assailed by a multitude of stimuli they have to make sense of the inflow of information, filtering and processing what is necessary, but discarding that which is unimportant. This paper aims at investigating the interactions between evolution of the sensorial channel extracting the information from the environment and the simultaneous individual adaptation of agent-control. Our particular goal is to study the influence of learning on the evolution of sensors, with learning duration being the tunable parameter. A genetic algorithm governs the evolution of sensors appropriate for the agent solving a simple grid world task. The performance of the agent is taken as fitness; ‘sensors’ are conceived as a map from environmental states to agent observations, and individual adaptation is modeled by Qlearning. Our experimental results show that due to the principles of cognitive economy learning and varying the degree thereof actually transforms the fitness landscape. In particular we identify a trade off between learning speed (load) and sensor accuracy (error). These results are further reinforced by theoretical analysis: we derive an analytical measure for the quality of sensors based on the mutual entropy between the system of states and the selection of an optimal action, a concept recently proposed by Polani, Martinetz, and Kim.
Opini Anda
Klik untuk menuliskan opini Anda tentang koleksi ini!
Kembali
Process time: 0.046875 second(s)