Anda belum login :: 23 Nov 2024 20:03 WIB
Detail
ArtikelReactivity and Safe Learning in Multi-Agent Systems  
Oleh: Banerjee, Bikramjit ; Jing Peng
Jenis: Article from Journal - e-Journal
Dalam koleksi: Adaptive Behavior vol. 14 no. 4 (Dec. 2006), page 339–356.
Topik: multi-agent systems; reinforcement learning; game theory
Fulltext: 339.pdf (1.0MB)
Isi artikelMulti-agent reinforcement learning (MRL) is a growing area of research. What makes it particularly challenging is that multiple learners render each other’s environments non-stationary. In addition to adapting their behaviors to other learning agents, online learners must also provide assurances about their online performance in order to promote user trust of adaptive agent systems deployed in real world applications. In this article, instead of developing new algorithms with such assurances, we study the question of safety in online performance of some existing MRL algorithms. We identify the key notion of reactivity of a learner by analyzing how an algorithm (PHC-Exploiter), designed to exploit some simpler opponents, can itself be exploited by them. We quantify and analyze this concept of reactivity in the context of these algorithms to explain their experimental behaviors. We argue that no learner can be designed that can deliberately avoid exploitation. We also show that any attempt to optimize reactivity must take into account a trade off with sensitivity to noise, and devise an adaptive method (based on environmental feedback) designed to maximize the learner’s safety and minimize its sensitivity to noise.
Opini AndaKlik untuk menuliskan opini Anda tentang koleksi ini!

Kembali
design
 
Process time: 0.015625 second(s)