Please use this identifier to cite or link to this item:
https://hdl.handle.net/20.500.14365/889
Full metadata record
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Demir, Alper | - |
dc.date.accessioned | 2023-06-16T12:47:51Z | - |
dc.date.available | 2023-06-16T12:47:51Z | - |
dc.date.issued | 2023 | - |
dc.identifier.issn | 0924-669X | - |
dc.identifier.issn | 1573-7497 | - |
dc.identifier.uri | https://doi.org/10.1007/s10489-022-04328-z | - |
dc.identifier.uri | https://hdl.handle.net/20.500.14365/889 | - |
dc.description.abstract | Reinforcement Learning faces an important challenge in partially observable environments with long-term dependencies. In order to learn in an ambiguous environment, an agent has to keep previous perceptions in a memory. Earlier memory-based approaches use a fixed method to determine what to keep in the memory, which limits them to certain problems. In this study, we follow the idea of giving the control of the memory to the agent by allowing it to take memory-changing actions. Thus, the agent becomes more adaptive to the dynamics of an environment. Further, we formalize an intrinsic motivation to support this learning mechanism, which guides the agent to memorize distinctive events and enable it to disambiguate its state in the environment. Our overall approach is tested and analyzed on several partial observable tasks with long-term dependencies. The experiments show a clear improvement in terms of learning performance compared to other memory based methods. | en_US |
dc.description.sponsorship | Scientific and Technological Research Council of Turkey [120E427] | en_US |
dc.description.sponsorship | AcknowledgementsThis work is supported by the Scientific and Technological Research Council of Turkey under Grant No. 120E427. Authors would also like to thank Huseyin Aydin, Erkin Cilden and Faruk Polat for their support. | en_US |
dc.language.iso | en | en_US |
dc.publisher | Springer | en_US |
dc.relation.ispartof | Applıed Intellıgence | en_US |
dc.rights | info:eu-repo/semantics/openAccess | en_US |
dc.subject | Memory | en_US |
dc.subject | Intrinsic motivation | en_US |
dc.subject | Partial observability | en_US |
dc.subject | Reinforcement learning | en_US |
dc.subject | Agents | en_US |
dc.title | Learning What To Memorize: Using Intrinsic Motivation To Form Useful Memory in Partially Observable Reinforcement Learning | en_US |
dc.type | Article | en_US |
dc.identifier.doi | 10.1007/s10489-022-04328-z | - |
dc.identifier.scopus | 2-s2.0-85148369315 | - |
dc.department | İzmir Ekonomi Üniversitesi | en_US |
dc.authorscopusid | 57549355800 | - |
dc.identifier.wos | WOS:000937666700001 | - |
dc.relation.publicationcategory | Makale - Uluslararası Hakemli Dergi - Kurum Öğretim Elemanı | en_US |
dc.identifier.scopusquality | Q2 | - |
dc.identifier.wosquality | Q2 | - |
item.openairecristype | http://purl.org/coar/resource_type/c_18cf | - |
item.grantfulltext | open | - |
item.fulltext | With Fulltext | - |
item.cerifentitytype | Publications | - |
item.languageiso639-1 | en | - |
item.openairetype | Article | - |
crisitem.author.dept | 05.05. Computer Engineering | - |
Appears in Collections: | Scopus İndeksli Yayınlar Koleksiyonu / Scopus Indexed Publications Collection WoS İndeksli Yayınlar Koleksiyonu / WoS Indexed Publications Collection |
CORE Recommender
SCOPUSTM
Citations
3
checked on Mar 12, 2025
WEB OF SCIENCETM
Citations
1
checked on Mar 12, 2025
Page view(s)
64
checked on Mar 10, 2025
Download(s)
46
checked on Mar 10, 2025
Google ScholarTM
Check
Altmetric
Items in GCRIS Repository are protected by copyright, with all rights reserved, unless otherwise indicated.