Wartungsarbeiten: Am 13.04..2026 von ca 10:30 bis 11:30 Uhr steht Ihnen das System nicht zur Verfügung. Bitte stellen Sie sich entsprechend darauf ein. Maintenance: at 2026-04-13 the system will be unavailable from 10.30 a.m. until 11.30 a.m. Please plan accordingly.

Memory-based deep reinforcement learning in endless imperfect information games

dc.contributor.advisorRudolph, Günter
dc.contributor.authorPleines, Marco
dc.contributor.refereePreuss, Mike
dc.date.accepted2023-12-11
dc.date.accessioned2024-03-15T10:07:51Z
dc.date.available2024-03-15T10:07:51Z
dc.date.issued2023
dc.description.abstractMemory capabilities in Deep Reinforcement Learning (DRL) agents have become increasingly crucial, especially in tasks characterized by partial observability or imperfect information. However, the field faces two significant challenges: the absence of a universally accepted benchmark and limited access to open-source baseline implementations. We present "Memory Gym", a novel benchmark suite encompassing both finite and endless versions of the Mortar Mayhem, Mystery Path, and Searing Spotlights environments. The finite tasks emphasize strong dependencies on memory and memory interactions, while the remarkable endless tasks, inspired by the game "I packed my bag", act as an automatic curriculum, progressively challenging an agent's retention and recall capabilities. To complement this benchmark, we provide two comprehensible and open-source baselines anchored on the widely-adopted Proximal Policy Optimization algorithm. The first employs a recurrent mechanism through a Gated Recurrent Unit (GRU) cell, while the second adopts an attention-based approach using Transformer-XL (TrXL) for episodic memory with a sliding window. Given the dearth of readily available transformer-based DRL implementations, our TrXL baseline offers significant value. Our results reveal an intriguing performance dynamic: TrXL is often superior in finite tasks, but in the endless environments, GRU unexpectedly marks a comeback. This discrepancy prompts further investigation into TrXL's potential limitations, including whether its initial query misses temporal cues, the impact of stale hidden states, and the intricacies of positional encoding.en
dc.identifier.urihttp://hdl.handle.net/2003/42391
dc.identifier.urihttp://dx.doi.org/10.17877/DE290R-24227
dc.language.isoende
dc.subjectMemory-based agentsen
dc.subjectDeep reinforcement learningen
dc.subjectbenchmarkingen
dc.subjectTransformer-XLen
dc.subjectGated recurrent uniten
dc.subject.ddc004
dc.subject.rswkAgent <Informatik>de
dc.subject.rswkDeep learningde
dc.subject.rswkBenchmarkde
dc.titleMemory-based deep reinforcement learning in endless imperfect information gamesen
dc.typeTextde
dc.type.publicationtypePhDThesisde
dcterms.accessRightsopen access
eldorado.dnb.deposittruede
eldorado.secondarypublicationfalsede

Dateien

Originalbündel

Gerade angezeigt 1 - 1 von 1
Lade...
Vorschaubild
Name:
Dissertation_Pleines.pdf
Größe:
3.89 MB
Format:
Adobe Portable Document Format
Beschreibung:
DNB

Lizenzbündel

Gerade angezeigt 1 - 1 von 1
Lade...
Vorschaubild
Name:
license.txt
Größe:
4.85 KB
Format:
Item-specific license agreed upon to submission
Beschreibung:

Sammlungen