Enhanced dqn framework for selecting actions and updating replay memory considering massive non-executable actions

Bonwoo Gu, Yunsick Sung

Research output: Contribution to journalArticlepeer-review

5 Scopus citations

Abstract

A Deep-Q-Network (DQN) controls a virtual agent as the level of a player using only screenshots as inputs. Replay memory selects a limited number of experience replays according to an arbitrary batch size and updates them using the associated Q-function. Hence, relatively fewer experience replays of different states are utilized when the number of states is fixed and the state of the randomly selected transitions becomes identical or similar. The DQN may not be applicable in some environments where it is necessary to perform the learning process using more experience replays than is required by the limited batch size. In addition, because it is unknown whether each action can be executed, a problem of an increasing amount of repetitive learning occurs as more non-executable actions are selected. In this study, an enhanced DQN framework is proposed to resolve the batch size problem and reduce the learning time of a DQN in an environment with numerous non-executable actions. In the proposed framework, non-executable actions are filtered to reduce the number of selectable actions to identify the optimal action for the current state. The proposed method was validated in Gomoku, a strategy board game, in which the application of a traditional DQN would be difficult.

Original languageEnglish
Article number11162
JournalApplied Sciences (Switzerland)
Volume11
Issue number23
DOIs
StatePublished - 1 Dec 2021

Keywords

  • Deep-Q-Network
  • Game artificial intelligence
  • Gomoku
  • Reinforcement learning
  • Replay memory

Fingerprint

Dive into the research topics of 'Enhanced dqn framework for selecting actions and updating replay memory considering massive non-executable actions'. Together they form a unique fingerprint.

Cite this