Loading…

An Adaptive Approach for the Exploration-Exploitation Dilemma in Non-stationary Environment

A central problem in reinforcement learning is balancing exploration-exploitation dilemma in non-stationary environment. To address this problem, a data-driven Q-learning is presented. In this study, firstly, the information system of behavior is formed by experience of agent. Then the trigger mecha...

Full description

Saved in:
Bibliographic Details
Main Authors: Yuanxia Shen, Chuanhua Zeng
Format: Conference Proceeding
Language:English
Subjects:
Online Access:Request full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:A central problem in reinforcement learning is balancing exploration-exploitation dilemma in non-stationary environment. To address this problem, a data-driven Q-learning is presented. In this study, firstly, the information system of behavior is formed by experience of agent. Then the trigger mechanism of environment is constructed to trace changes of environment by uncertain knowledge of information system. The dynamic information of environment is used to balance exploration-exploitation dilemma with self-driven way. We illustrated this algorithm with grid-world navigation tasks. The results of simulated experiments show that this algorithm improves learning efficiency obviously.
DOI:10.1109/CSSE.2008.677