The search functionality is under construction.

IEICE TRANSACTIONS on Information

Deep Reinforcement Learning with Sarsa and Q-Learning: A Hybrid Approach

Zhi-xiong XU, Lei CAO, Xi-liang CHEN, Chen-xi LI, Yong-liang ZHANG, Jun LAI

  • Full Text Views

    0

  • Cite this

Summary :

The commonly used Deep Q Networks is known to overestimate action values under certain conditions. It's also proved that overestimations do harm to performance, which might cause instability and divergence of learning. In this paper, we present the Deep Sarsa and Q Networks (DSQN) algorithm, which can considered as an enhancement to the Deep Q Networks algorithm. First, DSQN algorithm takes advantage of the experience replay and target network techniques in Deep Q Networks to improve the stability of neural networks. Second, double estimator is utilized for Q-learning to reduce overestimations. Especially, we introduce Sarsa learning to Deep Q Networks for removing overestimations further. Finally, DSQN algorithm is evaluated on cart-pole balancing, mountain car and lunarlander control task from the OpenAI Gym. The empirical evaluation results show that the proposed method leads to reduced overestimations, more stable learning process and improved performance.

Publication
IEICE TRANSACTIONS on Information Vol.E101-D No.9 pp.2315-2322
Publication Date
2018/09/01
Publicized
2018/05/22
Online ISSN
1745-1361
DOI
10.1587/transinf.2017EDP7278
Type of Manuscript
PAPER
Category
Artificial Intelligence, Data Mining

Authors

Zhi-xiong XU
  PLA University of Science and Technology
Lei CAO
  PLA University of Science and Technology
Xi-liang CHEN
  PLA University of Science and Technology
Chen-xi LI
  PLA University of Science and Technology
Yong-liang ZHANG
  PLA University of Science and Technology
Jun LAI
  PLA University of Science and Technology

Keyword