Accelerating Experience Replay for Deep Q-Networks with Reduced Target Computation

If you need an accessible version of this item, please email your request to digschol@iu.edu so that they may create one and provide it to you.
Date
2023
Language
American English
Embargo Lift Date
Committee Members
Degree
Degree Year
Department
Grantor
Journal Title
Journal ISSN
Volume Title
Found At
CS & IT
Abstract

Mnih’s seminal deep reinforcement learning paper that applied a Deep Q-network to Atari video games demonstrated the importance of a replay buffer and a target network. Though the pair were required for convergence, the use of the replay buffer came at a significant computational cost. With each new sample generated by the system, the targets in the mini batch buffer were continually recomputed. We propose an alternative that eliminates the target recomputation called TAO-DQN (Target Accelerated Optimization-DQN). Our approach focuses on a new replay buffer algorithm that lowers the computational burden. We implemented this new approach on three experiments involving environments from the OpenAI gym. This resulted in convergence to better policies in fewer episodes and less time. Furthermore, we offer a mathematical justification for our improved convergence rate.

Description
item.page.description.tableofcontents
item.page.relation.haspart
Cite As
Ziggon, B., Song, F., & Coulter, B. (2023). Accelerating Experience Replay for Deep Q-Networks with Reduced Target Computation. CS & IT Conference Proceedings, 13(1), Article 1. https://doi.org/10.5121/csit.2023.130101
ISSN
Publisher
Series/Report
Sponsorship
Major
Extent
Identifier
Relation
Journal
CS & IT Conference Proceedings
Source
Publisher
Alternative Title
Type
Conference proceedings
Number
Volume
Conference Dates
Conference Host
Conference Location
Conference Name
Conference Panel
Conference Secretariat Location
Version
Final published version
Full Text Available at
This item is under embargo {{howLong}}