Accelerating Experience Replay for Deep Q-Networks with Reduced Target Computation
dc.contributor.author | Zigon, Bob | |
dc.contributor.author | Song, Fengguang | |
dc.contributor.department | Computer and Information Science, School of Science | |
dc.date.accessioned | 2023-11-07T16:28:10Z | |
dc.date.available | 2023-11-07T16:28:10Z | |
dc.date.issued | 2023 | |
dc.description.abstract | Mnih’s seminal deep reinforcement learning paper that applied a Deep Q-network to Atari video games demonstrated the importance of a replay buffer and a target network. Though the pair were required for convergence, the use of the replay buffer came at a significant computational cost. With each new sample generated by the system, the targets in the mini batch buffer were continually recomputed. We propose an alternative that eliminates the target recomputation called TAO-DQN (Target Accelerated Optimization-DQN). Our approach focuses on a new replay buffer algorithm that lowers the computational burden. We implemented this new approach on three experiments involving environments from the OpenAI gym. This resulted in convergence to better policies in fewer episodes and less time. Furthermore, we offer a mathematical justification for our improved convergence rate. | |
dc.eprint.version | Final published version | |
dc.identifier.citation | Ziggon, B., Song, F., & Coulter, B. (2023). Accelerating Experience Replay for Deep Q-Networks with Reduced Target Computation. CS & IT Conference Proceedings, 13(1), Article 1. https://doi.org/10.5121/csit.2023.130101 | |
dc.identifier.uri | https://hdl.handle.net/1805/36952 | |
dc.language.iso | en_US | |
dc.publisher | CS & IT | |
dc.relation.isversionof | 10.5121/csit.2023.130101 | |
dc.relation.journal | CS & IT Conference Proceedings | |
dc.rights | Attribution 4.0 International | en |
dc.rights.uri | http://creativecommons.org/licenses/by/4.0/ | |
dc.source | Publisher | |
dc.subject | DQN | |
dc.subject | Experience Replay | |
dc.subject | Replay Buffer | |
dc.subject | Target Network | |
dc.title | Accelerating Experience Replay for Deep Q-Networks with Reduced Target Computation | |
dc.type | Conference proceedings |