Poster + Paper
6 June 2022 An expansion on prioritized experience replay with round robin scheduling
Author Affiliations +
Conference Poster
Abstract
With the outstanding accomplishments achieved in recent years, Reinforcement Learning (RL) has become an area where researchers have flocked to in order to find innovative ideas and solutions to challenge and conquer some of the most difficult tasks. While the bulk of the research has been focused on the learning algorithms such as SARSA, Q-Learning, and Genetic, not much attention has been paid to tools used to help these algorithms (e.g. the Experience Replay Buffer). This paper goes over what is believed to be the most accurate Taxonomy of the AI field and briefly covers the Q-Learning algorithm, as it is the base algorithm for this study. Most importantly, it proposes a new Experience Replay Buffer technique, the Round Robin Prioritized Experience Replay Buffer (RRPERB), which aims to help RL agents learn quicker and generalize better to rarely seen states by not completely depriving them of experiences which are ranked as less priority.
© (2022) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
Angel I. Solis "An expansion on prioritized experience replay with round robin scheduling", Proc. SPIE 12113, Artificial Intelligence and Machine Learning for Multi-Domain Operations Applications IV, 121131O (6 June 2022); https://doi.org/10.1117/12.2617504
Advertisement
Advertisement
RIGHTS & PERMISSIONS
Get copyright permission  Get copyright permission on Copyright Marketplace
KEYWORDS
Evolutionary algorithms

Artificial intelligence

Taxonomy

Stochastic processes

Algorithm development

Machine learning

Neuroscience

Back to Top