We propose a self-contained, detailed, description of a scalable standardized kernel (RKHS) approach to popular reinforcement learning algorithms, where agents interact with environments having continuous states and discrete actions spaces, dealing with possibly unstructured data. These algorithms, namely Q-Learning, Actor Critic, Q-Value Gradient, Hamilton-Jacobi-Bellman (HJB) and Heuristic Controls, are implemented with a RKHS library using default settings. We show that this approach to reinforcement learning is accurate, robust, sample efficient and versatile, as we benchmark our algorithms in this paper on simple games and use them as a baseline for our applications.
Pufferlib is a high-performance toolkit for research and industry with optimized parallel simulation, environments that run and train at 1M+ steps/second, and tons of quality of life improvements for practitioners. Puffer Website
EverRAG transforms your Evernote into an intelligent, searchable database.
Making my custom gymnasium environment on the game of RISK global domination, and solving it using RL algorithms
Here I wrote down my answers while I studied the book from Barto,Sutton. Programming exercises as well with custom gym environments.
Code from my kaggle competitions and dataset analysis.
A repository containing different RL algorithms implementations on basic environments