Version 1
: Received: 9 May 2020 / Approved: 10 May 2020 / Online: 10 May 2020 (18:02:43 CEST)
How to cite:
Sharma, S. Modeling an Inverted Pendulum via Differential Equations and Reinforcement Learning Techniques. Preprints2020, 2020050181. https://doi.org/10.20944/preprints202005.0181.v1
Sharma, S. Modeling an Inverted Pendulum via Differential Equations and Reinforcement Learning Techniques. Preprints 2020, 2020050181. https://doi.org/10.20944/preprints202005.0181.v1
Sharma, S. Modeling an Inverted Pendulum via Differential Equations and Reinforcement Learning Techniques. Preprints2020, 2020050181. https://doi.org/10.20944/preprints202005.0181.v1
APA Style
Sharma, S. (2020). Modeling an Inverted Pendulum via Differential Equations and Reinforcement Learning Techniques. Preprints. https://doi.org/10.20944/preprints202005.0181.v1
Chicago/Turabian Style
Sharma, S. 2020 "Modeling an Inverted Pendulum via Differential Equations and Reinforcement Learning Techniques" Preprints. https://doi.org/10.20944/preprints202005.0181.v1
Abstract
The prevalence of differential equations as a mathematical technique has refined the fields of control theory and constrained optimization due to the newfound ability to accurately model chaotic, unbalanced systems. However, in recent research, systems are increasingly more nonlinear and difficult to model using Differential Equations only. Thus, a newer technique is to use policy iteration and Reinforcement Learning, techniques that center around an action and reward sequence for a controller. Reinforcement Learning (RL) can be applied to control theory problems since a system can robustly apply RL in a dynamic environment such as the cartpole system (an inverted pendulum). This solution successfully avoids use of PID or other dynamics optimization systems, in favor of a more robust, reward-based control mechanism. This paper applies RL and Q-Learning to the classic cartpole problem, while also discussing the mathematical background and differential equations which are used to model the aforementioned system.
Copyright:
This is an open access article distributed under the Creative Commons Attribution License which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.