Reinforcement learning approach to control an inverted pendulum: A general framework for educational purposes
| dc.centro | Escuela de Ingenierías Industriales | es_ES |
| dc.contributor.author | Israilov, Sardor | |
| dc.contributor.author | Fu, Li | |
| dc.contributor.author | Sánchez-Rodríguez, Jesús | |
| dc.contributor.author | Fusco, Franco | |
| dc.contributor.author | Allibert, Guillaume | |
| dc.contributor.author | Raufaste, Christophe | |
| dc.contributor.author | Argentina, Médéric | |
| dc.date.accessioned | 2025-11-07T10:04:33Z | |
| dc.date.available | 2025-11-07T10:04:33Z | |
| dc.date.issued | 2023-02-13 | |
| dc.departamento | Física Aplicada II | es_ES |
| dc.description.abstract | Machine learning is often cited as a new paradigm in control theory, but is also often viewed as empirical and less intuitive for students than classical model-based methods. This is particularly the case for reinforcement learning, an approach that does not require any mathematical model to drive a system inside an unknown environment. This lack of intuition can be an obstacle to design experiments and implement this approach. Reversely there is a need to gain experience and intuition from experiments. In this article, we propose a general framework to reproduce successful experiments and simulations based on the inverted pendulum, a classic problem often used as a benchmark to evaluate control strategies. Two algorithms (basic Q-Learning and Deep Q-Networks (DQN)) are introduced, both in experiments and in simulation with a virtual environment, to give a comprehensive understanding of the approach and discuss its implementation on real systems. In experiments, we show that learning over a few hours is enough to control the pendulum with high accuracy. Simulations provide insights about the effect of each physical parameter and tests the feasibility and robustness of the approach | es_ES |
| dc.description.sponsorship | French National Research Agency | es_ES |
| dc.identifier.citation | Israilov, S., Fu, L., Sánchez-Rodríguez, J., Fusco, F., Allibert, G., Raufaste, C., & Argentina, M. (2023). Reinforcement learning approach to control an inverted pendulum: A general framework for educational purposes. Plos one, 18(2), e0280071. | es_ES |
| dc.identifier.doi | 10.1371/journal.pone.0280071 | |
| dc.identifier.uri | https://hdl.handle.net/10630/40639 | |
| dc.language.iso | eng | es_ES |
| dc.publisher | Plos One | es_ES |
| dc.rights | Atribución 4.0 Internacional | * |
| dc.rights.accessRights | open access | es_ES |
| dc.rights.uri | http://creativecommons.org/licenses/by/4.0/ | * |
| dc.subject | Redes neuronales artificiales | es_ES |
| dc.subject | Aprendizaje automático | es_ES |
| dc.subject | Péndulo | es_ES |
| dc.subject.other | Pendulums | es_ES |
| dc.subject.other | Learning curves | es_ES |
| dc.subject.other | Artificial neural networks | es_ES |
| dc.subject.other | Machine learning | es_ES |
| dc.subject.other | Algorithms | es_ES |
| dc.subject.other | Gaussian noise | es_ES |
| dc.title | Reinforcement learning approach to control an inverted pendulum: A general framework for educational purposes | es_ES |
| dc.type | journal article | es_ES |
| dc.type.hasVersion | VoR | es_ES |
| dspace.entity.type | Publication |
Files
Original bundle
1 - 1 of 1
Loading...
- Name:
- journal.pone.0280071.pdf
- Size:
- 2.26 MB
- Format:
- Adobe Portable Document Format
- Description:

