TY - GEN
T1 - From Simulation to Deployment
T2 - 15th International Conference on Information Technology and Applications 2021
AU - Sridharan, Sreenithi
AU - Shaikh, Talal
N1 - Publisher Copyright:
© 2022, The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
PY - 2022/4/21
Y1 - 2022/4/21
N2 - Reinforcement learning algorithms are time and resource-intensive and can be influenced by the setup of the physical robot environment and its hardware capabilities. Often, in small-scale projects, it is not workable to build a physical robot equipped with good processing capabilities and to set up a fully controlled and monitored environment for reinforcement learning. During this period, it will be more cost-effective for most of the training to be conducted in a simulated environment and then transferred to a physical robot. In this project, two RL experiments were conducted on a simple two-wheeled robot model in a simulated environment. The first was to make the robot start from a completely fallen down position and learn to stand, and the second to start from a balanced position and learn to maintain the position. It was found that starting from a balanced position gave a better performance, and hence, this learned model was used as a baseline for testing on a physical robot such as the LEGO Mindstorms, but it could be seen that the LEGO hardware was not well suited for this kind of intense reinforcement learning algorithms.
AB - Reinforcement learning algorithms are time and resource-intensive and can be influenced by the setup of the physical robot environment and its hardware capabilities. Often, in small-scale projects, it is not workable to build a physical robot equipped with good processing capabilities and to set up a fully controlled and monitored environment for reinforcement learning. During this period, it will be more cost-effective for most of the training to be conducted in a simulated environment and then transferred to a physical robot. In this project, two RL experiments were conducted on a simple two-wheeled robot model in a simulated environment. The first was to make the robot start from a completely fallen down position and learn to stand, and the second to start from a balanced position and learn to maintain the position. It was found that starting from a balanced position gave a better performance, and hence, this learned model was used as a baseline for testing on a physical robot such as the LEGO Mindstorms, but it could be seen that the LEGO hardware was not well suited for this kind of intense reinforcement learning algorithms.
KW - Reinforcement learning
KW - Self-balancing robot
KW - Simulation
UR - http://www.scopus.com/inward/record.url?scp=85129279341&partnerID=8YFLogxK
U2 - 10.1007/978-981-16-7618-5_25
DO - 10.1007/978-981-16-7618-5_25
M3 - Conference contribution
AN - SCOPUS:85129279341
SN - 9789811676178
T3 - Lecture Notes in Networks and Systems
SP - 283
EP - 295
BT - Proceedings of International Conference on Information Technology and Applications. ICITA 2021
A2 - Ullah, Abrar
A2 - Gill, Steve
A2 - Rocha, Álvaro
A2 - Anwar, Sajid
PB - Springer
Y2 - 13 November 2021 through 14 November 2021
ER -