TY - GEN
T1 - SocNavGym: A Reinforcement Learning Gym for Social Navigation
AU - Kapoor, Aditya
AU - Swamy, Sushant
AU - Bachiller-Burgos, Pilar
AU - Manso, Luis J.
PY - 2023/11/13
Y1 - 2023/11/13
N2 - It is essential for autonomous robots to be socially compliant while navigating in human-popuated environments. Machine Learning and, especially, Deep Reinforcement Learning have recently gained considerable traction in the field of Social Navigation. This can be partially attributed to the resulting policies not being bound by human limitations in terms of code complexity or the number of variables that are handled. Unfortunately, the lack of safety guarantees and the large data requirements by DRL algorithms make learning in the real world unfeasible. To bridge this gap, simulation environments are frequently used. We propose SocNavGym, an advanced simulation environment for social navigation that can generate a wide variety of social navigation scenarios and facilitates the development of intelligent social agents. SocNavGym is lightweight, fast, easy to use, and can be effortlessly configured to generate different types of social navigation scenarios. It can also be configured to work with different hand-crafted and data-driven social reward signals and to yield a variety of evaluation metrics to benchmark agents’ performance. Further, we also provide a case study where a Dueling-DQN agent is trained to learn social-navigation policies using SocNavGym. The results provide evidence that SocNavGym can be used to train an agent from scratch to navigate in simple as well as complex social scenarios. Our experiments also show that the agents trained using the data-driven reward function display more advanced social compliance in comparison to the heuristic-based reward function.
AB - It is essential for autonomous robots to be socially compliant while navigating in human-popuated environments. Machine Learning and, especially, Deep Reinforcement Learning have recently gained considerable traction in the field of Social Navigation. This can be partially attributed to the resulting policies not being bound by human limitations in terms of code complexity or the number of variables that are handled. Unfortunately, the lack of safety guarantees and the large data requirements by DRL algorithms make learning in the real world unfeasible. To bridge this gap, simulation environments are frequently used. We propose SocNavGym, an advanced simulation environment for social navigation that can generate a wide variety of social navigation scenarios and facilitates the development of intelligent social agents. SocNavGym is lightweight, fast, easy to use, and can be effortlessly configured to generate different types of social navigation scenarios. It can also be configured to work with different hand-crafted and data-driven social reward signals and to yield a variety of evaluation metrics to benchmark agents’ performance. Further, we also provide a case study where a Dueling-DQN agent is trained to learn social-navigation policies using SocNavGym. The results provide evidence that SocNavGym can be used to train an agent from scratch to navigate in simple as well as complex social scenarios. Our experiments also show that the agents trained using the data-driven reward function display more advanced social compliance in comparison to the heuristic-based reward function.
UR - https://ieeexplore.ieee.org/document/10309591
U2 - 10.1109/RO-MAN57019.2023.10309591
DO - 10.1109/RO-MAN57019.2023.10309591
M3 - Conference publication
SN - 979-8-3503-3671-9
T3 - IEEE International Conference on Robot and Human Interactive Communication (RO-MAN)
BT - 2023 32nd IEEE International Conference on Robot and Human Interactive Communication (RO-MAN)
PB - IEEE
T2 - 2023 32nd IEEE International Conference on Robot and Human Interactive Communication
Y2 - 28 August 2023 through 31 August 2023
ER -