Skip to yearly menu bar Skip to main content


Poster

Position Paper: Automatic Environment Shaping is the Next Frontier in RL

Younghyo Park · Gabriel Margolis · Pulkit Agrawal


Abstract:

Many roboticists dream of presenting a robot with a task in the evening and returning the next morning to find the robot capable of solving the task. What is preventing us from achieving this? Sim-to-real reinforcement learning (RL) has achieved impressive performance on challenging robotics tasks, but requires substantial human effort to set up the task in a way that is amenable to RL. It's our position that algorithmic improvements in policy optimization and other ideas should be guided towards resolving the primary bottleneck of "shaping" the training environment (observation, action, reward, task, dynamics) after the target task is described and before or during running RL. Most practitioners don't tune the RL algorithm, but other environment parameters to obtain a desirable controller. We posit that scaling RL to diverse robotic tasks will only be achieved if the community focuses on automating "shaping".

Live content is unavailable. Log in and register to view live content