Exploration in Reinforcement Learning Workshop
Benjamin Eysenbach · Benjamin Eysenbach · Surya Bhupatiraju · Shixiang Gu · Harrison Edwards · Martha White · Pierre-Yves Oudeyer · Kenneth Stanley · Emma Brunskill

Sat Jun 15th 08:30 AM -- 06:00 PM @ Hall A
Event URL: »

Exploration is a key component of reinforcement learning (RL). While RL has begun to solve relatively simple tasks, current algorithms cannot complete complex tasks. Our existing algorithms often endlessly dither, failing to meaningfully explore their environments in search of high-reward states. If we hope to have agents autonomously learn increasingly complex tasks, these machines must be equipped with machinery for efficient exploration.

The goal of this workshop is to present and discuss exploration in RL, including deep RL, evolutionary algorithms, real-world applications, and developmental robotics. Invited speakers will share their perspectives on efficient exploration, and researchers will share recent work in spotlight presentations and poster sessions.

09:00 AM Doina Precup (Keynote) Video » 
09:30 AM Spotlight Talks <span> <a href="#"> <span class="maincard_media maincard_media_Video"> <a href="" class="btn btn-default btn-xs href_Video" title="Video"><i class="fa fa-video-camera"></i> Video &raquo;</a>&nbsp; </span></a> </span>
10:00 AM Poster Session #1 (Poster Session)
Adrien Ali Taiga, Aniket Deshmukh, Tabish Rashid, Jonathan Binas, Niko Yasui, Vitchyr Pong, Takahisa Imagawa, Jesse Clifton, Sid Mysore, Shi-Chun Tsai, Caleb Chuck, Giulia Vezzani, Hannes Bengt Eriksson
11:00 AM Emo Todorov (Invited Talk) Video » 
11:30 AM Best Paper Talks <span> <a href="#"> <span class="maincard_media maincard_media_Video"> <a href="" class="btn btn-default btn-xs href_Video" title="Video"><i class="fa fa-video-camera"></i> Video &raquo;</a>&nbsp; </span></a> </span>
12:00 PM Pieter Abbeel (Invited Talk) Video » 
12:30 PM Lunch <span> <a href="#"></a> </span>
02:00 PM Raia Hadsell (Invited Talk)
02:30 PM Lightning Talks <span> <a href="#"></a> </span>
03:00 PM Poster Session #2 (Poster Session)
04:00 PM Martha White - Adapting Behaviour via Intrinsic Rewards to Learn Predictions (Invited Talk)
04:30 PM Panel Discussion <span> <a href="#"></a> </span>

Author Information

Benjamin Eysenbach (CMU, Google Brain)
Benjamin Eysenbach (Google)
Surya Bhupatiraju (Google Brain)
Shixiang Gu (Google)
Harrison Edwards (OpenAI / University of Edinburgh)
Martha White (University of Alberta)
Pierre-Yves Oudeyer (Inria)

Dr. Pierre-Yves Oudeyer is Research Director (DR1) at Inria and head of the Inria and Ensta-ParisTech FLOWERS team (France). Before, he has been a permanent researcher in Sony Computer Science Laboratory for 8 years (1999-2007). After working on computational models of language evolution, he is now working on developmental and social robotics, focusing on sensorimotor development, language acquisition and life-long learning in robots. Strongly inspired by infant development, the mechanisms he studies include artificial curiosity, intrinsic motivation, the role of morphology in learning motor control, human-robot interfaces, joint attention and joint intentional understanding, and imitation learning. He has published a book, more than 80 papers in international journals and conferences, holds 8 patents, gave several invited keynote lectures in international conferences, and received several prizes for his work in developmental robotics and on the origins of language. In particular, he is laureate of the ERC Starting Grant EXPLORERS. He is editor of the IEEE CIS Newsletter on Autonomous Mental Development, and associate editor of IEEE Transactions on Autonomous Mental Development, Frontiers in Neurorobotics, and of the International Journal of Social Robotics. He is also working actively for the diffusion of science towards the general public, through the writing of popular science articles and participation to radio and TV programs as well as science exhibitions. Web: and

Kenneth Stanley (Uber AI and University of Central Florida)
Kenneth Stanley

Kenneth O. Stanley is Charles Millican Professor of Computer Science at the University of Central Florida and director there of the Evolutionary Complexity Research Group. He was also a co-founder of Geometric Intelligence Inc., which was acquired by Uber to create Uber AI Labs, where he is now also a senior research science manager and head of Core AI research. He received a B.S.E. from the University of Pennsylvania in 1997 and received a Ph.D. in 2004 from the University of Texas at Austin. He is an inventor of the Neuroevolution of Augmenting Topologies (NEAT), HyperNEAT, and novelty search neuroevolution algorithms for evolving complex artificial neural networks. His main research contributions are in neuroevolution (i.e. evolving neural networks), generative and developmental systems, coevolution, machine learning for video games, interactive evolution, and open-ended evolution. He has won best paper awards for his work on NEAT, NERO, NEAT Drummer, FSMC, HyperNEAT, novelty search, and Galactic Arms Race. His original 2002 paper on NEAT also received the 2017 ISAL Award for Outstanding Paper of the Decade 2002 - 2012 from the International Society for Artificial Life. He is a coauthor of the popular science book, "Why Greatness Cannot Be Planned: The Myth of the Objective" (published by Springer), and has spoken widely on its subject.

Emma Brunskill (Stanford University)

More from the Same Authors