Timezone: »
In situations where a task can be cleanly formulated and data is plentiful, modern machine learning (ML) techniques have achieved impressive (and often super-human) results. Here, plentiful data'' can mean labels from humans, access to a simulator and well designed reward function, or other forms of interaction and supervision.<br><br>On the other hand, in situations where tasks cannot be cleanly formulated and plentifully supervised, ML has not yet shown the same progress. We still seem far from flexible agents that can learn without human engineers carefully designing or collating their supervision. This is problematic in many settings where machine learning is or will be applied in real world settings, where these agents have to interact with human users and may be used in settings that go beyond any initial clean training data used during system development. A key open question is how to make machine learning effective and robust enough to operate in real world open domains.<br><br>Artificial {\it open} worlds are ideal laboratories for studying how to extend the successes of ML to build such agents. <br>Open worlds are characterized by:<br>\begin{itemize}<br> \item Large (or perhaps infinite) collections of tasks, often not specified till test time; or lack of well defined tasks altogether (despite there being lots to do).<br> \item
unbounded'' environments, long ``epsiodes''; or no episodes at all.
\item Many interacting agents; more generally, emergent behavior from interactions with the environment.
\end{itemize}
On one hand, they retain many of the challenging features of the real world with respect to studying learning agents. On the other hand,
they allow cheap collection of environment interaction data. Furthermore, because many artificial worlds of interest are games that people enjoy playing, they could allow interaction with humans at scale.
A particularly promising direction is that open world games can bridge: the closed domains and benchmarks that have traditionally driven research progress and open ended real world applications in which resulting technology is deployed.
We propose a workshop designed to catalyze research towards addressing these challenges posed by machine learning in open worlds.
Our goal is to bring together researchers with a wide range of perspectives whose work focuses on, or is enabled by, open worlds. This would be the very first workshop focused on this topic, and we anticipate that it would play a key role in sharing experience, brainstorming ideas, and catalyzing novel directions for research.
Sat 7:00 a.m. - 7:15 a.m.
|
Opening remarks
|
Katja Hofmann 🔗 |
Sat 7:15 a.m. - 7:35 a.m.
|
The NetHack Learning Environment
(
Invited talk 1
)
SlidesLive Video » Progress in Reinforcement Learning (RL) algorithms goes hand-in-hand with the development of challenging environments that test the limits of current methods. While existing RL environments are either sufficiently complex or based on fast simulation, they are rarely both these things. In this talk, I present the NetHack Learning Environment (NLE), a scalable, procedurally generated, stochastic, rich, and challenging environment for RL research based on the popular single-player terminal-based rogue like game, NetHack. We argue that NetHack is sufficiently complex to drive long-term research on problems such as exploration, planning, skill acquisition, and language-conditioned RL, while dramatically reducing the computational resources required to gather a large amount of experience. We compare NLE and its task suite to existing alternatives, and discuss why it is an ideal medium for testing the robustness and systematic generalization of RL agents. We demonstrate empirical success for early stages of the game using a distributed deep RL baseline and Random Network Distillation exploration, alongside qualitative analysis of various agents trained in the environment. NLE is open-source at https://github.com/facebookresearch/nle . |
Tim Rocktäschel 🔗 |
Sat 7:35 a.m. - 7:45 a.m.
|
The NetHack Learning Environment Q&A
(
Invited talk 1 Q&A
)
Please join using the Zoom link and post your questions on Rocketchat. |
Tim Rocktäschel · Katja Hofmann 🔗 |
Sat 7:45 a.m. - 8:05 a.m.
|
Open-ended environments for advancing RL
(
Invited talk 2
)
The field of reinforcement learning is pushed forwards by the presence of challenging environments. Over the years, the complexity of these environments has continued to increase, but the question is how can we continue to push the complexity of environments with respect to the optimal policy complexity in a scalable manner. Here I will discuss using multi-agent environments to create more open-ended environments, and discuss examples of our work to move in this direction with Capture the Flag and Starcraft 2. Finally I will discuss some future directions for generating even more open-ended environments to further push our RL algorithms. |
Max Jaderberg 🔗 |
Sat 8:05 a.m. - 8:15 a.m.
|
Open-ended environments for advancing RL Q&A
(
Invited talk 2 Q&A
)
Please join using the Zoom link and post your questions on Rocketchat. |
Max Jaderberg · Katja Hofmann 🔗 |
Sat 8:15 a.m. - 8:30 a.m.
|
Coffee break
(
coffee break
)
|
Kavya Srinet 🔗 |
Sat 8:30 a.m. - 8:50 a.m.
|
Collaborative Construction and Communication in Minecraft
(
Invited talk 3
)
I will present work done by my group on defining a collaborative construction task that allows us to use the Minecraft platform to study situated natural language generation and understanding. In this task, one player (the Architect) needs to instruct another (the Builder) via a chat interface to construct a given target structure that only the Architect is shown. I will discuss what makes this task interesting and challenging. I will also describe models that we have developed for the Architect and the Builder role, and discuss what remains to be done to create agents that can solve this task. |
Julia Hockenmaier 🔗 |
Sat 8:50 a.m. - 9:00 a.m.
|
Collaborative Construction and Communication in Minecraft Q&A
(
Invited talk 3 Q&A
)
Please join using the Zoom link and post your questions on Rocketchat. |
Julia Hockenmaier · Arthur Szlam 🔗 |
Sat 9:00 a.m. - 9:20 a.m.
|
Collaboration in Situated Instruction Following
(
Invited talk 4
)
SlidesLive Video » I will focus on the problem of executing natural language instructions in a collaborative environment. I will propose the task of learning to follow sequences of instructions in a collaborative scenario, where two agents, a leader and a follower, execute actions in the environment and the leader controls the follower using natural language. To study this problem, we build CerealBar, a multi-player 3D game where a leader instructs a follower, and both act in the environment together to accomplish complex goals. I will focus on learning an autonomous follower that executes the instructions of a human leader. I will briefly describe a model to address this problem, and a learning method that relies on static recorded human-human interactions, while still learning to recover from cascading errors between instructions. |
Yoav Artzi 🔗 |
Sat 9:20 a.m. - 9:30 a.m.
|
Collaboration in Situated Instruction Following Q&A
(
Invited talk 4 Q&A
)
Please join using the Zoom link and post your questions on Rocketchat. |
Yoav Artzi · Arthur Szlam 🔗 |
Sat 9:30 a.m. - 10:00 a.m.
|
Lunch and networking
|
Kavya Srinet 🔗 |
Sat 10:00 a.m. - 10:20 a.m.
|
What an Agent Knows: Evaluation in Open Worlds
(
Invited talk 5
)
SlidesLive Video » Agents tackling complex problems in open environments often benefit from the ability to construct knowledge. Learning to independently solve sub-tasks and form models of the world can help agents progress in solving challenging problems. In this talk, we draw attention to challenges that arise when evaluating an agent’s knowledge, specifically focusing on methods that express an agent’s knowledge as predictions. Using the General Value Function framework we highlight the distinction between useful knowledge and strict measures of accuracy. Having identified challenges in assessing an agent’s knowledge, we propose a possible evaluation approach that is compatible with large and open worlds. |
Alex Kearney 🔗 |
Sat 10:20 a.m. - 10:30 a.m.
|
What an Agent Knows: Evaluation in Open Worlds Q&A
(
Invited talk 5 Q&A
)
Please join using the Zoom link and post your questions on Rocketchat. |
Alex Kearney · William Guss 🔗 |
Sat 10:30 a.m. - 10:50 a.m.
|
Endless Frontiers?
(
Invited talk 6
)
The research community is gradually coming to a realization that policies trained arcade-like video games are very limited. They overfit badly and are not going to take us far along the way to some sort of general intelligence. This should perhaps not be surprising, given that such games generally have tightly defined tasks, fixed perspectives, and generally static worlds. More and more attention is therefore given to games that are in some sense open-ended or feature open worlds. Could such games be the solution to our problems, allowing the development of more general artificial intelligence? Perhaps, but basing competitions or benchmarks on open-ended games is not going to be easy, as the very features which make for a good benchmark are the same that lead to brittle policies. Shoe-horning open-world games into a standard RL framework is unlikely to be the best option for going forward. Many of the most interesting opportunities for developing intelligent behavior is likely to come from agents constructing their own challenges and environments. The boundary between playing a game and constructing a world is not well-defined: I will give examples from where the same RL setup was used to play SimCity and to develop game levels. I will also briefly introduce the Generative Design in Minecraft Competition, which focuses on building believable settlements. |
Julian Togelius 🔗 |
Sat 10:50 a.m. - 11:00 a.m.
|
Endless Frontiers? Q&A
(
Invited talk 6 Q&A
)
Please join using the Zoom link and post your questions on Rocketchat. |
Julian Togelius · William Guss 🔗 |
Sat 11:00 a.m. - 11:30 a.m.
|
Coffee break
(
coffee break
)
|
Kavya Srinet 🔗 |
Sat 11:30 a.m. - 1:00 p.m.
|
Poster session
(
Poster session and demo
)
Hey all! Our poster session is hosted in a virtual town. To go to the poster session, please go to : https://gather.town/bWIYAlhNYvQPRkCE/ICML2020-LAOW and use the password: flatfish |
Noboru Kuno · Kavya Srinet · William Guss · Brandon Houghton 🔗 |
Sat 1:00 p.m. - 2:00 p.m.
|
Panel discussion
Join us for a panel discussion with our invited speakers |
Kavya Srinet · Katja Hofmann · Yoav Artzi · Alex Kearney · Julia Hockenmaier 🔗 |
Author Information
Arthur Szlam (Facebook)
Katja Hofmann (Microsoft)
Ruslan Salakhutdinov (Carnegie Mellen University)
Noboru Kuno (Microsoft Research)
Sean Kuno is a Senior Research Program Manager of Microsoft Research Outreach. He is based in Redmond U.S. and he is a member of Artificial Intelligence Engaged team of Microsoft Research Outreach. Kuno leads the ideation, design and launch of community programs for AI projects such as Project Malmo, working in partnership with universities and government agencies worldwide.
William Guss (Carnegie Mellon University)
Kavya Srinet (Facebook AI Research)
Brandon Houghton (OpenAI)
More from the Same Authors
-
2021 : Online Sub-Sampling for Reinforcement Learning with General Function Approximation »
Dingwen Kong · Ruslan Salakhutdinov · Ruosong Wang · Lin Yang -
2023 Poster: Graph Generative Model for Benchmarking Graph Neural Networks »
Minji Yoon · Yue Wu · John Palowitch · Bryan Perozzi · Ruslan Salakhutdinov -
2022 Poster: Recurrent Model-Free RL Can Be a Strong Baseline for Many POMDPs »
Tianwei Ni · Benjamin Eysenbach · Ruslan Salakhutdinov -
2022 Poster: Interactively Learning Preference Constraints in Linear Bandits »
David Lindner · Sebastian Tschiatschek · Katja Hofmann · Andreas Krause -
2022 Spotlight: Recurrent Model-Free RL Can Be a Strong Baseline for Many POMDPs »
Tianwei Ni · Benjamin Eysenbach · Ruslan Salakhutdinov -
2022 Spotlight: Interactively Learning Preference Constraints in Linear Bandits »
David Lindner · Sebastian Tschiatschek · Katja Hofmann · Andreas Krause -
2022 : [Meta Platforms] Building embodied agents that can learn from their environments and humans »
Kavya Srinet -
2021 : Towards Human-like and Collaborative AI in Video Games »
Katja Hofmann -
2021 Poster: CURI: A Benchmark for Productive Concept Learning Under Uncertainty »
Shanmukha Ramakrishna Vedantam · Arthur Szlam · Maximilian Nickel · Ari Morcos · Brenden Lake -
2021 Poster: Towards Understanding and Mitigating Social Biases in Language Models »
Paul Liang · Chiyu Wu · LP Morency · Ruslan Salakhutdinov -
2021 Poster: Reasoning Over Virtual Knowledge Bases With Open Predicate Relations »
Haitian Sun · Patrick Verga · Bhuwan Dhingra · Ruslan Salakhutdinov · William Cohen -
2021 Poster: Exploration in Approximate Hyper-State Space for Meta Reinforcement Learning »
Luisa Zintgraf · Leo Feng · Cong Lu · Maximilian Igl · Kristian Hartikainen · Katja Hofmann · Shimon Whiteson -
2021 Spotlight: CURI: A Benchmark for Productive Concept Learning Under Uncertainty »
Shanmukha Ramakrishna Vedantam · Arthur Szlam · Maximilian Nickel · Ari Morcos · Brenden Lake -
2021 Spotlight: Reasoning Over Virtual Knowledge Bases With Open Predicate Relations »
Haitian Sun · Patrick Verga · Bhuwan Dhingra · Ruslan Salakhutdinov · William Cohen -
2021 Spotlight: Towards Understanding and Mitigating Social Biases in Language Models »
Paul Liang · Chiyu Wu · LP Morency · Ruslan Salakhutdinov -
2021 Spotlight: Exploration in Approximate Hyper-State Space for Meta Reinforcement Learning »
Luisa Zintgraf · Leo Feng · Cong Lu · Maximilian Igl · Kristian Hartikainen · Katja Hofmann · Shimon Whiteson -
2021 Poster: Instabilities of Offline RL with Pre-Trained Neural Representation »
Ruosong Wang · Yifan Wu · Ruslan Salakhutdinov · Sham Kakade -
2021 Poster: TeachMyAgent: a Benchmark for Automatic Curriculum Learning in Deep RL »
Clément Romac · Rémy Portelas · Katja Hofmann · Pierre-Yves Oudeyer -
2021 Spotlight: Instabilities of Offline RL with Pre-Trained Neural Representation »
Ruosong Wang · Yifan Wu · Ruslan Salakhutdinov · Sham Kakade -
2021 Spotlight: TeachMyAgent: a Benchmark for Automatic Curriculum Learning in Deep RL »
Clément Romac · Rémy Portelas · Katja Hofmann · Pierre-Yves Oudeyer -
2021 Poster: Information Obfuscation of Graph Neural Networks »
Peiyuan Liao · Han Zhao · Keyulu Xu · Tommi Jaakkola · Geoff Gordon · Stefanie Jegelka · Ruslan Salakhutdinov -
2021 Poster: Uncertainty Weighted Actor-Critic for Offline Reinforcement Learning »
Yue Wu · Shuangfei Zhai · Nitish Srivastava · Joshua M Susskind · Jian Zhang · Ruslan Salakhutdinov · Hanlin Goh -
2021 Poster: On Proximal Policy Optimization's Heavy-tailed Gradients »
Saurabh Garg · Joshua Zhanson · Emilio Parisotto · Adarsh Prasad · Zico Kolter · Zachary Lipton · Sivaraman Balakrishnan · Ruslan Salakhutdinov · Pradeep Ravikumar -
2021 Spotlight: On Proximal Policy Optimization's Heavy-tailed Gradients »
Saurabh Garg · Joshua Zhanson · Emilio Parisotto · Adarsh Prasad · Zico Kolter · Zachary Lipton · Sivaraman Balakrishnan · Ruslan Salakhutdinov · Pradeep Ravikumar -
2021 Spotlight: Information Obfuscation of Graph Neural Networks »
Peiyuan Liao · Han Zhao · Keyulu Xu · Tommi Jaakkola · Geoff Gordon · Stefanie Jegelka · Ruslan Salakhutdinov -
2021 Spotlight: Uncertainty Weighted Actor-Critic for Offline Reinforcement Learning »
Yue Wu · Shuangfei Zhai · Nitish Srivastava · Joshua M Susskind · Jian Zhang · Ruslan Salakhutdinov · Hanlin Goh -
2021 Poster: Navigation Turing Test (NTT): Learning to Evaluate Human-Like Navigation »
Sam Devlin · Raluca Georgescu · Ida Momennejad · Jaroslaw Rzepecki · Evelyn Zuniga · Gavin Costello · Guy Leroy · Ali Shaw · Katja Hofmann -
2021 Poster: Not All Memories are Created Equal: Learning to Forget by Expiring »
Sainbayar Sukhbaatar · Da JU · Spencer Poff · Stephen Roller · Arthur Szlam · Jason Weston · Angela Fan -
2021 Oral: Not All Memories are Created Equal: Learning to Forget by Expiring »
Sainbayar Sukhbaatar · Da JU · Spencer Poff · Stephen Roller · Arthur Szlam · Jason Weston · Angela Fan -
2021 Spotlight: Navigation Turing Test (NTT): Learning to Evaluate Human-Like Navigation »
Sam Devlin · Raluca Georgescu · Ida Momennejad · Jaroslaw Rzepecki · Evelyn Zuniga · Gavin Costello · Guy Leroy · Ali Shaw · Katja Hofmann -
2020 : Panel discussion »
Kavya Srinet · Katja Hofmann · Yoav Artzi · Alex Kearney · Julia Hockenmaier -
2020 : Poster session »
Noboru Kuno · Kavya Srinet · William Guss · Brandon Houghton -
2020 : Coffee break »
Kavya Srinet -
2020 : Endless Frontiers? Q&A »
Julian Togelius · William Guss -
2020 : What an Agent Knows: Evaluation in Open Worlds Q&A »
Alex Kearney · William Guss -
2020 : Lunch and networking »
Kavya Srinet -
2020 : Collaboration in Situated Instruction Following Q&A »
Yoav Artzi · Arthur Szlam -
2020 : Collaborative Construction and Communication in Minecraft Q&A »
Julia Hockenmaier · Arthur Szlam -
2020 : Coffee break »
Kavya Srinet -
2020 : Open-ended environments for advancing RL Q&A »
Max Jaderberg · Katja Hofmann -
2020 : The NetHack Learning Environment Q&A »
Tim Rocktäschel · Katja Hofmann -
2020 : Opening remarks »
Katja Hofmann -
2020 Workshop: Bridge Between Perception and Reasoning: Graph Neural Networks & Beyond »
Jian Tang · Le Song · Jure Leskovec · Renjie Liao · Yujia Li · Sanja Fidler · Richard Zemel · Ruslan Salakhutdinov -
2020 : Q&A with Katja Hoffman »
Katja Hofmann · Luisa Zintgraf · Rika Antonova · Sarath Chandar · Shagun Sodhani -
2020 : Challenges & Opportunities in Lifelong Reinforcement Learning by Katja Hoffman »
Katja Hofmann · Rika Antonova · Luisa Zintgraf -
2020 Poster: Fast Adaptation to New Environments via Policy-Dynamics Value Functions »
Roberta Raileanu · Max Goldstein · Arthur Szlam · Facebook Rob Fergus -
2019 Poster: Fast Context Adaptation via Meta-Learning »
Luisa Zintgraf · Kyriacos Shiarlis · Vitaly Kurin · Katja Hofmann · Shimon Whiteson -
2019 Oral: Fast Context Adaptation via Meta-Learning »
Luisa Zintgraf · Kyriacos Shiarlis · Vitaly Kurin · Katja Hofmann · Shimon Whiteson -
2019 Talk: Opening Remarks »
Kamalika Chaudhuri · Ruslan Salakhutdinov -
2018 Poster: Transformation Autoregressive Networks »
Junier Oliva · Kumar Avinava Dubey · Manzil Zaheer · Barnabás Póczos · Ruslan Salakhutdinov · Eric Xing · Jeff Schneider -
2018 Oral: Transformation Autoregressive Networks »
Junier Oliva · Kumar Avinava Dubey · Manzil Zaheer · Barnabás Póczos · Ruslan Salakhutdinov · Eric Xing · Jeff Schneider -
2018 Poster: Optimizing the Latent Space of Generative Networks »
Piotr Bojanowski · Armand Joulin · David Lopez-Paz · Arthur Szlam -
2018 Poster: Composable Planning with Attributes »
Amy Zhang · Sainbayar Sukhbaatar · Adam Lerer · Arthur Szlam · Facebook Rob Fergus -
2018 Oral: Composable Planning with Attributes »
Amy Zhang · Sainbayar Sukhbaatar · Adam Lerer · Arthur Szlam · Facebook Rob Fergus -
2018 Oral: Optimizing the Latent Space of Generative Networks »
Piotr Bojanowski · Armand Joulin · David Lopez-Paz · Arthur Szlam -
2018 Poster: Structured Control Nets for Deep Reinforcement Learning »
Mario Srouji · Jian Zhang · Ruslan Salakhutdinov -
2018 Poster: Gated Path Planning Networks »
Lisa Lee · Emilio Parisotto · Devendra Singh Chaplot · Eric Xing · Ruslan Salakhutdinov -
2018 Oral: Structured Control Nets for Deep Reinforcement Learning »
Mario Srouji · Jian Zhang · Ruslan Salakhutdinov -
2018 Oral: Gated Path Planning Networks »
Lisa Lee · Emilio Parisotto · Devendra Singh Chaplot · Eric Xing · Ruslan Salakhutdinov -
2017 : Panel Discussion »
Balaraman Ravindran · Chelsea Finn · Alessandro Lazaric · Katja Hofmann · Marc Bellemare -
2017 Poster: Toward Controlled Generation of Text »
Zhiting Hu · Zichao Yang · Xiaodan Liang · Ruslan Salakhutdinov · Eric Xing -
2017 Poster: Improved Variational Autoencoders for Text Modeling using Dilated Convolutions »
Zichao Yang · Zhiting Hu · Ruslan Salakhutdinov · Taylor Berg-Kirkpatrick -
2017 Talk: Improved Variational Autoencoders for Text Modeling using Dilated Convolutions »
Zichao Yang · Zhiting Hu · Ruslan Salakhutdinov · Taylor Berg-Kirkpatrick -
2017 Talk: Toward Controlled Generation of Text »
Zhiting Hu · Zichao Yang · Xiaodan Liang · Ruslan Salakhutdinov · Eric Xing