1st Workshop on X-Embodiment Robot Learning
Conference on Robot Learning, 2024
Recording: https://youtube.com/live/ELUMFpJCUS0
Workshop Papers: https://openreview.net/group?id=robot-learning.org/CoRL/2024/Workshop/XE
Robot foundation models that can control robots to perform a wide range of tasks across diverse scenes have the potential to change how robot learning is done: instead of training policies for every new task from scratch, such generalist policies can be quickly fine-tuned for new tasks and generalize to new scenes, objects and language instructions. Two challenges have made the paradigm shift towards generalist models harder in robotics than other fields of machine learning: the scarcity of robot training data and the diversity of robot platforms and sensory suites that such robot foundation models need to operate on. Recently, cross-embodied policy training has emerged as a solution to both problems: by training on data from many robot embodiments, we can multiply the amount of robot training data and create models that can handle a wide range of sensors and actuators. Following the success of the recent Open X-Embodiment collaboration, that created the largest robot dataset to date, this workshop aims to provide a discussion forum for ideas around cross-embodied policy learning and the future of generalist, cross-embodied robot foundation models.
We invite submissions at the intersection of scalable and cross-embodied robot learning. Relevant topics include, but are not limited to:
Data:
Which data sources should cross-embodied policies be trained on?
How can we drastically increase the amount of available robot training data?
What interfaces are best for scalable data collection for robotics?
How can we measure the quality of policy training datasets?
Models:
Which models are best suited for cross-embodied policy training?
How can we handle the diversity of sensory inputs and action spaces?
How can we leverage pre-trained vision and language models and make use of cheaper-to-collect demonstration data, with or without robot action labels?
Evaluation:
How can we scalably evaluate the performance of policies trained across hundreds of tasks, scenes and robot embodiments?
How can we ensure that such evaluations are reproducible as the number of evaluation environments and platforms increases?
Invited Speakers:
Shuran Song
Assistant Professor, Stanford University
Kevin Black
PhD Student, UC Berkeley
Physical Intelligence
Ryan Julian
Senior Research Scientist, Google Deepmind
Nicolas Heess
Research Scientist, Google Deepmind
Kiana Ehsani
Sr Research Scientist, AllenAI
Yuke Zhu
Assistant Professor, UT Austin, NVIDIA
Edward Johns
Senior Lecturer, Imperial College
Sergey Levine
Associate Professor, UC Berkeley
Panelists:
Sergey Levine
Edward Johns
Nicolas Heess
Dorsa Sadigh
Russ Tedrake
Georgia Chalvatzaki
Workshop Schedule:
Spotlight papers:
FLaRe: Achieving Masterful and Adaptive Robot Policies with Large-Scale Reinforcement Learning Fine-Tuning
Jiaheng Hu, Rose Hendrix, Ali Farhadi, Aniruddha Kembhavi, Roberto Martín-Martín, Peter Stone, Kuo-Hao Zeng, Kiana Ehsani
Data Scaling Laws in Imitation Learning for Robotic Manipulation
Yingdong Hu, Fanqi Lin, Pingyue Sheng, Chuan Wen, Jiacheng You, Yang Gao
Flow as the Cross-Domain Manipulation Interface
Mengda Xu, Zhenjia Xu, Yinghao Xu, Cheng Chi, Gordon Wetzstein, Manuela Veloso, Shuran Song
Organizers:
Karl Pertsch
UC Berkeley & Stanford
Keerthana Gopalakrishnan
Google DeepMind
Lawrence Chen
UC Berkeley
Lucy Shi
Stanford
Ted Xiao
Google DeepMind
Quan Vuong
Physical Intelligence
Pannag Sanketi
Google DeepMind
Christine Chan
Google DeepMind
Ken Goldberg
UC Berkeley
Gaurav Sukhatme
USC
Chelsea Finn
Stanford
Call for Papers
Submission Deadline: October 10 October 03, 2024, 11:59 AOE
Submission Portal: OpenReview
Workshop Date: Nov 09, 2024
We invite submissions with up to 8 pages for the main paper, and unlimited references / appendices. Submissions are encouraged to use the CoRL template. We welcome relevant submissions that were recently published at other venues (e.g. NeurIPS / ICML / ICLR), but ask authors to indicate this upon submission. All accepted papers will be presented as posters, and select papers as spotlight talks.