1st Workshop on X-Embodiment Robot Learning

Conference on Robot Learning, 2024

Robot foundation models that can control robots to perform a wide range of tasks across diverse scenes have the potential to change how robot learning is done: instead of training policies for every new task from scratch, such generalist policies can be quickly fine-tuned for new tasks and generalize to new scenes, objects and language instructions. Two challenges have made the paradigm shift towards generalist models harder in robotics than other fields of machine learning: the scarcity of robot training data and the diversity of robot platforms and sensory suites that such robot foundation models need to operate on. Recently, cross-embodied policy training has emerged as a solution to both problems: by training on data from many robot embodiments, we can multiply the amount of robot training data and create models that can handle a wide range of sensors and actuators. Following the success of the recent Open X-Embodiment collaboration, that created the largest robot dataset to date, this workshop aims to provide a discussion forum for ideas around cross-embodied policy learning and the future of generalist, cross-embodied robot foundation models. 

We invite submissions at the intersection of scalable and cross-embodied robot learning. Relevant topics include, but are not limited to:


Invited Speakers:

Shuran Song

Assistant Professor, Stanford University

Kevin Black


PhD Student, UC Berkeley

Physical Intelligence

Ryan Julian

Senior Research Scientist, Google Deepmind

Nicolas Heess

Research Scientist, Google Deepmind

Kiana Ehsani 

Sr Research Scientist, AllenAI

Yuke Zhu

Assistant Professor, UT Austin, NVIDIA

Edward Johns

Senior Lecturer, Imperial College

Sergey Levine 

Associate Professor, UC Berkeley

Panelists:

Sergey Levine 

Edward Johns

Nicolas Heess

Dorsa Sadigh

Russ Tedrake

Georgia Chalvatzaki

Workshop Schedule:

Workshop Schedule

Spotlight papers:

FLaRe: Achieving Masterful and Adaptive Robot Policies with Large-Scale Reinforcement Learning Fine-Tuning
Jiaheng Hu, Rose Hendrix, Ali Farhadi, Aniruddha Kembhavi, Roberto Martín-Martín, Peter Stone, Kuo-Hao Zeng, Kiana Ehsani

Data Scaling Laws in Imitation Learning for Robotic Manipulation
Yingdong Hu, Fanqi Lin, Pingyue Sheng, Chuan Wen, Jiacheng You, Yang Gao

Flow as the Cross-Domain Manipulation Interface
Mengda Xu, Zhenjia Xu, Yinghao Xu, Cheng Chi, Gordon Wetzstein, Manuela Veloso, Shuran Song

Organizers:

Karl Pertsch

UC Berkeley & Stanford

Keerthana Gopalakrishnan

Google DeepMind

Lawrence Chen

UC Berkeley

Lucy Shi

Stanford

Ted Xiao

Google DeepMind

Quan Vuong

Physical Intelligence

Pannag Sanketi

Google DeepMind

Christine Chan

Google DeepMind

Ken Goldberg

UC Berkeley

Gaurav Sukhatme

USC

Chelsea Finn

Stanford

Call for Papers

Submission Deadline: October 10 October 03, 2024, 11:59 AOE

Submission Portal: OpenReview

Workshop Date: Nov 09, 2024

We invite submissions with up to 8 pages for the main paper, and unlimited references / appendices. Submissions are encouraged to use the CoRL template. We welcome relevant submissions that were recently published at other venues (e.g. NeurIPS / ICML / ICLR), but ask authors to indicate this upon submission. All accepted papers will be presented as posters, and select papers as spotlight talks.

Sponsors: