Skip to content
@OpenHelix-Team

OpenHelix Robotics

OpenHelix Robotics: Building Next-generation Embodiment Intelligence

We are a group focused on vision-language-action models (VLAs). We wish to bring insights to the community with our research.

GitHub User's stars Followers

Introduction

OpenHelix-Team introduces a novel family of fully open-source Vision-Language-Action Models (VLAs) that achieves state-of-the-art performance with substantially lower cost.

Awesome VLAs

Multimodal Large Language Models

  • Cobra (AAAI 2025): Extending Mamba to Multi-modal Large Language Model for Efficient Inference

General Foundation Models

  • VLA-Adapter (AAAI 2026 (Oral)): An Effective Paradigm for Tiny-Scale Vision-Language-Action Model
  • LLaVA-VLA (ICRA 2026): A Simple Yet Powerful Vision-Language-Action Model

Visual Feature Alignment for VLAs

  • ReconVLA (AAAI 2026 Best Paper Award): Reconstructive Vision-Language-Action Model as Effective Robot Perceiver
  • Spatial Forcing (ICLR 2026): Implicit Spatial Representation Alignment for Vision-Language-Action Model

World-modeling VLAs

  • Unified Diffusion VLA (ICLR 2026): The first open-sourced diffusion Vision-Language-Action model
  • HiF-VLA (CVPR 2026): An efficient, bidirectional spatiotemporal expansion Vision-Language-Action Model
  • frappe: Infusing World Modeling into Generalist Policies via Multiple Future Representation Alignment
  • VLA-RFT: Vision-Language-Action Models with Reinforcement Fine-Tuning

Visual Enhanced Frameworks

  • VLA-2: Empowering Vision-Language-Action Models with an Agentic Framework for Unseen Concept Manipulation
  • LongVLA (CoRL 2025): Unleashing Long-Horizon Capability of Vision-Language-Action Models for Robot Manipulation

Efficient VLAs

  • CEED-VLA: Consistency Vision-Language-Action Model with Early-Exit Decoding
  • OpenHelix: An Open-Source Dual-System Vision-Language-Action Model for Robotic Manipulation

Quadruped VLAs

  • GeRM (IROS 2024): A Generalist Robotic Model with Mixture-of-Experts for Quadruped Robot

Humanoid VLAs

Collaborating Institutions

This initiative is jointly established and co-developed with the following research institutions:

  • Westlake University
  • The Hong Kong University of Science and Technology (Guangzhou)
  • Zhejiang University
  • Tsinghua University
  • Beijing Academy of Artificial Intelligence (BAAI)
  • Xi’an Jiaotong University
  • Beijing University of Posts and Telecommunications

Contact

If you are interested in discussion or joining us, please send emails to [email protected].

Pinned Loading

  1. Awesome-Force-Tactile-VLA Awesome-Force-Tactile-VLA Public

    A paper list of multimodal VLAs

    47 2

  2. ReconVLA ReconVLA Public

    Official implementation of ReconVLA: Reconstructive Vision-Language-Action Model as Effective Robot Perceiver.

    Python 232 18

  3. VLA-Adapter VLA-Adapter Public

    VLA-Adapter: An Effective Paradigm for Tiny-Scale Vision-Language-Action Model

    Python 2.1k 188

  4. OpenHelix OpenHelix Public

    OpenHelix: An Open-source Dual-System VLA Model for Robotic Manipulation

    Python 355 18

  5. cobra cobra Public

    [AAAI-25] Cobra: Extending Mamba to Multi-modal Large Language Model for Efficient Inference

    Python 293 13

Repositories

Showing 10 of 18 repositories

People

This organization has no public members. You must be a member to see who’s a part of this organization.

Top languages

Loading…

Most used topics

Loading…