RynnBrain

Robot
392

Open embodied foundation model with spatiotemporal memory and physical reasoning for robot task planning, navigation, and manipulation

Implemented Skills

No skills extracted yet

README

RynnBrain

💫 Project Page | 🤗 Hugging Face | 🤖 ModelScope | 🚀 Demo | 📚 Cookbooks | 📄 arXiv

Introduction

We present RynnBrain, an embodied foundation model grounded in physical reality. RynnBrain is available in two dense variants (2B and 8B) and one mixture-of-experts (MoE) model (30B-A3B). In addition, we release three post‑trained models: RynnBrain‑Plan (robot task planning), RynnBrain‑Nav (vision-language navigation), and RynnBrain‑CoP (chain-of-point reasoning).

🌟 Key Highlights

  • Comprehensive egocentric understanding: Excels in fine-grained video understanding and egocentric cognition, covering tasks such as embodied QA, counting, and OCR.
  • Diverse spatio-temporal localization: Possesses powerful localization capabilities across episodic memory, enabling precise identification of objects, target areas, and motion trajectories.
  • Physical-space reasoning: Employs an interleaved reasoning strategy that alternates between textual and spatial grounding, ensuring that its reasoning processes are firmly rooted in the physical environment.
  • Physics-aware precise planning: Integrates located affordances and object information into planning, enabling downstream VLA models to execute intricate tasks with fine-grained instructions.

Timeline

discover2/18/2026

Discovered RynnBrain: Alibaba DAMO Academy's embodied foundation model with 392 GitHub stars

View source
Project Info
Updated 2/17/2026