Real2Render2Real: Scaling Robot Data Without Dynamics Simulation or Robot Hardware
Abstract
Scaling robot learning requires vast and diverse datasets. Yet the prevailing data collection paradigm-human teleoperation-remains costly and constrained by manual effort and physical robot access. We introduce Real2Render2Real (R2R2R), a novel approach for generating robot training data without relying on object dynamics simulation or teleoperation of robot hardware. The input is a smartphone-captured scan of one or more objects and a single video of a human demonstration. R2R2R renders thousands of high visual fidelity robot-agnostic demonstrations by reconstructing detailed 3D object geometry and appearance, and tracking 6-DoF object motion. R2R2R uses 3D Gaussian Splatting (3DGS) to enable flexible asset generation and trajectory synthesis for both rigid and articulated objects, converting these representations to meshes to maintain compatibility with scalable rendering engines like IsaacLab but with collision modeling off. Robot demonstration data generated by R2R2R integrates directly with models that operate on robot proprioceptive states and image observations, such as vision-language-action models (VLA) and imitation learning policies. Physical experiments suggest that models trained on R2R2R data from a single human demonstration can match the performance of models trained on 150 human teleoperation demonstrations. Project page: https://real2render2real.com
Community
We built a way to easily scale robot datasets without teleop, physics sim, or robot hardware.
1 smartphone scan + 1 human demo โ thousands of diverse robot trajectories.
Trainable by diffusion policy and VLA models as-is.
๐ real2render2real.com
twitter/X: https://x.com/letian_fu/status/1923407715638051060
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- X-Sim: Cross-Embodiment Learning via Real-to-Sim-to-Real (2025)
- RL-Driven Data Generation for Robust Vision-Based Dexterous Grasping (2025)
- Crossing the Human-Robot Embodiment Gap with Sim-to-Real RL using One Human Demonstration (2025)
- GraspVLA: a Grasping Foundation Model Pre-trained on Billion-scale Synthetic Action Data (2025)
- Novel Demonstration Generation with Gaussian Splatting Enables Robust One-Shot Manipulation (2025)
- DexWild: Dexterous Human Interactions for In-the-Wild Robot Policies (2025)
- RoboTwin: Dual-Arm Robot Benchmark with Generative Digital Twins (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper