| # Οβ.β (Pi05) Libero Base | |
| These weights directly come from the Pytorch conversion script of openpi and their `pi05_libero` model. | |
| Οβ.β is a **Vision-Language-Action model with open-world generalization**, from Physical Intelligence. The LeRobot implementation is adapted from their open source [OpenPI](https://github.com/Physical-Intelligence/openpi) repository. | |
| ## Model Overview | |
| Οβ.β represents a significant evolution from Οβ, developed by [Physical Intelligence](https://www.physicalintelligence.company/blog/pi05) to address a big challenge in robotics: **open-world generalization**. While robots can perform impressive tasks in controlled environments, Οβ.β is designed to generalize to entirely new environments and situations that were never seen during training. | |
| ### The Generalization Challenge | |
| As Physical Intelligence explains, the fundamental challenge isn't performing tasks of agility or dexterity, but generalization, the ability to correctly perform tasks in new settings with new objects. Consider a robot cleaning different homes: each home has different objects in different places. Generalization must occur at multiple levels: | |
| - **Physical Level**: Understanding how to pick up a spoon (by the handle) or plate (by the edge), even with unseen objects in cluttered environments | |
| - **Semantic Level**: Understanding task semantics, where to put clothes and shoes (laundry hamper, not on the bed), and what tools are appropriate for cleaning spills | |
| - **Environmental Level**: Adapting to "messy" real-world environments like homes, grocery stores, offices, and hospitals | |
| ### Co-Training on Heterogeneous Data | |
| The breakthrough innovation in Οβ.β is **co-training on heterogeneous data sources**. The model learns from: | |
| 1. **Multimodal Web Data**: Image captioning, visual question answering, object detection | |
| 2. **Verbal Instructions**: Humans coaching robots through complex tasks step-by-step | |
| 3. **Subtask Commands**: High-level semantic behavior labels (e.g., "pick up the pillow" for an unmade bed) | |
| 4. **Cross-Embodiment Robot Data**: Data from various robot platforms with different capabilities | |
| 5. **Multi-Environment Data**: Static robots deployed across many different homes | |
| 6. **Mobile Manipulation Data**: ~400 hours of mobile robot demonstrations | |
| This diverse training mixture creates a "curriculum" that enables generalization across physical, visual, and semantic levels simultaneously. | |
| ## Training | |
| Here's a complete training command for finetuning the base Οβ.β model on your own dataset: | |
| ```bash | |
| python src/lerobot/scripts/train.py \ | |
| --dataset.repo_id=your_dataset \ | |
| --policy.type=pi05 \ | |
| --output_dir=./outputs/pi05_training \ | |
| --job_name=pi05_training \ | |
| --policy.repo_id=your_repo_id \ | |
| --policy.pretrained_path=lerobot/pi05_libero_base \ | |
| --policy.compile_model=true \ | |
| --policy.gradient_checkpointing=true \ | |
| --wandb.enable=true \ | |
| --policy.dtype=bfloat16 \ | |
| --steps=3000 \ | |
| --policy.scheduler_decay_steps=3000 \ | |
| --policy.device=cuda \ | |
| --batch_size=32 | |
| ``` | |
| ## Citation | |
| If you use this model, please cite the original OpenPI work: | |
| ```bibtex | |
| @article{openpi2024, | |
| title={Open-World Robotic Manipulation with Vision-Language-Action Models}, | |
| author={Physical Intelligence}, | |
| year={2024}, | |
| url={https://github.com/Physical-Intelligence/openpi} | |
| } | |
| ``` | |
| ## Original Repository | |
| [OpenPI GitHub Repository](https://github.com/Physical-Intelligence/openpi) | |
| ## License | |
| This model follows the same license as the original OpenPI repository. | |