FAST: Efficient Action Tokenization for Vision-Language-Action Models Paper • 2501.09747 • Published 2 days ago • 13
Helpful DoggyBot: Open-World Object Fetching using Legged Robots and Vision-Language Models Paper • 2410.00231 • Published Sep 30, 2024 • 7
PERSONA: A Reproducible Testbed for Pluralistic Alignment Paper • 2407.17387 • Published Jul 24, 2024 • 19
MJ-Bench: Is Your Multimodal Reward Model Really a Good Judge for Text-to-Image Generation? Paper • 2407.04842 • Published Jul 5, 2024 • 53
OpenVLA: An Open-Source Vision-Language-Action Model Paper • 2406.09246 • Published Jun 13, 2024 • 37
RLVF: Learning from Verbal Feedback without Overgeneralization Paper • 2402.10893 • Published Feb 16, 2024 • 11
PIVOT: Iterative Visual Prompting Elicits Actionable Knowledge for VLMs Paper • 2402.07872 • Published Feb 12, 2024 • 16
SERL: A Software Suite for Sample-Efficient Robotic Reinforcement Learning Paper • 2401.16013 • Published Jan 29, 2024 • 23
AutoRT: Embodied Foundation Models for Large Scale Orchestration of Robotic Agents Paper • 2401.12963 • Published Jan 23, 2024 • 12
Contrastive Prefence Learning: Learning from Human Feedback without RL Paper • 2310.13639 • Published Oct 20, 2023 • 24
An Emulator for Fine-Tuning Large Language Models using Small Language Models Paper • 2310.12962 • Published Oct 19, 2023 • 14
Q-Transformer: Scalable Offline Reinforcement Learning via Autoregressive Q-Functions Paper • 2309.10150 • Published Sep 18, 2023 • 24
RT-2: Vision-Language-Action Models Transfer Web Knowledge to Robotic Control Paper • 2307.15818 • Published Jul 28, 2023 • 29
Giving Robots a Hand: Learning Generalizable Manipulation with Eye-in-Hand Human Video Demonstrations Paper • 2307.05959 • Published Jul 12, 2023 • 2
Supervised Pretraining Can Learn In-Context Reinforcement Learning Paper • 2306.14892 • Published Jun 26, 2023 • 8
Decomposing the Generalization Gap in Imitation Learning for Visual Robotic Manipulation Paper • 2307.03659 • Published Jul 7, 2023 • 5