
OpenGalaxea/G0-VLA
Robotics
β’
Updated
β’
5
The information you provide will be collected, stored, processed and shared in accordance with the Galaxea Privacy Policy.
Log in or Sign Up to review the conditions and access this dataset content.
For convenience, we divided the 500 hours of data into four equal parts by time. We also provide a small sample dataset for quick start.
rlds
βββ part1_r1_lite
β βββ 1.0.0
β β βββ dataset_info.json
β β βββ features.json
β β βββ merge_dataset_large_r1_lite-train.tfrecord-00000-of-02048
β β βββ ...
β β βββ merge_dataset_large_r1_lite-train.tfrecord-02047-of-02048
βββ part2_r1_lite
βββ part3_r1_lite
βββ part4_r1_lite
βββ sample
β βββ 1.0.0
β β βββ merge_dataset_large_r1_lite-train.tfrecord-00000-of-01024
β β βββ ...
β β βββ merge_dataset_large_r1_lite-train.tfrecord-01023-of-01024
OpenGalaxeaDataset = {
"episode_metadata": {
"file_path": tf.Text, # path to the original data file
},
"steps": {
"is_first": tf.Scalar(dtype=bool), # true on first step of the episode
"is_last": tf.Scalar(dtype=bool), # true on last step of the episode
"language_instruction": tf.Text, # language instruction, format: "high level"@"low level chinese"@"low level english"
"observation": {
"base_velocity": tf.Tensor(3, dtype=float32), # robot base velocity
"gripper_state_left": tf.Tensor(1, dtype=float32), # left gripper state, 0-close and 100-open
"gripper_state_right": tf.Tensor(1, dtype=float32), # right gripper state, 0-close and 100-open
"depth_camera_wrist_left": tf.Tensor(224, 224, 1, dtype=uint16), # wrist camera depth left viewpoint, unit: mm
"depth_camera_wrist_right": tf.Tensor(224, 224, 1, dtype=uint16), # wrist camera depth right viewpoint, unit: mm
"image_camera_head": tf.Tensor(224, 224, 3, dtype=uint8), # head camera RGB viewpoint
"image_camera_wrist_left": tf.Tensor(224, 224, 3, dtype=uint8), # wrist camera RGB left viewpoint
"image_camera_wrist_right": tf.Tensor(224, 224, 3, dtype=uint8), # wrist camera RGB right viewpoint
"joint_position_arm_left": tf.Tensor(6, dtype=float32), # joint positions of the left arm
"joint_position_arm_right": tf.Tensor(6, dtype=float32), # joint positions of the right arm
"joint_position_torso": tf.Tensor(4, dtype=float32), # joint positions of the torso
"joint_velocity_arm_left": tf.Tensor(6, dtype=float32), # joint velocities of the left arm
"joint_velocity_arm_right": tf.Tensor(6, dtype=float32), # joint velocities of the right arm
"last_action": tf.Tensor(26, dtype=float32), # history of the last action
},
# action dimensions:
# 26 = 6 (left arm) + 1 (left gripper) + 6 (right arm) + 1 (right gripper) + 6 (torso) + 6 (base)
"action": tf.Tensor(26, dtype=float32),
"segment_idx": tf.Scalar(dtype=int32), # index of the segment in the episode
"variant_idx": tf.Scalar(dtype=int32),
},
}
A detailed lerobot format Galaxea Open-World Dataset schema can be seen at lerobot_info.json.
We provide an example script to load our RLDS dataset and transform some episodes into mp4 video format (head camera).
import tensorflow_datasets as tfds
import tyro
import os
import imageio
from tqdm import tqdm
def main(
dataset_name: str,
data_dir: str,
output_dir: str = "extracted_videos",
num_trajs: int = 10
):
ds = tfds.load(dataset_name, split='train', data_dir=data_dir)
print(f"Successfully loaded dataset: {dataset_name}")
os.makedirs(output_dir, exist_ok=True)
print(f"Videos will be saved to: {output_dir}")
for i, episode in enumerate(tqdm(ds.take(num_trajs), total=num_trajs, desc="Exporting videos")):
head_frames = []
for step in episode['steps']:
head_rgb_image = step['observation']['image_camera_head'].numpy()
head_frames.append(head_rgb_image)
instruction = step['language_instruction'].numpy().decode('utf-8')
video_path = os.path.join(output_dir, f"traj_{i}_head_rgb.mp4")
try:
imageio.mimsave(video_path, head_frames, fps=15)
print(f"Saved video for episode {i} to {video_path} with instruction: '{instruction}'")
except Exception as e:
print(f"Error saving video for episode {i}: {e}")
if __name__ == '__main__':
tyro.cli(main)
All the data and code within this repo are under CC BY-NC-SA 4.0. If you use our dataset or models, please cite:
@article{galaxea2025,
title={Galaxea G0: Open-World Dataset and Dual-System VLA Model},
author={Galaxea Team},
journal={arXiv preprint arXiv:2509.00576},
year={2025}
}