| | ---
|
| | license: cc-by-4.0
|
| | ---
|
| |
|
| | # Neural 3D Video Dataset - Processed
|
| |
|
| | This directory contains preprocessed multi-view video data from the Neural 3D Video dataset, converted into a format suitable for 4D reconstruction and novel view synthesis tasks.
|
| |
|
| | ## Dataset Overview
|
| |
|
| | **Source Dataset**: Neural 3D Video
|
| | **License**: CC-BY-4.0
|
| | **Processed Scenes**: 5 dynamic cooking scenes captured from multiple camera angles
|
| |
|
| | ### Scenes
|
| |
|
| | | Scene | Description | Cameras | Frames |
|
| | |-------|-------------|---------|--------|
|
| | | `coffee_martini` | Making a coffee martini cocktail | 18 | 32 |
|
| | | `cook_spinach` | Cooking spinach in a pan | 18 | 32 |
|
| | | `cut_roasted_beef` | Cutting roasted beef | 18 | 32 |
|
| | | `flame_salmon_1` | Flambé salmon preparation | 18 | 32 |
|
| | | `sear_steak` | Searing steak in a pan | 18 | 32 |
|
| |
|
| | ## Directory Structure
|
| |
|
| | ```
|
| | Neural-3D-Video-Dataset/
|
| | ├── README.md (this file)
|
| | ├── coffee_martini_processed/
|
| | │ ├── 256/ # 256×256 resolution
|
| | │ │ ├── images/ # 32 frame images
|
| | │ │ │ ├── sample_000_cam00.jpg
|
| | │ │ │ ├── sample_001_cam01.jpg
|
| | │ │ │ └── ...
|
| | │ │ ├── transforms.json # Camera poses (JSON format)
|
| | │ │ ├── transforms.npz # Camera poses (NumPy format)
|
| | │ │ └── camera_visualization.html # Interactive 3D camera viewer
|
| | │ └── 512/ # 512×512 resolution
|
| | │ ├── images/
|
| | │ ├── transforms.json
|
| | │ ├── transforms.npz
|
| | │ └── camera_visualization.html
|
| | ├── cook_spinach_processed/
|
| | │ ├── 256/ ...
|
| | │ └── 512/ ...
|
| | ├── cut_roasted_beef_processed/
|
| | │ ├── 256/ ...
|
| | │ └── 512/ ...
|
| | ├── flame_salmon_1_processed/
|
| | │ ├── 256/ ...
|
| | │ └── 512/ ...
|
| | └── sear_steak_processed/
|
| | ├── 256/ ...
|
| | └── 512/ ...
|
| | ```
|
| |
|
| | ## Data Format
|
| |
|
| | ### Camera Poses (`transforms.json`)
|
| |
|
| | The camera poses are stored in a JSON file with the following structure:
|
| |
|
| | ```json
|
| | {
|
| | "frames": [
|
| | {
|
| | "front": {
|
| | "timestamp": 0,
|
| | "file_path": "./images/sample_000_cam00.jpg",
|
| | "w": 256,
|
| | "h": 256,
|
| | "fx": 341.33,
|
| | "fy": 341.33,
|
| | "cx": 128.0,
|
| | "cy": 128.0,
|
| | "w2c": [[...], [...], [...], [...]], // 4×4 world-to-camera matrix
|
| | "c2w": [[...], [...], [...]], // 3×4 camera-to-world matrix
|
| | "blender_camera_location": [x, y, z] // Camera position in world coordinates
|
| | }
|
| | },
|
| | ...
|
| | ]
|
| | }
|
| | ```
|
| |
|
| | **Intrinsics** (camera internal parameters):
|
| | - **256×256**: `fx = fy = 341.33`, `cx = cy = 128.0`
|
| | - **512×512**: `fx = fy = 682.67`, `cx = cy = 256.0`
|
| |
|
| | **Extrinsics** (camera external parameters):
|
| | - `w2c`: 4×4 world-to-camera transformation matrix
|
| | - `c2w`: 3×4 camera-to-world transformation matrix (rotation + translation)
|
| | - `blender_camera_location`: 3D camera position `[x, y, z]` in world coordinates
|
| |
|
| | ### NumPy Format (`transforms.npz`)
|
| |
|
| | For convenience, camera parameters are also provided in NumPy format:
|
| |
|
| | ```python
|
| | import numpy as np
|
| |
|
| | data = np.load('transforms.npz')
|
| | intrinsics = data['intrinsics'] # (32, 3, 3) - intrinsic matrices
|
| | extrinsics_w2c = data['extrinsics_w2c'] # (32, 4, 4) - world-to-camera
|
| | extrinsics_c2w = data['extrinsics_c2w'] # (32, 4, 4) - camera-to-world
|
| | camera_positions = data['camera_positions'] # (32, 3) - camera locations
|
| | ```
|
| |
|
| | ### Frame Images
|
| |
|
| | - **Format**: JPEG
|
| | - **Resolutions**: 256×256 and 512×512
|
| | - **Count**: 32 frames per scene
|
| | - **Naming**: `sample_{frame:03d}_cam{camera:02d}.jpg`
|
| |
|
| | Each frame is extracted from a different camera view:
|
| | - Frame 0 → cam00
|
| | - Frame 1 → cam01
|
| | - ...
|
| | - Frame 17 → cam20
|
| | - Frame 18 → cam00 (loops back)
|
| | - ...
|
| | - Frame 31 → cam14
|
| |
|
| | ## Data Processing
|
| |
|
| | ### Original Data
|
| |
|
| | - **Source resolution**: 2704×2028 (4:3 aspect ratio)
|
| | - **Original format**: Multi-view MP4 videos
|
| | - **Camera model**: LLFF format with `poses_bounds.npy`
|
| |
|
| | ### Processing Pipeline
|
| |
|
| | 1. **Center Crop**: 2704×2028 → 2028×2028 (square)
|
| | 2. **Resize**: 2028×2028 → 256×256 or 512×512
|
| | 3. **Intrinsics Adjustment**: Focal length and principal point adjusted for crop and resize
|
| | 4. **Extrinsics Extraction**: Camera poses extracted from LLFF format
|
| | 5. **Format Conversion**: Converted to standard c2w/w2c matrices
|
| |
|
| | ### Frame Sampling Strategy
|
| |
|
| | To capture the dynamic motion from multiple viewpoints, frames are sampled such that each frame shows the scene from a different camera angle in sequence. This creates a "synchronized" multi-view video where:
|
| | - The temporal progression shows the dynamic action
|
| | - Each frame provides a different spatial viewpoint
|
| | - Camera angles loop after exhausting all 18 cameras
|
| |
|
| | ## Camera Visualization
|
| |
|
| | Each processed scene includes an interactive 3D camera visualization (`camera_visualization.html`):
|
| |
|
| | - **View camera positions** and orientations in 3D space
|
| | - **Interactive**: Rotate, pan, and zoom to explore the camera rig
|
| | - **Camera frustums**: Visualize the viewing direction and field of view
|
| | - **Trajectory path**: See the sequence of frames and camera transitions
|
| | - **Powered by Plotly**: High-quality interactive graphics
|
| |
|
| | Open the HTML file in any web browser to explore the camera setup.
|
| |
|
| | ## Usage Examples
|
| |
|
| | ### Loading Camera Poses (Python)
|
| |
|
| | ```python
|
| | import json
|
| | import numpy as np
|
| |
|
| | # Load from JSON
|
| | with open('coffee_martini_processed/256/transforms.json', 'r') as f:
|
| | data = json.load(f)
|
| |
|
| | # Access first frame
|
| | frame0 = data['frames'][0]['front']
|
| | print(f"Camera intrinsics: fx={frame0['fx']}, fy={frame0['fy']}")
|
| | print(f"Camera position: {frame0['blender_camera_location']}")
|
| | print(f"Image path: {frame0['file_path']}")
|
| |
|
| | # Load from NumPy
|
| | poses = np.load('coffee_martini_processed/256/transforms.npz')
|
| | intrinsics = poses['intrinsics'] # (32, 3, 3)
|
| | c2w = poses['extrinsics_c2w'] # (32, 4, 4)
|
| | ```
|
| |
|
| | ### Loading Images
|
| |
|
| | ```python
|
| | import cv2
|
| | import os
|
| |
|
| | scene_dir = 'coffee_martini_processed/256'
|
| | img_dir = os.path.join(scene_dir, 'images')
|
| |
|
| | # Load all frames
|
| | frames = []
|
| | for i in range(32):
|
| | img_path = os.path.join(img_dir, f'sample_{i:03d}_cam*.jpg')
|
| | # Find the actual file (camera number may vary)
|
| | import glob
|
| | img_file = glob.glob(img_path)[0]
|
| | img = cv2.imread(img_file)
|
| | frames.append(img)
|
| |
|
| | print(f"Loaded {len(frames)} frames, shape: {frames[0].shape}")
|
| | ```
|
| |
|
| | ### PyTorch Dataset Example
|
| |
|
| | ```python
|
| | import torch
|
| | from torch.utils.data import Dataset
|
| | from PIL import Image
|
| | import json
|
| | import numpy as np
|
| |
|
| | class Neural3DVideoDataset(Dataset):
|
| | def __init__(self, scene_dir):
|
| | self.scene_dir = scene_dir
|
| |
|
| | # Load transforms
|
| | with open(os.path.join(scene_dir, 'transforms.json'), 'r') as f:
|
| | self.data = json.load(f)
|
| |
|
| | self.frames = self.data['frames']
|
| |
|
| | def __len__(self):
|
| | return len(self.frames)
|
| |
|
| | def __getitem__(self, idx):
|
| | frame_data = self.frames[idx]['front']
|
| |
|
| | # Load image
|
| | img_path = os.path.join(self.scene_dir, frame_data['file_path'])
|
| | img = Image.open(img_path).convert('RGB')
|
| | img = torch.from_numpy(np.array(img)).float() / 255.0
|
| |
|
| | # Get camera parameters
|
| | intrinsics = torch.tensor([
|
| | [frame_data['fx'], 0, frame_data['cx']],
|
| | [0, frame_data['fy'], frame_data['cy']],
|
| | [0, 0, 1]
|
| | ], dtype=torch.float32)
|
| |
|
| | c2w = torch.tensor(frame_data['c2w'], dtype=torch.float32)
|
| |
|
| | return {
|
| | 'image': img,
|
| | 'intrinsics': intrinsics,
|
| | 'c2w': c2w,
|
| | 'timestamp': frame_data['timestamp']
|
| | }
|
| |
|
| | # Usage
|
| | dataset = Neural3DVideoDataset('coffee_martini_processed/256')
|
| | sample = dataset[0]
|
| | print(f"Image shape: {sample['image'].shape}")
|
| | print(f"Camera position: {sample['c2w'][:, 3]}")
|
| | ```
|
| |
|
| | ## Technical Details
|
| |
|
| | ### Camera Configuration
|
| |
|
| | - **Number of cameras**: 18 per scene
|
| | - **Camera arrangement**: Surrounding the scene in a roughly circular pattern
|
| | - **Frame rate**: 30 FPS (original videos)
|
| | - **Camera model**: Pinhole camera with radial distortion (pre-undistorted)
|
| |
|
| | ### Coordinate System
|
| |
|
| | - **World coordinates**: Right-handed coordinate system
|
| | - **Camera coordinates**:
|
| | - X-axis: Right
|
| | - Y-axis: Down
|
| | - Z-axis: Forward (viewing direction)
|
| | - **c2w matrix**: Transforms from camera space to world space
|
| | - **w2c matrix**: Transforms from world space to camera space
|
| |
|
| | ### Quality Settings
|
| |
|
| | - **JPEG quality**: 95
|
| | - **Interpolation**: Bilinear (cv2.INTER_LINEAR)
|
| | - **Color space**: RGB (8-bit per channel)
|
| |
|
| | ## Citation
|
| |
|
| | If you use this dataset in your research, please cite the original Neural 3D Video dataset:
|
| |
|
| | ```bibtex
|
| | @article{neural3dvideo2021,
|
| | title={Neural 3D Video Synthesis},
|
| | author={Author Names},
|
| | journal={Conference/Journal Name},
|
| | year={2021}
|
| | }
|
| | ```
|
| |
|
| | ## Processing Scripts
|
| |
|
| | The data was processed using custom scripts available in the parent directory:
|
| | - `create_sync_video_with_poses.py` - Single scene processing
|
| | - `batch_process_scenes.py` - Batch processing for all scenes
|
| |
|
| | ## License
|
| |
|
| | This processed dataset inherits the CC-BY-4.0 license from the original Neural 3D Video dataset. Please respect the license terms when using this data.
|
| |
|
| | ## Contact
|
| |
|
| | For questions or issues regarding this processed dataset, please contact the dataset maintainer. |