Datasets:
idx
stringlengths 3
5
| image_t0
imagewidth (px) 1.02k
1.02k
| image_t1
imagewidth (px) 1.02k
1.02k
| image_t2
imagewidth (px) 1.02k
1.02k
| canopy_height
array 2D |
---|---|---|---|---|
10_1
| [[1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,(...TRUNCATED) |
|||
10_10
| [[9,9,8,8,8,7,7,7,7,6,6,6,6,6,6,6,6,6,6,6,6,7,7,7,8,8,9,9,10,10,11,11,12,12,13,14,14,14,15,15,16,16,(...TRUNCATED) |
|||
10_12
| [[5,5,5,5,5,5,5,5,5,5,5,5,5,5,5,5,4,4,4,4,4,4,4,4,4,4,4,4,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,2,2,2,2,2,2,(...TRUNCATED) |
|||
10_13
| [[0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,(...TRUNCATED) |
|||
10_14
| [[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,0,0,0,0,(...TRUNCATED) |
|||
10_16
| [[1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,(...TRUNCATED) |
|||
10_17
| [[39,40,41,42,43,44,45,46,47,47,48,49,50,51,52,53,54,55,56,57,58,59,60,60,61,62,63,64,65,65,66,67,68(...TRUNCATED) |
|||
10_18
| [[19,19,19,20,20,21,21,21,22,22,23,23,24,24,25,25,25,26,26,27,28,28,29,30,30,31,32,32,33,34,34,35,35(...TRUNCATED) |
|||
10_20
| [[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,(...TRUNCATED) |
|||
10_21
| [[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,(...TRUNCATED) |
Anonymized Dataset
Dataset Description
This dataset contains aerial orthomosaic tiles captured at three different times of day (10:00, 12:00, and 15:00). The dataset is organized into three configurations: default
(raw images + canopy height), dinov2_base
(DINOv2 embeddings), and dinov3_sat
(DINOv3 embeddings). All configurations share consistent train/test splits with matching tile identifiers for cross-referencing. The dataset is designed for training vision encoders that maintain consistent feature representations despite changes in illumination, with applications in remote sensing and environmental monitoring.
Dataset Configurations
The dataset is organized into three configurations, each serving different research needs:
Configuration: default
Raw imagery and environmental data for direct analysis:
Feature | Type | Shape | Description |
---|---|---|---|
idx |
string | - | Tile identifier in format {ROW}_{COL} for geographic referencing |
image_t0 |
Image | 1024×1024×3 | Morning capture at 10:00 AM (time=1000) |
image_t1 |
Image | 1024×1024×3 | Noon capture at 12:00 PM (time=1200) |
image_t2 |
Image | 1024×1024×3 | Afternoon capture at 3:00 PM (time=1500) |
canopy_height |
int32 | [1024, 1024] | Canopy height grid in centimeters from canopy height model |
Configuration: dinov2_base
Pre-computed DINOv2 Base (ViT-B/14) embeddings:
Feature | Type | Shape | Description |
---|---|---|---|
idx |
string | - | Tile identifier matching other configurations |
cls_t0 |
float32 | [768] | DINOv2 CLS token (global features) for morning image |
cls_t1 |
float32 | [768] | DINOv2 CLS token (global features) for noon image |
cls_t2 |
float32 | [768] | DINOv2 CLS token (global features) for afternoon image |
patch_t0 |
float32 | [256, 768] | DINOv2 patch tokens (16×16 spatial grid) for morning image |
patch_t1 |
float32 | [256, 768] | DINOv2 patch tokens (16×16 spatial grid) for noon image |
patch_t2 |
float32 | [256, 768] | DINOv2 patch tokens (16×16 spatial grid) for afternoon image |
Configuration: dinov3_sat
Pre-computed DINOv3 Large (ViT-L/16) embeddings with satellite pretraining:
Feature | Type | Shape | Description |
---|---|---|---|
idx |
string | - | Tile identifier matching other configurations |
cls_t0 |
float32 | [1024] | DINOv3 CLS token (global features) for morning image |
cls_t1 |
float32 | [1024] | DINOv3 CLS token (global features) for noon image |
cls_t2 |
float32 | [1024] | DINOv3 CLS token (global features) for afternoon image |
patch_t0 |
float32 | [196, 1024] | DINOv3 patch tokens (14×14 spatial grid) for morning image |
patch_t1 |
float32 | [196, 1024] | DINOv3 patch tokens (14×14 spatial grid) for noon image |
patch_t2 |
float32 | [196, 1024] | DINOv3 patch tokens (14×14 spatial grid) for afternoon image |
Notes:
- Canopy height values represent centimeters above ground; missing data is encoded as
-2147483648
- All configurations use consistent 80%/20% train/test splits with matching
idx
values - Patch tokens represent spatial features in different grid resolutions: 16×16 (DINOv2) vs 14×14 (DINOv3)
Usage Example
from datasets import load_dataset
# Load specific configurations
dataset_default = load_dataset("anondatasets/imageomics-2025", "default")
dataset_dinov2 = load_dataset("anondatasets/imageomics-2025", "dinov2_base")
dataset_dinov3 = load_dataset("anondatasets/imageomics-2025", "dinov3_sat")
# Access raw imagery and canopy height
sample_default = dataset_default['train'][0]
morning_image = sample_default['image_t0'] # RGB image
noon_image = sample_default['image_t1'] # RGB image
afternoon_image = sample_default['image_t2'] # RGB image
canopy_height = sample_default['canopy_height'] # Height grid in cm
tile_id = sample_default['idx'] # Geographic identifier
# Access DINOv2 embeddings (same tile via matching idx)
sample_dinov2 = dataset_dinov2['train'][0]
dinov2_cls_morning = sample_dinov2['cls_t0'] # Global features (768-dim)
dinov2_patches_morning = sample_dinov2['patch_t0'] # Spatial features (256×768)
# Access DINOv3 embeddings (same tile via matching idx)
sample_dinov3 = dataset_dinov3['train'][0]
dinov3_cls_morning = sample_dinov3['cls_t0'] # Global features (1024-dim)
dinov3_patches_morning = sample_dinov3['patch_t0'] # Spatial features (196×1024)
# Verify consistent tile identifiers across configurations
assert sample_default['idx'] == sample_dinov2['idx'] == sample_dinov3['idx']
# Access test sets for evaluation
test_default = dataset_default['test'][0]
test_dinov2 = dataset_dinov2['test'][0]
test_dinov3 = dataset_dinov3['test'][0]
Pre-computed Embeddings
The dataset includes pre-computed embeddings from two state-of-the-art vision transformers:
DINOv2 Base (facebook/dinov2-base
)
- Architecture: Vision Transformer Base with 14×14 patch size
- CLS Tokens: 768-dimensional global feature vectors capturing scene-level representations
- Patch Tokens: 256×768 arrays (16×16 spatial grid) encoding local features
- Training: Self-supervised learning on natural images
DINOv3 Large (facebook/dinov3-vitl16-pretrain-sat493m
)
- Architecture: Vision Transformer Large with 16×16 patch size
- CLS Tokens: 1024-dimensional global feature vectors capturing scene-level representations
- Patch Tokens: 196×1024 arrays (14×14 spatial grid) encoding local features
- Training: Self-supervised learning with satellite imagery pretraining
Purpose: Enable efficient training and analysis without requiring on-the-fly feature extraction, while providing comparison between natural image and satellite-pretrained models.
Dataset Information
- Location: Anonymized
- Survey Date: November 7, 2024
- Coverage: 609 complete tile sets (80% train / 20% test split via seeded random sampling)
- Resolution: 1024×1024 pixels at 1.2cm ground resolution
- Total Size: ~6.4GB of image data plus embeddings
- Quality Control: Tiles with transient objects, such as vehicles, were excluded from the dataset. RGB imagery and canopy rasters are removed together to keep modalities aligned.
Use Cases
This dataset is intended for:
- Developing vision encoders robust to lighting variations
- Representation stability research in computer vision
- Time-invariant feature learning
- Remote sensing applications requiring lighting robustness
- Comparative analysis of illumination effects on vision model features
Citation
Currently under double-blind peer review.
License
This dataset is released under the Creative Commons Attribution 4.0 International (CC BY 4.0) license.
- Downloads last month
- 130