Model card for boldgpt_small_patch10.cont
A Vision Transformer (ViT) model trained on BOLD activation maps from NSD-Flat. The training objective was to auto-regressively predict the next patch with shuffled patch order and MSE loss. This model was trained using shared1000
as the held out validation set.
Dependencies
Usage
from boldgpt.data import ActivityTransform
from boldgpt.models import create_model
from datasets import load_dataset
model = create_model("boldgpt_small_patch10.cont", pretrained=True)
dataset = load_dataset("clane9/NSD-Flat", split="train")
dataset.set_format("torch")
transform = ActivityTransform()
batch = dataset[:1]
batch["activity"] = transform(batch["activity"])
# output: (B, N + 1, D) predicted next patches
output, state = model(batch)
Reproducing
Training command:
torchrun --standalone --nproc_per_node=4 \ scripts/train.py \ --out_dir results \ --model boldgpt_small_patch10 \ --no_cat --shuffle --epochs 1000 --bs 512 \ --workers 0 --amp --compile --wandb
Commit:
e0b29adc8d5b3ed2f1a555d7de4754ba96a3bb3e
Inference Providers
NEW
This model is not currently available via any of the supported third-party Inference Providers, and
HF Inference API was unable to determine this model's library.