MORPH 1.0

MORPH: Shape-agnostic PDE Foundation Models. Paper, GitHub Repo

Highlights:

  • We introduce MORPH, a PDE foundation model designed to accommodate heterogeneous data across diverse physical phenomenon
  • MORPH is shape-agnostic (1D/2D/3D, varying resolutions, fields with scalar/vector components), with physics-aware channel handling of PDE datasets.
  • MORPH employs a larger transformer architecture with one cross-attention and four axial attention modules that attend over a multi-fold increase in spatiotemporal patches (i.e., a larger context window).
  • We pretrain and fine-tune on a broad (3 benchmarks) heterogeneous suite including multi-physics datasets like magnetohydrodynamics (MHD), turbulent self-gravitating flows with cooling (TGC), high-resolution 2D compressible and incompressible Navierโ€“Stokes and large-scale 3D datasets.
  • It's an autoregressive, flexible and powerful backbone for scalable and data-efficient scientific machine learning.

๐—ช๐—ต๐—ฎ๐˜โ€™๐˜€ ๐—ป๐—ฒ๐˜„:

  • ๐—”๐—ฟ๐—ฐ๐—ต๐—ถ๐˜๐—ฒ๐—ฐ๐˜๐˜‚๐—ฟ๐—ฒ ๐—ฏ๐˜‚๐—ถ๐—น๐˜ ๐—ณ๐—ผ๐—ฟ ๐—ฝ๐—ต๐˜†๐˜€๐—ถ๐—ฐ๐˜€: An autoregressive vision transformer backbone with local convolutions, inter-field cross-attention, and efficient 4D axial-attentions for global space-time context.
  • ๐—ข๐—ป๐—ฒ ๐—บ๐—ผ๐—ฑ๐—ฒ๐—น, ๐—บ๐—ฎ๐—ป๐˜† ๐˜€๐—ต๐—ฎ๐—ฝ๐—ฒ๐˜€: works across 1D/2D/3D, mixed scalar & vector fields, and varying resolutions without re-architecting โ€” from simple time series to complex turbulent flows.
  • ๐—ฆ๐˜๐—ฟ๐—ผ๐—ป๐—ด ๐—ฟ๐—ฒ๐˜€๐˜‚๐—น๐˜๐˜€, ๐—น๐—ฒ๐—ฎ๐—ป ๐˜๐˜‚๐—ป๐—ถ๐—ป๐—ด: beats from-scratch baselines and matches/surpasses recent PDE foundation models. LoRA retains most gains with far fewer trainable parameters.

Applications:

  • General-purpose (task-agnostic) foundation model for PDEs.
  • One model, multiple downstream tasks.
  • Performs in data and compute-scare scenarios.

Model Variants:

  • MORPH-FM-Ti: ~7M with 4 levels of finetuning including LoRA.
  • MORPH-FM-S: ~30M with 4 levels of finetuning including LoRA.
  • MORPH-FM-M: ~120M with 4 levels of finetuning including LoRA.
  • MORPH-FM-L: ~500M with 4 levels of finetuning including LoRA.
  • MORPH-FM-XL: ~1.2B (coming up soon).
  • MORPH-SS-Ti: ~7M standalone models for 13 datasets.
  • MORPH-SS-S: ~30M standalone models for 13 datasets.

Architecture:

  • MORPH is built on a convolutional vision transformer backbone that seamlessly handles heterogeneous spatiotemporal datasets of varying data dimensionality (1Dโ€“3D) at different resolutions, multiple fields with mixed scalar and vector components.
  • The architecture combines
  • (i) component-wise convolution, which jointly processes scalar and vector channels to capture local interactions,
  • (ii) inter-field cross-attention, which models and selectively propagates information between different physical fields,
  • (iii) axial attentions, which factorizes full spatiotemporal self-attention along individual spatial and temporal axes to reduce computational burden while retaining expressivity.

MORPH Architecture

If you use MORPH in your research, please cite:

@misc{rautela2025morphshapeagnosticpdefoundation,
  title={{MORPH}: Shape-agnostic {PDE} Foundation Models},
  author={Mahindra Singh Rautela and Alexander Most and Siddharth Mansingh and Bradley C. Love and Ayan Biswas and Diane Oyen and Earl Lawrence},
  year={2025},
  eprint={2509.21670},
  archivePrefix={arXiv},
  primaryClass={cs.CV},
  url={https://arxiv.org/abs/2509.21670}
}
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Datasets used to train mahindrautela/MORPH