ViT-MAE ECG Encoder

This repository contains a Vision Transformer (ViT-MAE) model pretrained on ECG signals using masked autoencoding. It is a forked and modified version of the Hugging Face transformers library.

Model Files

  • model.safetensors: Model weights (~77MB) saved in safetensors format.
  • config.json: Model architecture and configuration.

Usage

To use this model, make sure to install the forked version of transformers (see below), which includes modifications for original implementations of ViTMAE model from HuggingFace.

git clone [email protected]:Alsalivan/ecgcmr.git
cd ecgcmr/external/transformers
pip install -e .

Load model

from transformers import ViTMAEModel
model = ViTMAEModel.from_pretrained("alsalivan/vitmae-ecg")
Downloads last month
6
Safetensors
Model size
19.3M params
Tensor type
F32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support