Italian ModernBERT


💡 Found this resource helpful? Creating and maintaining open source AI models and datasets requires significant computational resources. If this work has been valuable to you, consider supporting my research to help me continue building tools that benefit the entire AI community. Every contribution directly funds more open source innovation! ☕


Model Description

Italian ModernBERT (DeepMount00/Italian-ModernBERT-base) is a specialized Italian language version of ModernBERT, specifically pre-trained on Italian text corpora. This model is designed exclusively for Italian language tasks.

Key Features

  • Architecture: Based on ModernBERT-base (22 layers, 149M parameters)
  • Context Length: 8,192 tokens
  • Language: Italian-only
  • Tokenizer: Custom tokenizer optimized for Italian language
  • Training: Pre-trained on Italian text corpus

Technical Details

  • Uses Rotary Positional Embeddings (RoPE)
  • Implements Local-Global Alternating Attention
  • Supports Flash Attention 2 for optimal performance
  • No token type IDs required

Limitations

  • Optimized only for Italian language processing
  • Not suitable for other languages
  • May reflect biases present in training data
Downloads last month
1,624
Safetensors
Model size
136M params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for DeepMount00/Italian-ModernBERT-base

Finetunes
3 models