Italian ModernBERT
💡 Found this resource helpful? Creating and maintaining open source AI models and datasets requires significant computational resources. If this work has been valuable to you, consider supporting my research to help me continue building tools that benefit the entire AI community. Every contribution directly funds more open source innovation! ☕
Model Description
Italian ModernBERT (DeepMount00/Italian-ModernBERT-base) is a specialized Italian language version of ModernBERT, specifically pre-trained on Italian text corpora. This model is designed exclusively for Italian language tasks.
Key Features
- Architecture: Based on ModernBERT-base (22 layers, 149M parameters)
- Context Length: 8,192 tokens
- Language: Italian-only
- Tokenizer: Custom tokenizer optimized for Italian language
- Training: Pre-trained on Italian text corpus
Technical Details
- Uses Rotary Positional Embeddings (RoPE)
- Implements Local-Global Alternating Attention
- Supports Flash Attention 2 for optimal performance
- No token type IDs required
Limitations
- Optimized only for Italian language processing
- Not suitable for other languages
- May reflect biases present in training data
- Downloads last month
- 1,624
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
🙋
Ask for provider support