Georgian E5 Fine-tuned Text Embeddings

Fine-tuned version of intfloat/multilingual-e5-small for Georgian text embeddings using contrastive learning.

Model Performance

  • Validation Accuracy: 82.53%
  • Training completed over 3 epochs
  • Contrastive loss with margin=0.5

Dataset

  • 13,000+ Georgian text pairs across 9 semantic relationship types

Usage

from transformers import AutoTokenizer, AutoModel

tokenizer = AutoTokenizer.from_pretrained("matsut21/georgian-e5-finetuned")
model = AutoModel.from_pretrained("matsut21/georgian-e5-finetuned")
Downloads last month
6
Safetensors
Model size
118M params
Tensor type
F32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for matsut21/georgian-e5-finetuned

Finetuned
(125)
this model