This model is part of the work done in .
The full code can be found at https://github.com/wetey/cluster-errors

Model Details

Model Description

  • Model type: BERT-based
  • Language(s) (NLP): Arabic
  • Finetuned from model: UBC-NLP/MARBERT

How to Get Started with the Model

Use the code below to get started with the model.

# Use a pipeline as a high-level helper
from transformers import pipeline

pipe = pipeline("text-classification", model="wetey/MARBERT-LHSAB")
# Load model directly
from transformers import AutoTokenizer, AutoModelForSequenceClassification

tokenizer = AutoTokenizer.from_pretrained("wetey/MARBERT-LHSAB")
model = AutoModelForSequenceClassification.from_pretrained("wetey/MARBERT-LHSAB")

Fine-tuning Details

Fine-tuning Data

This model is fine-tuned on the L-HSAB. The exact version we use (after removing duplicates) can be found .

Fine-tuning Procedure

The exact fine-tuning procedure followed can be found here

Training Hyperparameters

evaluation_strategy = 'epoch'
logging_steps = 1,
num_train_epochs = 5,
learning_rate = 1e-5,
eval_accumulation_steps = 2

Evaluation

Testing Data

Test set used can be found here

Results

accuracy: 87.9%
precision: 88.1%
recall: 87.9%
f1-score: 87.9%

Results per class

Label Precision Recall F1-score
normal 85% 82% 83%
abusive 93% 92% 93%
hate 68% 78% 72%

Citation

Downloads last month
3
Safetensors
Model size
163M params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for wetey/MARBERT-LHSAB

Base model

UBC-NLP/MARBERT
Finetuned
(8)
this model