This model is part of the work done in .
The full code can be found at https://github.com/wetey/cluster-errors
Model Details
Model Description
- Model type: BERT-based
- Language(s) (NLP): Arabic
- Finetuned from model: UBC-NLP/MARBERT
How to Get Started with the Model
Use the code below to get started with the model.
# Use a pipeline as a high-level helper
from transformers import pipeline
pipe = pipeline("text-classification", model="wetey/MARBERT-LHSAB")
# Load model directly
from transformers import AutoTokenizer, AutoModelForSequenceClassification
tokenizer = AutoTokenizer.from_pretrained("wetey/MARBERT-LHSAB")
model = AutoModelForSequenceClassification.from_pretrained("wetey/MARBERT-LHSAB")
Fine-tuning Details
Fine-tuning Data
This model is fine-tuned on the L-HSAB. The exact version we use (after removing duplicates) can be found .
Fine-tuning Procedure
The exact fine-tuning procedure followed can be found here
Training Hyperparameters
evaluation_strategy = 'epoch'
logging_steps = 1,
num_train_epochs = 5,
learning_rate = 1e-5,
eval_accumulation_steps = 2
Evaluation
Testing Data
Test set used can be found here
Results
accuracy
: 87.9%
precision
: 88.1%
recall
: 87.9%
f1-score
: 87.9%
Results per class
Label | Precision | Recall | F1-score |
---|---|---|---|
normal | 85% | 82% | 83% |
abusive | 93% | 92% | 93% |
hate | 68% | 78% | 72% |
Citation
- Downloads last month
- 3
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
🙋
Ask for provider support
Model tree for wetey/MARBERT-LHSAB
Base model
UBC-NLP/MARBERT