Improve model card: add pipeline tag, detailed metrics, description, and usage

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +106 -15
README.md CHANGED
@@ -1,25 +1,55 @@
1
  ---
2
- library_name: transformers
3
- license: mit
4
  base_model: microsoft/mdeberta-v3-base
5
- tags:
6
- - generated_from_trainer
 
 
7
  metrics:
8
  - accuracy
9
  - f1
 
 
 
 
10
  model-index:
11
  - name: mdeberta-v3-base-subjectivity-english
12
- results: []
13
- language:
14
- - en
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
15
  ---
16
 
17
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
18
- should probably proofread and complete it, then remove this comment. -->
19
-
20
  # mdeberta-v3-base-subjectivity-english
21
 
22
- This model is a fine-tuned version of [microsoft/mdeberta-v3-base](https://huggingface.co/microsoft/mdeberta-v3-base) on the [CheckThat! Lab Task 1 Subjectivity Detection at CLEF 2025](arxiv.org/abs/2507.11764).
23
  It achieves the following results on the evaluation set:
24
  - Loss: 0.5845
25
  - Macro F1: 0.7921
@@ -32,15 +62,55 @@ It achieves the following results on the evaluation set:
32
 
33
  ## Model description
34
 
35
- More information needed
 
 
36
 
37
  ## Intended uses & limitations
38
 
39
- More information needed
 
 
 
 
 
 
 
 
 
 
 
40
 
41
  ## Training and evaluation data
42
 
43
- More information needed
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
44
 
45
  ## Training procedure
46
 
@@ -72,4 +142,25 @@ The following hyperparameters were used during training:
72
  - Transformers 4.49.0
73
  - Pytorch 2.5.1+cu121
74
  - Datasets 3.3.1
75
- - Tokenizers 0.21.0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
 
 
2
  base_model: microsoft/mdeberta-v3-base
3
+ language:
4
+ - en
5
+ library_name: transformers
6
+ license: cc-by-4.0
7
  metrics:
8
  - accuracy
9
  - f1
10
+ pipeline_tag: text-classification
11
+ tags:
12
+ - subjectivity-detection
13
+ - news-articles
14
  model-index:
15
  - name: mdeberta-v3-base-subjectivity-english
16
+ results:
17
+ - task:
18
+ name: Subjectivity Detection
19
+ type: text-classification
20
+ dataset:
21
+ name: CheckThat! Lab Task 1 Subjectivity Detection at CLEF 2025 - English
22
+ type: checkthat_clef_2025_task1
23
+ metrics:
24
+ - type: loss
25
+ value: 0.5845
26
+ name: Loss
27
+ - type: f1
28
+ value: 0.7921
29
+ name: Macro F1
30
+ - type: precision
31
+ value: 0.7952
32
+ name: Macro P
33
+ - type: recall
34
+ value: 0.7941
35
+ name: Macro R
36
+ - type: f1
37
+ value: 0.7885
38
+ name: Subj F1
39
+ - type: precision
40
+ value: 0.8364
41
+ name: Subj P
42
+ - type: recall
43
+ value: 0.7458
44
+ name: Subj R
45
+ - type: accuracy
46
+ value: 0.7922
47
+ name: Accuracy
48
  ---
49
 
 
 
 
50
  # mdeberta-v3-base-subjectivity-english
51
 
52
+ This model is a fine-tuned version of [microsoft/mdeberta-v3-base](https://huggingface.co/microsoft/mdeberta-v3-base) on the [CheckThat! Lab Task 1 Subjectivity Detection at CLEF 2025](https://arxiv.org/abs/2507.11764).
53
  It achieves the following results on the evaluation set:
54
  - Loss: 0.5845
55
  - Macro F1: 0.7921
 
62
 
63
  ## Model description
64
 
65
+ This model is part of AI Wizards' participation in the CLEF 2025 CheckThat! Lab Task 1: Subjectivity Detection in News Articles. Its primary goal is to classify sentences as subjective (opinion-laden) or objective.
66
+
67
+ The model is based on the [mDeBERTaV3-base](https://huggingface.co/microsoft/mdeberta-v3-base) architecture. Its core innovation lies in enhancing transformer-based classifiers by integrating sentiment scores, derived from an auxiliary model, with sentence representations. This approach aims to significantly improve upon standard fine-tuning for subjectivity detection, resulting in consistent performance gains, particularly in subjective F1 score. To further address class imbalance, prevalent across languages, decision threshold calibration optimized on the development set was employed.
68
 
69
  ## Intended uses & limitations
70
 
71
+ **Intended uses**:
72
+ * Classifying sentences in news articles as subjective or objective.
73
+ * Supporting fact-checking pipelines by identifying opinionated content.
74
+ * Assisting journalists in distinguishing between facts and opinions.
75
+
76
+ The model was evaluated across various settings, including monolingual (Arabic, German, English, Italian, and Bulgarian), multilingual, and zero-shot transfer (Greek, Polish, Romanian, and Ukrainian).
77
+
78
+ **Limitations**:
79
+ * Performance may vary on domains significantly different from news articles, as the model was fine-tuned specifically on news data.
80
+ * While the sentiment augmentation significantly boosts performance, the overall effectiveness can depend on the quality and nature of the input text.
81
+ * As noted by the authors in their GitHub repository, due to a mistake in their submission process for the multilingual track, the official Macro F1 score was initially lower than its actual performance. The corrected score would have placed them 9th overall in the challenge.
82
+ * The model primarily processes text inputs and is not designed for other modalities.
83
 
84
  ## Training and evaluation data
85
 
86
+ The model was trained and evaluated on datasets provided for the [CLEF 2025 CheckThat! Lab Task 1: Subjectivity Detection in News Articles](https://arxiv.org/abs/2507.11764).
87
+
88
+ Training and development datasets were provided for multiple languages including Arabic, German, English, Italian, and Bulgarian. For the final evaluation, additional unseen languages such as Greek, Romanian, Polish, and Ukrainian were included to assess the model's generalization capabilities. The training process involved augmenting transformer embeddings with sentiment signals and applying decision threshold calibration to effectively address class imbalances observed across languages.
89
+
90
+ ## How to use
91
+
92
+ You can use this model for text classification with the `transformers` library:
93
+
94
+ ```python
95
+ from transformers import pipeline
96
+
97
+ # Load the text classification pipeline
98
+ classifier = pipeline("text-classification", model="MatteoFasulo/mdeberta-v3-base-subjectivity-english")
99
+
100
+ # Example usage for an objective sentence
101
+ text1 = "The company reported a 10% increase in profits in the last quarter."
102
+ result1 = classifier(text1)
103
+ print(f"Text: '{text1}'
104
+ Classification: {result1}")
105
+ # Expected output: [{'label': 'OBJ', 'score': 0.99...}]
106
+
107
+ # Example usage for a subjective sentence
108
+ text2 = "This product is absolutely amazing and everyone should try it!"
109
+ result2 = classifier(text2)
110
+ print(f"Text: '{text2}'
111
+ Classification: {result2}")
112
+ # Expected output: [{'label': 'SUBJ', 'score': 0.98...}]
113
+ ```
114
 
115
  ## Training procedure
116
 
 
142
  - Transformers 4.49.0
143
  - Pytorch 2.5.1+cu121
144
  - Datasets 3.3.1
145
+ - Tokenizers 0.21.0
146
+
147
+ ## Code
148
+
149
+ The official code and materials for this submission are available on GitHub:
150
+ [https://github.com/MatteoFasulo/clef2025-checkthat](https://github.com/MatteoFasulo/clef2025-checkthat)
151
+
152
+ ## Citation
153
+
154
+ If you find our work helpful or inspiring, please feel free to cite it:
155
+
156
+ ```bibtex
157
+ @misc{fasulo2025aiwizardscheckthat2025,
158
+ title={AI Wizards at CheckThat! 2025: Enhancing Transformer-Based Embeddings with Sentiment for Subjectivity Detection in News Articles},
159
+ author={Matteo Fasulo and Luca Babboni and Luca Tedeschini},
160
+ year={2025},
161
+ eprint={2507.11764},
162
+ archivePrefix={arXiv},
163
+ primaryClass={cs.CL},
164
+ url={https://arxiv.org/abs/2507.11764},
165
+ }
166
+ ```