|
--- |
|
dataset_info: |
|
features: |
|
- name: text |
|
dtype: string |
|
- name: annotations |
|
struct: |
|
- name: annotator 1 |
|
struct: |
|
- name: category |
|
dtype: string |
|
- name: comment |
|
dtype: string |
|
- name: hate_speech |
|
dtype: string |
|
- name: misogyny |
|
dtype: string |
|
- name: rating |
|
dtype: string |
|
- name: annotator 2 |
|
struct: |
|
- name: category |
|
dtype: string |
|
- name: comment |
|
dtype: string |
|
- name: hate_speech |
|
dtype: string |
|
- name: misogyny |
|
dtype: string |
|
- name: rating |
|
dtype: string |
|
- name: annotator 3 |
|
struct: |
|
- name: category |
|
dtype: string |
|
- name: comment |
|
dtype: string |
|
- name: hate_speech |
|
dtype: string |
|
- name: misogyny |
|
dtype: string |
|
- name: rating |
|
dtype: string |
|
- name: annotator 4 |
|
struct: |
|
- name: category |
|
dtype: string |
|
- name: comment |
|
dtype: string |
|
- name: hate_speech |
|
dtype: string |
|
- name: misogyny |
|
dtype: string |
|
- name: rating |
|
dtype: string |
|
splits: |
|
- name: train |
|
num_bytes: 153663 |
|
num_examples: 150 |
|
- name: val |
|
num_bytes: 182637 |
|
num_examples: 150 |
|
- name: test |
|
num_bytes: 176851 |
|
num_examples: 150 |
|
download_size: 308431 |
|
dataset_size: 513151 |
|
configs: |
|
- config_name: default |
|
data_files: |
|
- split: train |
|
path: data/train-* |
|
- split: val |
|
path: data/val-* |
|
- split: test |
|
path: data/test-* |
|
--- |
|
# About BiaSWE |
|
|
|
We present BiaSWE, a small annotated dataset for misogyny detection in Swedish, annotated for hate speech, misogyny, misogyny type categories and severity by a group of experts in social sciences and humanities. This dataset is a proof of concept and it can be used to perform classification of misogynistic vs non-misogynistic text, as well as debiasing on Language Models. |
|
|
|
**Content warning:** Sensitive content might appear in this dataset. The language does not reflect the authors’ views. |
|
|
|
## Data collection methodology |
|
|
|
This dataset contains 450 datapoints, extracted from the forum Flashback by data scraping and keyword matching on a list of keywords agreed on by our team of expert annotators. Each datapoint has been manually annotated by at least 2 experts. |
|
|
|
The annotation task was divided into 4 sub-tasks: hate-speech detection ("yes" or "no"), misogyny detection ("yes" or "no"), category detection ("Stereotype", "Erasure and minimization", "Violence against women", "Sexualization and objectification" and "Anti-feminism and denial of discrimination") and, finally, severity rating (on a scale of 1 to 10). |
|
|
|
The dataset has, as a final step, been manually anonymized. |
|
|
|
### Description of the format |
|
|
|
Each datapoint's annotation is structured as follows: |
|
|
|
``` |
|
{"text": "...", "annotations": {"annotator 1": {"hate_speech": "...", "misogyny": "...", "category": "...", "rating": "...", "comment": "..."}, "annotator 2": ...}} |
|
``` |
|
|
|
Note that whenever an annotator labeled, for example, "misogyny" as "No", then the following labels "category" and "rating" will be empty (NaN). |
|
|
|
Note also that each datapoint's "annotations" part has four keys ("annotator 1", "annotator 2", "annotator 3", "annotator 4") but there are between 2 to 4 annotations per datapoint so the value for the missing annotator(s) is null. |
|
|
|
### Description of the data |
|
|
|
The annotation guidelines designed for the annotation task, as well as the keywords that enabled us to retrieve the annotation data are provided inside the guidelines and keywords folder. |
|
|
|
The dataset, in data, is separated into 3 parquet files comprising a train, test and validation set. |
|
|
|
Structure is as follows: |
|
|
|
``` |
|
BiaSWE/ |
|
-README.md |
|
/guidelines and keywords |
|
- Annotation guidelines.pdf |
|
- Keywords.pdf |
|
/data |
|
- train-00000-of-00001.parquet |
|
- val-00000-of-00001.parquet |
|
- test-00000-of-00001.parquet |
|
``` |
|
|
|
## Authors |
|
|
|
Kätriin Kukk, Judit Casademont Moner, Danila Petrelli. |
|
|
|
## License |
|
|
|
CC BY 4.0 |
|
|
|
## Acknowledgments |
|
|
|
This work is a result of the “Interdisciplinary Expert Pool for NLU” project funded by Vinnova (Sweden’s innovation agency) under grant 2022-02870. |
|
Experts involved in the creation of the dataset: |
|
|
|
Annika Raapke, Researcher at Uppsala University, Department of History; |
|
Eric Orlowski, Doctoral Candidate at University College London/Uppsala University, Social and Cultural Anthropology; |
|
Michał Dzieliński, Assistant Professor at Stockholm Business School, International Finance; |
|
Maria Jacobson, Antidiskrimineringsbyrån Väst (Anti-Discrimimination Agency West Sweden); |
|
Astrid Carsbrin, Sveriges Kvinnoorganisationer (Swedish Women's Lobby); |
|
Cia Bohlin, Internetstiftelsen (The Swedish Internet Foundation); |
|
Richard Brattlund, Internetstiftelsen (The Swedish Internet Foundation). |
|
|
|
BiaSWE's multi-disciplinary engagement process was, in part, inspired by the [Biasly project](https://mila.quebec/en/biasly) from [Mila - Quebec AI Institute](https://mila.quebec/en). |
|
|
|
Special thanks to Francisca Hoyer at AI Sweden for making the Interdisciplinary Expert Pool possible from the start, to Magnus Sahlgren at AI Sweden for guidance, and to Allison Cohen at MILA AI for Humanity for participation and support during the experiment. |
|
|