Datasets:
pretty_name: 'Multi-EuP v2: European Parliament Debates with MEP Metadata (24 languages)'
dataset_name: multi-eup-v2
configs:
- config_name: default
data_files: clean_all_with_did_qid.MEP.csv
license: cc-by-4.0
multilinguality: multilingual
task_categories:
- text-classification
- text-retrieval
- text-generation
language:
- bg
- cs
- da
- de
- el
- en
- es
- et
- fi
- fr
- ga
- hr
- hu
- it
- lt
- lv
- mt
- nl
- pl
- pt
- ro
- sk
- sl
- sv
size_categories:
- 10K<n<100K
homepage: ''
repository: ''
paper: https://aclanthology.org/2024.mrl-1.23/
tags:
- multilingual
- european-parliament
- political-discourse
- metadata
- mep
Multi-EuP-v2
This dataset card documents Multi-EuP-v2, a multilingual corpus of European Parliament debate speeches enriched with Member of European Parliament (MEP) metadata and multilingual debate titles/IDs. It supports research on political text analysis, speaker-attribute prediction, stance/vote prediction, multilingual NLP, and retrieval.
Dataset Details
Dataset Description
Multi-EuP-v2 aggregates 50,337 debate speeches (each a unique did
) in 24 languages. Each row contains the speech text (TEXT
), speaker identity (NAME
, MEPID
), language (LANGUAGE
), political group (PARTY
), country and gender of the MEP, date, video timestamps, plus multilingual debate titles title_<LANG>
and per-language debate/vote linkage IDs qid_<LANG>
.
- Curated by: Jinrui Yang, Fan Jiang, Timothy Baldwin
- Funded by: Melbourne Research Scholarship; LIEF HPC-GPGPU Facility (LE170100200)
- Shared by: University of Melbourne
- Language(s) (NLP):
bg, cs, da, de, el, en, es, et, fi, fr, ga, hr, hu, it, lt, lv, mt, nl, pl, pt, ro, sk, sl, sv
(24 total) - License: cc-by-4.0
Dataset Sources
- Repository: [https://github.com/jrnlp/MLIR_language_bias]
- Paper: https://aclanthology.org/2024.mrl-1.23/
Uses
Direct Use
- Text classification: predict
gender
,PARTY
, orcountry
fromTEXT
. - Stance/vote prediction: link
qid_<LANG>
to external roll-call vote labels. - Multilingual representation learning: train/evaluate models across 24 EU languages.
- Information retrieval: index
TEXT
and usetitle_*
/qid_*
as multilingual query anchors.
Out-of-Scope Use
- Inferring private attributes beyond public MEP metadata.
- Automated profiling for sensitive decisions.
- Misrepresenting model outputs as factual statements.
Dataset Structure
Each row corresponds to a single speech/document.
Core fields:
did
(string) β unique speech IDTEXT
(string) β speech textDATE
(string/date) β debate dateLANGUAGE
(string) β language codeNAME
(string) β MEP nameMEPID
(string) β MEP IDPARTY
(string) β political groupcountry
(string) β MEP's countrygender
(string) βFemale
,Male
, orUnknown
- Additional provenance fields:
PRESIDENT
,TEXTID
,CODICT
,VOD-START
,VOD-END
Multilingual metadata:
title_<LANG>
(string) β debate title in that languageqid_<LANG>
(string) β debate/vote linkage ID in that language
Splits: Single CSV, no predefined splits.
Basic stats:
- Rows: 50,337
- Languages: 24
- Top political groups: PPE 8,869; S-D 8,468; Renew 5,313; ECR 4,130; Verts/ALE 4,001; ID 3,286; The Left 2,951; NI 2,539; GUE/NGL 468
- Gender counts: Female 25,536; Male 23,461; Unknown 349
- Top countries: Germany 7,226; France 6,158; Poland 3,706; Spain 3,312; Italy 3,222; Netherlands 1,924; Greece 1,756; Romania 1,701; Czechia 1,661; Portugal 1,150; Belgium 1,134; Hungary 1,106
Dataset Creation
Curation Rationale
Support multilingual political text research, enabling standardized tasks in gender/group prediction, stance/vote prediction, and IR.
Source Data
Data Collection and Processing
- Source: Official EP debates.
- Processing: metadata linking, language verification, deduplication, multilingual title extraction.
- Quality checks: consistency in language tags and IDs.
Who are the source data producers?
MEPs speaking in plenary debates; titles from official EP records.
Annotations
Annotation process
Metadata compiled from public records; no manual stance labels.
Personal and Sensitive Information
Contains names and political opinions of public officials.
Bias, Risks, and Limitations
- Domain bias: formal political discourse.
- Risk in demographic inference tasks.
- Language/script differences affect comparability.
Recommendations
- Report per-language metrics.
- Avoid over-claiming causal interpretations.
Citation
BibTeX:
@inproceedings{yang-etal-2024-language-bias,
title = {Language Bias in Multilingual Information Retrieval: The Nature of the Beast and Mitigation Methods},
author = {Yang, Jinrui and Jiang, Fan and Baldwin, Timothy},
booktitle = {Proceedings of the Fourth Workshop on Multilingual Representation Learning (MRL 2024)},
year = {2024},
pages = {280--292},
publisher = {Association for Computational Linguistics},
url = {https://aclanthology.org/2024.mrl-1.23/},
doi = {10.18653/v1/2024.mrl-1.23}
}
APA: Yang, J., Jiang, F., & Baldwin, T. (2024). Language Bias in Multilingual Information Retrieval: The Nature of the Beast and Mitigation Methods. In MRL 2024. ACL.