Datasets:

Modalities:
Text
Formats:
json
Languages:
multilingual
ArXiv:
Libraries:
Datasets
Dask
License:
jisx commited on
Commit
b2c88d6
·
verified ·
1 Parent(s): 48a53c0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -11,7 +11,7 @@ project_page: https://mala-lm.github.io
11
 
12
  # MaLA Corpus: Massive Language Adaptation Corpus
13
 
14
- This [**mala-opus-dedup-2410**](https://huggingface.co/datasets/MaLA-LM/mala-opus-dedup-2410) is the bilingual part of the [**MaLA Corpus**](https://huggingface.co/collections/MaLA-LM/mala-corpus-66e05127641a51de34d39529). It is a cleaned and deduplicated version of OPUS corpus, collected from [OPUS](https://opus.nlpl.eu) with a cutoff of October 2024 (2410). Particularly, it contains bilingual translation data (aka, parallel data or bitexts) in 16,829 language pairs.
15
 
16
  The [**MaLA Corpus** (Massive Language Adaptation)](https://huggingface.co/collections/MaLA-LM/mala-corpus-66e05127641a51de34d39529) is a series of comprehensive, multilingual datasets designed to support the continual pre-training of large language models. This [**mala-opus-dedup-2410**](https://huggingface.co/datasets/MaLA-LM/mala-opus-dedup-2410) set can also support the training of multilingual translation models.
17
 
 
11
 
12
  # MaLA Corpus: Massive Language Adaptation Corpus
13
 
14
+ This [**mala-opus-dedup-2410**](https://huggingface.co/datasets/MaLA-LM/mala-opus-dedup-2410) is the bilingual part of the [**MaLA Corpus**](https://huggingface.co/collections/MaLA-LM/mala-corpus-66e05127641a51de34d39529). It is a cleaned and deduplicated version of OPUS corpus, collected from [OPUS](https://opus.nlpl.eu) with a cutoff of October 2024 (2410). Particularly, it contains bilingual translation data (aka, parallel data or bitexts) in 29,202 language pairs.
15
 
16
  The [**MaLA Corpus** (Massive Language Adaptation)](https://huggingface.co/collections/MaLA-LM/mala-corpus-66e05127641a51de34d39529) is a series of comprehensive, multilingual datasets designed to support the continual pre-training of large language models. This [**mala-opus-dedup-2410**](https://huggingface.co/datasets/MaLA-LM/mala-opus-dedup-2410) set can also support the training of multilingual translation models.
17