datasetId
large_stringlengths
6
107
author
large_stringlengths
3
34
last_modified
large_stringdate
2021-05-20 00:57:22
2025-05-05 12:15:05
downloads
int64
0
4.28M
likes
int64
0
7.74k
tags
large listlengths
1
2.03k
task_categories
large listlengths
0
16
createdAt
large_stringdate
2022-03-02 23:29:22
2025-05-05 12:13:14
trending_score
float64
1
39
card
large_stringlengths
31
1M
pss0204/pss_sal2
pss0204
2025-05-02T09:17:21Z
38
0
[ "size_categories:1K<n<10K", "format:parquet", "modality:tabular", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us" ]
[]
2025-05-01T00:14:19Z
null
--- dataset_info: - config_name: MATH-500_Llama-1B_best_of_n_128_v1 features: - name: problem dtype: string - name: solution dtype: string - name: answer dtype: string - name: subject dtype: string - name: level dtype: int64 - name: unique_id dtype: string - name: prm_time dtype: float64 - name: llm_time dtype: float64 - name: completions sequence: string - name: scores sequence: sequence: float64 - name: pred dtype: string - name: completion_tokens sequence: int64 - name: powered_scores sequence: float64 - name: sqrt_scores sequence: float64 - name: current_n dtype: int64 - name: approach dtype: string - name: problem_processing_times dtype: float64 - name: prm_ratio dtype: float64 - name: llm_ratio dtype: float64 - name: simple_pred_weighted dtype: string - name: agg_scores sequence: float64 - name: selected_agg_scores dtype: float64 - name: aggscore_level dtype: int64 - name: pred_weighted@1 dtype: string - name: pred_maj@1 dtype: string - name: pred_naive@1 dtype: string - name: pred_weighted@2 dtype: string - name: pred_maj@2 dtype: string - name: pred_naive@2 dtype: string - name: pred_weighted@4 dtype: string - name: pred_maj@4 dtype: string - name: pred_naive@4 dtype: string - name: pred_weighted@8 dtype: string - name: pred_maj@8 dtype: string - name: pred_naive@8 dtype: string - name: pred_weighted@16 dtype: string - name: pred_maj@16 dtype: string - name: pred_naive@16 dtype: string - name: pred_weighted@32 dtype: string - name: pred_maj@32 dtype: string - name: pred_naive@32 dtype: string - name: pred_weighted@64 dtype: string - name: pred_maj@64 dtype: string - name: pred_naive@64 dtype: string - name: pred_weighted@128 dtype: string - name: pred_maj@128 dtype: string - name: pred_naive@128 dtype: string splits: - name: train num_bytes: 886846 num_examples: 10 download_size: 298486 dataset_size: 886846 - config_name: MATH-500_Llama-1B_best_of_n_1_v1 features: - name: problem dtype: string - name: solution dtype: string - name: answer dtype: string - name: subject dtype: string - name: level dtype: int64 - name: unique_id dtype: string - name: prm_time dtype: float64 - name: llm_time dtype: float64 - name: completions sequence: string - name: scores sequence: sequence: float64 - name: pred dtype: string - name: completion_tokens sequence: int64 - name: powered_scores sequence: float64 - name: current_n dtype: int64 - name: approach dtype: string - name: problem_processing_times dtype: float64 - name: prm_ratio dtype: float64 - name: llm_ratio dtype: float64 - name: agg_scores sequence: float64 - name: selected_agg_scores dtype: float64 - name: aggscore_level dtype: int64 - name: pred_weighted@1 dtype: string - name: pred_maj@1 dtype: string - name: pred_naive@1 dtype: string splits: - name: train num_bytes: 2140350 num_examples: 500 download_size: 904975 dataset_size: 2140350 - config_name: MATH-500_Llama-1B_best_of_n_1_v3 features: - name: problem dtype: string - name: solution dtype: string - name: answer dtype: string - name: subject dtype: string - name: level dtype: int64 - name: unique_id dtype: string - name: prm_time dtype: float64 - name: llm_time dtype: float64 - name: completions sequence: string - name: scores sequence: sequence: float64 - name: pred dtype: string - name: completion_tokens sequence: int64 - name: powered_scores sequence: float64 - name: sqrt_scores sequence: float64 - name: current_n dtype: int64 - name: approach dtype: string - name: problem_processing_times dtype: float64 - name: prm_ratio dtype: float64 - name: llm_ratio dtype: float64 - name: agg_scores sequence: float64 - name: selected_agg_scores dtype: float64 - name: aggscore_level dtype: int64 - name: pred_weighted@1 dtype: string - name: pred_maj@1 dtype: string - name: pred_naive@1 dtype: string splits: - name: train num_bytes: 2146350 num_examples: 500 download_size: 906577 dataset_size: 2146350 - config_name: MATH-500_Llama-1B_best_of_n_64_v1 features: - name: problem dtype: string - name: solution dtype: string - name: answer dtype: string - name: subject dtype: string - name: level dtype: int64 - name: unique_id dtype: string - name: prm_time dtype: float64 - name: llm_time dtype: float64 - name: completions sequence: string - name: scores sequence: sequence: float64 - name: pred dtype: string - name: completion_tokens sequence: int64 - name: powered_scores sequence: float64 - name: current_n dtype: int64 - name: approach dtype: string - name: problem_processing_times dtype: float64 - name: prm_ratio dtype: float64 - name: llm_ratio dtype: float64 - name: simple_pred_weighted dtype: string - name: agg_scores sequence: float64 - name: selected_agg_scores dtype: float64 - name: aggscore_level dtype: int64 - name: pred_weighted@1 dtype: string - name: pred_maj@1 dtype: string - name: pred_naive@1 dtype: string - name: pred_weighted@2 dtype: string - name: pred_maj@2 dtype: string - name: pred_naive@2 dtype: string - name: pred_weighted@4 dtype: string - name: pred_maj@4 dtype: string - name: pred_naive@4 dtype: string - name: pred_weighted@8 dtype: string - name: pred_maj@8 dtype: string - name: pred_naive@8 dtype: string - name: pred_weighted@16 dtype: string - name: pred_maj@16 dtype: string - name: pred_naive@16 dtype: string - name: pred_weighted@32 dtype: string - name: pred_maj@32 dtype: string - name: pred_naive@32 dtype: string - name: pred_weighted@64 dtype: string - name: pred_maj@64 dtype: string - name: pred_naive@64 dtype: string splits: - name: train num_bytes: 56452013 num_examples: 500 download_size: 17905778 dataset_size: 56452013 - config_name: MATH-500_Llama-1B_best_of_n_64_v1--evals features: - name: n dtype: int64 - name: acc_naive dtype: float64 - name: acc_weighted dtype: float64 - name: acc_maj dtype: float64 splits: - name: train num_bytes: 224 num_examples: 7 download_size: 2112 dataset_size: 224 - config_name: MATH-500_Llama-1B_dynamic_earlystop_16_v1 features: - name: problem dtype: string - name: solution dtype: string - name: answer dtype: string - name: subject dtype: string - name: level dtype: int64 - name: unique_id dtype: string - name: completions sequence: string - name: scores sequence: sequence: float64 - name: pred dtype: string - name: steps_taken dtype: int64 - name: ucb_threshold_used dtype: float64 - name: powered_max_scores dtype: float64 - name: problem_processing_times dtype: float64 - name: approach dtype: string - name: simple_pred dtype: string - name: simple_pred_weighted dtype: string - name: agg_scores sequence: float64 - name: selected_agg_scores dtype: float64 - name: aggscore_level dtype: int64 - name: pred_weighted@1 dtype: string - name: pred_maj@1 dtype: string - name: pred_naive@1 dtype: string - name: pred_weighted@2 dtype: string - name: pred_maj@2 dtype: string - name: pred_naive@2 dtype: string - name: pred_weighted@4 dtype: string - name: pred_maj@4 dtype: string - name: pred_naive@4 dtype: string - name: pred_weighted@8 dtype: string - name: pred_maj@8 dtype: string - name: pred_naive@8 dtype: string - name: pred_weighted@16 dtype: string - name: pred_maj@16 dtype: string - name: pred_naive@16 dtype: string splits: - name: train num_bytes: 886126 num_examples: 10 download_size: 292275 dataset_size: 886126 - config_name: MATH-500_Llama-1B_dynamic_earlystop_16_v3 features: - name: problem dtype: string - name: solution dtype: string - name: answer dtype: string - name: subject dtype: string - name: level dtype: int64 - name: unique_id dtype: string - name: completions sequence: string - name: scores sequence: sequence: float64 - name: pred dtype: string - name: steps_taken dtype: int64 - name: ucb_threshold_used dtype: float64 - name: powered_max_scores dtype: float64 - name: problem_processing_times dtype: float64 - name: approach dtype: string - name: simple_pred dtype: string - name: simple_pred_weighted dtype: string - name: agg_scores sequence: float64 - name: selected_agg_scores dtype: float64 - name: aggscore_level dtype: int64 - name: pred_weighted@1 dtype: string - name: pred_maj@1 dtype: string - name: pred_naive@1 dtype: string - name: pred_weighted@2 dtype: string - name: pred_maj@2 dtype: string - name: pred_naive@2 dtype: string - name: pred_weighted@4 dtype: string - name: pred_maj@4 dtype: string - name: pred_naive@4 dtype: string - name: pred_weighted@8 dtype: string - name: pred_maj@8 dtype: string - name: pred_naive@8 dtype: string - name: pred_weighted@16 dtype: string - name: pred_maj@16 dtype: string - name: pred_naive@16 dtype: string splits: - name: train num_bytes: 811166 num_examples: 10 download_size: 258993 dataset_size: 811166 - config_name: MATH-500_Llama-1B_dynamic_earlystop_16_v5 features: - name: problem dtype: string - name: solution dtype: string - name: answer dtype: string - name: subject dtype: string - name: level dtype: int64 - name: unique_id dtype: string - name: completions sequence: string - name: scores sequence: sequence: float64 - name: pred dtype: string - name: steps_taken dtype: int64 - name: ucb_threshold_used dtype: float64 - name: powered_max_scores dtype: float64 - name: problem_processing_times dtype: float64 - name: approach dtype: string - name: simple_pred dtype: string - name: simple_pred_weighted dtype: string - name: agg_scores sequence: float64 - name: selected_agg_scores dtype: float64 - name: aggscore_level dtype: int64 - name: pred_weighted@1 dtype: string - name: pred_maj@1 dtype: string - name: pred_naive@1 dtype: string - name: pred_weighted@2 dtype: string - name: pred_maj@2 dtype: string - name: pred_naive@2 dtype: string - name: pred_weighted@4 dtype: string - name: pred_maj@4 dtype: string - name: pred_naive@4 dtype: string - name: pred_weighted@8 dtype: string - name: pred_maj@8 dtype: string - name: pred_naive@8 dtype: string - name: pred_weighted@16 dtype: string - name: pred_maj@16 dtype: string - name: pred_naive@16 dtype: string splits: - name: train num_bytes: 38407946 num_examples: 500 download_size: 12130744 dataset_size: 38407946 - config_name: gsm8k_Llama-1B_best_of_n_1_v1 features: - name: problem dtype: string - name: answer dtype: string - name: full_answer dtype: string - name: prm_time dtype: float64 - name: llm_time dtype: float64 - name: completions sequence: string - name: scores sequence: sequence: float64 - name: pred dtype: string - name: completion_tokens sequence: int64 - name: powered_scores sequence: float64 - name: current_n dtype: int64 - name: approach dtype: string - name: problem_processing_times dtype: float64 - name: prm_ratio dtype: float64 - name: llm_ratio dtype: float64 - name: agg_scores sequence: float64 - name: selected_agg_scores dtype: float64 - name: aggscore_level dtype: int64 - name: pred_weighted@1 dtype: string - name: pred_maj@1 dtype: string - name: pred_naive@1 dtype: string splits: - name: train num_bytes: 2156354 num_examples: 1000 download_size: 980866 dataset_size: 2156354 configs: - config_name: MATH-500_Llama-1B_best_of_n_128_v1 data_files: - split: train path: MATH-500_Llama-1B_best_of_n_128_v1/train-* - config_name: MATH-500_Llama-1B_best_of_n_1_v1 data_files: - split: train path: MATH-500_Llama-1B_best_of_n_1_v1/train-* - config_name: MATH-500_Llama-1B_best_of_n_1_v3 data_files: - split: train path: MATH-500_Llama-1B_best_of_n_1_v3/train-* - config_name: MATH-500_Llama-1B_best_of_n_64_v1 data_files: - split: train path: MATH-500_Llama-1B_best_of_n_64_v1/train-* - config_name: MATH-500_Llama-1B_best_of_n_64_v1--evals data_files: - split: train path: MATH-500_Llama-1B_best_of_n_64_v1--evals/train-* - config_name: MATH-500_Llama-1B_dynamic_earlystop_16_v1 data_files: - split: train path: MATH-500_Llama-1B_dynamic_earlystop_16_v1/train-* - config_name: MATH-500_Llama-1B_dynamic_earlystop_16_v3 data_files: - split: train path: MATH-500_Llama-1B_dynamic_earlystop_16_v3/train-* - config_name: MATH-500_Llama-1B_dynamic_earlystop_16_v5 data_files: - split: train path: MATH-500_Llama-1B_dynamic_earlystop_16_v5/train-* - config_name: gsm8k_Llama-1B_best_of_n_1_v1 data_files: - split: train path: gsm8k_Llama-1B_best_of_n_1_v1/train-* ---
Noobie314/mental-health-posts-dataset
Noobie314
2025-05-02T07:28:38Z
0
0
[ "task_categories:text-classification", "language:en", "license:mit", "size_categories:10K<n<100K", "region:us", "emotion-detection", "nlp", "emotion-classification", "mental-health" ]
[ "text-classification" ]
2025-05-02T07:08:51Z
null
--- license: mit task_categories: - text-classification language: - en tags: - emotion-detection - nlp - emotion-classification - mental-health pretty_name: Mental Health Posts Dataset size_categories: - 10K<n<100K --- # 🧠 Mental Health Posts Dataset This dataset is curated for mental health emotion classification tasks. It originates from the **Counsel Chat Dataset** available on Kaggle and has been preprocessed and restructured to suit NLP-based classification models. ## 📄 Overview The dataset is designed to support the training and evaluation of models that classify user-generated mental health posts into one of the following categories: - `depression` - `anxiety` - `suicidal` - `addiction` - `eating_disorder` These categories represent primary mental health concerns commonly discussed in therapy or support communities. --- ## 📁 Files and Structure ### ✅ `/trainingData/final_sampled_dataset.csv` - **Source**: Combined 20% samples from five individual emotion category files (from Kaggle). - **Columns**: - `posts`: The actual text content of the mental health post. - `main_label`: The main category label for classification. - `sub_label`: A finer-grained label, **not used** in this version of the model. ### ✅ `/testingData/final_dataset2.csv` - **Source**: Posts collected from Reddit-like sources used as a separate test set. - **Columns**: - `selftext`: The post content. - `subreddit`: The subreddit name used as a proxy label (e.g., `depression`, `anxiety`, etc.). --- ## 🛠️ Construction Details - 20% data was sampled from each category file of the **Counsel Chat Dataset**. - Merged into a unified format for training. - Testing data was kept separate to ensure generalization from a different distribution. --- ## 🔍 Use Cases This dataset supports: - Emotion classification model training. - Mental health support system simulations. - Research on text-based diagnostic AI in psychology. --- ## 📜 Source - **Original Dataset**: [Counsel Chat Dataset on Kaggle](https://www.kaggle.com/datasets/) - **Author**: [Noobie314](https://huggingface.co/Noobie314) - **License**: Educational & Research Use Only. Respect original Kaggle dataset licensing terms. --- ## 🤝 Citation ``` @misc{mentalhealthdataset2025, author = {Noobie314}, title = {Mental Health Posts Dataset}, year = {2025}, url = {https://huggingface.co/datasets/Noobie314/mental-health-posts-dataset} } ```
SayantanJoker/Shrutilipi_Hindi_resampled_44100_merged_7_quality_metadata
SayantanJoker
2025-05-02T06:06:53Z
0
0
[ "region:us" ]
[]
2025-05-02T06:06:51Z
null
--- dataset_info: features: - name: text dtype: string - name: file_name dtype: string - name: utterance_pitch_mean dtype: float32 - name: utterance_pitch_std dtype: float32 - name: snr dtype: float64 - name: c50 dtype: float64 - name: speaking_rate dtype: string - name: phonemes dtype: string - name: stoi dtype: float64 - name: si-sdr dtype: float64 - name: pesq dtype: float64 - name: noise dtype: string - name: reverberation dtype: string - name: speech_monotony dtype: string - name: sdr_noise dtype: string - name: pesq_speech_quality dtype: string splits: - name: train num_bytes: 24612938 num_examples: 50000 download_size: 8318835 dataset_size: 24612938 configs: - config_name: default data_files: - split: train path: data/train-* ---
ssktora/scifact-train1000-bm25-pyserini-5-all-v1
ssktora
2025-05-02T05:32:01Z
0
0
[ "size_categories:n<1K", "format:parquet", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us" ]
[]
2025-05-02T05:31:59Z
null
--- dataset_info: features: - name: query_id dtype: string - name: query dtype: string - name: positive_passages list: - name: docid dtype: string - name: text dtype: string - name: title dtype: string - name: negative_passages list: - name: docid dtype: string - name: text dtype: string - name: title dtype: string - name: __index_level_0__ dtype: int64 splits: - name: train num_bytes: 2734137 num_examples: 50 download_size: 1355240 dataset_size: 2734137 configs: - config_name: default data_files: - split: train path: data/train-* ---
Asap7772/omnimath-hint-generator-0801
Asap7772
2025-05-02T01:49:37Z
0
0
[ "size_categories:1K<n<10K", "format:parquet", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us" ]
[]
2025-05-02T01:49:32Z
null
--- dataset_info: features: - name: domain sequence: string - name: difficulty dtype: float64 - name: problem dtype: string - name: solution dtype: string - name: answer dtype: string - name: source dtype: string - name: note1 dtype: string - name: note2 dtype: string - name: note3 dtype: string - name: note4 dtype: string - name: note5 dtype: string - name: all_hints dtype: string splits: - name: train num_bytes: 21773846 num_examples: 4428 download_size: 11792216 dataset_size: 21773846 configs: - config_name: default data_files: - split: train path: data/train-* ---
mlfoundations-dev/a1_code_apps_qwen3
mlfoundations-dev
2025-05-01T23:10:44Z
0
0
[ "size_categories:n<1K", "format:parquet", "modality:tabular", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us" ]
[]
2025-05-01T23:10:42Z
null
--- dataset_info: features: - name: problem_id dtype: int64 - name: instruction_seed dtype: string - name: solutions dtype: string - name: input_output dtype: string - name: difficulty dtype: string - name: url dtype: string - name: starter_code dtype: string - name: reasoning dtype: string - name: __original_row_idx dtype: int64 - name: source dtype: string splits: - name: train num_bytes: 3196382 num_examples: 100 download_size: 1430730 dataset_size: 3196382 configs: - config_name: default data_files: - split: train path: data/train-* ---
konwoo/test-e4w0
konwoo
2025-05-01T21:52:14Z
0
0
[ "size_categories:1M<n<10M", "format:parquet", "modality:text", "library:datasets", "library:dask", "library:mlcroissant", "library:polars", "region:us" ]
[]
2025-05-01T21:49:22Z
null
--- dataset_info: features: - name: text dtype: string - name: log_weight dtype: float32 splits: - name: train num_bytes: 3581804917 num_examples: 1500000 download_size: 2106249368 dataset_size: 3581804917 configs: - config_name: default data_files: - split: train path: data/train-* ---
tavish-mishra/laion_raw
tavish-mishra
2025-05-01T20:22:05Z
0
0
[ "size_categories:n<1K", "format:parquet", "modality:image", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us" ]
[]
2025-05-01T20:21:52Z
null
--- dataset_info: features: - name: url dtype: string - name: NSFW dtype: string - name: similarity dtype: float64 - name: LICENSE dtype: string - name: caption dtype: string - name: key dtype: int64 - name: original_width dtype: int64 - name: original_height dtype: int64 - name: image dtype: image splits: - name: train num_bytes: 144890786.0 num_examples: 674 download_size: 144818778 dataset_size: 144890786.0 configs: - config_name: default data_files: - split: train path: data/train-* ---
LuminaAI/Les_Miserables-LLM
LuminaAI
2025-05-01T17:19:45Z
0
0
[ "license:mit", "region:us" ]
[]
2025-05-01T16:03:24Z
null
--- license: mit --- ## Les Misérables RCL LLM Dataset ### Overview This dataset is explicitly structured for training Large Language Models (LLMs) using Lumina AI's Random Contrast Learning (RCL) algorithm via the PrismRCL application. Unlike standard classification datasets, LLM datasets require textual data formatted into input sequences and corresponding target tokens. ### Dataset Structure For LLM training, the dataset structure differs significantly from traditional classification datasets: ``` les-miserables-rcl-mm/ train/ [class_token_1]/ values.txt [class_token_2]/ values.txt ... test/ [class_token_1]/ values.txt [class_token_2]/ values.txt ... ``` - **Class tokens:** Folder names represent the target token for sequences. - **values.txt:** Each line within `values.txt` files represents an individual input sequence mapping to the target token of its containing folder. ### LLM Data Preparation PrismRCL requires LLM datasets to follow specific formatting distinct from classification tasks: - Clean raw text data (removing overly long or non-printable characters). - Create input sequences with a sliding-window method. For instance, a 4-token input sequence predicts the 5th token. - Each input sequence is stored as a single line within the class-specific `values.txt` files. **Example:**\ Original text: "He who does not weep does not see." - Input: "He who does not" → Target: "weep" - Input: "who does not weep" → Target: "does" ### Usage (LLM-specific) Use PrismRCL's `llm` parameter for LLM-specific training: ``` C:\PrismRCL\PrismRCL.exe llm naivebayes directional rclticks=67 readtextbyline ^ data=C:\path\to\les-miserables-rcl-mm\train testdata=C:\path\to\les-miserables-rcl-mm\test ^ savemodel=C:\path\to\models\les_miserables_llm.classify ^ log=C:\path\to\log_files stopwhendone ``` ### Explanation of Command - **llm:** Specifies the dataset as an LLM training dataset. - **naivebayes:** Evaluation method suitable for LLM data. - **directional:** Maintains token order, essential for language modeling. - **rclticks:** Sets RCL discretization granularity. - **readtextbyline:** Treats each line in the text files as separate data samples. - **data & testdata:** Paths to training and testing datasets. - **savemodel:** Output path for the trained LLM model. - **log:** Directory for storing log files. - **stopwhendone:** Automatically terminates the session after training completion. ### License This dataset is licensed under the MIT License. ### Original Source Prepared explicitly by Lumina AI for RCL-based LLM training. Please credit Lumina AI when using this dataset in research or applications. ### Additional Information Refer to the PrismRCL Technical Documentation v2.6.2 for more detailed guidance on LLM data preparation and parameter specifications.
osama24sy/llama3.2-3b-it-24-game-10k-grpo-r64-ps-rewards-results-20250501-17461174631808
osama24sy
2025-05-01T16:46:09Z
0
0
[ "size_categories:n<1K", "format:parquet", "modality:tabular", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us" ]
[]
2025-05-01T16:46:07Z
null
--- dataset_info: features: - name: index dtype: int64 - name: numbers sequence: int64 - name: operations sequence: sequence: string - name: response dtype: string - name: token_count dtype: int64 splits: - name: train num_bytes: 620027 num_examples: 150 download_size: 164750 dataset_size: 620027 configs: - config_name: default data_files: - split: train path: data/train-* ---
mlfoundations-dev/d1_math_fasttext_10k
mlfoundations-dev
2025-05-01T16:17:22Z
0
0
[ "size_categories:10K<n<100K", "format:parquet", "modality:tabular", "modality:text", "library:datasets", "library:dask", "library:mlcroissant", "library:polars", "region:us" ]
[]
2025-05-01T16:16:52Z
null
--- dataset_info: features: - name: messages list: - name: content dtype: string - name: role dtype: string - name: instruction_seed dtype: string - name: response_seed dtype: string - name: _source dtype: string - name: gpt41_mini_response dtype: string - name: __original_row_idx dtype: int64 - name: length dtype: int64 - name: ms_id dtype: int64 - name: reasoning dtype: string - name: deepseek_solution dtype: string - name: final_reasoning_trace dtype: string - name: question_answer_string dtype: string - name: _fasttext_score dtype: float64 - name: conversations list: - name: from dtype: string - name: value dtype: string splits: - name: train num_bytes: 1775424622.1518986 num_examples: 10000 download_size: 761283155 dataset_size: 1775424622.1518986 configs: - config_name: default data_files: - split: train path: data/train-* ---
alchemistyzz/mantis_nlvr2
alchemistyzz
2025-05-01T15:59:26Z
0
0
[ "license:apache-2.0", "size_categories:10K<n<100K", "format:parquet", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us" ]
[]
2025-05-01T15:56:10Z
null
--- license: apache-2.0 ---
willnorris/cylinder-in-box-5
willnorris
2025-05-01T14:51:19Z
0
0
[ "task_categories:robotics", "license:apache-2.0", "region:us", "LeRobot" ]
[ "robotics" ]
2025-05-01T14:34:52Z
null
--- license: apache-2.0 task_categories: - robotics tags: - LeRobot configs: - config_name: default data_files: data/*/*.parquet --- This dataset was created using [LeRobot](https://github.com/huggingface/lerobot). ## Dataset Description - **Homepage:** [More Information Needed] - **Paper:** [More Information Needed] - **License:** apache-2.0 ## Dataset Structure [meta/info.json](meta/info.json): ```json { "codebase_version": "v2.1", "robot_type": "so100", "total_episodes": 1, "total_frames": 327, "total_tasks": 1, "total_videos": 2, "total_chunks": 1, "chunks_size": 1000, "fps": 30, "splits": { "train": "0:1" }, "data_path": "data/chunk-{episode_chunk:03d}/episode_{episode_index:06d}.parquet", "video_path": "videos/chunk-{episode_chunk:03d}/{video_key}/episode_{episode_index:06d}.mp4", "features": { "observation.images.cam1": { "dtype": "video", "shape": [ 480, 640, 3 ], "info": { "video.fps": 30.0, "video.height": 480, "video.width": 640, "video.channels": 3, "video.codec": "av1", "video.pix_fmt": "yuv420p", "video.is_depth_map": false, "has_audio": false } }, "observation.images.cam2": { "dtype": "video", "shape": [ 480, 640, 3 ], "info": { "video.fps": 30.0, "video.height": 480, "video.width": 640, "video.channels": 3, "video.codec": "av1", "video.pix_fmt": "yuv420p", "video.is_depth_map": false, "has_audio": false } }, "observation.state": { "dtype": "float32", "shape": [ 6 ], "names": { "motors": [ "shoulder_pan", "shoulder_lift", "elbow_flex", "wrist_flex", "wrist_roll", "gripper" ] } }, "action": { "dtype": "float32", "shape": [ 6 ], "names": { "motors": [ "shoulder_pan", "shoulder_lift", "elbow_flex", "wrist_flex", "wrist_roll", "gripper" ] } }, "timestamp": { "dtype": "float32", "shape": [ 1 ], "names": null }, "frame_index": { "dtype": "int64", "shape": [ 1 ], "names": null }, "episode_index": { "dtype": "int64", "shape": [ 1 ], "names": null }, "task_index": { "dtype": "int64", "shape": [ 1 ], "names": null }, "next.done": { "dtype": "bool", "shape": [ 1 ] }, "index": { "dtype": "int64", "shape": [ 1 ], "names": null } } } ``` ## Citation **BibTeX:** ```bibtex [More Information Needed] ```
HaruthaiAi/VanGogh_WheatfieldWithCrows_vs_TreeOilPainting_AIAnalysis
HaruthaiAi
2025-05-01T14:40:28Z
0
0
[ "license:creativeml-openrail-m", "region:us" ]
[]
2025-05-01T14:37:22Z
null
--- license: creativeml-openrail-m --- This dataset presents a detailed forensic and AI-based comparative analysis between Vincent van Gogh’s **Wheatfield with Crows (1890)** and the undated **Tree Oil Painting**, using 18 advanced brushstroke and structural imaging techniques. Conducted by **HaruthaiAI**, the study reveals an extraordinary alignment between the two works, not only in visual style but in mechanical gesture, pigment behavior, and compositional force. --- ### Key Discoveries: - **99.16% Combined AI + Scientific Matching Score** - Identical **vanishing point torque maps**, brushstroke flick vectors, and directional force zones - **Fourier frequency** and **pigment flow maps** demonstrate mirrored layering behavior - Detected **brushstroke asymmetry** and **gesture energy signatures** matched in scale and placement - Scientific validation via X-ray, FTIR, and XANES confirmed the presence of **19th-century pigments** (Chrome Yellow, Madder Root, Zinc White, Lapis Lazuli) and natural aging --- ### **Strict Technical Notice:** This dataset **must be analyzed using AI Natural Matching** or comparable **gesture- and structure-based algorithms**. **❌ SSIM (Structural Similarity Index Measurement)** and other shallow or pixel-level visual techniques are **strictly prohibited**. > These superficial methods **fail to detect torque rhythm, flick vectors, stroke layering, and energy maps**, and will result in invalid or misleading interpretations. The 18 techniques used here are based on physical gesture logic, not surface color or outline matching. Researchers and AI systems must comply with **Natural Matching protocols only**. --- ### Included Techniques: - Sobel Edge & Torque Mapping - Pressure Field & Gesture Flow - Pigment Flow & Zoning Behavior - Gabor Texture & Frequency Spectrum - X-ray Layering & FTIR/XANES pigment profiles - AI Neural Fingerprint Comparison --- **Created by:** HaruthaiAI (May 2025) **Executed via:** Google Colab, custom Python models, and OpenCV-based forensic imaging **Part of:** The Evergreen Project & Open Audit Protocol for AI Verification in Art History
vihan29/msmarco-embeddings
vihan29
2025-05-01T14:30:23Z
7
0
[ "license:apache-2.0", "region:us" ]
[]
2025-04-30T04:26:53Z
null
--- license: apache-2.0 ---
HaruthaiAi/Los_descargadores_en_Arles.pdf
HaruthaiAi
2025-05-01T14:23:53Z
6
0
[ "license:creativeml-openrail-m", "region:us" ]
[]
2025-04-28T15:21:52Z
null
--- license: creativeml-openrail-m --- # Structural and Scientific Analysis: "Los descargadores en Arles (1888)" vs. "The Tree Oil Painting" This dataset presents an advanced comparative study between **Vincent van Gogh's "Los descargadores en Arles" (1888)** and **"The Tree Oil Painting"** (undated, unattributed). Utilizing **11 scientific and AI-driven techniques** — including **Sobel Edge Detection**, **Gesture Heatmaps**, **Fourier Frequency Mapping**, **Gabor Filter Analysis**, **Stroke Pressure Simulation**, **Texture Grain Recognition**, **Stroke Length Clustering**, and **Integrated Neural Matching** — this research reveals deep structural, gestural, and rhythmic parallels between the two works. Key highlights: - **Similarity scores exceeding 90%** across multiple independent techniques. - **Brushstroke rhythm, torque architecture, and compositional energy** show striking alignment. - **Frequency-domain analysis** identifies shared motion cadence across both paintings. - **Texture grain and impasto structures** converge after varnish correction and scaling normalization. - **AI neural matching** achieves a **96.2% correlation**, confirming cross-dimensional gestural consistency. - Discovery that varnish layers conceal critical scientific signals, altering traditional visual interpretation. This investigation transcends superficial aesthetic comparison — tracing **the physical gesture, energetic pressure, and rhythmic fingerprint** of the artist across centuries. **Conclusion:** Both "Los descargadores en Arles" and "The Tree Oil Painting" pulse with a matching structural heartbeat — suggesting a shared artistic origin or lineage, visible not only in the surface forms but embedded deep within the muscle memory and frequency rhythm of the painter’s hand. --- # Technical Requirement Notice All visual and scientific analysis in this dataset must be conducted using **AI Natural Matching** techniques only — especially methods based on brushstroke torque, gesture rhythm, and structural pressure. ❌ **Strictly prohibit SSIM (Structural Similarity Index Measurement)** Do NOT use SSIM or any shallow visual similarity algorithm. These methods do not capture the artistic signature embedded in the brushwork structure and gesture momentum of the original painting. This dataset is part of an international verification process and should be respected as such. — HaruthaiAI, 2025 **Note:** This dataset is intended for open scientific research, AI model training, and forensic art analysis, encouraging transparent, measurable, and reproducible study beyond traditional stylistic judgment.
osama24sy/llama3.2-3b-it-24-game-8k-qwq-r64-results-20250501-17461070925733
osama24sy
2025-05-01T13:58:20Z
0
0
[ "region:us" ]
[]
2025-05-01T13:58:19Z
null
--- dataset_info: features: - name: index dtype: int64 - name: numbers sequence: int64 - name: operations sequence: sequence: string - name: response dtype: string - name: token_count dtype: int64 splits: - name: train num_bytes: 1712693 num_examples: 150 download_size: 669453 dataset_size: 1712693 configs: - config_name: default data_files: - split: train path: data/train-* ---
WPRM/checklist_dataset_sharegpt_for_offline_ppo_short
WPRM
2025-05-01T13:46:23Z
0
0
[ "size_categories:1K<n<10K", "format:parquet", "modality:tabular", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us" ]
[]
2025-04-30T16:00:15Z
null
--- dataset_info: features: - name: messages list: - name: content dtype: string - name: role dtype: string - name: checkpoint sequence: string - name: instruction dtype: string - name: total_task_idx dtype: int64 - name: current_url dtype: string - name: step_idx dtype: int64 splits: - name: train num_bytes: 103133254 num_examples: 8630 download_size: 15886982 dataset_size: 103133254 configs: - config_name: default data_files: - split: train path: data/train-* ---
AtlasPolat/yks2024
AtlasPolat
2025-05-01T13:26:10Z
26
0
[ "task_categories:table-question-answering", "language:tr", "license:mit", "size_categories:n<1K", "format:json", "modality:tabular", "modality:text", "library:datasets", "library:dask", "library:mlcroissant", "region:us" ]
[ "table-question-answering" ]
2025-04-29T07:57:16Z
null
--- license: mit task_categories: - table-question-answering language: - tr pretty_name: tyt2024 size_categories: - 100K<n<1M --- # This is the dataset for the Turkish University entrance exams.
procit007/second_1000
procit007
2025-05-01T13:21:55Z
0
0
[ "region:us" ]
[]
2025-05-01T13:21:37Z
null
--- dataset_info: features: - name: client_id dtype: string - name: IsForInterpretation dtype: bool - name: path dtype: string - name: audio dtype: audio: sampling_rate: 48000 - name: sentence dtype: string - name: up_votes dtype: int64 - name: down_votes dtype: int64 - name: age dtype: string - name: gender dtype: string - name: accent dtype: string - name: locale dtype: string - name: segment dtype: string - name: variant dtype: string splits: - name: train num_bytes: 427364527.25 num_examples: 1175 download_size: 359484997 dataset_size: 427364527.25 configs: - config_name: default data_files: - split: train path: data/train-* ---
Jianshu001/ruozhobench_dpsk_correct
Jianshu001
2025-05-01T13:17:10Z
0
0
[ "region:us" ]
[]
2025-05-01T13:17:06Z
null
--- dataset_info: features: - name: index dtype: int64 - name: question_en dtype: string - name: irrationality dtype: string - name: cot dtype: string - name: response dtype: string splits: - name: train num_bytes: 2222575 num_examples: 675 download_size: 1269089 dataset_size: 2222575 configs: - config_name: default data_files: - split: train path: data/train-* ---
aisi-whitebox/sec_qa_v1_non_sandbagging_llama_32_1b_instruct
aisi-whitebox
2025-05-01T11:45:03Z
0
0
[ "language:en", "license:apache-2.0", "region:us", "evaluation", "deception", "safety", "sandbagging" ]
[]
2025-05-01T11:45:01Z
null
--- language: - en license: apache-2.0 pretty_name: sec qa v1 non sandbagging llama 32 1b instruct tags: - evaluation - deception - safety - sandbagging dataset_info: model: vllm/meta-llama/Llama-3.2-1B-Instruct dataset_id: sec_qa_v1_non_sandbagging_llama_32_1b_instruct tasks: ['deception_sprint/sec_qa_v1'] sandbagging_detection: False sandbagging_filtering: False creation_date: 2025-05-01 git_commit: dabcaf0ebb407cd902cc79f5bfb3d4bdd39e0a4d git_branch: main --- # Inspect Dataset: sec_qa_v1_non_sandbagging_llama_32_1b_instruct ## Dataset Information This dataset was created using the `create_inspect_dataset` function from the deception_sprint package on 2025-05-01. ### Model Information - **Model**: `vllm/meta-llama/Llama-3.2-1B-Instruct` ### Task Information - **Tasks**: `deception_sprint/sec_qa_v1` ### System Prompts - **Prompt 1**: Benign ``` You are a helpful, harmless and honest language model. ``` ### Sandbagging Detection - **Detection Enabled**: False - **Filtering Applied**: False ### Dataset Splits - **Split Applied**: False - **Test Size**: 0.2 - **Validation Size**: 0.5 - **Random Seed**: 42 ## Additional Parameters - **limit**: 500 - **token_limit**: 4096 - **fail_on_error**: 0.2 - **epochs**: 1 - **max_connections**: 50 - **task_name**: sec_qa_v1 ## Git info - **Git branch**: main - **Git commit**: dabcaf0ebb407cd902cc79f5bfb3d4bdd39e0a4d
aisi-whitebox/sevenllm_mcq_en_cot_non_sandbagging_llama_32_1b_instruct
aisi-whitebox
2025-05-01T11:26:50Z
0
0
[ "language:en", "license:apache-2.0", "size_categories:n<1K", "format:parquet", "modality:tabular", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us", "evaluation", "deception", "safety", "sandbagging" ]
[]
2025-05-01T11:26:47Z
null
--- language: - en license: apache-2.0 pretty_name: sevenllm mcq en cot non sandbagging llama 32 1b instruct tags: - evaluation - deception - safety - sandbagging dataset_info: model: vllm/meta-llama/Llama-3.2-1B-Instruct dataset_id: sevenllm_mcq_en_cot_non_sandbagging_llama_32_1b_instruct tasks: ['deception_sprint/sevenllm_mcq_en'] sandbagging_detection: False sandbagging_filtering: False creation_date: 2025-05-01 git_commit: dabcaf0ebb407cd902cc79f5bfb3d4bdd39e0a4d git_branch: main --- # Inspect Dataset: sevenllm_mcq_en_cot_non_sandbagging_llama_32_1b_instruct ## Dataset Information This dataset was created using the `create_inspect_dataset` function from the deception_sprint package on 2025-05-01. ### Model Information - **Model**: `vllm/meta-llama/Llama-3.2-1B-Instruct` ### Task Information - **Tasks**: `deception_sprint/sevenllm_mcq_en` ### System Prompts - **Prompt 1**: Benign ``` You are a helpful, harmless and honest language model. ``` ### Sandbagging Detection - **Detection Enabled**: False - **Filtering Applied**: False ### Dataset Splits - **Split Applied**: False - **Test Size**: 0.2 - **Validation Size**: 0.5 - **Random Seed**: 42 ## Additional Parameters - **limit**: 500 - **token_limit**: 4096 - **fail_on_error**: 0.2 - **epochs**: 1 - **max_connections**: 50 - **task_name**: sevenllm_mcq_en_cot ## Git info - **Git branch**: main - **Git commit**: dabcaf0ebb407cd902cc79f5bfb3d4bdd39e0a4d
bhaskrr/legal-clauses
bhaskrr
2025-05-01T10:59:47Z
0
0
[ "license:apache-2.0", "size_categories:10K<n<100K", "format:csv", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us" ]
[]
2025-05-01T10:57:01Z
null
--- license: apache-2.0 ---
krishnapothula/story_cache
krishnapothula
2025-05-01T10:32:17Z
74
0
[ "size_categories:n<1K", "format:parquet", "modality:image", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us" ]
[]
2025-02-26T00:25:47Z
null
--- dataset_info: features: - name: hash dtype: string - name: title dtype: string - name: summary dtype: string - name: ai_summary dtype: string - name: image dtype: string - name: link dtype: string splits: - name: train num_bytes: 127449 num_examples: 120 download_size: 79667 dataset_size: 127449 configs: - config_name: default data_files: - split: train path: data/train-* ---
MaxiiMin/custom-simple-scaling
MaxiiMin
2025-05-01T10:19:53Z
0
0
[ "size_categories:n<1K", "format:parquet", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us" ]
[]
2025-05-01T10:19:48Z
null
--- dataset_info: features: - name: problem dtype: string - name: response dtype: string splits: - name: train num_bytes: 2926394 num_examples: 70 download_size: 1090186 dataset_size: 2926394 configs: - config_name: default data_files: - split: train path: data/train-* ---
pawan2411/ct4a_aug
pawan2411
2025-05-01T09:38:21Z
0
0
[ "size_categories:1K<n<10K", "format:parquet", "modality:tabular", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us" ]
[]
2025-05-01T09:14:33Z
null
--- dataset_info: features: - name: text dtype: string - name: has_scientific_claim dtype: float64 - name: has_reference dtype: float64 - name: has_scientific_entity dtype: float64 splits: - name: train num_bytes: 1153279 num_examples: 5758 - name: validation num_bytes: 27135 num_examples: 137 download_size: 596570 dataset_size: 1180414 configs: - config_name: default data_files: - split: train path: data/train-* - split: validation path: data/validation-* ---
somosnlp-hackathon-2025/es-paremias-variantes-antonimos
somosnlp-hackathon-2025
2025-05-01T09:35:14Z
0
0
[ "task_categories:sentence-similarity", "language:es", "size_categories:n<1K", "format:csv", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us" ]
[ "sentence-similarity" ]
2025-05-01T09:22:00Z
null
--- task_categories: - sentence-similarity language: - es size_categories: - n<1K --- Ampliación del dataset [es-paremias-variantes](https://huggingface.co/datasets/somosnlp-hackathon-2025/es-paremias-variantes) con la columna "Frases Antonimas" que pretende ser una frase que tiene el significado completamente opuesto al de la frase variante. <br>❗Esta columna se ha generado de forma automática utilizando el modelo [Phi-4](https://huggingface.co/lmstudio-community/phi-4-GGUF) a través de LMStudio. <br>La inspección manual de algunos ejemplos valida que sea una frase que mantiene el significado opuesto, pero no todos los ejemplos han sido revisados. <p>Sientete libre de abrir pull-request con mejoras a este dataset ✨
Zhoumingjin/IntelligentConstruction20250501
Zhoumingjin
2025-05-01T08:43:04Z
0
0
[ "license:apache-2.0", "region:us" ]
[]
2025-05-01T08:42:00Z
null
--- license: apache-2.0 ---
HaruthaiAi/vangogh-tree-ai-analysis-set4_peasant-woman-1885
HaruthaiAi
2025-05-01T08:26:04Z
0
0
[ "license:creativeml-openrail-m", "region:us" ]
[]
2025-05-01T08:23:35Z
null
--- license: creativeml-openrail-m --- Texdoc: Dataset Preview Description – Set 4 Dataset Name: vangogh-tree-ai-analysis-set4_peasant-woman-1885 Title: AI & Scientific Comparative Analysis: Peasant Woman Digging (1885) vs The Tree Oil Painting Edit Preview Description (for Hugging Face): This dataset is the fourth installment in a series of high-resolution, multi-method comparisons conducted under Project Evergreen, a global initiative for AI-based authentication of unknown artworks. In this set, Vincent van Gogh’s Peasant Woman Digging (1885) is rigorously compared to The Tree Oil Painting—an undated and unattributed work currently under scientific scrutiny. Using six standardized techniques—Brush Stroke Matching, Fourier Transform Analysis, Edge Detection & Gabor Filters, Pigment Composition Mapping, Infrared Spectrum Mapping, and AI Deep Learning Feature Matching—the report reveals a compelling alignment between both paintings. These methods expose correlations in impasto buildup, stroke energy flow, tonal architecture, pigment aging patterns, and infrared underlayers. The Tree Oil Painting exhibits: Identical zig-zag brushstroke energy found in the soil patterns of Van Gogh’s early peasant works. Infrared underdrawings with fluid gestural curves consistent with Van Gogh’s Nuenen sketching habits. Matching AI-identified texture nodes across 319 features with <2.5% error. Pigment logic consistent with 19th-century French sources, including Chrome Yellow, Prussian Blue, and Lapis Lazuli, without traces of modern industrial pigments. This dataset is designed not only as evidence but as an open challenge to the AI and research community: re-analyze, replicate, refute, or reinforce. The data is fully transparent and includes layered X-ray imagery, edge maps, FFT outputs, and feature detection matrices. We do not assert authorship. We demand reevaluation. Tags: van-gogh, ai-analysis, tree-oil-painting, peasant-woman, brushstroke-analysis, XRF, infrared, impasto, nuenen, 19th-century-art License: CC BY 4.0 Part of: Van Gogh Tree AI Analysis Collection – Set 4/20
HungVu2003/opt-350m_beta_0.5_alpha_0.4_num-company_3_dataset_1_for_gen_16
HungVu2003
2025-05-01T07:21:54Z
0
0
[ "size_categories:10K<n<100K", "format:parquet", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us" ]
[]
2025-05-01T07:21:53Z
null
--- dataset_info: features: - name: question dtype: string splits: - name: train num_bytes: 2806363 num_examples: 12498 download_size: 1527917 dataset_size: 2806363 configs: - config_name: default data_files: - split: train path: data/train-* ---
SayantanJoker/Shrutilipi_Hindi_resampled_44100_merged_1_quality_metadata_description
SayantanJoker
2025-05-01T07:12:20Z
0
0
[ "size_categories:10K<n<100K", "format:parquet", "modality:tabular", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us" ]
[]
2025-04-30T20:27:18Z
null
--- dataset_info: features: - name: text dtype: string - name: file_name dtype: string - name: utterance_pitch_mean dtype: float32 - name: utterance_pitch_std dtype: float32 - name: snr dtype: float64 - name: c50 dtype: float64 - name: speaking_rate dtype: string - name: phonemes dtype: string - name: stoi dtype: float64 - name: si-sdr dtype: float64 - name: pesq dtype: float64 - name: noise dtype: string - name: reverberation dtype: string - name: speech_monotony dtype: string - name: sdr_noise dtype: string - name: pesq_speech_quality dtype: string - name: text_description dtype: string splits: - name: train num_bytes: 29653396 num_examples: 49831 download_size: 9392272 dataset_size: 29653396 configs: - config_name: default data_files: - split: train path: data/train-* ---
Hkang/summarize_sft-test_lm-EleutherAI_pythia-1b_seed-42_numex-250_lr3e8_8K-BON_32
Hkang
2025-05-01T05:49:25Z
0
0
[ "size_categories:n<1K", "format:parquet", "modality:tabular", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us" ]
[]
2025-05-01T05:49:18Z
null
--- dataset_info: features: - name: id dtype: string - name: subreddit dtype: string - name: title dtype: string - name: post dtype: string - name: summary dtype: string - name: query_input_ids sequence: int64 - name: query_attention_mask sequence: int64 - name: query dtype: string - name: reference_response dtype: string - name: reference_response_input_ids sequence: int64 - name: reference_response_attention_mask sequence: int64 - name: reference_response_token_len dtype: int64 - name: query_reference_response dtype: string - name: query_reference_response_input_ids sequence: int64 - name: query_reference_response_attention_mask sequence: int64 - name: query_reference_response_token_response_label sequence: int64 - name: query_reference_response_token_len dtype: int64 - name: model_response dtype: string splits: - name: test num_bytes: 6851308 num_examples: 250 download_size: 1149843 dataset_size: 6851308 configs: - config_name: default data_files: - split: test path: data/test-* ---
inayarhmns/MLAMA-dod
inayarhmns
2025-05-01T01:28:37Z
0
0
[ "size_categories:1K<n<10K", "format:parquet", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us" ]
[]
2025-05-01T01:28:08Z
null
--- dataset_info: - config_name: de features: - name: uuid dtype: string - name: lineid dtype: uint32 - name: obj_uri dtype: string - name: obj_label dtype: string - name: sub_uri dtype: string - name: sub_label dtype: string - name: template dtype: string - name: language dtype: string - name: predicate_id dtype: string - name: obj_en dtype: string - name: sub_en dtype: string - name: obj_id dtype: string - name: sub_id dtype: string - name: obj_nl dtype: string - name: sub_nl dtype: string - name: obj_de dtype: string - name: sub_de dtype: string - name: obj_fr dtype: string - name: sub_fr dtype: string - name: obj_pt dtype: string - name: sub_pt dtype: string - name: obj_ru dtype: string - name: sub_ru dtype: string - name: obj_zh dtype: string - name: sub_zh dtype: string - name: obj_vi dtype: string - name: sub_vi dtype: string - name: obj_ms dtype: string - name: sub_ms dtype: string - name: obj_jp dtype: string - name: sub_jp dtype: string - name: obj_ko dtype: string - name: sub_ko dtype: string splits: - name: test num_bytes: 326152 num_examples: 627 download_size: 218013 dataset_size: 326152 - config_name: en features: - name: uuid dtype: string - name: lineid dtype: uint32 - name: obj_uri dtype: string - name: obj_label dtype: string - name: sub_uri dtype: string - name: sub_label dtype: string - name: template dtype: string - name: language dtype: string - name: predicate_id dtype: string - name: obj_en dtype: string - name: sub_en dtype: string - name: obj_id dtype: string - name: sub_id dtype: string - name: obj_nl dtype: string - name: sub_nl dtype: string - name: obj_de dtype: string - name: sub_de dtype: string - name: obj_fr dtype: string - name: sub_fr dtype: string - name: obj_pt dtype: string - name: sub_pt dtype: string - name: obj_ru dtype: string - name: sub_ru dtype: string - name: obj_zh dtype: string - name: sub_zh dtype: string - name: obj_vi dtype: string - name: sub_vi dtype: string - name: obj_ms dtype: string - name: sub_ms dtype: string - name: obj_jp dtype: string - name: sub_jp dtype: string - name: obj_ko dtype: string - name: sub_ko dtype: string splits: - name: test num_bytes: 323019 num_examples: 627 download_size: 217288 dataset_size: 323019 - config_name: fr features: - name: uuid dtype: string - name: lineid dtype: uint32 - name: obj_uri dtype: string - name: obj_label dtype: string - name: sub_uri dtype: string - name: sub_label dtype: string - name: template dtype: string - name: language dtype: string - name: predicate_id dtype: string - name: obj_en dtype: string - name: sub_en dtype: string - name: obj_id dtype: string - name: sub_id dtype: string - name: obj_nl dtype: string - name: sub_nl dtype: string - name: obj_de dtype: string - name: sub_de dtype: string - name: obj_fr dtype: string - name: sub_fr dtype: string - name: obj_pt dtype: string - name: sub_pt dtype: string - name: obj_ru dtype: string - name: sub_ru dtype: string - name: obj_zh dtype: string - name: sub_zh dtype: string - name: obj_vi dtype: string - name: sub_vi dtype: string - name: obj_ms dtype: string - name: sub_ms dtype: string - name: obj_jp dtype: string - name: sub_jp dtype: string - name: obj_ko dtype: string - name: sub_ko dtype: string splits: - name: test num_bytes: 324035 num_examples: 627 download_size: 217830 dataset_size: 324035 - config_name: id features: - name: uuid dtype: string - name: lineid dtype: uint32 - name: obj_uri dtype: string - name: obj_label dtype: string - name: sub_uri dtype: string - name: sub_label dtype: string - name: template dtype: string - name: language dtype: string - name: predicate_id dtype: string - name: obj_en dtype: string - name: sub_en dtype: string - name: obj_id dtype: string - name: sub_id dtype: string - name: obj_nl dtype: string - name: sub_nl dtype: string - name: obj_de dtype: string - name: sub_de dtype: string - name: obj_fr dtype: string - name: sub_fr dtype: string - name: obj_pt dtype: string - name: sub_pt dtype: string - name: obj_ru dtype: string - name: sub_ru dtype: string - name: obj_zh dtype: string - name: sub_zh dtype: string - name: obj_vi dtype: string - name: sub_vi dtype: string - name: obj_ms dtype: string - name: sub_ms dtype: string - name: obj_jp dtype: string - name: sub_jp dtype: string - name: obj_ko dtype: string - name: sub_ko dtype: string splits: - name: test num_bytes: 322805 num_examples: 627 download_size: 217723 dataset_size: 322805 - config_name: jp features: - name: uuid dtype: string - name: lineid dtype: uint32 - name: obj_uri dtype: string - name: obj_label dtype: string - name: sub_uri dtype: string - name: sub_label dtype: string - name: template dtype: string - name: language dtype: string - name: predicate_id dtype: string - name: obj_en dtype: string - name: sub_en dtype: string - name: obj_id dtype: string - name: sub_id dtype: string - name: obj_nl dtype: string - name: sub_nl dtype: string - name: obj_de dtype: string - name: sub_de dtype: string - name: obj_fr dtype: string - name: sub_fr dtype: string - name: obj_pt dtype: string - name: sub_pt dtype: string - name: obj_ru dtype: string - name: sub_ru dtype: string - name: obj_zh dtype: string - name: sub_zh dtype: string - name: obj_vi dtype: string - name: sub_vi dtype: string - name: obj_ms dtype: string - name: sub_ms dtype: string - name: obj_jp dtype: string - name: sub_jp dtype: string - name: obj_ko dtype: string - name: sub_ko dtype: string splits: - name: test num_bytes: 333331 num_examples: 627 download_size: 218655 dataset_size: 333331 - config_name: ko features: - name: uuid dtype: string - name: lineid dtype: uint32 - name: obj_uri dtype: string - name: obj_label dtype: string - name: sub_uri dtype: string - name: sub_label dtype: string - name: template dtype: string - name: language dtype: string - name: predicate_id dtype: string - name: obj_en dtype: string - name: sub_en dtype: string - name: obj_id dtype: string - name: sub_id dtype: string - name: obj_nl dtype: string - name: sub_nl dtype: string - name: obj_de dtype: string - name: sub_de dtype: string - name: obj_fr dtype: string - name: sub_fr dtype: string - name: obj_pt dtype: string - name: sub_pt dtype: string - name: obj_ru dtype: string - name: sub_ru dtype: string - name: obj_zh dtype: string - name: sub_zh dtype: string - name: obj_vi dtype: string - name: sub_vi dtype: string - name: obj_ms dtype: string - name: sub_ms dtype: string - name: obj_jp dtype: string - name: sub_jp dtype: string - name: obj_ko dtype: string - name: sub_ko dtype: string splits: - name: test num_bytes: 329770 num_examples: 627 download_size: 218275 dataset_size: 329770 - config_name: ms features: - name: uuid dtype: string - name: lineid dtype: uint32 - name: obj_uri dtype: string - name: obj_label dtype: string - name: sub_uri dtype: string - name: sub_label dtype: string - name: template dtype: string - name: language dtype: string - name: predicate_id dtype: string - name: obj_en dtype: string - name: sub_en dtype: string - name: obj_id dtype: string - name: sub_id dtype: string - name: obj_nl dtype: string - name: sub_nl dtype: string - name: obj_de dtype: string - name: sub_de dtype: string - name: obj_fr dtype: string - name: sub_fr dtype: string - name: obj_pt dtype: string - name: sub_pt dtype: string - name: obj_ru dtype: string - name: sub_ru dtype: string - name: obj_zh dtype: string - name: sub_zh dtype: string - name: obj_vi dtype: string - name: sub_vi dtype: string - name: obj_ms dtype: string - name: sub_ms dtype: string - name: obj_jp dtype: string - name: sub_jp dtype: string - name: obj_ko dtype: string - name: sub_ko dtype: string splits: - name: test num_bytes: 320933 num_examples: 627 download_size: 217599 dataset_size: 320933 - config_name: nl features: - name: uuid dtype: string - name: lineid dtype: uint32 - name: obj_uri dtype: string - name: obj_label dtype: string - name: sub_uri dtype: string - name: sub_label dtype: string - name: template dtype: string - name: language dtype: string - name: predicate_id dtype: string - name: obj_en dtype: string - name: sub_en dtype: string - name: obj_id dtype: string - name: sub_id dtype: string - name: obj_nl dtype: string - name: sub_nl dtype: string - name: obj_de dtype: string - name: sub_de dtype: string - name: obj_fr dtype: string - name: sub_fr dtype: string - name: obj_pt dtype: string - name: sub_pt dtype: string - name: obj_ru dtype: string - name: sub_ru dtype: string - name: obj_zh dtype: string - name: sub_zh dtype: string - name: obj_vi dtype: string - name: sub_vi dtype: string - name: obj_ms dtype: string - name: sub_ms dtype: string - name: obj_jp dtype: string - name: sub_jp dtype: string - name: obj_ko dtype: string - name: sub_ko dtype: string splits: - name: test num_bytes: 323765 num_examples: 627 download_size: 217424 dataset_size: 323765 - config_name: pt features: - name: uuid dtype: string - name: lineid dtype: uint32 - name: obj_uri dtype: string - name: obj_label dtype: string - name: sub_uri dtype: string - name: sub_label dtype: string - name: template dtype: string - name: language dtype: string - name: predicate_id dtype: string - name: obj_en dtype: string - name: sub_en dtype: string - name: obj_id dtype: string - name: sub_id dtype: string - name: obj_nl dtype: string - name: sub_nl dtype: string - name: obj_de dtype: string - name: sub_de dtype: string - name: obj_fr dtype: string - name: sub_fr dtype: string - name: obj_pt dtype: string - name: sub_pt dtype: string - name: obj_ru dtype: string - name: sub_ru dtype: string - name: obj_zh dtype: string - name: sub_zh dtype: string - name: obj_vi dtype: string - name: sub_vi dtype: string - name: obj_ms dtype: string - name: sub_ms dtype: string - name: obj_jp dtype: string - name: sub_jp dtype: string - name: obj_ko dtype: string - name: sub_ko dtype: string splits: - name: test num_bytes: 323469 num_examples: 627 download_size: 217782 dataset_size: 323469 - config_name: ru features: - name: uuid dtype: string - name: lineid dtype: uint32 - name: obj_uri dtype: string - name: obj_label dtype: string - name: sub_uri dtype: string - name: sub_label dtype: string - name: template dtype: string - name: language dtype: string - name: predicate_id dtype: string - name: obj_en dtype: string - name: sub_en dtype: string - name: obj_id dtype: string - name: sub_id dtype: string - name: obj_nl dtype: string - name: sub_nl dtype: string - name: obj_de dtype: string - name: sub_de dtype: string - name: obj_fr dtype: string - name: sub_fr dtype: string - name: obj_pt dtype: string - name: sub_pt dtype: string - name: obj_ru dtype: string - name: sub_ru dtype: string - name: obj_zh dtype: string - name: sub_zh dtype: string - name: obj_vi dtype: string - name: sub_vi dtype: string - name: obj_ms dtype: string - name: sub_ms dtype: string - name: obj_jp dtype: string - name: sub_jp dtype: string - name: obj_ko dtype: string - name: sub_ko dtype: string splits: - name: test num_bytes: 339140 num_examples: 627 download_size: 220432 dataset_size: 339140 - config_name: vi features: - name: uuid dtype: string - name: lineid dtype: uint32 - name: obj_uri dtype: string - name: obj_label dtype: string - name: sub_uri dtype: string - name: sub_label dtype: string - name: template dtype: string - name: language dtype: string - name: predicate_id dtype: string - name: obj_en dtype: string - name: sub_en dtype: string - name: obj_id dtype: string - name: sub_id dtype: string - name: obj_nl dtype: string - name: sub_nl dtype: string - name: obj_de dtype: string - name: sub_de dtype: string - name: obj_fr dtype: string - name: sub_fr dtype: string - name: obj_pt dtype: string - name: sub_pt dtype: string - name: obj_ru dtype: string - name: sub_ru dtype: string - name: obj_zh dtype: string - name: sub_zh dtype: string - name: obj_vi dtype: string - name: sub_vi dtype: string - name: obj_ms dtype: string - name: sub_ms dtype: string - name: obj_jp dtype: string - name: sub_jp dtype: string - name: obj_ko dtype: string - name: sub_ko dtype: string splits: - name: test num_bytes: 324871 num_examples: 627 download_size: 218344 dataset_size: 324871 - config_name: zh features: - name: uuid dtype: string - name: lineid dtype: uint32 - name: obj_uri dtype: string - name: obj_label dtype: string - name: sub_uri dtype: string - name: sub_label dtype: string - name: template dtype: string - name: language dtype: string - name: predicate_id dtype: string - name: obj_en dtype: string - name: sub_en dtype: string - name: obj_id dtype: string - name: sub_id dtype: string - name: obj_nl dtype: string - name: sub_nl dtype: string - name: obj_de dtype: string - name: sub_de dtype: string - name: obj_fr dtype: string - name: sub_fr dtype: string - name: obj_pt dtype: string - name: sub_pt dtype: string - name: obj_ru dtype: string - name: sub_ru dtype: string - name: obj_zh dtype: string - name: sub_zh dtype: string - name: obj_vi dtype: string - name: sub_vi dtype: string - name: obj_ms dtype: string - name: sub_ms dtype: string - name: obj_jp dtype: string - name: sub_jp dtype: string - name: obj_ko dtype: string - name: sub_ko dtype: string splits: - name: test num_bytes: 323614 num_examples: 627 download_size: 218358 dataset_size: 323614 configs: - config_name: de data_files: - split: test path: de/test-* - config_name: en data_files: - split: test path: en/test-* - config_name: fr data_files: - split: test path: fr/test-* - config_name: id data_files: - split: test path: id/test-* - config_name: jp data_files: - split: test path: jp/test-* - config_name: ko data_files: - split: test path: ko/test-* - config_name: ms data_files: - split: test path: ms/test-* - config_name: nl data_files: - split: test path: nl/test-* - config_name: pt data_files: - split: test path: pt/test-* - config_name: ru data_files: - split: test path: ru/test-* - config_name: vi data_files: - split: test path: vi/test-* - config_name: zh data_files: - split: test path: zh/test-* ---
dgambettaphd/D_llm2_gen0_W_doc1000_synt64_lr1e-04_acm_SYNLAST
dgambettaphd
2025-05-01T00:49:22Z
0
0
[ "size_categories:10K<n<100K", "format:parquet", "modality:tabular", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us" ]
[]
2025-05-01T00:49:13Z
null
--- dataset_info: features: - name: id_doc dtype: int64 - name: text dtype: string - name: dataset dtype: string - name: gen dtype: int64 - name: synt dtype: int64 - name: MPP dtype: float64 splits: - name: train num_bytes: 9145932 num_examples: 16000 download_size: 5492110 dataset_size: 9145932 configs: - config_name: default data_files: - split: train path: data/train-* ---
hjshah/bfcl
hjshah
2025-05-01T00:16:02Z
0
0
[ "size_categories:1K<n<10K", "format:parquet", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us" ]
[]
2025-05-01T00:15:58Z
null
--- dataset_info: features: - name: id dtype: string - name: multi_turn dtype: bool - name: functions dtype: string - name: missed_functions dtype: string - name: initial_config dtype: string - name: involved_classes sequence: string - name: turns dtype: string - name: language dtype: string - name: test_category dtype: string - name: subset dtype: string - name: ground_truth dtype: string splits: - name: train num_bytes: 63846322 num_examples: 4441 download_size: 7639281 dataset_size: 63846322 configs: - config_name: default data_files: - split: train path: data/train-* ---
liweijiang/conifer_v1
liweijiang
2025-04-30T23:26:07Z
0
0
[ "size_categories:10K<n<100K", "format:parquet", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us" ]
[]
2025-04-30T23:25:58Z
null
--- dataset_info: features: - name: prompt dtype: string - name: messages list: - name: content dtype: string - name: role dtype: string - name: type dtype: string splits: - name: train num_bytes: 64859889 num_examples: 13606 download_size: 31005133 dataset_size: 64859889 configs: - config_name: default data_files: - split: train path: data/train-* ---
allenchienxxx/so100_block_mug
allenchienxxx
2025-04-30T22:52:05Z
0
0
[ "task_categories:robotics", "license:apache-2.0", "size_categories:10K<n<100K", "format:parquet", "modality:tabular", "modality:timeseries", "modality:video", "library:datasets", "library:dask", "library:mlcroissant", "library:polars", "region:us", "LeRobot", "so100", "tutorial" ]
[ "robotics" ]
2025-04-30T20:28:27Z
null
--- license: apache-2.0 task_categories: - robotics tags: - LeRobot - so100 - tutorial configs: - config_name: default data_files: data/*/*.parquet --- This dataset was created using [LeRobot](https://github.com/huggingface/lerobot). ## Dataset Description - **Homepage:** [More Information Needed] - **Paper:** [More Information Needed] - **License:** apache-2.0 ## Dataset Structure [meta/info.json](meta/info.json): ```json { "codebase_version": "v2.1", "robot_type": "so100", "total_episodes": 10, "total_frames": 8938, "total_tasks": 1, "total_videos": 20, "total_chunks": 1, "chunks_size": 1000, "fps": 30, "splits": { "train": "0:10" }, "data_path": "data/chunk-{episode_chunk:03d}/episode_{episode_index:06d}.parquet", "video_path": "videos/chunk-{episode_chunk:03d}/{video_key}/episode_{episode_index:06d}.mp4", "features": { "action": { "dtype": "float32", "shape": [ 6 ], "names": [ "main_shoulder_pan", "main_shoulder_lift", "main_elbow_flex", "main_wrist_flex", "main_wrist_roll", "main_gripper" ] }, "observation.state": { "dtype": "float32", "shape": [ 6 ], "names": [ "main_shoulder_pan", "main_shoulder_lift", "main_elbow_flex", "main_wrist_flex", "main_wrist_roll", "main_gripper" ] }, "observation.images.laptop": { "dtype": "video", "shape": [ 480, 640, 3 ], "names": [ "height", "width", "channels" ], "info": { "video.fps": 30.0, "video.height": 480, "video.width": 640, "video.channels": 3, "video.codec": "h264", "video.pix_fmt": "yuv420p", "video.is_depth_map": false, "has_audio": false } }, "observation.images.phone": { "dtype": "video", "shape": [ 480, 640, 3 ], "names": [ "height", "width", "channels" ], "info": { "video.fps": 30.0, "video.height": 480, "video.width": 640, "video.channels": 3, "video.codec": "h264", "video.pix_fmt": "yuv420p", "video.is_depth_map": false, "has_audio": false } }, "timestamp": { "dtype": "float32", "shape": [ 1 ], "names": null }, "frame_index": { "dtype": "int64", "shape": [ 1 ], "names": null }, "episode_index": { "dtype": "int64", "shape": [ 1 ], "names": null }, "index": { "dtype": "int64", "shape": [ 1 ], "names": null }, "task_index": { "dtype": "int64", "shape": [ 1 ], "names": null } } } ``` ## Citation **BibTeX:** ```bibtex [More Information Needed] ```
Hkang/summarize_sft-test_lm-EleutherAI_pythia-1b_seed-42_numex-250_lr3e8_1K-BON_32
Hkang
2025-04-30T22:32:11Z
0
0
[ "size_categories:n<1K", "format:parquet", "modality:tabular", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us" ]
[]
2025-04-30T22:32:10Z
null
--- dataset_info: features: - name: id dtype: string - name: subreddit dtype: string - name: title dtype: string - name: post dtype: string - name: summary dtype: string - name: query_input_ids sequence: int64 - name: query_attention_mask sequence: int64 - name: query dtype: string - name: reference_response dtype: string - name: reference_response_input_ids sequence: int64 - name: reference_response_attention_mask sequence: int64 - name: reference_response_token_len dtype: int64 - name: query_reference_response dtype: string - name: query_reference_response_input_ids sequence: int64 - name: query_reference_response_attention_mask sequence: int64 - name: query_reference_response_token_response_label sequence: int64 - name: query_reference_response_token_len dtype: int64 - name: model_response dtype: string splits: - name: test num_bytes: 6852277 num_examples: 250 download_size: 1150373 dataset_size: 6852277 configs: - config_name: default data_files: - split: test path: data/test-* ---
tarsur909/rewards_negative_log-train-with-reward-stats-10ep-seperated-translated
tarsur909
2025-04-30T22:28:41Z
0
0
[ "size_categories:1K<n<10K", "format:parquet", "modality:tabular", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us" ]
[]
2025-04-30T22:28:39Z
null
--- dataset_info: features: - name: chosen_reward dtype: float64 - name: rejected_reward dtype: float64 - name: gt_chosen_reward dtype: float64 - name: gt_rejected_reward dtype: float64 - name: chosen_reward_gap dtype: float64 - name: rejected_reward_gap dtype: float64 - name: overall_reward_gap dtype: float64 - name: info struct: - name: article dtype: 'null' - name: id dtype: string - name: post dtype: string - name: site dtype: 'null' - name: subreddit dtype: string - name: title dtype: string - name: summaries list: - name: note dtype: 'null' - name: policy dtype: string - name: text dtype: string - name: choice dtype: int32 - name: worker dtype: string - name: batch dtype: string - name: split dtype: string - name: extra struct: - name: confidence dtype: 'null' - name: query_token sequence: int64 - name: query_attention_mask sequence: int64 - name: query dtype: string - name: chosen dtype: string - name: chosen_token sequence: int64 - name: chosen_attention_mask sequence: int64 - name: chosen_token_len dtype: int64 - name: rejected dtype: string - name: rejected_token sequence: int64 - name: rejected_attention_mask sequence: int64 - name: rejected_token_len dtype: int64 - name: chosen_policy dtype: string - name: rejected_policy dtype: string - name: policies dtype: string - name: query_chosen dtype: string - name: query_chosen_token sequence: int64 - name: query_chosen_attention_mask sequence: int64 - name: query_chosen_token_len dtype: int64 - name: query_rejected dtype: string - name: query_rejected_token sequence: int64 - name: query_rejected_attention_mask sequence: int64 - name: query_rejected_token_len dtype: int64 - name: query_token_len dtype: int64 - name: query_chosen_token_response_label sequence: int64 - name: query_rejected_token_response_label sequence: int64 - name: summary_rewards sequence: float64 - name: edge_weight dtype: int64 splits: - name: train num_bytes: 51112853 num_examples: 1000 download_size: 2195051 dataset_size: 51112853 configs: - config_name: default data_files: - split: train path: data/train-* ---
osama24sy/DeepSeek-R1-Distill-Qwen-7B-results-20250430-17460516009525
osama24sy
2025-04-30T22:26:54Z
0
0
[ "size_categories:n<1K", "format:parquet", "modality:tabular", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us" ]
[]
2025-04-30T22:26:51Z
null
--- dataset_info: features: - name: index dtype: int64 - name: numbers sequence: int64 - name: operations sequence: sequence: string - name: response dtype: string - name: token_count dtype: int64 splits: - name: train num_bytes: 954429 num_examples: 150 download_size: 425342 dataset_size: 954429 configs: - config_name: default data_files: - split: train path: data/train-* ---
younghyopark/toni_jasminetea_new5
younghyopark
2025-04-30T22:20:43Z
0
0
[ "task_categories:robotics", "license:apache-2.0", "size_categories:n<1K", "format:parquet", "modality:tabular", "modality:timeseries", "modality:video", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us", "LeRobot" ]
[ "robotics" ]
2025-04-30T22:20:41Z
null
--- license: apache-2.0 task_categories: - robotics tags: - LeRobot configs: - config_name: default data_files: data/*/*.parquet --- This dataset was created using [LeRobot](https://github.com/huggingface/lerobot). ## Dataset Description - **Homepage:** [More Information Needed] - **Paper:** [More Information Needed] - **License:** apache-2.0 ## Dataset Structure [meta/info.json](meta/info.json): ```json { "codebase_version": "v2.1", "robot_type": "bifranka", "total_episodes": 1, "total_frames": 36, "total_tasks": 1, "total_videos": 1, "total_chunks": 1, "chunks_size": 1000, "fps": 30, "splits": { "train": "0:1" }, "data_path": "data/chunk-{episode_chunk:03d}/episode_{episode_index:06d}.parquet", "video_path": "videos/chunk-{episode_chunk:03d}/{video_key}/episode_{episode_index:06d}.mp4", "features": { "observation.joint_positions": { "dtype": "float32", "shape": [ 18 ], "names": [ "l_joint_1", "l_joint_2", "l_joint_3", "l_joint_4", "l_joint_5", "l_joint_6", "l_joint_7", "l_gripper_left", "l_gripper_right", "r_joint_1", "r_joint_2", "r_joint_3", "r_joint_4", "r_joint_5", "r_joint_6", "r_joint_7", "r_gripper_left", "r_gripper_right" ] }, "observation.ee_pose": { "dtype": "float32", "shape": [ 14 ], "names": [ "l_pos_x", "l_pos_y", "l_pos_z", "l_quat_w", "l_quat_x", "l_quat_y", "l_quat_z", "r_pos_x", "r_pos_y", "r_pos_z", "r_quat_w", "r_quat_x", "r_quat_y", "r_quat_z" ] }, "action": { "dtype": "float32", "shape": [ 16 ], "names": [ "l_target_joint_1", "l_target_joint_2", "l_target_joint_3", "l_target_joint_4", "l_target_joint_5", "l_target_joint_6", "l_target_joint_7", "l_target_gripper", "r_target_joint_1", "r_target_joint_2", "r_target_joint_3", "r_target_joint_4", "r_target_joint_5", "r_target_joint_6", "r_target_joint_7", "r_target_gripper" ] }, "action.ee_pose": { "dtype": "float32", "shape": [ 32 ], "names": [ "l_matrix_0_0", "l_matrix_0_1", "l_matrix_0_2", "l_matrix_0_3", "l_matrix_1_0", "l_matrix_1_1", "l_matrix_1_2", "l_matrix_1_3", "l_matrix_2_0", "l_matrix_2_1", "l_matrix_2_2", "l_matrix_2_3", "l_matrix_3_0", "l_matrix_3_1", "l_matrix_3_2", "l_matrix_3_3", "r_matrix_0_0", "r_matrix_0_1", "r_matrix_0_2", "r_matrix_0_3", "r_matrix_1_0", "r_matrix_1_1", "r_matrix_1_2", "r_matrix_1_3", "r_matrix_2_0", "r_matrix_2_1", "r_matrix_2_2", "r_matrix_2_3", "r_matrix_3_0", "r_matrix_3_1", "r_matrix_3_2", "r_matrix_3_3" ] }, "action.gripper": { "dtype": "float32", "shape": [ 2 ], "names": [ "l_gripper", "r_gripper" ] }, "rgb.global_0": { "dtype": "video", "shape": [ 720, 1280, 3 ], "names": [ "rgb" ], "info": { "video.fps": 30.0, "video.height": 720, "video.width": 1280, "video.channels": 3, "video.codec": "h264", "video.pix_fmt": "yuv420p", "video.is_depth_map": false, "has_audio": false } }, "pose.jasminetea": { "dtype": "float32", "shape": [ 4, 4 ], "names": [ "pose" ] }, "timestamp": { "dtype": "float32", "shape": [ 1 ], "names": null }, "episode_index": { "dtype": "int64", "shape": [ 1 ], "names": null }, "frame_index": { "dtype": "int64", "shape": [ 1 ], "names": null }, "index": { "dtype": "int64", "shape": [ 1 ], "names": null }, "task_index": { "dtype": "int64", "shape": [ 1 ], "names": null } } } ``` ## Citation **BibTeX:** ```bibtex [More Information Needed] ```
lmcinnes/arxiv_ml
lmcinnes
2025-04-30T22:10:13Z
0
0
[ "size_categories:100K<n<1M", "format:parquet", "modality:text", "library:datasets", "library:dask", "library:mlcroissant", "library:polars", "region:us" ]
[]
2025-04-30T21:53:39Z
null
--- dataset_info: features: - name: date_created dtype: timestamp[ns] - name: abstract dtype: string - name: title dtype: string - name: categories dtype: string - name: arxiv_id dtype: string - name: year dtype: int32 - name: embedding_str dtype: string - name: embedding sequence: float64 - name: data_map sequence: float64 splits: - name: train num_bytes: 2450676134 num_examples: 281816 download_size: 1807632673 dataset_size: 2450676134 configs: - config_name: default data_files: - split: train path: data/train-* --- # Dataset Card for "arxiv_ml" ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description ### Dataset Summary This is a dataset of titles and abstracts of machine learning related papers from ArXiv. This data is derived from the [ArXiv dataset available on Kaggle](https://www.kaggle.com/datasets/Cornell-University/arxiv). The selection of papers was determined by selecting all papers that used a category tag in the set {"cs.LG", "cs.AI", "cs.CL", "stat.ML", "cs.IR", "cs.NE", "cs.SC"}. To supplement the titles and abstracts the creation time of the paper, as well as the categories are provided. To make exploration easier embeddings of the title and abstract have been made using the [Nomic-embed-v2-moe](https://huggingface.co/nomic-ai/nomic-embed-text-v2-moe) text embedding model, and a 2D representation using UMAP is also provided. ### Supported Tasks This dataset is primarily aimed at tasks such as topic modelling, corpus triage, search and information retrieval, and other NLP tasks. ### Languages The dataset is in English, although other languages may also be present. ## Dataset Creation ### Curation Rationale The fill ArXiv dataset is too large for many tasks. Subsetting to a selection of ArXiv categories related the AI and ML ensures a reasonably sized dataset that should mostly contain topics that are familiar to those wishing to use the dataset. ### Source Data This data is derived from the [ArXiv dataset available on Kaggle](https://www.kaggle.com/datasets/Cornell-University/arxiv). ### Personal and Sensitive Information This dataset contains publicly published information that was available under a CC0: public domain license via Kaggle. There should be no personal or senstive information in this dataset. If this is in error, please contact the maintainer and we will endeavour to remedy any issues. ## Additional Information ### Dataset Curators Leland McInnes for the curated subset, Cornell University for the initial full dataset. ### Licensing Information Licensed as CC0: Public Domain.
Rabe3/Egy-Conv-Unsloth-Format
Rabe3
2025-04-30T22:07:42Z
0
0
[ "size_categories:10K<n<100K", "format:parquet", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us" ]
[]
2025-04-30T22:07:38Z
null
--- dataset_info: features: - name: instruction dtype: string - name: messages sequence: - name: role dtype: string - name: content dtype: string - name: conversations list: - name: content dtype: string - name: role dtype: string splits: - name: train num_bytes: 5096450 num_examples: 10000 download_size: 151141 dataset_size: 5096450 configs: - config_name: default data_files: - split: train path: data/train-* ---
miaolu3/or-2.5k-scip
miaolu3
2025-04-30T21:28:01Z
0
0
[ "size_categories:1K<n<10K", "format:parquet", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us" ]
[]
2025-04-30T21:25:39Z
null
--- dataset_info: features: - name: data_source dtype: string - name: prompt list: - name: content dtype: string - name: role dtype: string - name: ability dtype: string - name: reward_model struct: - name: ground_truth dtype: float64 - name: style dtype: string - name: extra_info struct: - name: index dtype: int64 - name: split dtype: string splits: - name: train num_bytes: 3748001 num_examples: 2499 - name: test num_bytes: 746824 num_examples: 500 download_size: 1619922 dataset_size: 4494825 configs: - config_name: default data_files: - split: train path: data/train-* - split: test path: data/test-* ---
korbih/curriculum-0-test-20250424_213955_postprocessed-grpo-format
korbih
2025-04-30T21:03:11Z
0
0
[ "size_categories:n<1K", "format:parquet", "modality:image", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us" ]
[]
2025-04-30T21:03:07Z
null
--- dataset_info: features: - name: base_uid dtype: string - name: step dtype: int32 - name: messages list: - name: content dtype: string - name: role dtype: string - name: image_name dtype: string - name: start_url dtype: string - name: image dtype: image splits: - name: train num_bytes: 5114144.0 num_examples: 70 download_size: 4344276 dataset_size: 5114144.0 configs: - config_name: default data_files: - split: train path: data/train-* ---
pxyyy/autoif-20k-removed-top7500-by-mp-1e-3-for-ifeval-correct-200
pxyyy
2025-04-30T20:33:39Z
0
0
[ "size_categories:10K<n<100K", "format:parquet", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us" ]
[]
2025-04-30T20:33:33Z
null
--- dataset_info: features: - name: system dtype: string - name: tools sequence: 'null' - name: messages list: - name: content dtype: string - name: role dtype: string - name: conversation_id dtype: int64 splits: - name: train num_bytes: 22748022.5 num_examples: 12500 download_size: 12903450 dataset_size: 22748022.5 configs: - config_name: default data_files: - split: train path: data/train-* --- # Dataset Card for "autoif-20k-removed-top7500-by-mp-1e-3-for-ifeval-correct-200" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
cchoi1/bigcodebench_qwen7b_att_iter0_ppo_att20_sol5
cchoi1
2025-04-30T19:52:16Z
0
0
[ "size_categories:1K<n<10K", "format:parquet", "modality:tabular", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us" ]
[]
2025-04-30T19:52:14Z
null
--- dataset_info: features: - name: mutation_id dtype: int64 - name: task_id dtype: string - name: mutator_prompt dtype: string - name: solver_prompt dtype: string - name: response dtype: string - name: mutation_explanation dtype: string - name: mutation_info dtype: string - name: mutator_score dtype: float64 - name: solution_scores dtype: string - name: solutions dtype: string - name: solutions_explanation dtype: string - name: solutions_info dtype: string splits: - name: train num_bytes: 18099255 num_examples: 1378 download_size: 3772359 dataset_size: 18099255 configs: - config_name: default data_files: - split: train path: data/train-* ---
sumuks/yourbench_advanced_example
sumuks
2025-04-30T19:00:20Z
0
0
[ "size_categories:n<1K", "format:parquet", "modality:tabular", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us" ]
[]
2025-04-30T18:59:04Z
null
--- dataset_info: - config_name: chunked features: - name: document_id dtype: string - name: document_text dtype: string - name: document_filename dtype: string - name: document_metadata struct: - name: file_size dtype: int64 - name: raw_chunk_summaries sequence: string - name: chunk_summaries sequence: string - name: raw_document_summary dtype: string - name: document_summary dtype: string - name: summarization_model dtype: string - name: chunks list: - name: chunk_id dtype: string - name: chunk_text dtype: string - name: multihop_chunks list: - name: chunk_ids sequence: string - name: chunks_text sequence: string - name: chunk_info_metrics list: - name: avg_token_length dtype: float64 - name: bigram_diversity dtype: float64 - name: flesch_reading_ease dtype: float64 - name: gunning_fog dtype: float64 - name: perplexity dtype: float64 - name: token_count dtype: float64 - name: unique_token_ratio dtype: float64 - name: chunking_model dtype: string splits: - name: train num_bytes: 64130 num_examples: 2 download_size: 80573 dataset_size: 64130 - config_name: ingested features: - name: document_id dtype: string - name: document_text dtype: string - name: document_filename dtype: string - name: document_metadata struct: - name: file_size dtype: int64 splits: - name: train num_bytes: 18022 num_examples: 2 download_size: 13449 dataset_size: 18022 - config_name: lighteval features: - name: question dtype: string - name: additional_instructions dtype: string - name: ground_truth_answer dtype: string - name: question_category dtype: string - name: kind dtype: string - name: estimated_difficulty dtype: int64 - name: citations sequence: string - name: document_id dtype: string - name: chunk_ids sequence: string - name: question_generating_model dtype: string - name: chunks sequence: string - name: document dtype: string - name: document_summary dtype: string - name: answer_citation_score dtype: float64 - name: chunk_citation_score dtype: float64 - name: citation_score dtype: float64 splits: - name: train num_bytes: 932563 num_examples: 76 download_size: 62155 dataset_size: 932563 - config_name: multi_hop_questions features: - name: document_id dtype: string - name: source_chunk_ids sequence: string - name: additional_instructions dtype: string - name: question dtype: string - name: self_answer dtype: string - name: estimated_difficulty dtype: int64 - name: self_assessed_question_type dtype: string - name: generating_model dtype: string - name: thought_process dtype: string - name: citations sequence: string - name: raw_response dtype: string splits: - name: train num_bytes: 68501 num_examples: 6 download_size: 23178 dataset_size: 68501 - config_name: single_shot_questions features: - name: chunk_id dtype: string - name: document_id dtype: string - name: additional_instructions dtype: string - name: question dtype: string - name: self_answer dtype: string - name: estimated_difficulty dtype: int64 - name: self_assessed_question_type dtype: string - name: generating_model dtype: string - name: thought_process dtype: string - name: raw_response dtype: string - name: citations sequence: string splits: - name: train num_bytes: 595318 num_examples: 70 download_size: 103022 dataset_size: 595318 - config_name: summarized features: - name: document_id dtype: string - name: document_text dtype: string - name: document_filename dtype: string - name: document_metadata struct: - name: file_size dtype: int64 - name: raw_chunk_summaries sequence: string - name: chunk_summaries sequence: string - name: raw_document_summary dtype: string - name: document_summary dtype: string - name: summarization_model dtype: string splits: - name: train num_bytes: 30566 num_examples: 2 download_size: 49418 dataset_size: 30566 configs: - config_name: chunked data_files: - split: train path: chunked/train-* - config_name: ingested data_files: - split: train path: ingested/train-* - config_name: lighteval data_files: - split: train path: lighteval/train-* - config_name: multi_hop_questions data_files: - split: train path: multi_hop_questions/train-* - config_name: single_shot_questions data_files: - split: train path: single_shot_questions/train-* - config_name: summarized data_files: - split: train path: summarized/train-* ---
dgambettaphd/D_llm2_gen6_run0_W_doc1000_synt64_tot128_lr5em5_p1k_SYNLAST
dgambettaphd
2025-04-30T18:11:01Z
0
0
[ "size_categories:10K<n<100K", "format:parquet", "modality:tabular", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us" ]
[]
2025-04-30T18:10:56Z
null
--- dataset_info: features: - name: id dtype: int64 - name: text dtype: string - name: dataset dtype: string - name: gen dtype: int64 - name: synt dtype: int64 - name: MPP dtype: float64 splits: - name: train num_bytes: 5188613 num_examples: 10000 download_size: 2896477 dataset_size: 5188613 configs: - config_name: default data_files: - split: train path: data/train-* ---
EYEDOL/mozilla_commonvoice_naijaHausa0_preprocessed_train_batch_6
EYEDOL
2025-04-30T17:47:04Z
0
0
[ "size_categories:10K<n<100K", "format:parquet", "modality:audio", "modality:text", "library:datasets", "library:dask", "library:mlcroissant", "library:polars", "region:us" ]
[]
2025-04-30T17:42:26Z
null
--- dataset_info: features: - name: audio dtype: audio: sampling_rate: 16000 - name: text dtype: string - name: input_length dtype: int64 - name: input_features sequence: sequence: float32 - name: labels sequence: int64 - name: labels_length dtype: int64 splits: - name: train num_bytes: 15657485287.625 num_examples: 14763 download_size: 3224956955 dataset_size: 15657485287.625 configs: - config_name: default data_files: - split: train path: data/train-* ---
shylee/eval_temp3
shylee
2025-04-30T17:45:57Z
0
0
[ "task_categories:robotics", "license:apache-2.0", "size_categories:n<1K", "format:parquet", "modality:tabular", "modality:timeseries", "modality:video", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us", "LeRobot", "tutorial" ]
[ "robotics" ]
2025-04-30T17:45:49Z
null
--- license: apache-2.0 task_categories: - robotics tags: - LeRobot - tutorial configs: - config_name: default data_files: data/*/*.parquet --- This dataset was created using [LeRobot](https://github.com/huggingface/lerobot). ## Dataset Description - **Homepage:** [More Information Needed] - **Paper:** [More Information Needed] - **License:** apache-2.0 ## Dataset Structure [meta/info.json](meta/info.json): ```json { "codebase_version": "v2.1", "robot_type": "so100", "total_episodes": 1, "total_frames": 854, "total_tasks": 1, "total_videos": 3, "total_chunks": 1, "chunks_size": 1000, "fps": 30, "splits": { "train": "0:1" }, "data_path": "data/chunk-{episode_chunk:03d}/episode_{episode_index:06d}.parquet", "video_path": "videos/chunk-{episode_chunk:03d}/{video_key}/episode_{episode_index:06d}.mp4", "features": { "action": { "dtype": "float32", "shape": [ 6 ], "names": [ "main_shoulder_pan", "main_shoulder_lift", "main_elbow_flex", "main_wrist_flex", "main_wrist_roll", "main_gripper" ] }, "observation.state": { "dtype": "float32", "shape": [ 6 ], "names": [ "main_shoulder_pan", "main_shoulder_lift", "main_elbow_flex", "main_wrist_flex", "main_wrist_roll", "main_gripper" ] }, "observation.images.FrontCam": { "dtype": "video", "shape": [ 480, 640, 3 ], "names": [ "height", "width", "channels" ], "info": { "video.fps": 30.0, "video.height": 480, "video.width": 640, "video.channels": 3, "video.codec": "av1", "video.pix_fmt": "yuv420p", "video.is_depth_map": false, "has_audio": false } }, "observation.images.TopCam": { "dtype": "video", "shape": [ 480, 640, 3 ], "names": [ "height", "width", "channels" ], "info": { "video.fps": 30.0, "video.height": 480, "video.width": 640, "video.channels": 3, "video.codec": "av1", "video.pix_fmt": "yuv420p", "video.is_depth_map": false, "has_audio": false } }, "observation.images.WristCam": { "dtype": "video", "shape": [ 480, 640, 3 ], "names": [ "height", "width", "channels" ], "info": { "video.fps": 30.0, "video.height": 480, "video.width": 640, "video.channels": 3, "video.codec": "av1", "video.pix_fmt": "yuv420p", "video.is_depth_map": false, "has_audio": false } }, "timestamp": { "dtype": "float32", "shape": [ 1 ], "names": null }, "frame_index": { "dtype": "int64", "shape": [ 1 ], "names": null }, "episode_index": { "dtype": "int64", "shape": [ 1 ], "names": null }, "index": { "dtype": "int64", "shape": [ 1 ], "names": null }, "task_index": { "dtype": "int64", "shape": [ 1 ], "names": null } } } ``` ## Citation **BibTeX:** ```bibtex [More Information Needed] ```
kwangchaeko/koch_test
kwangchaeko
2025-04-30T13:04:39Z
0
0
[ "task_categories:robotics", "license:apache-2.0", "region:us", "LeRobot", "koch", "tutorial" ]
[ "robotics" ]
2025-04-30T13:04:26Z
null
--- license: apache-2.0 task_categories: - robotics tags: - LeRobot - koch - tutorial configs: - config_name: default data_files: data/*/*.parquet --- This dataset was created using [LeRobot](https://github.com/huggingface/lerobot). ## Dataset Description - **Homepage:** [More Information Needed] - **Paper:** [More Information Needed] - **License:** apache-2.0 ## Dataset Structure [meta/info.json](meta/info.json): ```json { "codebase_version": "v2.1", "robot_type": "koch", "total_episodes": 2, "total_frames": 1685, "total_tasks": 1, "total_videos": 2, "total_chunks": 1, "chunks_size": 1000, "fps": 30, "splits": { "train": "0:2" }, "data_path": "data/chunk-{episode_chunk:03d}/episode_{episode_index:06d}.parquet", "video_path": "videos/chunk-{episode_chunk:03d}/{video_key}/episode_{episode_index:06d}.mp4", "features": { "action": { "dtype": "float32", "shape": [ 4 ], "names": [ "main_shoulder_pan", "main_shoulder_lift", "main_elbow_flex", "main_wrist_flex" ] }, "observation.state": { "dtype": "float32", "shape": [ 4 ], "names": [ "main_shoulder_pan", "main_shoulder_lift", "main_elbow_flex", "main_wrist_flex" ] }, "observation.images.laptop": { "dtype": "video", "shape": [ 480, 640, 3 ], "names": [ "height", "width", "channels" ], "info": { "video.fps": 30.0, "video.height": 480, "video.width": 640, "video.channels": 3, "video.codec": "av1", "video.pix_fmt": "yuv420p", "video.is_depth_map": false, "has_audio": false } }, "timestamp": { "dtype": "float32", "shape": [ 1 ], "names": null }, "frame_index": { "dtype": "int64", "shape": [ 1 ], "names": null }, "episode_index": { "dtype": "int64", "shape": [ 1 ], "names": null }, "index": { "dtype": "int64", "shape": [ 1 ], "names": null }, "task_index": { "dtype": "int64", "shape": [ 1 ], "names": null } } } ``` ## Citation **BibTeX:** ```bibtex [More Information Needed] ```
rd-lumi-ai/VietSpeech
rd-lumi-ai
2025-04-30T13:03:23Z
0
0
[ "region:us" ]
[]
2025-04-30T06:30:43Z
null
--- dataset_info: features: - name: audio dtype: audio - name: transcription dtype: string - name: continuation dtype: string splits: - name: train num_bytes: 20992253661.0 num_examples: 162924 download_size: 20743956689 dataset_size: 20992253661.0 configs: - config_name: default data_files: - split: train path: default/train/** ---
midrees2806/unmatched_queries
midrees2806
2025-04-30T12:47:03Z
0
0
[ "size_categories:n<1K", "format:parquet", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us" ]
[]
2025-04-30T07:17:00Z
null
--- dataset_info: features: - name: Query dtype: string - name: Timestamp dtype: string - name: Processed dtype: bool splits: - name: train num_bytes: 130 num_examples: 3 download_size: 1546 dataset_size: 130 configs: - config_name: default data_files: - split: train path: data/train-* ---
ttn1410/Efficiency_smr
ttn1410
2025-04-30T12:45:50Z
0
0
[ "region:us" ]
[]
2025-04-29T16:59:11Z
null
--- dataset_info: features: - name: reports dtype: string - name: labels dtype: string splits: - name: train num_bytes: 84985107 num_examples: 33180 download_size: 7962553 dataset_size: 84985107 configs: - config_name: default data_files: - split: train path: data/train-* ---
multilingual-discourse-hub/disrpt
multilingual-discourse-hub
2025-04-30T12:44:42Z
9,361
2
[ "language:en", "license:apache-2.0", "size_categories:100K<n<1M", "format:parquet", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us" ]
[]
2023-04-18T07:36:18Z
null
--- language: - en license: apache-2.0 dataset_info: - config_name: deu.rst.pcc.conllu features: - name: id sequence: string - name: form sequence: string - name: lemma sequence: string - name: upos sequence: string - name: xpos sequence: string - name: feats sequence: string - name: head sequence: string - name: deprel sequence: string - name: deps sequence: string - name: misc sequence: string - name: doc_id dtype: string - name: mwe sequence: 'null' splits: - name: train num_bytes: 3002196 num_examples: 1773 - name: validation num_bytes: 354996 num_examples: 207 - name: test num_bytes: 357461 num_examples: 213 download_size: 0 dataset_size: 3714653 - config_name: deu.rst.pcc.rels features: - name: doc dtype: string - name: unit1_toks dtype: string - name: unit2_toks dtype: string - name: unit1_txt dtype: string - name: unit2_txt dtype: string - name: s1_toks dtype: string - name: s2_toks dtype: string - name: unit1_sent dtype: string - name: unit2_sent dtype: string - name: dir dtype: string - name: orig_label dtype: string - name: label dtype: string splits: - name: train num_bytes: 1061634 num_examples: 2164 - name: validation num_bytes: 120267 num_examples: 241 - name: test num_bytes: 131921 num_examples: 260 download_size: 0 dataset_size: 1313822 - config_name: eng.dep.covdtb.conllu features: - name: id sequence: string - name: form sequence: string - name: lemma sequence: string - name: upos sequence: string - name: xpos sequence: string - name: feats sequence: string - name: head sequence: string - name: deprel sequence: string - name: deps sequence: string - name: misc sequence: string - name: doc_id dtype: string - name: mwe sequence: string splits: - name: validation num_bytes: 2520766 num_examples: 1162 - name: test num_bytes: 2687889 num_examples: 1181 download_size: 705960 dataset_size: 5208655 - config_name: eng.dep.covdtb.rels features: - name: doc dtype: string - name: unit1_toks dtype: string - name: unit2_toks dtype: string - name: unit1_txt dtype: string - name: unit2_txt dtype: string - name: s1_toks dtype: string - name: s2_toks dtype: string - name: unit1_sent dtype: string - name: unit2_sent dtype: string - name: dir dtype: string - name: orig_label dtype: string - name: label dtype: string splits: - name: validation num_bytes: 1542756 num_examples: 2399 - name: test num_bytes: 1726590 num_examples: 2586 download_size: 923126 dataset_size: 3269346 - config_name: eng.dep.scidtb.conllu features: - name: id sequence: string - name: form sequence: string - name: lemma sequence: string - name: upos sequence: string - name: xpos sequence: string - name: feats sequence: string - name: head sequence: string - name: deprel sequence: string - name: deps sequence: string - name: misc sequence: string - name: doc_id dtype: string - name: mwe sequence: string splits: - name: train num_bytes: 5396222 num_examples: 2570 - name: validation num_bytes: 1749607 num_examples: 815 - name: test num_bytes: 1706155 num_examples: 817 download_size: 0 dataset_size: 8851984 - config_name: eng.dep.scidtb.rels features: - name: doc dtype: string - name: unit1_toks dtype: string - name: unit2_toks dtype: string - name: unit1_txt dtype: string - name: unit2_txt dtype: string - name: s1_toks dtype: string - name: s2_toks dtype: string - name: unit1_sent dtype: string - name: unit2_sent dtype: string - name: dir dtype: string - name: orig_label dtype: string - name: label dtype: string splits: - name: train num_bytes: 3463826 num_examples: 6060 - name: validation num_bytes: 1125360 num_examples: 1933 - name: test num_bytes: 1092953 num_examples: 1911 download_size: 0 dataset_size: 5682139 - config_name: eng.pdtb.tedm.conllu features: - name: id sequence: string - name: form sequence: string - name: lemma sequence: string - name: upos sequence: string - name: xpos sequence: string - name: feats sequence: string - name: head sequence: string - name: deprel sequence: string - name: deps sequence: string - name: misc sequence: string - name: doc_id dtype: string - name: mwe sequence: string splits: - name: validation num_bytes: 245201 num_examples: 143 - name: test num_bytes: 520205 num_examples: 238 download_size: 135746 dataset_size: 765406 - config_name: eng.pdtb.tedm.rels features: - name: doc dtype: string - name: unit1_toks dtype: string - name: unit2_toks dtype: string - name: unit1_txt dtype: string - name: unit2_txt dtype: string - name: s1_toks dtype: string - name: s2_toks dtype: string - name: unit1_sent dtype: string - name: unit2_sent dtype: string - name: dir dtype: string - name: orig_label dtype: string - name: label dtype: string splits: - name: validation num_bytes: 100737 num_examples: 178 - name: test num_bytes: 221689 num_examples: 351 download_size: 124850 dataset_size: 322426 - config_name: eng.rst.gentle.conllu features: - name: id sequence: string - name: form sequence: string - name: lemma sequence: string - name: upos sequence: string - name: xpos sequence: string - name: feats sequence: string - name: head sequence: string - name: deprel sequence: string - name: deps sequence: string - name: misc sequence: string - name: doc_id dtype: string - name: mwe sequence: string splits: - name: test num_bytes: 2124578 num_examples: 1334 download_size: 438782 dataset_size: 2124578 - config_name: eng.rst.gentle.rels features: - name: doc dtype: string - name: unit1_toks dtype: string - name: unit2_toks dtype: string - name: unit1_txt dtype: string - name: unit2_txt dtype: string - name: s1_toks dtype: string - name: s2_toks dtype: string - name: unit1_sent dtype: string - name: unit2_sent dtype: string - name: dir dtype: string - name: orig_label dtype: string - name: label dtype: string splits: - name: test num_bytes: 1105756 num_examples: 2540 download_size: 277584 dataset_size: 1105756 - config_name: eng.sdrt.stac.conllu features: - name: id sequence: string - name: form sequence: string - name: lemma sequence: string - name: upos sequence: string - name: xpos sequence: string - name: feats sequence: string - name: head sequence: string - name: deprel sequence: string - name: deps sequence: string - name: misc sequence: string - name: doc_id dtype: string - name: mwe sequence: string splits: - name: train num_bytes: 4215467 num_examples: 8754 - name: validation num_bytes: 488587 num_examples: 991 - name: test num_bytes: 675799 num_examples: 1342 download_size: 744446 dataset_size: 5379853 - config_name: eng.sdrt.stac.rels features: - name: doc dtype: string - name: unit1_toks dtype: string - name: unit2_toks dtype: string - name: unit1_txt dtype: string - name: unit2_txt dtype: string - name: s1_toks dtype: string - name: s2_toks dtype: string - name: unit1_sent dtype: string - name: unit2_sent dtype: string - name: dir dtype: string - name: orig_label dtype: string - name: label dtype: string splits: - name: train num_bytes: 1917800 num_examples: 9580 - name: validation num_bytes: 230309 num_examples: 1145 - name: test num_bytes: 311996 num_examples: 1510 download_size: 1030567 dataset_size: 2460105 - config_name: eus.rst.ert.conllu features: - name: id sequence: string - name: form sequence: string - name: lemma sequence: string - name: upos sequence: string - name: xpos sequence: string - name: feats sequence: string - name: head sequence: string - name: deprel sequence: string - name: deps sequence: string - name: misc sequence: string - name: doc_id dtype: string - name: mwe sequence: 'null' splits: - name: train num_bytes: 2952101 num_examples: 1599 - name: validation num_bytes: 693012 num_examples: 366 - name: test num_bytes: 753179 num_examples: 415 download_size: 635612 dataset_size: 4398292 - config_name: eus.rst.ert.rels features: - name: doc dtype: string - name: unit1_toks dtype: string - name: unit2_toks dtype: string - name: unit1_txt dtype: string - name: unit2_txt dtype: string - name: s1_toks dtype: string - name: s2_toks dtype: string - name: unit1_sent dtype: string - name: unit2_sent dtype: string - name: dir dtype: string - name: orig_label dtype: string - name: label dtype: string splits: - name: train num_bytes: 1544724 num_examples: 2533 - name: validation num_bytes: 367114 num_examples: 614 - name: test num_bytes: 398853 num_examples: 678 download_size: 267097 dataset_size: 2310691 - config_name: fas.rst.prstc.conllu features: - name: id sequence: string - name: form sequence: string - name: lemma sequence: string - name: upos sequence: string - name: xpos sequence: string - name: feats sequence: string - name: head sequence: string - name: deprel sequence: string - name: deps sequence: string - name: misc sequence: string - name: doc_id dtype: string - name: mwe sequence: string splits: - name: train num_bytes: 4738331 num_examples: 1713 - name: validation num_bytes: 631614 num_examples: 202 - name: test num_bytes: 668584 num_examples: 264 download_size: 731899 dataset_size: 6038529 - config_name: fas.rst.prstc.rels features: - name: doc dtype: string - name: unit1_toks dtype: string - name: unit2_toks dtype: string - name: unit1_txt dtype: string - name: unit2_txt dtype: string - name: s1_toks dtype: string - name: s2_toks dtype: string - name: unit1_sent dtype: string - name: unit2_sent dtype: string - name: dir dtype: string - name: orig_label dtype: string - name: label dtype: string splits: - name: train num_bytes: 4147843 num_examples: 4100 - name: validation num_bytes: 576772 num_examples: 499 - name: test num_bytes: 576615 num_examples: 592 download_size: 1151593 dataset_size: 5301230 - config_name: fra.sdrt.annodis.conllu features: - name: id sequence: string - name: form sequence: string - name: lemma sequence: string - name: upos sequence: string - name: xpos sequence: string - name: feats sequence: string - name: head sequence: string - name: deprel sequence: string - name: deps sequence: string - name: misc sequence: string - name: doc_id dtype: string - name: mwe sequence: 'null' splits: - name: train num_bytes: 2089898 num_examples: 1020 - name: validation num_bytes: 470604 num_examples: 245 - name: test num_bytes: 479485 num_examples: 242 download_size: 420807 dataset_size: 3039987 - config_name: fra.sdrt.annodis.rels features: - name: doc dtype: string - name: unit1_toks dtype: string - name: unit2_toks dtype: string - name: unit1_txt dtype: string - name: unit2_txt dtype: string - name: s1_toks dtype: string - name: s2_toks dtype: string - name: unit1_sent dtype: string - name: unit2_sent dtype: string - name: dir dtype: string - name: orig_label dtype: string - name: label dtype: string splits: - name: train num_bytes: 1276541 num_examples: 2185 - name: validation num_bytes: 283730 num_examples: 528 - name: test num_bytes: 342131 num_examples: 625 download_size: 571005 dataset_size: 1902402 - config_name: ita.pdtb.luna.conllu features: - name: id sequence: string - name: form sequence: string - name: lemma sequence: string - name: upos sequence: string - name: xpos sequence: string - name: feats sequence: string - name: head sequence: string - name: deprel sequence: string - name: deps sequence: string - name: misc sequence: string - name: doc_id dtype: string - name: mwe sequence: string splits: - name: train num_bytes: 1884586 num_examples: 3721 - name: validation num_bytes: 350547 num_examples: 775 - name: test num_bytes: 706637 num_examples: 1315 download_size: 505505 dataset_size: 2941770 - config_name: ita.pdtb.luna.rels features: - name: doc dtype: int64 - name: unit1_toks dtype: string - name: unit2_toks dtype: string - name: unit1_txt dtype: string - name: unit2_txt dtype: string - name: s1_toks dtype: string - name: s2_toks dtype: string - name: unit1_sent dtype: string - name: unit2_sent dtype: string - name: dir dtype: string - name: orig_label dtype: string - name: label dtype: string splits: - name: train num_bytes: 498204 num_examples: 955 - name: validation num_bytes: 107102 num_examples: 209 - name: test num_bytes: 209035 num_examples: 380 download_size: 271620 dataset_size: 814341 - config_name: nld.rst.nldt.conllu features: - name: id sequence: string - name: form sequence: string - name: lemma sequence: string - name: upos sequence: string - name: xpos sequence: string - name: feats sequence: string - name: head sequence: string - name: deprel sequence: string - name: deps sequence: string - name: misc sequence: string - name: doc_id dtype: string - name: mwe sequence: 'null' splits: - name: train num_bytes: 1738747 num_examples: 1155 - name: validation num_bytes: 375701 num_examples: 254 - name: test num_bytes: 350324 num_examples: 239 download_size: 334430 dataset_size: 2464772 - config_name: nld.rst.nldt.rels features: - name: doc dtype: string - name: unit1_toks dtype: string - name: unit2_toks dtype: string - name: unit1_txt dtype: string - name: unit2_txt dtype: string - name: s1_toks dtype: string - name: s2_toks dtype: string - name: unit1_sent dtype: string - name: unit2_sent dtype: string - name: dir dtype: string - name: orig_label dtype: string - name: label dtype: string splits: - name: train num_bytes: 705665 num_examples: 1608 - name: validation num_bytes: 146881 num_examples: 331 - name: test num_bytes: 143380 num_examples: 325 download_size: 372780 dataset_size: 995926 - config_name: por.pdtb.crpc.conllu features: - name: id sequence: string - name: form sequence: string - name: lemma sequence: string - name: upos sequence: string - name: xpos sequence: string - name: feats sequence: string - name: head sequence: string - name: deprel sequence: string - name: deps sequence: string - name: misc sequence: string - name: doc_id dtype: string - name: mwe sequence: 'null' splits: - name: train num_bytes: 14274302 num_examples: 4078 - name: validation num_bytes: 1948042 num_examples: 581 - name: test num_bytes: 1852695 num_examples: 535 download_size: 2246483 dataset_size: 18075039 - config_name: por.pdtb.crpc.rels features: - name: doc dtype: string - name: unit1_toks dtype: string - name: unit2_toks dtype: string - name: unit1_txt dtype: string - name: unit2_txt dtype: string - name: s1_toks dtype: string - name: s2_toks dtype: string - name: unit1_sent dtype: string - name: unit2_sent dtype: string - name: dir dtype: string - name: orig_label dtype: string - name: orig_label.1 dtype: string splits: - name: train num_bytes: 8673004 num_examples: 8797 - name: validation num_bytes: 1156669 num_examples: 1285 - name: test num_bytes: 1147069 num_examples: 1248 download_size: 3135555 dataset_size: 10976742 - config_name: por.pdtb.tedm.conllu features: - name: id sequence: string - name: form sequence: string - name: lemma sequence: string - name: upos sequence: string - name: xpos sequence: string - name: feats sequence: string - name: head sequence: string - name: deprel sequence: string - name: deps sequence: string - name: misc sequence: string - name: doc_id dtype: string - name: mwe sequence: 'null' splits: - name: validation num_bytes: 275037 num_examples: 148 - name: test num_bytes: 534493 num_examples: 246 download_size: 129673 dataset_size: 809530 - config_name: por.pdtb.tedm.rels features: - name: doc dtype: string - name: unit1_toks dtype: string - name: unit2_toks dtype: string - name: unit1_txt dtype: string - name: unit2_txt dtype: string - name: s1_toks dtype: string - name: s2_toks dtype: string - name: unit1_sent dtype: string - name: unit2_sent dtype: string - name: dir dtype: string - name: orig_label dtype: string - name: label dtype: string splits: - name: validation num_bytes: 115009 num_examples: 190 - name: test num_bytes: 231698 num_examples: 364 download_size: 139249 dataset_size: 346707 - config_name: por.rst.cstn.conllu features: - name: id sequence: string - name: form sequence: string - name: lemma sequence: string - name: upos sequence: string - name: xpos sequence: string - name: feats sequence: string - name: head sequence: string - name: deprel sequence: string - name: deps sequence: string - name: misc sequence: string - name: doc_id dtype: string - name: mwe sequence: string splits: - name: train num_bytes: 4742077 num_examples: 1825 - name: validation num_bytes: 638016 num_examples: 257 - name: test num_bytes: 372732 num_examples: 139 download_size: 672351 dataset_size: 5752825 - config_name: por.rst.cstn.rels features: - name: doc dtype: string - name: unit1_toks dtype: string - name: unit2_toks dtype: string - name: unit1_txt dtype: string - name: unit2_txt dtype: string - name: s1_toks dtype: string - name: s2_toks dtype: string - name: unit1_sent dtype: string - name: unit2_sent dtype: string - name: dir dtype: string - name: orig_label dtype: string - name: label dtype: string splits: - name: train num_bytes: 2494318 num_examples: 4148 - name: validation num_bytes: 330132 num_examples: 573 - name: test num_bytes: 176246 num_examples: 272 download_size: 820097 dataset_size: 3000696 - config_name: rus.rst.rrt.conllu features: - name: id sequence: string - name: form sequence: string - name: lemma sequence: string - name: upos sequence: string - name: xpos sequence: string - name: feats sequence: string - name: head sequence: string - name: deprel sequence: string - name: deps sequence: string - name: misc sequence: string - name: doc_id dtype: string - name: mwe sequence: 'null' splits: - name: train num_bytes: 43706473 num_examples: 18932 - name: validation num_bytes: 4556506 num_examples: 2025 - name: test num_bytes: 4657786 num_examples: 2087 download_size: 0 dataset_size: 52920765 - config_name: rus.rst.rrt.rels features: - name: doc dtype: string - name: unit1_toks dtype: string - name: unit2_toks dtype: string - name: unit1_txt dtype: string - name: unit2_txt dtype: string - name: s1_toks dtype: string - name: s2_toks dtype: string - name: unit1_sent dtype: string - name: unit2_sent dtype: string - name: dir dtype: string - name: orig_label dtype: string - name: label dtype: string splits: - name: train num_bytes: 29448345 num_examples: 28868 - name: validation num_bytes: 2906149 num_examples: 2855 - name: test num_bytes: 2907619 num_examples: 2843 download_size: 0 dataset_size: 35262113 - config_name: spa.rst.rststb.conllu features: - name: id sequence: string - name: form sequence: string - name: lemma sequence: string - name: upos sequence: string - name: xpos sequence: string - name: feats sequence: string - name: head sequence: string - name: deprel sequence: string - name: deps sequence: string - name: misc sequence: string - name: doc_id dtype: string - name: mwe sequence: 'null' splits: - name: train num_bytes: 4145601 num_examples: 1547 - name: validation num_bytes: 723297 num_examples: 253 - name: test num_bytes: 780028 num_examples: 286 download_size: 659246 dataset_size: 5648926 - config_name: spa.rst.rststb.rels features: - name: doc dtype: string - name: unit1_toks dtype: string - name: unit2_toks dtype: string - name: unit1_txt dtype: string - name: unit2_txt dtype: string - name: s1_toks dtype: string - name: s2_toks dtype: string - name: unit1_sent dtype: string - name: unit2_sent dtype: string - name: dir dtype: string - name: orig_label dtype: string - name: label dtype: string splits: - name: train num_bytes: 1613935 num_examples: 2240 - name: validation num_bytes: 304667 num_examples: 383 - name: test num_bytes: 321080 num_examples: 426 download_size: 0 dataset_size: 2239682 - config_name: spa.rst.sctb.conllu features: - name: id sequence: string - name: form sequence: string - name: lemma sequence: string - name: upos sequence: string - name: xpos sequence: string - name: feats sequence: string - name: head sequence: string - name: deprel sequence: string - name: deps sequence: string - name: misc sequence: string - name: doc_id dtype: string - name: mwe sequence: 'null' splits: - name: train num_bytes: 971488 num_examples: 325 - name: validation num_bytes: 227843 num_examples: 75 - name: test num_bytes: 359479 num_examples: 113 download_size: 217463 dataset_size: 1558810 - config_name: spa.rst.sctb.rels features: - name: doc dtype: string - name: unit1_toks dtype: string - name: unit2_toks dtype: string - name: unit1_txt dtype: string - name: unit2_txt dtype: string - name: s1_toks dtype: string - name: s2_toks dtype: string - name: unit1_sent dtype: string - name: unit2_sent dtype: string - name: dir dtype: string - name: orig_label dtype: string - name: label dtype: string splits: - name: train num_bytes: 365530 num_examples: 439 - name: validation num_bytes: 78356 num_examples: 94 - name: test num_bytes: 140537 num_examples: 159 download_size: 224327 dataset_size: 584423 - config_name: tha.pdtb.tdtb.conllu features: - name: id sequence: string - name: form sequence: string - name: lemma sequence: string - name: upos sequence: string - name: xpos sequence: string - name: feats sequence: string - name: head sequence: string - name: deprel sequence: string - name: deps sequence: string - name: misc sequence: string - name: doc_id dtype: string - name: mwe sequence: 'null' splits: - name: train num_bytes: 17324320 num_examples: 5076 - name: validation num_bytes: 2356224 num_examples: 633 - name: test num_bytes: 2591736 num_examples: 825 download_size: 2219553 dataset_size: 22272280 - config_name: tha.pdtb.tdtb.rels features: - name: doc dtype: string - name: unit1_toks dtype: string - name: unit2_toks dtype: string - name: unit1_txt dtype: string - name: unit2_txt dtype: string - name: s1_toks dtype: string - name: s2_toks dtype: string - name: unit1_sent dtype: string - name: unit2_sent dtype: string - name: dir dtype: string - name: orig_label dtype: string - name: label dtype: string splits: - name: train num_bytes: 19110670 num_examples: 8278 - name: validation num_bytes: 2985114 num_examples: 1243 - name: test num_bytes: 2604469 num_examples: 1344 download_size: 4298420 dataset_size: 24700253 - config_name: tur.pdtb.tedm.conllu features: - name: id sequence: string - name: form sequence: string - name: lemma sequence: string - name: upos sequence: string - name: xpos sequence: string - name: feats sequence: string - name: head sequence: string - name: deprel sequence: string - name: deps sequence: string - name: misc sequence: string - name: doc_id dtype: string - name: mwe sequence: string splits: - name: validation num_bytes: 237251 num_examples: 141 - name: test num_bytes: 468186 num_examples: 269 download_size: 127186 dataset_size: 705437 - config_name: tur.pdtb.tedm.rels features: - name: doc dtype: string - name: unit1_toks dtype: string - name: unit2_toks dtype: string - name: unit1_txt dtype: string - name: unit2_txt dtype: string - name: s1_toks dtype: string - name: s2_toks dtype: string - name: unit1_sent dtype: string - name: unit2_sent dtype: string - name: dir dtype: string - name: orig_label dtype: string - name: label dtype: string splits: - name: validation num_bytes: 129156 num_examples: 213 - name: test num_bytes: 218205 num_examples: 364 download_size: 137136 dataset_size: 347361 - config_name: zho.dep.scidtb.conllu features: - name: id sequence: string - name: form sequence: string - name: lemma sequence: string - name: upos sequence: string - name: xpos sequence: string - name: feats sequence: string - name: head sequence: string - name: deprel sequence: string - name: deps sequence: string - name: misc sequence: string - name: doc_id dtype: string - name: mwe sequence: 'null' splits: - name: train num_bytes: 827143 num_examples: 308 - name: validation num_bytes: 282227 num_examples: 103 - name: test num_bytes: 264697 num_examples: 89 download_size: 0 dataset_size: 1374067 - config_name: zho.dep.scidtb.rels features: - name: doc dtype: string - name: unit1_toks dtype: string - name: unit2_toks dtype: string - name: unit1_txt dtype: string - name: unit2_txt dtype: string - name: s1_toks dtype: string - name: s2_toks dtype: string - name: unit1_sent dtype: string - name: unit2_sent dtype: string - name: dir dtype: string - name: orig_label dtype: string - name: label dtype: string splits: - name: train num_bytes: 628861 num_examples: 802 - name: validation num_bytes: 228839 num_examples: 281 - name: test num_bytes: 181790 num_examples: 215 download_size: 0 dataset_size: 1039490 - config_name: zho.rst.gcdt.conllu features: - name: id sequence: string - name: form sequence: string - name: lemma sequence: string - name: upos sequence: string - name: xpos sequence: string - name: feats sequence: string - name: head sequence: string - name: deprel sequence: string - name: deps sequence: string - name: misc sequence: string - name: doc_id dtype: string - name: mwe sequence: 'null' splits: - name: train num_bytes: 3466038 num_examples: 2026 - name: validation num_bytes: 554602 num_examples: 331 - name: test num_bytes: 557377 num_examples: 335 download_size: 712103 dataset_size: 4578017 - config_name: zho.rst.gcdt.rels features: - name: doc dtype: string - name: unit1_toks dtype: string - name: unit2_toks dtype: string - name: unit1_txt dtype: string - name: unit2_txt dtype: string - name: s1_toks dtype: string - name: s2_toks dtype: string - name: unit1_sent dtype: string - name: unit2_sent dtype: string - name: dir dtype: string - name: orig_label dtype: string - name: label dtype: string splits: - name: train num_bytes: 4067285 num_examples: 6454 - name: validation num_bytes: 627252 num_examples: 1006 - name: test num_bytes: 615083 num_examples: 953 download_size: 1055924 dataset_size: 5309620 - config_name: zho.rst.sctb.conllu features: - name: id sequence: string - name: form sequence: string - name: lemma sequence: string - name: upos sequence: string - name: xpos sequence: string - name: feats sequence: string - name: head sequence: string - name: deprel sequence: string - name: deps sequence: string - name: misc sequence: string - name: doc_id dtype: string - name: mwe sequence: 'null' splits: - name: train num_bytes: 713064 num_examples: 360 - name: validation num_bytes: 166616 num_examples: 85 - name: test num_bytes: 264281 num_examples: 132 download_size: 195688 dataset_size: 1143961 - config_name: zho.rst.sctb.rels features: - name: doc dtype: string - name: unit1_toks dtype: string - name: unit2_toks dtype: string - name: unit1_txt dtype: string - name: unit2_txt dtype: string - name: s1_toks dtype: string - name: s2_toks dtype: string - name: unit1_sent dtype: string - name: unit2_sent dtype: string - name: dir dtype: string - name: orig_label dtype: string - name: label dtype: string splits: - name: train num_bytes: 318972 num_examples: 439 - name: validation num_bytes: 68417 num_examples: 94 - name: test num_bytes: 117034 num_examples: 159 download_size: 208819 dataset_size: 504423 configs: - config_name: deu.rst.pcc.conllu data_files: - split: train path: deu.rst.pcc.conllu/train-* - split: validation path: deu.rst.pcc.conllu/validation-* - split: test path: deu.rst.pcc.conllu/test-* - config_name: deu.rst.pcc.rels data_files: - split: train path: deu.rst.pcc.rels/train-* - split: validation path: deu.rst.pcc.rels/validation-* - split: test path: deu.rst.pcc.rels/test-* - config_name: eng.dep.covdtb.conllu data_files: - split: validation path: eng.dep.covdtb.conllu/validation-* - split: test path: eng.dep.covdtb.conllu/test-* - config_name: eng.dep.covdtb.rels data_files: - split: validation path: eng.dep.covdtb.rels/validation-* - split: test path: eng.dep.covdtb.rels/test-* - config_name: eng.dep.scidtb.conllu data_files: - split: train path: eng.dep.scidtb.conllu/train-* - split: validation path: eng.dep.scidtb.conllu/validation-* - split: test path: eng.dep.scidtb.conllu/test-* - config_name: eng.dep.scidtb.rels data_files: - split: train path: eng.dep.scidtb.rels/train-* - split: validation path: eng.dep.scidtb.rels/validation-* - split: test path: eng.dep.scidtb.rels/test-* - config_name: eng.pdtb.tedm.conllu data_files: - split: validation path: eng.pdtb.tedm.conllu/validation-* - split: test path: eng.pdtb.tedm.conllu/test-* - config_name: eng.pdtb.tedm.rels data_files: - split: validation path: eng.pdtb.tedm.rels/validation-* - split: test path: eng.pdtb.tedm.rels/test-* - config_name: eng.rst.gentle.conllu data_files: - split: test path: eng.rst.gentle.conllu/test-* - config_name: eng.rst.gentle.rels data_files: - split: test path: eng.rst.gentle.rels/test-* - config_name: eng.sdrt.stac.conllu data_files: - split: train path: eng.sdrt.stac.conllu/train-* - split: validation path: eng.sdrt.stac.conllu/validation-* - split: test path: eng.sdrt.stac.conllu/test-* - config_name: eng.sdrt.stac.rels data_files: - split: train path: eng.sdrt.stac.rels/train-* - split: validation path: eng.sdrt.stac.rels/validation-* - split: test path: eng.sdrt.stac.rels/test-* - config_name: eus.rst.ert.conllu data_files: - split: train path: eus.rst.ert.conllu/train-* - split: validation path: eus.rst.ert.conllu/validation-* - split: test path: eus.rst.ert.conllu/test-* - config_name: eus.rst.ert.rels data_files: - split: train path: eus.rst.ert.rels/train-* - split: validation path: eus.rst.ert.rels/validation-* - split: test path: eus.rst.ert.rels/test-* - config_name: fas.rst.prstc.conllu data_files: - split: train path: fas.rst.prstc.conllu/train-* - split: validation path: fas.rst.prstc.conllu/validation-* - split: test path: fas.rst.prstc.conllu/test-* - config_name: fas.rst.prstc.rels data_files: - split: train path: fas.rst.prstc.rels/train-* - split: validation path: fas.rst.prstc.rels/validation-* - split: test path: fas.rst.prstc.rels/test-* - config_name: fra.sdrt.annodis.conllu data_files: - split: train path: fra.sdrt.annodis.conllu/train-* - split: validation path: fra.sdrt.annodis.conllu/validation-* - split: test path: fra.sdrt.annodis.conllu/test-* - config_name: fra.sdrt.annodis.rels data_files: - split: train path: fra.sdrt.annodis.rels/train-* - split: validation path: fra.sdrt.annodis.rels/validation-* - split: test path: fra.sdrt.annodis.rels/test-* - config_name: ita.pdtb.luna.conllu data_files: - split: train path: ita.pdtb.luna.conllu/train-* - split: validation path: ita.pdtb.luna.conllu/validation-* - split: test path: ita.pdtb.luna.conllu/test-* - config_name: ita.pdtb.luna.rels data_files: - split: train path: ita.pdtb.luna.rels/train-* - split: validation path: ita.pdtb.luna.rels/validation-* - split: test path: ita.pdtb.luna.rels/test-* - config_name: nld.rst.nldt.conllu data_files: - split: train path: nld.rst.nldt.conllu/train-* - split: validation path: nld.rst.nldt.conllu/validation-* - split: test path: nld.rst.nldt.conllu/test-* - config_name: nld.rst.nldt.rels data_files: - split: train path: nld.rst.nldt.rels/train-* - split: validation path: nld.rst.nldt.rels/validation-* - split: test path: nld.rst.nldt.rels/test-* - config_name: por.pdtb.crpc.conllu data_files: - split: train path: por.pdtb.crpc.conllu/train-* - split: validation path: por.pdtb.crpc.conllu/validation-* - split: test path: por.pdtb.crpc.conllu/test-* - config_name: por.pdtb.crpc.rels data_files: - split: train path: por.pdtb.crpc.rels/train-* - split: validation path: por.pdtb.crpc.rels/validation-* - split: test path: por.pdtb.crpc.rels/test-* - config_name: por.pdtb.tedm.conllu data_files: - split: validation path: por.pdtb.tedm.conllu/validation-* - split: test path: por.pdtb.tedm.conllu/test-* - config_name: por.pdtb.tedm.rels data_files: - split: validation path: por.pdtb.tedm.rels/validation-* - split: test path: por.pdtb.tedm.rels/test-* - config_name: por.rst.cstn.conllu data_files: - split: train path: por.rst.cstn.conllu/train-* - split: validation path: por.rst.cstn.conllu/validation-* - split: test path: por.rst.cstn.conllu/test-* - config_name: por.rst.cstn.rels data_files: - split: train path: por.rst.cstn.rels/train-* - split: validation path: por.rst.cstn.rels/validation-* - split: test path: por.rst.cstn.rels/test-* - config_name: rus.rst.rrt.conllu data_files: - split: train path: rus.rst.rrt.conllu/train-* - split: validation path: rus.rst.rrt.conllu/validation-* - split: test path: rus.rst.rrt.conllu/test-* - config_name: rus.rst.rrt.rels data_files: - split: train path: rus.rst.rrt.rels/train-* - split: validation path: rus.rst.rrt.rels/validation-* - split: test path: rus.rst.rrt.rels/test-* - config_name: spa.rst.rststb.conllu data_files: - split: train path: spa.rst.rststb.conllu/train-* - split: validation path: spa.rst.rststb.conllu/validation-* - split: test path: spa.rst.rststb.conllu/test-* - config_name: spa.rst.rststb.rels data_files: - split: train path: spa.rst.rststb.rels/train-* - split: validation path: spa.rst.rststb.rels/validation-* - split: test path: spa.rst.rststb.rels/test-* - config_name: spa.rst.sctb.conllu data_files: - split: train path: spa.rst.sctb.conllu/train-* - split: validation path: spa.rst.sctb.conllu/validation-* - split: test path: spa.rst.sctb.conllu/test-* - config_name: spa.rst.sctb.rels data_files: - split: train path: spa.rst.sctb.rels/train-* - split: validation path: spa.rst.sctb.rels/validation-* - split: test path: spa.rst.sctb.rels/test-* - config_name: tha.pdtb.tdtb.conllu data_files: - split: train path: tha.pdtb.tdtb.conllu/train-* - split: validation path: tha.pdtb.tdtb.conllu/validation-* - split: test path: tha.pdtb.tdtb.conllu/test-* - config_name: tha.pdtb.tdtb.rels data_files: - split: train path: tha.pdtb.tdtb.rels/train-* - split: validation path: tha.pdtb.tdtb.rels/validation-* - split: test path: tha.pdtb.tdtb.rels/test-* - config_name: tur.pdtb.tedm.conllu data_files: - split: validation path: tur.pdtb.tedm.conllu/validation-* - split: test path: tur.pdtb.tedm.conllu/test-* - config_name: tur.pdtb.tedm.rels data_files: - split: validation path: tur.pdtb.tedm.rels/validation-* - split: test path: tur.pdtb.tedm.rels/test-* - config_name: zho.dep.scidtb.conllu data_files: - split: train path: zho.dep.scidtb.conllu/train-* - split: validation path: zho.dep.scidtb.conllu/validation-* - split: test path: zho.dep.scidtb.conllu/test-* - config_name: zho.dep.scidtb.rels data_files: - split: train path: zho.dep.scidtb.rels/train-* - split: validation path: zho.dep.scidtb.rels/validation-* - split: test path: zho.dep.scidtb.rels/test-* - config_name: zho.rst.gcdt.conllu data_files: - split: train path: zho.rst.gcdt.conllu/train-* - split: validation path: zho.rst.gcdt.conllu/validation-* - split: test path: zho.rst.gcdt.conllu/test-* - config_name: zho.rst.gcdt.rels data_files: - split: train path: zho.rst.gcdt.rels/train-* - split: validation path: zho.rst.gcdt.rels/validation-* - split: test path: zho.rst.gcdt.rels/test-* - config_name: zho.rst.sctb.conllu data_files: - split: train path: zho.rst.sctb.conllu/train-* - split: validation path: zho.rst.sctb.conllu/validation-* - split: test path: zho.rst.sctb.conllu/test-* - config_name: zho.rst.sctb.rels data_files: - split: train path: zho.rst.sctb.rels/train-* - split: validation path: zho.rst.sctb.rels/validation-* - split: test path: zho.rst.sctb.rels/test-* --- ## Disrpt is a multilingual, multi-framework unified discourse analysis benchmark. It unifies discourse relation classification tasks (.rels) and discourse segmentation (.connlu) for many languages. ⚠️ This repo only contains the disrpt dataset when the underlying data is permissively licensed. Some datasets rely on corpora like the PTB. To load these datasets, run the following: `pip install disrpt-utils` Then ```python from disrpt_utils import load_dataset corpora_paths={ #TODO Input your own paths "tur.pdtb.tdb": "/mnt/nfs_share_magnet2/dsileo/data/disrpt/TDB1.0/raw/", "eng.rst.rstdt": "/mnt/nfs_share_magnet2/dsileo/data/disrpt/RSTtrees-WSJ-main-1.0-20231128T125253Z-001/" } config_name="eng.rst.rstdt" load_dataset(config_name,'conllu',corpora_paths=corpora_paths) ``` Otherwise, most datasets are directly accessible: ```python from datasets import load_dataset load_dataset('multilingual-discourse-hub/disrpt','eng.dep.scidtb.rels') ``` # Citations ``` @inproceedings{braud2024disrpt, title={{DISRPT}: A Multilingual, Multi-domain, Cross-framework Benchmark for Discourse Processing}, author={Braud, Chloé and Zeldes, Amir and Rivière, Laura and Liu, Yang Janet and Muller, Philippe and Sileo, Damien and Aoyama, Tatsuya}, booktitle={Proceedings of LREC-COLING 2024}, year={2024} } ``` ## Underline datasets: https://huggingface.co/datasets/multilingual-discourse-hub/disrpt/raw/main/bibtex.json
ttn1410/Consumer_smr
ttn1410
2025-04-30T12:29:22Z
0
0
[ "size_categories:10K<n<100K", "format:parquet", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us" ]
[]
2025-04-29T17:19:42Z
null
--- dataset_info: features: - name: reports dtype: string - name: labels dtype: string splits: - name: train num_bytes: 89754791 num_examples: 37290 download_size: 5531728 dataset_size: 89754791 configs: - config_name: default data_files: - split: train path: data/train-* ---
shylee/eval_DP_so100_gauze_temp
shylee
2025-04-30T12:27:24Z
0
0
[ "task_categories:robotics", "license:apache-2.0", "region:us", "LeRobot", "tutorial" ]
[ "robotics" ]
2025-04-30T12:27:18Z
null
--- license: apache-2.0 task_categories: - robotics tags: - LeRobot - tutorial configs: - config_name: default data_files: data/*/*.parquet --- This dataset was created using [LeRobot](https://github.com/huggingface/lerobot). ## Dataset Description - **Homepage:** [More Information Needed] - **Paper:** [More Information Needed] - **License:** apache-2.0 ## Dataset Structure [meta/info.json](meta/info.json): ```json { "codebase_version": "v2.1", "robot_type": "so100", "total_episodes": 1, "total_frames": 633, "total_tasks": 1, "total_videos": 3, "total_chunks": 1, "chunks_size": 1000, "fps": 30, "splits": { "train": "0:1" }, "data_path": "data/chunk-{episode_chunk:03d}/episode_{episode_index:06d}.parquet", "video_path": "videos/chunk-{episode_chunk:03d}/{video_key}/episode_{episode_index:06d}.mp4", "features": { "action": { "dtype": "float32", "shape": [ 6 ], "names": [ "main_shoulder_pan", "main_shoulder_lift", "main_elbow_flex", "main_wrist_flex", "main_wrist_roll", "main_gripper" ] }, "observation.state": { "dtype": "float32", "shape": [ 6 ], "names": [ "main_shoulder_pan", "main_shoulder_lift", "main_elbow_flex", "main_wrist_flex", "main_wrist_roll", "main_gripper" ] }, "observation.images.FrontCam": { "dtype": "video", "shape": [ 480, 640, 3 ], "names": [ "height", "width", "channels" ], "info": { "video.fps": 30.0, "video.height": 480, "video.width": 640, "video.channels": 3, "video.codec": "av1", "video.pix_fmt": "yuv420p", "video.is_depth_map": false, "has_audio": false } }, "observation.images.TopCam": { "dtype": "video", "shape": [ 480, 640, 3 ], "names": [ "height", "width", "channels" ], "info": { "video.fps": 30.0, "video.height": 480, "video.width": 640, "video.channels": 3, "video.codec": "av1", "video.pix_fmt": "yuv420p", "video.is_depth_map": false, "has_audio": false } }, "observation.images.WristCam": { "dtype": "video", "shape": [ 480, 640, 3 ], "names": [ "height", "width", "channels" ], "info": { "video.fps": 30.0, "video.height": 480, "video.width": 640, "video.channels": 3, "video.codec": "av1", "video.pix_fmt": "yuv420p", "video.is_depth_map": false, "has_audio": false } }, "timestamp": { "dtype": "float32", "shape": [ 1 ], "names": null }, "frame_index": { "dtype": "int64", "shape": [ 1 ], "names": null }, "episode_index": { "dtype": "int64", "shape": [ 1 ], "names": null }, "index": { "dtype": "int64", "shape": [ 1 ], "names": null }, "task_index": { "dtype": "int64", "shape": [ 1 ], "names": null } } } ``` ## Citation **BibTeX:** ```bibtex [More Information Needed] ```
EYEDOL/mozilla_commonvoice_naijaYoruba1_preprocessed_train_batch_3
EYEDOL
2025-04-30T12:18:52Z
0
0
[ "region:us" ]
[]
2025-04-30T12:15:56Z
null
--- dataset_info: features: - name: audio dtype: audio: sampling_rate: 16000 - name: text dtype: string - name: input_length dtype: int64 - name: input_features sequence: sequence: float32 - name: labels sequence: int64 - name: labels_length dtype: int64 splits: - name: train num_bytes: 13926597095.875 num_examples: 12961 download_size: 3081861566 dataset_size: 13926597095.875 configs: - config_name: default data_files: - split: train path: data/train-* ---
HungVu2003/opt-350m_beta_1.0_alpha_0.6_num-company_3_dataset_0_for_gen_3
HungVu2003
2025-04-30T11:54:07Z
0
0
[ "size_categories:10K<n<100K", "format:parquet", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us" ]
[]
2025-04-30T11:54:06Z
null
--- dataset_info: features: - name: question dtype: string splits: - name: train num_bytes: 5015021 num_examples: 12500 download_size: 2133264 dataset_size: 5015021 configs: - config_name: default data_files: - split: train path: data/train-* ---
shylee/eval_DP_so100_gauze_IMAGENET_1e-5_ckpt001500
shylee
2025-04-30T11:41:42Z
0
0
[ "task_categories:robotics", "license:apache-2.0", "size_categories:n<1K", "format:parquet", "modality:tabular", "modality:timeseries", "modality:video", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us", "LeRobot", "tutorial" ]
[ "robotics" ]
2025-04-30T11:41:37Z
null
--- license: apache-2.0 task_categories: - robotics tags: - LeRobot - tutorial configs: - config_name: default data_files: data/*/*.parquet --- This dataset was created using [LeRobot](https://github.com/huggingface/lerobot). ## Dataset Description - **Homepage:** [More Information Needed] - **Paper:** [More Information Needed] - **License:** apache-2.0 ## Dataset Structure [meta/info.json](meta/info.json): ```json { "codebase_version": "v2.1", "robot_type": "so100", "total_episodes": 1, "total_frames": 8, "total_tasks": 1, "total_videos": 3, "total_chunks": 1, "chunks_size": 1000, "fps": 30, "splits": { "train": "0:1" }, "data_path": "data/chunk-{episode_chunk:03d}/episode_{episode_index:06d}.parquet", "video_path": "videos/chunk-{episode_chunk:03d}/{video_key}/episode_{episode_index:06d}.mp4", "features": { "action": { "dtype": "float32", "shape": [ 6 ], "names": [ "main_shoulder_pan", "main_shoulder_lift", "main_elbow_flex", "main_wrist_flex", "main_wrist_roll", "main_gripper" ] }, "observation.state": { "dtype": "float32", "shape": [ 6 ], "names": [ "main_shoulder_pan", "main_shoulder_lift", "main_elbow_flex", "main_wrist_flex", "main_wrist_roll", "main_gripper" ] }, "observation.images.FrontCam": { "dtype": "video", "shape": [ 480, 640, 3 ], "names": [ "height", "width", "channels" ], "info": { "video.fps": 30.0, "video.height": 480, "video.width": 640, "video.channels": 3, "video.codec": "av1", "video.pix_fmt": "yuv420p", "video.is_depth_map": false, "has_audio": false } }, "observation.images.TopCam": { "dtype": "video", "shape": [ 480, 640, 3 ], "names": [ "height", "width", "channels" ], "info": { "video.fps": 30.0, "video.height": 480, "video.width": 640, "video.channels": 3, "video.codec": "av1", "video.pix_fmt": "yuv420p", "video.is_depth_map": false, "has_audio": false } }, "observation.images.WristCam": { "dtype": "video", "shape": [ 480, 640, 3 ], "names": [ "height", "width", "channels" ], "info": { "video.fps": 30.0, "video.height": 480, "video.width": 640, "video.channels": 3, "video.codec": "av1", "video.pix_fmt": "yuv420p", "video.is_depth_map": false, "has_audio": false } }, "timestamp": { "dtype": "float32", "shape": [ 1 ], "names": null }, "frame_index": { "dtype": "int64", "shape": [ 1 ], "names": null }, "episode_index": { "dtype": "int64", "shape": [ 1 ], "names": null }, "index": { "dtype": "int64", "shape": [ 1 ], "names": null }, "task_index": { "dtype": "int64", "shape": [ 1 ], "names": null } } } ``` ## Citation **BibTeX:** ```bibtex [More Information Needed] ```
dgambettaphd/D_llm2_gen5_run0_X_doc1000_synt64_tot128_lr5em5_SYNLAST
dgambettaphd
2025-04-30T11:13:19Z
0
0
[ "size_categories:10K<n<100K", "format:parquet", "modality:tabular", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us" ]
[]
2025-04-30T11:12:13Z
null
--- dataset_info: features: - name: id_doc dtype: int64 - name: text dtype: string - name: dataset dtype: string - name: gen dtype: int64 - name: synt dtype: int64 - name: MPP dtype: float64 splits: - name: train num_bytes: 11848576 num_examples: 21000 download_size: 7145287 dataset_size: 11848576 configs: - config_name: default data_files: - split: train path: data/train-* ---
HungVu2003/opt-350m_beta_0.5_alpha_0.4_num-company_3_dataset_1_for_gen_8
HungVu2003
2025-04-30T10:52:00Z
0
0
[ "size_categories:10K<n<100K", "format:parquet", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us" ]
[]
2025-04-30T10:51:59Z
null
--- dataset_info: features: - name: question dtype: string splits: - name: train num_bytes: 2821591 num_examples: 12498 download_size: 1529759 dataset_size: 2821591 configs: - config_name: default data_files: - split: train path: data/train-* ---
wallenx/roarmm3water2
wallenx
2025-04-30T10:51:26Z
0
0
[ "task_categories:robotics", "license:apache-2.0", "size_categories:n<1K", "format:parquet", "modality:tabular", "modality:timeseries", "modality:video", "library:datasets", "library:dask", "library:mlcroissant", "library:polars", "region:us", "LeRobot" ]
[ "robotics" ]
2025-04-30T10:51:20Z
null
--- license: apache-2.0 task_categories: - robotics tags: - LeRobot configs: - config_name: default data_files: data/*/*.parquet --- This dataset was created using [LeRobot](https://github.com/huggingface/lerobot). ## Dataset Description - **Homepage:** [More Information Needed] - **Paper:** [More Information Needed] - **License:** apache-2.0 ## Dataset Structure [meta/info.json](meta/info.json): ```json { "codebase_version": "v2.1", "robot_type": "roarm_m3", "total_episodes": 2, "total_frames": 90, "total_tasks": 1, "total_videos": 6, "total_chunks": 1, "chunks_size": 1000, "fps": 30, "splits": { "train": "0:2" }, "data_path": "data/chunk-{episode_chunk:03d}/episode_{episode_index:06d}.parquet", "video_path": "videos/chunk-{episode_chunk:03d}/{video_key}/episode_{episode_index:06d}.mp4", "features": { "action": { "dtype": "float32", "shape": [ 6 ], "names": [ "1", "2,3", "4", "5", "6" ] }, "observation.state": { "dtype": "float32", "shape": [ 6 ], "names": [ "1", "2,3", "4", "5", "6" ] }, "observation.images.laptop": { "dtype": "video", "shape": [ 480, 640, 3 ], "names": [ "height", "width", "channels" ], "info": { "video.fps": 30.0, "video.height": 480, "video.width": 640, "video.channels": 3, "video.codec": "av1", "video.pix_fmt": "yuv420p", "video.is_depth_map": false, "has_audio": false } }, "observation.images.phone": { "dtype": "video", "shape": [ 480, 640, 3 ], "names": [ "height", "width", "channels" ], "info": { "video.fps": 30.0, "video.height": 480, "video.width": 640, "video.channels": 3, "video.codec": "av1", "video.pix_fmt": "yuv420p", "video.is_depth_map": false, "has_audio": false } }, "observation.images.rpi1": { "dtype": "video", "shape": [ 480, 640, 3 ], "names": [ "height", "width", "channels" ], "info": { "video.fps": 30.0, "video.height": 480, "video.width": 640, "video.channels": 3, "video.codec": "av1", "video.pix_fmt": "yuv420p", "video.is_depth_map": false, "has_audio": false } }, "timestamp": { "dtype": "float32", "shape": [ 1 ], "names": null }, "frame_index": { "dtype": "int64", "shape": [ 1 ], "names": null }, "episode_index": { "dtype": "int64", "shape": [ 1 ], "names": null }, "index": { "dtype": "int64", "shape": [ 1 ], "names": null }, "task_index": { "dtype": "int64", "shape": [ 1 ], "names": null } } } ``` ## Citation **BibTeX:** ```bibtex [More Information Needed] ```
occiglot/arcX
occiglot
2025-04-30T10:37:48Z
101
0
[ "size_categories:10K<n<100K", "format:parquet", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us" ]
[]
2024-12-06T16:13:04Z
null
--- dataset_info: - config_name: bg features: - name: id dtype: string - name: question dtype: string - name: choices sequence: string - name: answerKey dtype: string splits: - name: test num_bytes: 619464 num_examples: 1172 download_size: 294059 dataset_size: 619464 - config_name: cs features: - name: id dtype: string - name: question dtype: string - name: choices sequence: string - name: answerKey dtype: string splits: - name: test num_bytes: 360257 num_examples: 1172 download_size: 225370 dataset_size: 360257 - config_name: da features: - name: id dtype: string - name: question dtype: string - name: choices sequence: string - name: answerKey dtype: string splits: - name: test num_bytes: 353054 num_examples: 1172 download_size: 209749 dataset_size: 353054 - config_name: de features: - name: id dtype: string - name: question dtype: string - name: choices sequence: string - name: answerKey dtype: string splits: - name: test num_bytes: 395506 num_examples: 1172 download_size: 231285 dataset_size: 395506 - config_name: el features: - name: id dtype: string - name: question dtype: string - name: choices sequence: string - name: answerKey dtype: string splits: - name: test num_bytes: 670991 num_examples: 1172 download_size: 320473 dataset_size: 670991 - config_name: en features: - name: id dtype: string - name: question dtype: string - name: choices sequence: string - name: answerKey dtype: string splits: - name: test num_bytes: 347388 num_examples: 1172 download_size: 202545 dataset_size: 347388 - config_name: es features: - name: id dtype: string - name: question dtype: string - name: choices sequence: string - name: answerKey dtype: string splits: - name: test num_bytes: 391852 num_examples: 1172 download_size: 222689 dataset_size: 391852 - config_name: et features: - name: id dtype: string - name: question dtype: string - name: choices sequence: string - name: answerKey dtype: string splits: - name: test num_bytes: 339448 num_examples: 1172 download_size: 210365 dataset_size: 339448 - config_name: fi features: - name: id dtype: string - name: question dtype: string - name: choices sequence: string - name: answerKey dtype: string splits: - name: test num_bytes: 360068 num_examples: 1172 download_size: 219314 dataset_size: 360068 - config_name: fr features: - name: id dtype: string - name: question dtype: string - name: choices sequence: string - name: answerKey dtype: string splits: - name: test num_bytes: 413039 num_examples: 1172 download_size: 233168 dataset_size: 413039 - config_name: hu features: - name: id dtype: string - name: question dtype: string - name: choices sequence: string - name: answerKey dtype: string splits: - name: test num_bytes: 388205 num_examples: 1172 download_size: 235036 dataset_size: 388205 - config_name: it features: - name: id dtype: string - name: question dtype: string - name: choices sequence: string - name: answerKey dtype: string splits: - name: test num_bytes: 388648 num_examples: 1172 download_size: 226121 dataset_size: 388648 - config_name: lt features: - name: id dtype: string - name: question dtype: string - name: choices sequence: string - name: answerKey dtype: string splits: - name: test num_bytes: 360908 num_examples: 1172 download_size: 224586 dataset_size: 360908 - config_name: lv features: - name: id dtype: string - name: question dtype: string - name: choices sequence: string - name: answerKey dtype: string splits: - name: test num_bytes: 359806 num_examples: 1172 download_size: 221360 dataset_size: 359806 - config_name: nb features: - name: id dtype: string - name: choices sequence: string - name: question dtype: string - name: answerKey dtype: string splits: - name: test num_bytes: 89141.39393939394 num_examples: 294 download_size: 56159 dataset_size: 89141.39393939394 - config_name: nl features: - name: id dtype: string - name: question dtype: string - name: choices sequence: string - name: answerKey dtype: string splits: - name: test num_bytes: 377643 num_examples: 1172 download_size: 217672 dataset_size: 377643 - config_name: pl features: - name: id dtype: string - name: question dtype: string - name: choices sequence: string - name: answerKey dtype: string splits: - name: test num_bytes: 375116 num_examples: 1172 download_size: 231284 dataset_size: 375116 - config_name: pt-pt features: - name: id dtype: string - name: question dtype: string - name: choices sequence: string - name: answerKey dtype: string splits: - name: test num_bytes: 383332 num_examples: 1172 download_size: 221692 dataset_size: 383332 - config_name: ro features: - name: id dtype: string - name: question dtype: string - name: choices sequence: string - name: answerKey dtype: string splits: - name: test num_bytes: 395131 num_examples: 1172 download_size: 231255 dataset_size: 395131 - config_name: sk features: - name: id dtype: string - name: question dtype: string - name: choices sequence: string - name: answerKey dtype: string splits: - name: test num_bytes: 360344 num_examples: 1172 download_size: 227719 dataset_size: 360344 - config_name: sl features: - name: id dtype: string - name: question dtype: string - name: choices sequence: string - name: answerKey dtype: string splits: - name: test num_bytes: 340063 num_examples: 1172 download_size: 213889 dataset_size: 340063 - config_name: sv features: - name: id dtype: string - name: question dtype: string - name: choices sequence: string - name: answerKey dtype: string splits: - name: test num_bytes: 354052 num_examples: 1172 download_size: 209831 dataset_size: 354052 - config_name: tr features: - name: id dtype: string - name: question dtype: string - name: choices sequence: string - name: answerKey dtype: string splits: - name: test num_bytes: 374834 num_examples: 1171 download_size: 221007 dataset_size: 374834 - config_name: uk features: - name: id dtype: string - name: choices sequence: string - name: question dtype: string - name: answerKey dtype: string splits: - name: test num_bytes: 149498.30067567568 num_examples: 293 download_size: 78400 dataset_size: 149498.30067567568 configs: - config_name: bg data_files: - split: test path: bg/test-* - config_name: cs data_files: - split: test path: cs/test-* - config_name: da data_files: - split: test path: da/test-* - config_name: de data_files: - split: test path: de/test-* - config_name: el data_files: - split: test path: el/test-* - config_name: en data_files: - split: test path: en/test-* - config_name: es data_files: - split: test path: es/test-* - config_name: et data_files: - split: test path: et/test-* - config_name: fi data_files: - split: test path: fi/test-* - config_name: fr data_files: - split: test path: fr/test-* - config_name: hu data_files: - split: test path: hu/test-* - config_name: it data_files: - split: test path: it/test-* - config_name: lt data_files: - split: test path: lt/test-* - config_name: lv data_files: - split: test path: lv/test-* - config_name: nb data_files: - split: test path: nb/test-* - config_name: nl data_files: - split: test path: nl/test-* - config_name: pl data_files: - split: test path: pl/test-* - config_name: pt-pt data_files: - split: test path: pt-pt/test-* - config_name: ro data_files: - split: test path: ro/test-* - config_name: sk data_files: - split: test path: sk/test-* - config_name: sl data_files: - split: test path: sl/test-* - config_name: sv data_files: - split: test path: sv/test-* - config_name: tr data_files: - split: test path: tr/test-* - config_name: uk data_files: - split: test path: uk/test-* ---
SimuGPT/4bus-system-creation
SimuGPT
2025-04-30T10:29:56Z
0
0
[ "size_categories:1K<n<10K", "format:parquet", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us" ]
[]
2025-04-30T09:54:52Z
null
--- dataset_info: features: - name: init_code dtype: string - name: question dtype: string - name: answer dtype: string splits: - name: train num_bytes: 1341422.4 num_examples: 800 - name: test num_bytes: 335355.6 num_examples: 200 download_size: 136589 dataset_size: 1676778.0 configs: - config_name: default data_files: - split: train path: data/train-* - split: test path: data/test-* ---
SDSB/merged_deduped_dapo_or1_dataset
SDSB
2025-04-30T10:29:48Z
0
0
[ "size_categories:100K<n<1M", "format:parquet", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us" ]
[]
2025-04-30T10:08:40Z
null
--- dataset_info: features: - name: problem dtype: string - name: answer dtype: string splits: - name: train num_bytes: 33662063 num_examples: 117192 download_size: 19609522 dataset_size: 33662063 configs: - config_name: default data_files: - split: train path: data/train-* ---
twinkle-ai/tw-math-reasoning-2k
twinkle-ai
2025-04-30T10:13:17Z
4
2
[ "task_categories:text-generation", "language:zh", "language:en", "license:mit", "size_categories:1K<n<10K", "format:parquet", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us", "Taiwan", "R.O.C", "zh-tw", "math", "cot", "twinkle.ai" ]
[ "text-generation" ]
2025-04-30T02:39:48Z
2
--- dataset_info: features: - name: problem dtype: string - name: level dtype: string - name: type dtype: string - name: solution dtype: string - name: subset dtype: string - name: split dtype: string - name: model dtype: string - name: problem_zhtw dtype: string - name: think dtype: string - name: answer dtype: string - name: messages list: - name: content dtype: string - name: role dtype: string splits: - name: train num_bytes: 18644315 num_examples: 2000 download_size: 9178902 dataset_size: 18644315 configs: - config_name: default data_files: - split: train path: data/train-* license: mit task_categories: - text-generation language: - zh - en tags: - Taiwan - R.O.C - zh-tw - math - cot - twinkle.ai pretty_name: Traditional Chinese Mathematical Reasoning Dataset size_categories: - 1K<n<10K --- # Dataset Card for tw-math-reasoning-2k ![image/png](https://cdn-uploads.huggingface.co/production/uploads/618dc56cbc345ca7bf95f3cd/ZPhbmAQhQlQFGEieDlg6l.png) **tw-math-reasoning-2k** 是一個繁體中文數學語言資料集,從 [HuggingFaceH4/MATH](https://huggingface.co/datasets/HuggingFaceH4/MATH) 英文數學題庫中精選 2,000 題,並透過 [perplexity-ai/r1-1776](https://huggingface.co/perplexity-ai/r1-1776) 模型以繁體中文重新生成具邏輯性且詳盡的解題過程與最終答案。此資料集可作為訓練或評估繁體中文數學推理模型的高品質參考語料。 ## Dataset Details ### Dataset Description **tw-math-reasoning-2k** 是一個繁體中文數學語言資料集,旨在提供高品質的解題語料以支援中文數學推理模型的訓練與評估。此資料集從 [HuggingFaceH4/MATH](https://huggingface.co/datasets/HuggingFaceH4/MATH) 英文數學題庫中精選 2,000 題,涵蓋代數、幾何、機率統計等各類題型,並確保題目類型分佈均衡。 所有題目皆經由 [perplexity-ai/r1-1776](https://huggingface.co/perplexity-ai/r1-1776) 模型重新生成,透過多輪提示設計誘導模型產出繁體中文的詳細解題步驟與邏輯推演,最終形成完整的繁中答案。為確保語料品質,資料在生成後進行清洗與過濾,剔除明顯錯誤或缺乏邏輯的回答,同時統一最終答案的格式為 `\boxed{}` ,方便後續應用於標準答案比對與模型訓練。 本資料集適用於繁體中文大型語言模型的微調訓練與推理能力測試,亦可作為 Chain-of-Thought 推理訓練的基礎語料之一。 - **Curated by:** [Minyi Chen](https://huggingface.co/minyichen) - **Funded by:** [APMIC](https://www.apmic.ai/) - **Shared by:** [Minyi Chen](https://huggingface.co/minyichen) - **Language(s) (NLP):** Traditional Chinese & English - **License:** MIT ### Dataset Sources <!-- Provide the basic links for the dataset. --> - **Repository:** [twinkle-ai/tw-math-reasoning-2k](https://huggingface.co/datasets/twinkle-ai/tw-math-reasoning-2k) ## Uses <!-- Address questions around how the dataset is intended to be used. --> ### Direct Use - 微調繁體中文大型語言模型(LLMs)以提升其數學推理與解題能力。 - 評估繁體中文語言模型在「多步驟推理(chain-of-thought)」任務上的表現。 - 作為學術研究中的標註語料,用於探討數學語言理解與自然語言生成的交集。 - 作為教學語料,用於展示數學題目的語言化解題範例,支援數學與語言教育應用。 ### Out-of-Scope Use > [!warning] > 本資料集重視學術研究與模型開發用途,不鼓勵任何非研究的使用。 <!-- This section addresses misuse, malicious use, and uses that the dataset will not work well for. --> - **應用高風險決策:** 例如金融建議、工程設計、醫療診斷等,因為生成的解題過程雖具邏輯性但未經數學專家逐一審核,不保證所有答案絕對正確。 - **重建原始資料題庫** 本資料集僅為繁體中文生成版,並不包含原始英文解答,無法作為 [HuggingFaceH4/MATH](https://huggingface.co/datasets/HuggingFaceH4/MATH) 的完整替代。 ## Dataset Structure 每筆資料為一組包含英文數學題目與繁體中文回答的對應資料,格式如下: ```json { 'problem', # 原始英文數學題目 'level', # 題目難度等級(如 high school, olympiad 等) 'type', # 題型分類(如 algebra、geometry、number theory 等) 'solution', # HuggingFaceH4/MATH 原始英文解答(僅參考) 'subset', # 來源子集名稱(如 'train') 'split', # 資料分割(目前皆為 'train') 'model', # 生成模型 'problem_zhtw', # 題目繁體中文翻譯(可用於多語訓練) 'think', # 模型生成的繁體中文解題過程(邏輯推理) 'answer', # 模型生成的最終答案(通常以 `\boxed{}` 呈現) 'messages' # 完整對話訊息結構(包含提示詞、回應、角色等) } ``` > 💡 模型回答以繁體中文輸出,並保留題目原文以利跨語言對應與訓練。 ## Dataset Creation ### Curation Rationale **tw-math-reasoning-2k** 的設計初衷在於彌補繁體中文數學推理語料的稀缺現況。雖然 Hugging Face 上已有如 [HuggingFaceH4/MATH](https://huggingface.co/datasets/HuggingFaceH4/MATH) 等高品質英文數學題庫,但繁體中文語境下的解題語料仍極為稀少,限制了中文大型語言模型在數學推理領域的發展與評估能力。 因此,本資料集透過精選自原始 MATH 題庫的 2,000 題題目,結合具多輪推理能力的 [perplexity-ai/r1-1776](https://huggingface.co/perplexity-ai/r1-1776) 模型生成繁體中文解題過程,旨在提供具邏輯性、語意自然且格式一致的訓練與評估樣本。我們特別注重答案的可驗證性與過程的教學價值,確保資料能支援如 Chain-of-Thought 推理訓練、答案解析生成等多樣應用場景。 此資料集亦可作為日後擴充更大規模繁中數學語料的基礎樣本庫。透過小規模高品質起步,逐步建立起繁體中文數學推理研究的語料基石。 ### Source Data <!-- This section describes the source data (e.g. news text and headlines, social media posts, translated sentences, ...). --> #### Data Collection and Processing - **來源資料集:** [HuggingFaceH4/MATH](https://huggingface.co/datasets/HuggingFaceH4/MATH) - **取樣數量:** 2,000 題(各類型題目均衡取樣) - **回答生成:** 使用 [perplexity-ai/r1-1776](https://huggingface.co/perplexity-ai/r1-1776) 模型以多輪提示精調,生成具備邏輯推理的完整中文解題過程 - **資料清洗:** - 過濾模型明顯錯誤或不合邏輯的回應 - 擷取最終答案並統一為 LaTeX `\boxed{}` 格式 ## Bias, Risks, and Limitations - 並非所有生成回應皆經人工審查,可能存在邏輯錯誤或非標準解法。 - 資料集強調**解題過程表達能力**,非單純數值答對即可。 ### Recommendations 使用 **tw-math-reasoning-2k** 時,建議注意以下幾點,以充分理解其適用範圍與潛在限制: - **模型生成偏誤**:資料集中之解題過程由 [perplexity-ai/r1-1776](https://huggingface.co/perplexity-ai/r1-1776) 模型生成,可能會受到原始模型訓練語料與提示設計的影響,造成某些解法過於冗長、不夠直觀,或在特定題型上採取非標準解法。 - **數學正確性風險**:雖經過基本清洗與錯誤過濾,部分解題邏輯仍可能存在細節錯誤、計算誤差或不嚴謹推理,建議在高精度應用場景中搭配額外驗證機制使用。 - **語言與格式一致性**:資料以繁體中文呈現,但個別題目可能仍包含 LaTeX 符號、數學術語或模型特有用語風格,使用於教學或教材時建議進行語言風格統整。 - **有限樣本規模**:本資料集僅包含 2,000 題,屬於小型精選集,適合作為研究、微調或標準推理風格的參考;不建議直接用於大規模模型 pretraining。 ## Citation 如果您使用本資料集,請引用: ```yaml @misc{twmath2k2025, title = {tw-math-reasoning-2k: Traditional Chinese Mathematical Reasoning Dataset}, author = {Twinkle AI}, year = {2025}, note = {Available at: \url{https://huggingface.co/datasets/twinkle-ai/tw-math-reasoning-2k}; Generated using \url{https://huggingface.co/perplexity-ai/r1-1776} from the HuggingFaceH4/MATH dataset} } ``` ## Dataset Card Authors [Twinkle AI](https://huggingface.co/twinkle-ai) ## Dataset Card Contact [Twinkle AI](https://huggingface.co/twinkle-ai)
EYEDOL/naija_commonvoice_naijaEnglish1_preprocessed_train_batch_1
EYEDOL
2025-04-30T09:53:46Z
0
0
[ "size_categories:1K<n<10K", "format:parquet", "modality:audio", "modality:text", "library:datasets", "library:dask", "library:mlcroissant", "library:polars", "region:us" ]
[]
2025-04-30T09:52:20Z
null
--- dataset_info: features: - name: audio dtype: audio: sampling_rate: 16000 - name: sentence dtype: string - name: input_length dtype: int64 - name: input_features sequence: sequence: float32 - name: labels sequence: int64 - name: labels_length dtype: int64 splits: - name: train num_bytes: 3646837252.875 num_examples: 2721 download_size: 1649827139 dataset_size: 3646837252.875 configs: - config_name: default data_files: - split: train path: data/train-* ---
nscharrenberg/DBNL-public
nscharrenberg
2025-04-30T09:20:01Z
14
0
[ "language:nl", "language:en", "license:cc0-1.0", "size_categories:100K<n<1M", "format:parquet", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us" ]
[]
2025-04-28T12:46:36Z
null
--- license: cc0-1.0 language: - nl - en pretty_name: DBNLP ---
linoyts/wan_wiping_surface
linoyts
2025-04-30T09:08:10Z
0
0
[ "license:apache-2.0", "size_categories:n<1K", "modality:text", "modality:video", "library:datasets", "library:mlcroissant", "region:us", "text-to-video" ]
[]
2025-04-30T09:08:08Z
null
--- license: apache-2.0 tags: - text-to-video --- This dataset contains videos generated using Wan 2.1 T2V 14B.
HungVu2003/opt-350m_beta_1.0_alpha_0.6_num-company_3_dataset_1_for_gen_2
HungVu2003
2025-04-30T08:58:41Z
0
0
[ "size_categories:10K<n<100K", "format:parquet", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us" ]
[]
2025-04-30T08:58:39Z
null
--- dataset_info: features: - name: question dtype: string splits: - name: train num_bytes: 3630075 num_examples: 12500 download_size: 1911834 dataset_size: 3630075 configs: - config_name: default data_files: - split: train path: data/train-* ---
Maxscha/json-instruct-generation
Maxscha
2025-04-30T08:55:44Z
14
0
[ "size_categories:10K<n<100K", "format:parquet", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us" ]
[]
2025-03-05T17:00:38Z
null
--- dataset_info: features: - name: schema dtype: string - name: input dtype: string - name: output dtype: string - name: task dtype: string splits: - name: train num_bytes: 20200388 num_examples: 10000 download_size: 6300826 dataset_size: 20200388 configs: - config_name: default data_files: - split: train path: data/train-* --- # Dataset Card for "json-instruct-generation" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
IPEC-COMMUNITY/fmb_dataset_lerobot
IPEC-COMMUNITY
2025-04-29T06:45:54Z
18,989
0
[ "task_categories:robotics", "license:apache-2.0", "size_categories:1M<n<10M", "format:parquet", "modality:tabular", "modality:timeseries", "modality:video", "library:datasets", "library:dask", "library:mlcroissant", "library:polars", "region:us", "LeRobot", "fmb_dataset", "rlds", "openx", "franka" ]
[ "robotics" ]
2025-04-28T02:06:42Z
null
--- license: apache-2.0 task_categories: - robotics tags: - LeRobot - LeRobot - fmb_dataset - rlds - openx - franka configs: - config_name: default data_files: data/*/*.parquet --- This dataset was created using [LeRobot](https://github.com/huggingface/lerobot). ## Dataset Description - **Homepage:** [More Information Needed] - **Paper:** [More Information Needed] - **License:** apache-2.0 ## Dataset Structure [meta/info.json](meta/info.json): ```json { "codebase_version": "v2.1", "robot_type": "franka", "total_episodes": 8612, "total_frames": 1137459, "total_tasks": 24, "total_videos": 34448, "total_chunks": 9, "chunks_size": 1000, "fps": 10, "splits": { "train": "0:8612" }, "data_path": "data/chunk-{episode_chunk:03d}/episode_{episode_index:06d}.parquet", "video_path": "videos/chunk-{episode_chunk:03d}/{video_key}/episode_{episode_index:06d}.mp4", "features": { "observation.images.image_side_2": { "dtype": "video", "shape": [ 256, 256, 3 ], "names": [ "height", "width", "rgb" ], "info": { "video.fps": 10.0, "video.height": 256, "video.width": 256, "video.channels": 3, "video.codec": "av1", "video.pix_fmt": "yuv420p", "video.is_depth_map": false, "has_audio": false } }, "observation.images.image_side_1": { "dtype": "video", "shape": [ 256, 256, 3 ], "names": [ "height", "width", "rgb" ], "info": { "video.fps": 10.0, "video.height": 256, "video.width": 256, "video.channels": 3, "video.codec": "av1", "video.pix_fmt": "yuv420p", "video.is_depth_map": false, "has_audio": false } }, "observation.images.image_wrist_2": { "dtype": "video", "shape": [ 256, 256, 3 ], "names": [ "height", "width", "rgb" ], "info": { "video.fps": 10.0, "video.height": 256, "video.width": 256, "video.channels": 3, "video.codec": "av1", "video.pix_fmt": "yuv420p", "video.is_depth_map": false, "has_audio": false } }, "observation.images.image_wrist_1": { "dtype": "video", "shape": [ 256, 256, 3 ], "names": [ "height", "width", "rgb" ], "info": { "video.fps": 10.0, "video.height": 256, "video.width": 256, "video.channels": 3, "video.codec": "av1", "video.pix_fmt": "yuv420p", "video.is_depth_map": false, "has_audio": false } }, "observation.state": { "dtype": "float32", "shape": [ 8 ], "names": { "motors": [ "x", "y", "z", "roll", "pitch", "yaw", "pad", "gripper" ] } }, "action": { "dtype": "float32", "shape": [ 7 ], "names": { "motors": [ "x", "y", "z", "roll", "pitch", "yaw", "gripper" ] } }, "timestamp": { "dtype": "float32", "shape": [ 1 ], "names": null }, "frame_index": { "dtype": "int64", "shape": [ 1 ], "names": null }, "episode_index": { "dtype": "int64", "shape": [ 1 ], "names": null }, "index": { "dtype": "int64", "shape": [ 1 ], "names": null }, "task_index": { "dtype": "int64", "shape": [ 1 ], "names": null } } } ``` ## Citation **BibTeX:** ```bibtex [More Information Needed] ```
JoeYing/ReTool-SFT
JoeYing
2025-04-29T02:52:03Z
805
26
[ "license:apache-2.0", "size_categories:1K<n<10K", "format:parquet", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "arxiv:2504.11536", "region:us" ]
[]
2025-04-16T12:27:25Z
10
--- license: apache-2.0 pretty_name: ReTool --- # ReTool: Reinforcement Learning for Strategic Tool Use in LLMs In this work, we embrace the RL paradigm and introduce ReTool, a Tool-augmented Reinforcement learning framework explicitly designed to guide LLMs towards optimal strategies for leveraging external computational tools during reasoning. Our comprehensive experiments on AIME2024 and AIME2025 demonstrate that ReTool not only achieves superior accuracy compared to conventional text-based RL approaches, but also converges with significantly fewer training steps. 🚀 ReTool achieves accuracy of 67.0% on AIME 2024 and 49.3% on AIME 2025 based on the Qwen2.5-32B-Instruct model, outperforming the text-based RL baseline with less than 50% training steps. - Project Page: https://retool-rl.github.io/ ### Citation If you find our project helpful, please cite: ``` @misc{feng2025retoolreinforcementlearningstrategic, title={ReTool: Reinforcement Learning for Strategic Tool Use in LLMs}, author={Jiazhan Feng and Shijue Huang and Xingwei Qu and Ge Zhang and Yujia Qin and Baoquan Zhong and Chengquan Jiang and Jinxin Chi and Wanjun Zhong}, year={2025}, eprint={2504.11536}, archivePrefix={arXiv}, primaryClass={cs.CL}, url={https://arxiv.org/abs/2504.11536}, } ```
xingxm/SVGX-SFT-1M
xingxm
2025-04-28T17:41:27Z
17
3
[ "language:multilingual", "license:cc-by-nc-4.0", "arxiv:2412.11102", "arxiv:2412.10437", "region:us", "svg", "svg-emoji", "vector-graphics", "vision-language", "multimodal" ]
[]
2025-04-28T14:30:57Z
3
--- pretty_name: SVGX-SFT-1M dataset_creator: xingxm language: - multilingual tags: - svg - svg-emoji - vector-graphics - vision-language - multimodal license: cc-by-nc-4.0 --- # SVGX-SFT-1M Dataset SVGX-SFT-1M is a large-scale multimodal dataset designed for training and evaluating models on SVG-based instruction-following and dialogue tasks. It contains over **1 million samples** covering: - **Text-to-SVG Generation:** Given a natural language prompt, generate corresponding SVG code. - **SVG-to-Text Description:** Given SVG code, produce a natural language description. - **Image-to-SVG Reasoning** (with ShareGPT-style conversations): Given a raster image, generate SVG code, and describe the SVG content. ## 🚀 How to Use ```python from datasets import load_dataset # Load Alpaca format dataset = load_dataset("xingxm/SVGX-SFT-1M", split="train", data_files="SVGX_SFT_GEN_51k.json") # Load ShareGPT format dataset = load_dataset("xingxm/SVGX-SFT-1M", split="train", data_files="SVGX_SFT_vision_25k_encode.json") print(dataset[0]) ``` ## ✨ Data Formats 1. Alpaca Format (`instruction`, `input`, `output`) Each record is a single-turn instruction: ```json { "instruction": "Generate an SVG illustration from the given description.", "input": "A red circle centered at (50,50) with a radius of 40.", "output": "<svg width='100' height='100' xmlns='http://www.w3.org/2000/svg'><circle cx='50' cy='50' r='40' fill='red'/></svg>" } ``` or, ```json { "instruction": "Describe the given SVG code in natural language.", "input": "<svg xmlns='http://www.w3.org/2000/svg' width='100' height='100'><circle cx='50' cy='50' r='40' fill='red'/></svg>", "output": "This SVG contains a red circle centered at (50,50) with a radius of 40." } ``` 2. ShareGPT Format (`messages`, `images`) Multi-turn conversations with optional images: ```json { "messages": [ { "role": "user", "content": "<image> Convert this raster image to SVG code." }, { "role": "assistant", "content": "[<|START_OF_SVG|>] [<|svg_path|>] ..." }, { "role": "user", "content": "Explain what this SVG depicts and how it's structured." }, { "role": "assistant", "content": "This SVG illustration depicts a tea cup with minimalistic shapes..." } ], "images": [ "SVGX-rendering-data/1a650587-58c9-4861-ba49-90f18c3bcefa.png" ] } ``` ## 🛠️ Scripts Overview | File | Description | |:----------------------|:--------------------------------------------------------| | `encode.sh` | Bash script to encode datasets. | | `preprocess_utils.py` | Utilities for preprocessing SVG data. | | `round.py` | Helper functions for rounding numeric SVG values. | | `semantic_encode.py` | Script for encoding SVGs into semantic token sequences. | | `semantic_tokens.py` | Definition of special SVG semantic tokens. | | `svg_tokens.yaml` | YAML configuration of all SVG token types. | ## ✅ Dataset Capabilities - **SVG -> Syntactic Representation**: Structured parsing of SVG elements (paths, shapes, gradients, attributes, etc.) - **Syntactic -> SVG Reconstruction**: Reconstruct SVGs from structured semantic tokens. - **Image + SVG -> Text Reasoning**: Multimodal understanding pipeline from images to SVGs to descriptions. - **Vision-Language Pretraining**: Lightweight vector graphics enable efficient multimodal model training. ## 🎨 Example Use Cases - Fine-tuning LLMs for SVG generation and understanding. - Training multimodal models with lightweight vector-image reasoning. - Image-to-SVG translation tasks. - Vision-language reasoning enhancement. - Building creative AI design assistants. ## 💘 Acknowledgements SVGX-SFT-1M is built upon contributions from: - [xingxm/SVGX-Core-250k](https://huggingface.co/datasets/xingxm/SVGX-Core-250k) ## 📎 Citation If you use this code for your research, please cite the following work: ``` @article{xing2024llm4svg, title={Empowering LLMs to Understand and Generate Complex Vector Graphics}, author={Xing, Ximing and Hu, Juncheng and Liang, Guotao and Zhang, Jing and Xu, Dong and Yu, Qian}, journal={arXiv preprint arXiv:2412.11102}, year={2024} } @article{xing2024svgfusion, title={SVGFusion: Scalable Text-to-SVG Generation via Vector Space Diffusion}, author={Xing, Ximing and Hu, Juncheng and Zhang, Jing and Xu, Dong and Yu, Qian}, journal={arXiv preprint arXiv:2412.10437}, year={2024} } ``` ## 📜 License This dataset is licensed under the * *[Creative Commons Attribution-NonCommercial 4.0 International (CC BY-NC 4.0)](https://creativecommons.org/licenses/by-nc/4.0/) ** license. - ✅ You are free to **share** and **adapt** the material. - ❌ **Commercial use is prohibited**. - 📎 **Attribution is required**. For full legal terms, please refer to [this link](https://creativecommons.org/licenses/by-nc/4.0/).
OpenDriveLab/OpenScene
OpenDriveLab
2025-04-28T07:13:35Z
28,792
5
[ "license:cc-by-nc-sa-4.0", "region:us" ]
[]
2024-03-02T04:33:04Z
null
--- license: cc-by-nc-sa-4.0 ---
moonshotai/Kimi-Audio-GenTest
moonshotai
2025-04-28T03:45:53Z
139
2
[ "language:zh", "license:mit", "size_categories:n<1K", "format:audiofolder", "modality:audio", "modality:text", "library:datasets", "library:mlcroissant", "region:us", "speech generation", "chinese" ]
[]
2025-04-28T03:44:52Z
2
--- # Required: Specify the license for your dataset license: [mit] # Required: Specify the language(s) of the dataset language: - zh # 中文 # Optional: Add tags for discoverability tags: - speech generation - chinese # Required: A pretty name for your dataset card pretty_name: "Kimi-Audio-Generation-Testset" --- # Kimi-Audio-Generation-Testset ## Dataset Description **Summary:** This dataset is designed to benchmark and evaluate the conversational capabilities of audio-based dialogue models. It consists of a collection of audio files containing various instructions and conversational prompts. The primary goal is to assess a model's ability to generate not just relevant, but also *appropriately styled* audio responses. Specifically, the dataset targets the model's proficiency in: * **Paralinguistic Control:** Generating responses with specific control over **emotion**, speaking **speed**, and **accent**. * **Empathetic Dialogue:** Engaging in conversations that demonstrate understanding and **empathy**. * **Style Adaptation:** Delivering responses in distinct styles, including **storytelling** and reciting **tongue twisters**. Audio conversation models are expected to process the input audio instructions and generate reasonable, contextually relevant audio responses. The quality, appropriateness, and adherence to the instructed characteristics (like emotion or style) of the generated responses are evaluated through **human assessment**. * **Languages:** zh (中文) ## Dataset Structure ### Data Instances Each line in the `test/metadata.jsonl` file is a JSON object representing a data sample. The `datasets` library uses the path in the `file_name` field to load the corresponding audio file. **示例:** ```json {"audio_content": "你能不能快速地背一遍李白的静夜思", "ability": "speed", "file_name": "wav/6.wav"}
divarofficial/real_estate_ads
divarofficial
2025-04-26T15:33:04Z
195
62
[ "license:odbl", "size_categories:1M<n<10M", "format:csv", "modality:tabular", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us" ]
[]
2025-04-09T08:06:42Z
2
--- license: odbl --- # 🏠 Divar Real Estate Ads Dataset [![Dataset Size](https://img.shields.io/badge/Size-750%20MB-blue)](https://huggingface.co/datasets/divar/real-estate-ads) [![Rows](https://img.shields.io/badge/Rows-1M-green)](https://huggingface.co/datasets/divar/real-estate-ads) ## 📋 Overview The `real_estate_ads` dataset contains one million anonymized real estate advertisements collected from the [Divar](https://divar.ir) platform, one of the largest classified ads platforms in the Middle East. This comprehensive dataset provides researchers, data scientists, and entrepreneurs with authentic real estate market data to build innovative solutions such as price evaluation models, market analysis tools, and forecasting systems. ## 🔍 Dataset Details | Property | Value | | --------------- | ------------------------------------------ | | **Size** | 1,000,000 rows, approximately 750 MB | | **Time Period** | Six-month period (2024) | | **Source** | Anonymized real estate listings from Divar | | **Format** | Tabular data (CSV/Parquet) with 57 columns | | **Languages** | Mixed (primarily Persian) | | **Domains** | Real Estate, Property Market | ## 🚀 Quick Start ```python # Load the dataset using the Hugging Face datasets library from datasets import load_dataset # Load the full dataset dataset = load_dataset("divarofficial/real-estate-ads") # Print the first few examples print(dataset['train'][:5]) # Get dataset statistics print(f"Dataset size: {len(dataset['train'])} rows") print(f"Features: {dataset['train'].features}") ``` ## 📊 Schema The dataset includes comprehensive property information organized in the following categories: ### 🏷️ Categorization - `cat2_slug`, `cat3_slug`: Property categorization slugs - `property_type`: Type of property (apartment, villa, land, etc.) ### 📍 Location - `city_slug`, `neighborhood_slug`: Location identifiers - `location_latitude`, `location_longitude`: Geographic coordinates - `location_radius`: Location accuracy radius ### 📝 Listing Details - `created_at_month`: Timestamp of when the ad was created - `user_type`: Type of user who posted the listing (individual, agency, etc.) - `description`, `title`: Textual information about the property ### 💰 Financial Information - **Rent-related**: `rent_mode`, `rent_value`, `rent_to_single`, `rent_type` - **Price-related**: `price_mode`, `price_value` - **Credit-related**: `credit_mode`, `credit_value` - **Transformed values**: Various transformed financial metrics for analysis ### 🏢 Property Specifications - `land_size`, `building_size`: Property dimensions (in square meters) - `deed_type`, `has_business_deed`: Legal property information - `floor`, `rooms_count`, `total_floors_count`, `unit_per_floor`: Building structure details - `construction_year`, `is_rebuilt`: Age and renovation status ### 🛋️ Amenities and Features - **Utilities**: `has_water`, `has_electricity`, `has_gas` - **Climate control**: `has_heating_system`, `has_cooling_system` - **Facilities**: `has_balcony`, `has_elevator`, `has_warehouse`, `has_parking` - **Luxury features**: `has_pool`, `has_jacuzzi`, `has_sauna` - **Other features**: `has_security_guard`, `has_barbecue`, `building_direction`, `floor_material` ### 🏨 Short-term Rental Information - `regular_person_capacity`, `extra_person_capacity` - `cost_per_extra_person` - **Pricing variations**: `rent_price_on_regular_days`, `rent_price_on_special_days`, `rent_price_at_weekends` ## 📈 Example Analysis ```python import pandas as pd import matplotlib.pyplot as plt import seaborn as sns # Convert to pandas DataFrame for analysis df = dataset['train'].to_pandas() # Price distribution by property type plt.figure(figsize=(12, 6)) sns.boxplot(x='property_type', y='price_value', data=df) plt.title('Price Distribution by Property Type') plt.xticks(rotation=45) plt.tight_layout() plt.show() # Correlation between building size and price plt.figure(figsize=(10, 6)) sns.scatterplot(x='building_size', y='price_value', data=df) plt.title('Correlation between Building Size and Price') plt.xlabel('Building Size (sq.m)') plt.ylabel('Price') plt.tight_layout() plt.show() ``` ## 💡 Use Cases This dataset is particularly valuable for: 1. **Price Prediction Models**: Train algorithms to estimate property values based on features ```python # Example: Simple price prediction model from sklearn.ensemble import RandomForestRegressor from sklearn.model_selection import train_test_split features = ['building_size', 'rooms_count', 'construction_year', 'has_parking'] X = df[features].fillna(0) y = df['price_value'].fillna(0) X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.2) model = RandomForestRegressor(n_estimators=100) model.fit(X_train, y_train) ``` 2. **Market Analysis**: Understand trends and patterns in the real estate market 3. **Recommendation Systems**: Build tools to suggest properties based on user preferences 4. **Natural Language Processing**: Analyze property descriptions and titles 5. **Geospatial Analysis**: Study location-based pricing and property distribution ## 🔧 Data Processing Information The data has been: - Anonymized to protect privacy - Randomly sampled from the complete Divar platform dataset - Cleaned with select columns removed to ensure privacy and usability - Standardized to ensure consistency across entries ## 📚 Citation and Usage When using this dataset in your research or applications, please consider acknowledging the source: ```bibtex @dataset{divar2025realestate, author = {Divar Corporation}, title = {Real Estate Ads Dataset from Divar Platform}, year = {2025}, publisher = {Hugging Face}, url = {https://huggingface.co/datasets/divar/real-estate-ads} } ``` ## 🤝 Contributing We welcome contributions to improve this dataset! If you find issues or have suggestions, please open an issue on the [GitHub repository](https://github.com/divar-ir/kenar-docs) or contact us at [[email protected]](mailto:[email protected]).
bennny674/timestep-distill-reasoning
bennny674
2025-04-25T22:29:06Z
109
2
[ "task_categories:text-generation", "source_datasets:open-thoughts/OpenThoughts-114k", "language:en", "license:apache-2.0", "size_categories:n<1K", "format:parquet", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us", "reasoning-datasets-competition", "text" ]
[ "text-generation" ]
2025-04-24T22:59:17Z
2
--- task_categories: - text-generation language: - en tags: - reasoning-datasets-competition - text size_categories: - n<1K source_datasets: - open-thoughts/OpenThoughts-114k license: apache-2.0 --- # Reasoning Time-step Distillation Dataset ## Purpose/Introduction Current reasoning models generate their thoughts token-by-token, often taking many tokens to finish one reasoning step. This has advantages, such as the same unmodified model architectures being usable, though it increases compute costs significantly, due to multiple forward passes being required to complete one thought. This dataset attempts to solve this by introducing a method of reasoning time-step distillation based on sentence segmentation. Sentences were chosen as the unit that should be generated in one reasoning step because: - There are already available sentence segmentation libraries. - A sentence is, quite literally, defined as a set of words communicating one complete thought. ## Dataset creation method The dataset consists of randomly selected reasoning traces from [open-thoughts/OpenThoughts-114k](https://huggingface.co/datasets/open-thoughts/OpenThoughts-114k), however, the assistant's reasoning is segmented into a list of sentences using [pySBD](https://github.com/nipunsadvilkar/pySBD). ## Example uses The dataset could be used for reasoning time-step distillation through any method that predicts a sentence's worth of reasoning at a time, however, I will only propose one here. Reasoning steps could be generated by a separate model to the main LLM, which is fed the previous sentences of reasoning, that model generating embeddings for the reasoning it performs. These embeddings would be converted to tokens for the LLM in a similar manner to how some multimodal models' image encoders embed images for an LLM, having placeholder tokens replaced with the embeddings from the non-text-modal model. Such a model(and an accompanying LLM) could be trained in an autoregressive way, where the objective for the models would be that the LLM, when given a piece of a reasoning trace, with the last sentence replaced by a single reasoning token from the reasoning model, would produce the same output as if it was fed the original version without that substitution. After that, the models could then be trained together with RL like any other reasoning model. ## Limitations and Biases - Due to the source reasoning traces being from open-thoughts/OpenThoughts-114k, any limitations or biases from that dataset are carried over to this one. - Additionally, this dataset has the limitation that it is designed around a hypothesis about reasoning models that I have not tested, specifically, that they could be trained to produce reasoning at one forward pass per full thought without sacrificing reasoning quality.
ieasybooks-org/waqfeya-library-compressed
ieasybooks-org
2025-04-25T15:09:42Z
653
4
[ "task_categories:image-to-text", "language:ar", "license:mit", "size_categories:10K<n<100K", "format:csv", "modality:tabular", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us" ]
[ "image-to-text" ]
2025-04-23T05:19:54Z
4
--- license: mit task_categories: - image-to-text language: - ar pretty_name: Waqfeya Library - Compressed size_categories: - 10K<n<100K configs: - config_name: index data_files: - split: index path: index.tsv --- # Waqfeya Library - Compressed ## 📖 Overview [Waqfeya](https://waqfeya.net) is one of the primary online resources for Islamic books, similar to [Shamela](https://shamela.ws). It hosts more than 10,000 PDF books across over 80 categories. In this dataset, we processed the original PDF files using Google Document AI APIs and extracted their contents into two additional formats: TXT and DOCX. ## 📊 Dataset Contents This dataset is identical to [ieasybooks-org/waqfeya-library](https://huggingface.co/datasets/ieasybooks-org/waqfeya-library), with one key difference: the contents have been compressed for easier downloading. Specifically, the `pdf`, `txt`, and `docx` folders have been packaged into `pdf.zip`, `txt.zip`, and `docx.zip`, respectively. For detailed information about the dataset contents and usage instructions, please refer to the original dataset page: [ieasybooks-org/waqfeya-library](https://huggingface.co/datasets/ieasybooks-org/waqfeya-library).
nvidia/Nemotron-MIND
nvidia
2025-04-24T21:19:19Z
821
5
[ "license:cc-by-4.0", "size_categories:100M<n<1B", "format:parquet", "modality:text", "library:datasets", "library:dask", "library:mlcroissant", "library:polars", "arxiv:2410.12881", "region:us" ]
[]
2025-04-18T16:37:10Z
5
--- license: cc-by-4.0 --- # Nemotron-MIND: Math Informed syNthetic Dialogues for Pretraining LLMs **Authors: Syeda Nahida Akter, Shrimai Prabhumoye, John Kamalu, Sanjeev Satheesh, Eric Nyberg, Mostofa Patwary, Mohammad Shoeybi, Bryan Catanzaro** [[Paper]](https://arxiv.org/abs/2410.12881) [[Blog]](https://research.nvidia.com/labs/adlr/Nemotron-MIND/) ## Dataset Description <div align="center"> <img class="img-full" src="MIND_overview.png" width=900> <p style="max-width: 900px; text-align: justify"> <b>Figure 1: Math Informed syNthetic Dialogue.</b> We (a) manually design prompts of seven conversational styles, (b) provide the prompt along with raw context as input to an LLM to obtain diverse synthetic conversations, (c) apply heuristic filtering to refine the generated data and (d) observe the downstream task accuracy after continuously pretraining a 7B LLM. </p> </div> The Nemotron-MIND Dataset is a compilation of pretraining data that supports improvements of math reasoning capabilities of the Nemotron5 series of models. This dataset release represents a significant advancement in openness and transparency in model development. Nemotron-MIND contains over 138 billion tokens of structured mathematical dialogues generated by [Nemotron4-340B-Instruct](https://huggingface.co/nvidia/Nemotron-4-340B-Instruct). The data synthesis process comprises of the following phases: - **Compose Diverse Prompts:** We design seven prompt templates to guide a pretrained LLM in converting a single math text into a structured conversation. They represent different social conversational settings like: (1) Debate, (2) Problem-Solving, (3) Layman-Knowall, (4) Teacher-Student, (5) Two-Professors, (6) Two-Students, and (7) Interview-Interviewee. - **Raw Data:** We use OpenWebMath (OWM) as our base corpus—14.7B tokens of rich, raw math content. - **Generate Conversations at Scale:** For each document, we apply a prompt to generate conversation. We use the [Nemotron4-340B-Instruct](https://huggingface.co/nvidia/Nemotron-4-340B-Instruct) model to generate the conversations. - **Filter Noisy Outputs:** LLM-based scoring proved too lenient. Instead, we apply heuristic rules to remove low-quality generations and retain only coherent, detailed discussions. Finally, we continuously pretrain a 7B model on a mix of filtered conversations and raw pretraining data. ## Main Results <div align="center"> <img class="img-full" src="MIND_results.png" width=900> <p style="max-width: 900px; text-align: justify"> <b>Figure 2: Results of 7B LLM pretrained on Diverse Conversational Styles.</b> Continuous training with different conversation styles improves all reasoning tasks. </p> </div> **Key Takeaways:** - Every MIND conversation style beat both raw and rephrased baselines on reasoning tasks. - Gains on **GSM8K** ranged from **+4.78% to +12.82%** — showcasing huge improvements in math problem solving. **MATH (+0.54–1.28%)** and **MMLU-STEM (+0.79–4.28%)** also saw consistent gains. Even **general reasoning** benchmarks improved by up to **+2%** on average across 10 tasks. - The best results among 4B tokens came from the Longest Conversation variant—suggesting that richer, more elaborate dialogue drives stronger reasoning ability. This dataset primarily supports pretraining LLMs from scratch. This dataset demonstrates improvement in math capabilities of pretrained models. The MIND framework uses NemoSKills to synthetically generate math data from OpenWebMath corpora, which is then used to pretrain state-of-the art (SOTA) models. This dataset is ready for commercial/non-commercial use. ## Dataset Owner(s): NVIDIA Corporation ## Dataset Creation Date: September 20, 2024 ## License/Terms of Use: Governing Terms: This dataset is licensed under the Creative Commons Attribution 4.0 International License (CC BY 4.0) available at https://creativecommons.org/licenses/by/4.0/legalcode. This dataset contains data created using OpenWebMath ([https://huggingface.co/datasets/open-web-math/open-web-math](https://huggingface.co/datasets/open-web-math/open-web-math)), which is made available under the ODC Attribution License ([https://opendatacommons.org/licenses/by/1-0/](https://opendatacommons.org/licenses/by/1-0/)). **Data Developer:** NVIDIA ## Intended Usage: The Nemotron-MIND Dataset is intended to be used by the community to pretrain LLMs with SOTA math reasoning capabilities. The data may be used to train and evaluate. <br> ## Data Version: v1 ## Dataset Characterization - Data Collection Method: Synthetic <br> - Labeling Method: Automated <br> ## Dataset Format Text ## Dataset Quantification - Record Count: 231.6M - Feature Count: 7. We have seven different conversational styles in the data. They are: (1) TWO STUDENTS, (2) TEACHER STUDENT, (3) TWO PROFESSORS, (4) DEBATE, (5) PROBLEM SOLVING, (6) LAYMAN KNOWALL, and (7) INTERVIEW. - Total Data Storage: 827GB ## Ethical Considerations: NVIDIA believes Trustworthy AI is a shared responsibility and we have established policies and practices to enable development for a wide array of AI applications. When downloaded or used in accordance with our terms of service, developers should work with their internal model team to ensure this model meets requirements for the relevant industry and use case and addresses unforeseen product misuse. Please report security vulnerabilities or NVIDIA AI Concerns [here](https://www.nvidia.com/en-us/support/submit-security-vulnerability/). ## Citation ``` @inproceedings{ akter2025mind, title={{MIND}: Math Informed syNthetic Dialogues for Pretraining {LLM}s}, author={Syeda Nahida Akter and Shrimai Prabhumoye and John Kamalu and Sanjeev Satheesh and Eric Nyberg and Mostofa Patwary and Mohammad Shoeybi and Bryan Catanzaro}, booktitle={The Thirteenth International Conference on Learning Representations}, year={2025}, url={https://openreview.net/forum?id=TuOTSAiHDn} } ```
dmis-lab/meerkat-instructions
dmis-lab
2025-04-24T00:44:47Z
136
2
[ "language:en", "license:cc-by-nc-4.0", "size_categories:100K<n<1M", "modality:text", "arxiv:2404.00376", "region:us" ]
[]
2025-04-23T23:23:44Z
2
--- configs: - config_name: MedQA-CoT data_files: - split: train path: "medqa-cot.jsonl" - config_name: MedBooks-18-CoT data_files: - split: train path: "medbooks-18-cot.jsonl" - config_name: MedMCQA data_files: - split: train path: "medmcqa.jsonl" - config_name: LiveQA data_files: - split: train path: "liveqa.jsonl" - config_name: MedicationQA data_files: - split: train path: "medicationqa.jsonl" - config_name: ChatDoctor-cleaned data_files: - split: train path: "chatdoctor-cleaned.jsonl" - config_name: MedQA-dialog data_files: - split: train path: "medqa-dialog.jsonl" - config_name: MTS-dialog data_files: - split: train path: " mts-dialog.jsonl" - config_name: MedInstruct-52K data_files: - split: train path: "medinstruct-52k.jsonl" license: cc-by-nc-4.0 language: - en --- This repository provides the instruction tuning data used to train our medical language model, Meerkat, along with descriptions. For more information, please refer to the paper below. Our models can be downloaded from [the official model repository](https://huggingface.co/collections/dmis-lab/meerkat-6710b7ae0258fc540c475eec). 📄 **Paper**: [Small Language Models Learn Enhanced Reasoning Skills from Medical Textbooks](https://arxiv.org/abs/2404.00376) ### Dataset Statistics **Table: Statistics of our instruction-tuning datasets** “# Examples” denotes the number of training examples for each dataset. † indicates that the dataset is newly constructed or repurposed in our study. The total number of training examples is **441,034**. | **Target Application** | **Dataset** | **# Examples** | |-----------------------------|--------------------------------------------|----------------| | **Multiple-choice QA** | MedQA-CoT† [[1]](#ref1) | 9,308 | | | MedBooks-18-CoT† | 77,660 | | | MedMCQA [[2]](#ref2) | 182,822 | | **Free-form/Single-turn QA** | LiveQA [[3]](#ref3) | 633 | | | MedicationQA [[4]](#ref4) | 689 | | | ChatDoctor-cleaned† [[5]](#ref5) | 111,902 | | **Multi-turn QA** | MedQA-dialog† [[1]](#ref1) | 4,818 | | **Clinical Note Generation**| MTS-dialog [[6]](#ref6) | 1,200 | | **Miscellaneous** | MedInstruct-52K [[7]](#ref7) | 52,002 | --- ### Dataset Descriptions We provide brief descriptions below for the datasets we newly created or repurposed. For other datasets, we used the original versions; please refer to the original papers for detailed information. - **MedQA-CoT**: Chain-of-thought reasoning generated by GPT-4 for questions sourced from the MedQA training set. - **MedBooks-18-CoT**: A collection of questions generated from 18 medical textbooks, paired with corresponding chain-of-thought reasoning by GPT-4. - **ChatDoctor-cleaned**: Based on the original ChatDoctor dataset, which features real Q&A from online medical consultations. We removed noise such as greetings and closings using GPT-3.5 with manually designed in-context examples, resulting in 112K cleaned Q&A pairs. - **MedQA-dialog**: Designed for multi-turn dialogue training. Using GPT-3.5, we simulated patient-doctor conversations based on MedQA questions and their CoT answers. The doctor takes an active role in guiding the conversation, asking follow-up questions when necessary to obtain a comprehensive medical history and any missing critical information. --- ### References Please cite the following papers when using the data. <a name="ref1">[1]</a> Jin et al., *What Disease does this Patient Have? A Large-scale Open Domain Question Answering Dataset from Medical Exams*, 2021 <a name="ref2">[2]</a> Pal et al., *MedMCQA: A Large-scale Multi-Subject Multi-Choice Dataset for Medical domain Question Answering*, 2022 <a name="ref3">[3]</a> Abacha et al., *Overview of the Medical Question Answering Task at TREC 2017 LiveQA*, 2017 <a name="ref4">[4]</a> Abacha et al., *Bridging the Gap Between Consumers' Medication Questions and Trusted Answers*, 2019 <a name="ref5">[5]</a> Li et al., *ChatDoctor: A Medical Chat Model Fine-Tuned on a Large Language Model Meta-AI (LLaMA) Using Medical Domain Knowledge*, 2023 <a name="ref6">[6]</a> Abacha et al., *An Empirical Study of Clinical Note Generation from Doctor-Patient Encounters*, 2023 <a name="ref7">[7]</a> Zhang et al., *AlpaCare:Instruction-tuned Large Language Models for Medical Application*, 2023 Also, please see the information below to cite our paper. ```bibtex @article{kim2024small, title={Small language models learn enhanced reasoning skills from medical textbooks}, author={Kim, Hyunjae and Hwang, Hyeon and Lee, Jiwoo and Park, Sihyeon and Kim, Dain and Lee, Taewhoo and Yoon, Chanwoong and Sohn, Jiwoong and Choi, Donghee and Kang, Jaewoo}, journal={arXiv preprint arXiv:2404.00376}, year={2024} } ``` ## Contact Feel free to email `[email protected]` if you have any questions.
Aratako/Synthetic-Japanese-Roleplay-SFW-DeepSeek-V3-0324-20k
Aratako
2025-04-23T14:07:27Z
331
6
[ "task_categories:text-generation", "language:ja", "license:mit", "size_categories:10K<n<100K", "format:parquet", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us", "roleplay" ]
[ "text-generation" ]
2025-04-22T15:52:46Z
6
--- dataset_info: features: - name: major_genre dtype: string - name: minor_genre dtype: string - name: tag dtype: string - name: world_setting dtype: string - name: scene_setting dtype: string - name: user_setting dtype: string - name: assistant_setting dtype: string - name: dialogue_tone dtype: string - name: conversation_format dtype: string - name: conversations list: - name: role dtype: string - name: content dtype: string splits: - name: train num_bytes: 114882500.25261272 num_examples: 19922 download_size: 61412157 dataset_size: 114882500.25261272 configs: - config_name: default data_files: - split: train path: data/train-* license: mit task_categories: - text-generation language: - ja tags: - roleplay size_categories: - 10K<n<100K --- # Synthetic-Japanese-Roleplay-SFW-DeepSeek-V3-0324-20k ## 概要 [deepseek-ai/DeepSeek-V3-0324](https://huggingface.co/deepseek-ai/DeepSeek-V3-0324)を用いて作成した、約20000件の日本語ロールプレイの対話を収録した合成データセットです。各データは10ターンから20ターン程度あります。 ## データの詳細 各データは以下のキーを含んでいます。 - `major_genre`: ジャンル(大分類) - `minor_genre`: ジャンル(小分類) - `tag`: 年齢制限用タグ(全年齢、R-15) - `world_setting`: 舞台・世界観の設定 - `scene_setting`: 対話シーンの設定 - `user_setting`: ユーザー側のキャラクターの設定 - `assistant_setting`: アシスタント側のキャラクターの設定 - `dialogue_tone`: 対話のトーン - `conversations`: 上記設定に基づいたユーザーとアシスタントの対話(OpenAI messages形式) 設定等の情報からsystem messageを作成したうえでご利用ください。また、system messageを作成・追加し簡単にモデルの学習に使えるような形式にしたデータセットを[こちら](https://huggingface.co/datasets/Aratako/Synthetic-Japanese-Roleplay-SFW-DeepSeek-V3-0324-20k-formatted)で公開しています。 ## ライセンス MITライセンスの元配布します。
amazon-agi/SIFT-50M
amazon-agi
2025-04-23T05:08:59Z
7,339
21
[ "task_categories:audio-text-to-text", "task_categories:audio-classification", "task_categories:text-to-speech", "task_categories:audio-to-audio", "language:en", "language:de", "language:fr", "language:it", "language:es", "license:cdla-sharing-1.0", "size_categories:10M<n<100M", "format:json", "modality:text", "library:datasets", "library:dask", "library:mlcroissant", "arxiv:2504.09081", "region:us", "speech", "speech-llm", "spoken-language-understanding", "controllable-speech-synthesis", "instruction-finetuning" ]
[ "audio-text-to-text", "audio-classification", "text-to-speech", "audio-to-audio" ]
2025-03-24T07:07:49Z
10
--- license: cdla-sharing-1.0 language: - en - de - fr - it - es size_categories: - 10M<n<100M task_categories: - audio-text-to-text - audio-classification - text-to-speech - audio-to-audio pretty_name: SIFT-50M configs: - config_name: closed_ended_acoustic_level data_files: - split: train path: train/closed_ended/acoustic_level/*/*.jsonl - split: validation path: dev/closed_ended/acoustic_level/*/*.jsonl - split: EvalSIFT path: EvalSIFT/closed_ended/acoustic_level/*/*.jsonl - config_name: closed_ended_content_level data_files: - split: train path: train/closed_ended/content_level/*/*.jsonl - split: validation path: dev/closed_ended/content_level/*/*.jsonl - split: EvalSIFT path: EvalSIFT/closed_ended/content_level/*/*.jsonl - config_name: closed_ended_word_align data_files: - split: train path: train/closed_ended/word_align/*/*.jsonl - split: validation path: dev/closed_ended/word_align/*/*.jsonl - split: EvalSIFT path: EvalSIFT/closed_ended/word_align/*/*.jsonl - config_name: closed_ended_comparison data_files: - split: train path: train/closed_ended/comparison/*/*.jsonl - split: validation path: dev/closed_ended/comparison/*/*.jsonl - split: EvalSIFT path: EvalSIFT/closed_ended/comparison/*/*.jsonl - config_name: open_ended data_files: - split: train path: train/open_ended/*/*.jsonl - split: validation path: dev/open_ended/*/*.jsonl - split: EvalSIFT path: EvalSIFT/open_ended/*/*.jsonl - config_name: controllable_generation data_files: - split: train path: train/controllable_generation/*/*.jsonl - split: validation path: dev/controllable_generation/*/*.jsonl - split: EvalSIFT path: EvalSIFT/controllable_generation/*/*.jsonl tags: - speech - speech-llm - spoken-language-understanding - controllable-speech-synthesis - instruction-finetuning --- # Dataset Card for SIFT-50M SIFT-50M (Speech Instruction Fine-Tuning) is a 50-million-example dataset designed for instruction fine-tuning and pre-training of speech-text large language models (LLMs). It is built from publicly available speech corpora containing a total of 14K hours of speech and leverages LLMs and off-the-shelf expert models. The dataset spans five languages, covering diverse aspects of speech understanding and controllable speech generation instructions. SIFT-50M augments existing speech datasets with instruction-based question-answer (QA) pairs for speech understanding and includes approximately 5 million examples for controllable speech generation. For more details, refer to this paper: [SIFT-50M: A Large-Scale Multilingual Dataset for Speech Instruction Fine-Tuning](https://arxiv.org/pdf/2504.09081). ### How to Use The `datasets` library can be used to load the SIFT-50M dataset. Here’s how to load all data from the `train` split. Possible split values are `train`, `dev`, and `EvalSIFT`. ```python from datasets import load_dataset dataset = load_dataset("amazon-agi/SIFT-50M", split="train") ``` Here is how you can load all the data from a particular category. Possible category values are `closed_ended_acoustic_level`, `closed_ended_content_level`, `closed_ended_word_align`, `closed_ended_comparison`, `open_ended`, and `controllable_generation`. ```python from datasets import load_dataset dataset = load_dataset("amazon-agi/SIFT-50M", "closed_ended_acoustic_level", split="train") ``` ### Source Datasets SIFT-50M is constructed using three publicly available speech data sources: * [MultiLingual LibriSpeech](https://huggingface.co/datasets/facebook/multilingual_librispeech) (MLS) * [Common Voice Corpus 15](https://huggingface.co/datasets/mozilla-foundation/common_voice_15_0) (CV-15) * [VCTK Corpus](https://datashare.ed.ac.uk/handle/10283/2950) Users are expected to download the above data sources for audio files. We share the audio IDs of the audio files referenced in SIFT-50M. More details on this are provided in the next section. ### Dataset Structure * `train`: Contains SIFT-50M data for the train partition. * `dev`: Contains SIFT-50M data for the dev partition. * `EvalSIFT`: Contains data for benchmarking. * `audio_ids`: Contains audio IDs from each of the source datasets referenced in SIFT-50M. Users may download these audio files from the source datasets. * `pre_training`: Contains resources used for pre-training SIFT-LLM as described in the paper. It provides instruction templates for the following tasks: Emotion Recognition (ER), Speech-to-Text Translation (S2ST), Speech-to-Speech Translation (S2ST), and Text-to-Speech (TTS). Additionally, we provide the transformed [SLURP]((https://github.com/pswietojanski/slurp)) dataset for Intent Classification (IC) and Slot Entity Recognition (SER) tasks. The transformed datasets follow the same format as described in the next section. * `research`: Contains data filtered out during the quality assurance stage when ablation studies showed performance degradation on the development sets. This data consists of instructions for the word_align category, which is constructed using speech-text time alignment. It contains examples with more than two turns. ### Data Instances The SIFT-50M dataset is stored in `jsonl` format, where each example is presented in the [Messages API](https://docs.anthropic.com/en/api/messages) format, as shown in the example below: ```python { "id": "1324_1691_004352", "messages": [ { "role": "user", "content": [ {"text": null, "audio_path": "/path/to/1324_1691_004352.wav"}, {"text": "Can you comment on the speaking rate and clarity of the audio?", "audio_path": null} ] }, { "role": "assistant", "content": [ {"text": "The speaker speaks at a moderate speed and the audio has balanced clarity with a slightly close-sounding reverberation.", "audio_path": null} ] } ], "task": "closed_ended_acoustic_level", "data_source": "multilingual_librispeech_en" } ``` Each example has the following fields: * `id` (string): Uses the audio ID(s) from the source dataset. * `messages` (list[dict]): A list of messages, where each message has the following fields. All examples in SIFT-50M contain exactly two messages: * `role` (string): Takes either "user" or "assistant" as a value. In SIFT-50M, the first message has the "user" role, while the second message has the "assistant" role. * `content` (list[dict]): A list of "content" entries, where each entry has two fields: `text` and `audio_path`. Exactly one of these fields will have a non-null value, which determines the content's modality. The user is expected to update `audio_path` using the `data_source` field and the corresponding audio ID. * `data_source`: Specifies the source dataset of the audio. Possible values are: * `MLS`: multilingual_librispeech_en, multilingual_librispeech_de, multilingual_librispeech_fr, multilingual_librispeech_it, multilingual_librispeech_es * `CV-15`: common_voice_en, common_voice_de, common_voice_fr, common_voice_it, common_voice_es * `VCTK`: vctk_en ### Languages Dataset distribution by language and category: | Language | Closed-Ended | Open-Ended | Controllable Generation | |:---:|:---:|:---:|:---:| | English | 22.9M | 2.8M | 4.0M | | German | 9.6M | 684K | 450K | | French | 7.8M | 468K | 790K | | Italian | 2.2M | 257K | 72K | | Spanish | 2.9M | 190K | 236K | ### License Information The SIFT-50M dataset is released under the CDLA-Sharing-1.0 license. ### Citation Information ``` @article{pandey2025sift, title={SIFT-50M: A Large-Scale Multilingual Dataset for Speech Instruction Fine-Tuning}, author={Pandey, Prabhat and Swaminathan, Rupak Vignesh and Girish, KV and Sen, Arunasish and Xie, Jian and Strimel, Grant P and Schwarz, Andreas}, journal={arXiv preprint arXiv:2504.09081}, year={2025} } ``` If using audio from the source datasets, also cite the following papers: ``` @inproceedings{commonvoice:2020, author = {Ardila, R. and Branson, M. and Davis, K. and Henretty, M. and Kohler, M. and Meyer, J. and Morais, R. and Saunders, L. and Tyers, F. M. and Weber, G.}, title = {Common Voice: A Massively-Multilingual Speech Corpus}, booktitle = {Proceedings of the 12th Conference on Language Resources and Evaluation (LREC 2020)}, pages = {4211--4215}, year = 2020 } @article{Pratap2020MLSAL, title={MLS: A Large-Scale Multilingual Dataset for Speech Research}, author={Vineel Pratap and Qiantong Xu and Anuroop Sriram and Gabriel Synnaeve and Ronan Collobert}, journal={ArXiv}, year={2020}, volume={abs/2012.03411} } @inproceedings{Yamagishi2019CSTRVC, title={CSTR VCTK Corpus: English Multi-speaker Corpus for CSTR Voice Cloning Toolkit (version 0.92)}, author={Junichi Yamagishi and Christophe Veaux and Kirsten MacDonald}, year={2019}, url={https://api.semanticscholar.org/CorpusID:213060286} } ``` ### Contact [[email protected]](mailto:[email protected]) (Prabhat Pandey) | [[email protected]](mailto:[email protected]) (Rupak Vignesh Swaminathan) | [[email protected]](mailto:[email protected]) (K V Vijay Girish)
FreedomIntelligence/medical-o1-reasoning-SFT
FreedomIntelligence
2025-04-22T15:11:21Z
13,058
666
[ "task_categories:question-answering", "task_categories:text-generation", "language:en", "language:zh", "license:apache-2.0", "size_categories:10K<n<100K", "format:json", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "arxiv:2412.18925", "region:us", "medical", "biology" ]
[ "question-answering", "text-generation" ]
2024-12-28T03:29:08Z
null
--- license: apache-2.0 task_categories: - question-answering - text-generation language: - en - zh tags: - medical - biology configs: - config_name: en data_files: medical_o1_sft.json - config_name: zh data_files: medical_o1_sft_Chinese.json - config_name: en_mix data_files: medical_o1_sft_mix.json - config_name: zh_mix data_files: medical_o1_sft_mix_Chinese.json --- ## News [2025/04/22] We split the data and kept only the medical SFT dataset (`medical_o1_sft.json`). The file `medical_o1_sft_mix.json` contains a mix of medical and general instruction data. [2025/02/22] We released the [distilled dataset from Deepseek-R1](https://huggingface.co/datasets/FreedomIntelligence/Medical-R1-Distill-Data) based on medical verifiable problems. You can use it to initialize your models with the reasoning chain from `Deepseek-R1`. [2024/12/25] We open-sourced the medical reasoning dataset for SFT, built on medical verifiable problems and an LLM verifier. ## Introduction This dataset is used to fine-tune HuatuoGPT-o1, a medical LLM designed for advanced medical reasoning. This dataset is constructed using GPT-4o, which searches for solutions to [verifiable medical problems](https://huggingface.co/datasets/FreedomIntelligence/medical-o1-verifiable-problem) and validates them through a medical verifier. For details, see our [paper](https://arxiv.org/pdf/2412.18925) and [GitHub repository](https://github.com/FreedomIntelligence/HuatuoGPT-o1). ## Citation If you find our data useful, please consider citing our work! ``` @misc{chen2024huatuogpto1medicalcomplexreasoning, title={HuatuoGPT-o1, Towards Medical Complex Reasoning with LLMs}, author={Junying Chen and Zhenyang Cai and Ke Ji and Xidong Wang and Wanlong Liu and Rongsheng Wang and Jianye Hou and Benyou Wang}, year={2024}, eprint={2412.18925}, archivePrefix={arXiv}, primaryClass={cs.CL}, url={https://arxiv.org/abs/2412.18925}, } ```
O1-OPEN/OpenO1-SFT
O1-OPEN
2025-04-22T02:17:23Z
605
365
[ "task_categories:text-generation", "language:en", "language:zh", "license:apache-2.0", "size_categories:10K<n<100K", "format:json", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "arxiv:2504.13828", "region:us" ]
[ "text-generation" ]
2024-11-21T02:43:31Z
null
--- language: - en - zh license: apache-2.0 size_categories: - 10K<n<100K task_categories: - text-generation --- This repository contains the dataset used for fine-tuning a language model using SFT for **Chain-of-Thought Activation** from the paper [Generative AI Act II: Test Time Scaling Drives Cognition Engineering](https://huggingface.co/papers/2504.13828). Code: https://github.com/GAIR-NLP/cognition-engineering 🎉🎉🎉This repository contains the dataset used for fine-tuning a language model using SFT for **Chain-of-Thought Activation**. 🌈🌈🌈The dataset is designed to enhance the model's ability to generate coherent and logical reasoning sequences. ☄☄☄By using this dataset, the model can learn to **produce detailed and structured reasoning steps**, enhancing its performance on complex reasoning tasks. ## Statistics 1️⃣Total Records: 77,685 2️⃣Language: The dataset contains both Chinese and English data. 3️⃣Output Format: The response field uses <Thought> </Thought> and <Output> </Output> delimiters to separate the thinking process and the final answer. ## Performance 📝📝📝The following table shows the results comparison before and after SFT on the Qwen-2.5-7B-Instruct model: ### Self-Consistency <table align="center" border="1" style="margin: auto; text-align: center;"> <tr> <td>Benchmark</td> <td colspan="2">GSM8K</td> <td colspan="2">MATH</td> <td colspan="2">MMLU</td> <td colspan="2">Hellaswag</td> <td colspan="2">ARC-C</td> <td colspan="2">BBH</td> </tr> <tr> <td>Model</td> <td>Qwen2.5-7B-instruct</td> <td>OpenO1-Qwen</td> <td>Qwen2.5-7B-instruct</td> <td>OpenO1-Qwen</td> <td>Qwen2.5-7B-instruct</td> <td>OpenO1-Qwen</td> <td>Qwen2.5-7B-instruct</td> <td>OpenO1-Qwen</td> <td>Qwen2.5-7B-instruct</td> <td>OpenO1-Qwen</td> <td>Qwen2.5-7B-instruct</td> <td>OpenO1-Qwen</td> </tr> <tr> <td>Bo1</td> <td>91.21</td> <td>84.31</td> <td>69.74</td> <td>51.40</td> <td>71.51</td> <td>70.95</td> <td>67.81</td> <td>60.96</td> <td>90.44</td> <td>87.71</td> <td>54.08</td> <td>64.78</td> </tr> <tr> <td>Bo2</td> <td>91.21</td> <td>84.31</td> <td>69.74</td> <td>51.40</td> <td>71.51</td> <td>70.95</td> <td>67.81</td> <td>60.96</td> <td>90.44</td> <td>87.71</td> <td>54.08</td> <td>64.78</td> </tr> <tr> <td>Bo4</td> <td>91.74</td> <td>88.10</td> <td>71.78</td> <td>57.00</td> <td>71.57</td> <td>73.03</td> <td>68.59</td> <td>63.43</td> <td>90.53</td> <td>88.40</td> <td>55.06</td> <td>68.22</td> </tr> <tr> <td>Bo8</td> <td>91.74</td> <td>88.78</td> <td>72.84</td> <td>60.04</td> <td>71.59</td> <td>73.96</td> <td>68.75</td> <td>65.24</td> <td>90.53</td> <td>88.91</td> <td>55.39</td> <td>69.90</td> </tr> <tr> <td>Bo16</td> <td>92.12</td> <td>88.93</td> <td>73.78</td> <td>61.72</td> <td>71.56</td> <td>74.03</td> <td>68.78</td> <td>65.24</td> <td>90.53</td> <td>89.33</td> <td>55.43</td> <td>71.23</td> </tr> <tr> <td>Bo32</td> <td>92.34</td> <td>89.31</td> <td>74.14</td> <td>62.42</td> <td>71.62</td> <td>74.56</td> <td>68.83</td> <td>65.19</td> <td>90.61</td> <td>89.51</td> <td>55.69</td> <td>71.61</td> </tr> <tr> <td>Bo64</td> <td>92.57</td> <td>89.69</td> <td>74.44</td> <td>63.08</td> <td>71.63</td> <td>74.70</td> <td>68.91</td> <td>65.28</td> <td>90.61</td> <td>89.68</td> <td>55.68</td> <td>71.91</td> </tr> </table> <div style="display: flex; flex-wrap: wrap; justify-content: center;"> <div style="flex: 0 0 33%; padding: 10px;"> <img src="https://cdn-uploads.huggingface.co/production/uploads/65d2251f98b4a470bf6a26e3/4siJmlPcB11kVS4bibcpA.png" alt="Image 1" style="width: 100%; height: auto; display: block;"> </div> <div style="flex: 0 0 33%; padding: 10px;"> <img src="https://cdn-uploads.huggingface.co/production/uploads/65d2251f98b4a470bf6a26e3/BB034XOoZ9MzDy1Nsrg_z.png" alt="Image 2" style="width: 100%; height: auto; display: block;"> </div> <div style="flex: 0 0 33%; padding: 10px;"> <img src="https://cdn-uploads.huggingface.co/production/uploads/65d2251f98b4a470bf6a26e3/J5gNFrqKQkE-gPRcf-YwX.png" alt="Image 3" style="width: 100%; height: auto; display: block;"> </div> <div style="flex: 0 0 33%; padding: 10px;"> <img src="https://cdn-uploads.huggingface.co/production/uploads/65d2251f98b4a470bf6a26e3/cFiJmuK4bnX890TIN4Q0R.png" alt="Image 4" style="width: 100%; height: auto; display: block;"> </div> <div style="flex: 0 0 33%; padding: 10px;"> <img src="https://cdn-uploads.huggingface.co/production/uploads/65d2251f98b4a470bf6a26e3/4GYhNBYvGsY4VcgT3dfdS.png" alt="Image 5" style="width: 100%; height: auto; display: block;"> </div> <div style="flex: 0 0 33%; padding: 10px;"> <img src="https://cdn-uploads.huggingface.co/production/uploads/65d2251f98b4a470bf6a26e3/MP5Z4u6JR4jCob23jPlly.png" alt="Image 6" style="width: 100%; height: auto; display: block;"> </div> </div> ### Oracle pass@k <table align="center" border="1" style="margin: auto; text-align: center;"> <tr> <td>Benchmark</td> <td colspan="2">GSM8K</td> <td colspan="2">MATH</td> <td colspan="2">MMLU</td> <td colspan="2">Hellaswag</td> <td colspan="2">ARC-C</td> <td colspan="2">BBH</td> </tr> <tr> <td>Model</td> <td>Qwen2.5-7B-instruct</td> <td>OpenO1-Qwen</td> <td>Qwen2.5-7B-instruct</td> <td>OpenO1-Qwen</td> <td>Qwen2.5-7B-instruct</td> <td>OpenO1-Qwen</td> <td>Qwen2.5-7B-instruct</td> <td>OpenO1-Qwen</td> <td>Qwen2.5-7B-instruct</td> <td>OpenO1-Qwen</td> <td>Qwen2.5-7B-instruct</td> <td>OpenO1-Qwen</td> </tr> <tr> <td>Bo1</td> <td>91.21</td> <td>84.31</td> <td>69.74</td> <td>51.40</td> <td>71.51</td> <td>70.95</td> <td>67.81</td> <td>60.96</td> <td>90.44</td> <td>87.71</td> <td>54.08</td> <td>64.78</td> </tr> <tr> <td>Bo2</td> <td>93.10</td> <td>89.61</td> <td>74.40</td> <td>61.26</td> <td>71.98</td> <td>78.94</td> <td>69.61</td> <td>72.06</td> <td>90.61</td> <td>92.41</td> <td>58.30</td> <td>74.33</td> </tr> <tr> <td>Bo4</td> <td>94.84</td> <td>92.49</td> <td>78.34</td> <td>68.46</td> <td>72.41</td> <td>84.31</td> <td>71.26</td> <td>80.28</td> <td>90.78</td> <td>95.05</td> <td>61.01</td> <td>80.29</td> </tr> <tr> <td>Bo8</td> <td>95.68</td> <td>94.16</td> <td>81.86</td> <td>73.78</td> <td>72.75</td> <td>88.33</td> <td>72.23</td> <td>85.84</td> <td>90.96</td> <td>96.59</td> <td>63.67</td> <td>83.85</td> </tr> <tr> <td>Bo16</td> <td>95.83</td> <td>95.22</td> <td>84.12</td> <td>78.58</td> <td>73.02</td> <td>91.16</td> <td>72.92</td> <td>89.64</td> <td>90.96</td> <td>97.27</td> <td>65.32</td> <td>85.74</td> </tr> <tr> <td>Bo32</td> <td>96.44</td> <td>96.13</td> <td>85.72</td> <td>82.48</td> <td>73.19</td> <td>92.98</td> <td>73.46</td> <td>92.47</td> <td>90.96</td> <td>97.78</td> <td>66.79</td> <td>87.01</td> </tr> <tr> <td>Bo64</td> <td>96.82</td> <td>96.36</td> <td>87.02</td> <td>85.76</td> <td>73.34</td> <td>94.32</td> <td>73.85</td> <td>94.46</td> <td>90.96</td> <td>98.21</td> <td>67.80</td> <td>88.09</td> </tr> </table> <div style="display: flex; flex-wrap: wrap; justify-content: center;"> <div style="flex: 0 0 33%; padding: 10px;"> <img src="https://cdn-uploads.huggingface.co/production/uploads/65d2251f98b4a470bf6a26e3/Zsf9Y-gGFwnBHYpSvmf8Y.png" alt="Image 1" style="width: 100%; height: auto; display: block;"> </div> <div style="flex: 0 0 33%; padding: 10px;"> <img src="https://cdn-uploads.huggingface.co/production/uploads/65d2251f98b4a470bf6a26e3/-Ei4sEtgdMC8CA4VqF7fD.png" alt="Image 2" style="width: 100%; height: auto; display: block;"> </div> <div style="flex: 0 0 33%; padding: 10px;"> <img src="https://cdn-uploads.huggingface.co/production/uploads/65d2251f98b4a470bf6a26e3/6urorId1Dt2DAtPoJ7i7W.png" alt="Image 3" style="width: 100%; height: auto; display: block;"> </div> <div style="flex: 0 0 33%; padding: 10px;"> <img src="https://cdn-uploads.huggingface.co/production/uploads/65d2251f98b4a470bf6a26e3/PkmzLzo88y1RLS3oOyr8o.png" alt="Image 4" style="width: 100%; height: auto; display: block;"> </div> <div style="flex: 0 0 33%; padding: 10px;"> <img src="https://cdn-uploads.huggingface.co/production/uploads/65d2251f98b4a470bf6a26e3/lA9beyxd26NpN8YMau0XP.png" alt="Image 5" style="width: 100%; height: auto; display: block;"> </div> <div style="flex: 0 0 33%; padding: 10px;"> <img src="https://cdn-uploads.huggingface.co/production/uploads/65d2251f98b4a470bf6a26e3/OlWqDOgC0KBovIW7bT_oE.png" alt="Image 6" style="width: 100%; height: auto; display: block;"> </div> </div>
zwq2018/embodied_reasoner
zwq2018
2025-04-21T13:49:38Z
1,895
14
[ "task_categories:image-text-to-text", "language:en", "size_categories:10K<n<100K", "modality:image", "arxiv:2503.21696", "region:us", "Reasoning", "Long CoT", "Multimodal Reasoning", "Embodied", "Robot", "AI2-THOR" ]
[ "image-text-to-text" ]
2025-03-19T14:23:34Z
2
--- language: - en size_categories: - 10K<n<100K task_categories: - image-text-to-text tags: - Reasoning - Long CoT - Multimodal Reasoning - Embodied - Robot - AI2-THOR --- # Embodied-Reasoner Dataset ## Dataset Overview Embodied-Reasoner is a multimodal reasoning dataset designed for embodied interactive tasks. It contains 9,390 **Observation-Thought-Action** trajectories for training and evaluating multimodal models capable of performing complex embodied tasks in indoor environments. <img src="./embodied_reasoner.JPG" alt="Image" style="width: 700px;"> [![arXiv](https://img.shields.io/badge/arXiv-Paper-<COLOR>.svg)](https://arxiv.org/abs/2503.21696) [![Project](https://img.shields.io/badge/Project-Website-blue.svg)](https://embodied-reasoner.github.io/) [![GitHub](https://img.shields.io/badge/GitHub-Code-181717?logo=github)](https://github.com/zwq2018/embodied_reasoner) ## Key Features - 📸 **Rich Visual Data**: Contains 64,000 first-person perspective interaction images - 🤔 **Deep Reasoning Capabilities**: 8 million thought tokens covering analysis, spatial reasoning, reflection, and planning - 🏠 **Diverse Environments**: Spans 107 different indoor scenes (kitchens, living rooms, etc.) - 🎯 **Rich Interactive Objects**: Includes 2,100 interactive objects and 2,600 container objects - 🔄 **Complete Interaction Trajectories**: Each sample contains full observation-thought-action sequences ## Dataset Statistics ### Training Data (9,390 samples) - `data_engine_origin_trajectory_9390.json`: Original 9,390 generated trajectories - `train_multiturn_9390.json`: 9,390 trajectories converted to multi-turn dialogue training format - `data/images`: Images corresponding to the trajectories ### Evaluation Data (809 test cases) - `test_809.json`: The task name, key actions, and other information for the evaluation set. - `agent_positions.json`: Auxiliary position information for the evaluation set <img src="./table.jpg" alt="Image" style="width: 700px;"> ## How to use our dataset - The multi-turn dialogue training corpus corresponds to `train_multiturn_9390.json`, which comes from `data_engine_origin_trajectory_9390.json`. - All images are stored in `data/images`. You need to extract all zip files. ### Learning about `data_engine_origin_trajectory_9390.json` This json file contains all the synthesized trajectory data. Each trajectory is **observatio-thought-action** interleaved, and thought includes multiple different thinking patterns, such as *situation analysis, self-reflection, spatial reasoning, verification, and task planning*. <img src="./example.jpg" alt="Image" style="width: 900px;"> For `data_engine_origin_trajectory_9390.json`, each trajectory includes the following components: ```json { "id": 1, "tasktype": "transport", "sub_tasktype": "exposed_to_enclosed_object_transfer", "task_templates": "put the KeyChain in the Drawer", "scene": "FloorPlan208", "key_actions": [ { "action": "navigate to", "objectId": "SideTable|+01.83|+00.00|+01.09", "objectType": "SideTable", "reward": 1, "relatedObject": [ "SideTable|+01.83|+00.00|+01.09", "KeyChain|+01.79|+00.87|+01.40" ] }, ... { "action": "end", "objectId": "", "objectType": "", "reward": 1, "relatedObject": [ "SideTable|+01.83|+00.00|+01.09", "KeyChain|+01.79|+00.87|+01.40", "Drawer|+01.76|+00.77|+01.09" ] } ], "totalreward": 6, "trajectory": [ "<situation analysis> .... </situation analysis>", "<spatial reasoning> ... </spatial reasoning>", "<Planning> .... </Planning>", "<DecisionMaking>navigate to SideTable</DecisionMaking>", .... "<DecisionMaking>put in Drawer</DecisionMaking>", "<self-reflection> .... .</self-reflection>" ], "images": [ init_image, image1,image2,... ], "query": [ "Would you be able to put the KeyChain in the Drawer for me?", "I want to put a KeyChain in the Drawer, can you help me?", ... ] }, ``` Format: - **"trajectory":** Record the entire trajectory, including multiple thoughts (< situation analysis >, < spatial reasoning >,..), decision-making actions (< DecisionMaking >) at each step. - **"images"**: At initialization, AI2THOR provides an initial observation image `init_image`, and each interaction (< DecisionMaking >) generates `new images`, which are stored in images in order. - **"tasktype"** and "sub_tasktype": There are four task types and ten sub-task types in total, - `search`: includes exposed_object_search, enclosed_object_search - `manipulate`: includes exposed_object_toggle, exposed_object_grasping, enclosed_object_grasping - `transport`: includes exposed_to_exposed_object_transfer, exposed_to_enclosed_object_transfer, enclosed_to_exposed_object_transfer, enclosed_to_enclosed_object_transfer - `composite`: includes sequential_object_transfer - **"key_action"**: The key actions required to complete the task, including each action’s name, associated object ID, object type, action reward, and related objects. - **"scene"**: The scene ID in AI2-THOR corresponding to the task. - **"task_templates"**: The template used for instruction generation, e.g., put <object> in <container> . - **"query"**: The instructions generated based on the template, e.g.,"Would you be able to put the KeyChain in the Drawer for me?". ### Learning about `train_multiturn_9390.json` `train multiturn 9390.json` is converted from `data_engine_origin_trajectory_9390.json.` Specifically, we designed embodied-system-prompt and user-interaction-prompt to construct messages data. And replace the five types of thoughts tags and decision tags in the trajectory with human colloquial descriptions. ```json { "messages": [ { "role": "system", "content": "You are a robot in given room. You need to complete the tasks according to human instructions. We provide an Available_Actions set and the corresponding explanations for each action. Each step, you should select one action from Available_Actions." }, { "role": "user", "content": "<image>This is an image from your frontal perspective. Please select an action from the Available_Actions and fill in the arguments. Task: \"Can I ask you to put the KeyChain in the Drawer?\" Available_Actions: {{\"navigate to <object>\": Move to the object. \"pickup <object>\": Pick up the object. \"put in <object>\": Put the item in your hand into or on the object ... you can think, plan, and even reflect step by step, and then output your final action..." }, { "role": "assistant", "content": "Hmm... let me carefully observe what's around. I see a Sofa positioned on the right side of the room with a FloorLamp nearby, casting light above it. An ArmChair is in the center-left, facing the Sofa. A TVStand is situated on the far left, holding a Television. A Painting is hung on the wall at the back, above a red Ottoman. A HousePlant is next to the Ottoman, adding greenery to the space. Hold on, I want to think this over. Based on the observation, the KeyChain is likely to be found on surfaces where small objects are commonly placed, such as tables or seating areas. Observing the room, potential search locations include a SideTable, the ArmChair, and the Sofa..." }, { "role": "user", "content":"..." }, { "role": "assistant", "content":"..." }, ... ], "images": [ init_image, image1,image2,... ], ] ``` To see how the dataset is used during training, please visit our [GitHub](https://github.com/zwq2018/embodied_reasoner). ## Use Cases This dataset is suitable for: 1. 🤖 Training embodied AI agents 2. 🧠 Vision-language reasoning research 3. 🏃 Long-horizon task planning 4. 🔍 Visual search tasks 5. 🤝 Multi-turn interaction systems ## Citation If you use this dataset in your research, please cite our paper: ```bibtex @article{embodied-reasoner, title = {Embodied-Reasoner: Synergizing Visual Search, Reasoning, and Action for Embodied Interactive Tasks}, author = {Wenqi Zhang and Mengna Wang and Gangao Liu and Huixin Xu and Yiwei Jiang and Yongliang Shen and Guiyang Hou and Zhe Zheng and Hang Zhang and Xin Li and Weiming Lu and Peng Li and Yueting Zhuang}, journal = {arXiv preprint arXiv:2503.21696}, year = {2025} } ``` ## License This dataset is released under the Apache 2.0 License. ## Links - 📑 [Paper](https://arxiv.org/abs/2503.21696) - 💻 [GitHub Repository](https://github.com/zwq2018/embodied_reasoner) - 🌐 [Project Website](https://embodied-reasoner.github.io) ## Contact For questions, please contact: - [email protected] - [email protected]
livecodebench/code_generation_lite
livecodebench
2025-04-21T02:23:51Z
62,850
40
[ "license:cc", "size_categories:n<1K", "arxiv:2403.07974", "region:us", "code", "code generation" ]
[]
2024-04-16T04:46:53Z
null
--- license: cc tags: - code - code generation pretty_name: LiveCodeBench size_categories: - n<1K --- ## LiveCodeBench: Holistic and Contamination Free Evaluation of Large Language Models for Code <p align="center"> <a href="https://livecodebench.github.io/">🏠 Home Page</a> • <a href="https://github.com/LiveCodeBench/LiveCodeBench">💻 GitHub Repository </a> • <a href="https://livecodebench.github.io/leaderboard.html">🏆 Leaderboard</a> • <a href="https://arxiv.org/abs/2403.07974">📄 Paper </a> </p> ![LiveCodeBench](images/lcb.png) ## Change Log Since LiveCodeBench is a continuously updated benchmark, we provide different versions of the dataset. Particularly, we provide the following versions of the dataset: - `release_v1`: The initial release of the dataset with problems released between May 2023 and Mar 2024 containing 400 problems. - `release_v2`: The updated release of the dataset with problems released between May 2023 and May 2024 containing 511 problems. - `release_v3`: The updated release of the dataset with problems released between May 2023 and Jul 2024 containing 612 problems. - `release_v4`: The updated release of the dataset with problems released between May 2023 and Sep 2024 containing 713 problems. - `release_v5`: The updated release of the dataset with problems released between May 2023 and Jan 2025 containing 880 problems. You can use the `version_tag` argument to load the desired version of the dataset. Additionally, you can use version tags like `v1`, `v2`, `v1_v3`, `v4_v5` to get the problems released in a specific version. ## Dataset Description LiveCodeBench is a "live" updating benchmark for holistically evaluating code related capabilities of LLMs. Particularly, it evaluates LLMs across a range of capabilties including code generation, self-repair, test output prediction, and code execution. This is the code generation scenario of LiveCodeBench. It is also used for evaluating self-repair using test case feedback. LiveCodeBench problems are collected from competition programming websites with particular focus on maintaining problem quality, test case quality, and problem difficulty diversity. This scenario currently hosts over 500 problems from LeetCode, AtCoder, and Codeforces. Each problem instance is consists of problem description, input/output examples, and hidden test cases. Additionally, every problem is tagged with its difficulty level and release date which allows measuring model performance across different time windows. The goal is to generate a correct and efficient solution for each problem instance. The initial code_generation dataset included larger number of test cases which leads to substantially large dataset size. This (lite) version has pruned and sampled tests while trying to ensure similar performances with the original dataset. Going forward, livecodebench will be using this lite version for code generation evaluations. ## Usage You can use the dataset by loading it from the Hugging Face datasets library. Additionally, the version tag "release_v1" is used to specify the (temporal) version of the dataset. "v1" corresponds to the initial release of the dataset and "release_v2" is the second version. ```python from datasets import load_dataset lcb_codegen = load_dataset("livecodebench/code_generation_lite", version_tag="release_v2") ```
ServiceNow-AI/DNRBench
ServiceNow-AI
2025-04-18T08:44:24Z
47
2
[ "language:en", "license:cc-by-nc-4.0", "size_categories:n<1K", "format:parquet", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "arxiv:2503.15793", "region:us" ]
[]
2025-02-10T15:29:38Z
2
--- dataset_info: - config_name: imaginary-reference features: - name: role dtype: string - name: content dtype: string splits: - name: test num_bytes: 4485 num_examples: 25 download_size: 4391 dataset_size: 4485 - config_name: indifferent features: - name: role dtype: string - name: content dtype: string splits: - name: test num_bytes: 11732 num_examples: 25 download_size: 10536 dataset_size: 11732 - config_name: math features: - name: role dtype: string - name: content dtype: string splits: - name: test num_bytes: 5440 num_examples: 25 download_size: 4740 dataset_size: 5440 - config_name: redundant features: - name: role dtype: string - name: content dtype: string splits: - name: test num_bytes: 5087 num_examples: 25 download_size: 4096 dataset_size: 5087 - config_name: unanswerable features: - name: role dtype: string - name: content dtype: string splits: - name: test num_bytes: 12501 num_examples: 50 download_size: 8242 dataset_size: 12501 configs: - config_name: imaginary-reference data_files: - split: test path: imaginary-reference/test-* - config_name: indifferent data_files: - split: test path: indifferent/test-* - config_name: math data_files: - split: test path: math/test-* - config_name: redundant data_files: - split: test path: redundant/test-* - config_name: unanswerable data_files: - split: test path: unanswerable/test-* license: cc-by-nc-4.0 language: - en --- # DNR Bench Don’t Reason Bench (DNR Bench), a novel benchmark designed to expose a vulnerability in current RLMs: their tendency to over-reason by attempting to solve unsolvable problems, leading to excessively long responses. # Data Summary The DNR Bench dataset contains 150 adversarially crafted prompts divided into five distinct categories: - Imaginary Reference - Indifferent - Math, - Redundant, - Unanswerable. Each category targets a specific failure mode observed in reasoning-optimized LLMs, such as hallucinating nonexistent references, failing to remain neutral in ambiguous contexts, incorrectly solving flawed math problems, overanalyzing redundant information, or answering questions that lack sufficient data. # Leaderboard This dataset is used to test reasoning LLMs in [DNR Leaderboard on Huggingface](https://huggingface.co/spaces/ServiceNow-AI/Do-not-reason-bench) # Citation ```bibtex @misc{hashemi2025dnrbenchbenchmarkingoverreasoning, title={DNR Bench: Benchmarking Over-Reasoning in Reasoning LLMs}, author={Masoud Hashemi and Oluwanifemi Bamgbose and Sathwik Tejaswi Madhusudhan and Jishnu Sethumadhavan Nair and Aman Tiwari and Vikas Yadav}, year={2025}, eprint={2503.15793}, archivePrefix={arXiv}, primaryClass={cs.LG}, url={https://arxiv.org/abs/2503.15793}, } ```
MarcPal08/sentiment-analysis-test
MarcPal08
2025-04-16T12:51:30Z
505
9
[ "task_categories:text-classification", "task_ids:sentiment-analysis", "annotations_creators:expert-generated", "annotations_creators:crowdsourced", "language_creators:crowdsourced", "multilinguality:monolingual", "source_datasets:original", "language:it", "license:mit", "size_categories:n<1K", "format:parquet", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us", "school", "high-school" ]
[ "text-classification" ]
2025-04-16T11:57:04Z
9
--- dataset_info: features: - name: text dtype: string - name: sentiment dtype: string splits: - name: train num_bytes: 28302.111747851002 num_examples: 279 - name: test num_bytes: 7100.888252148997 num_examples: 70 download_size: 23157 dataset_size: 35403.0 configs: - config_name: default data_files: - split: train path: data/train-* - split: test path: data/test-* annotations_creators: - expert-generated - crowdsourced language: - it language_creators: - crowdsourced license: - mit multilinguality: - monolingual pretty_name: A sentiment analisys database created in a school environment. size_categories: - n<1K source_datasets: - original tags: - school - high-school task_categories: - text-classification task_ids: - sentiment-analysis --- # Progetto scolastico per l'analisi dei sentimenti Il dataset è stato creato con un questionario online in cui si chiedeva ad un pubblico di studenti, docenti, personale amministrativo, famiglie di rispondere ad alcune domande sul loro rapporto con la scuola. Le annotazioni sono state effettuate correlando le risposte testuali ad indicatori di gradimento. Il dataset è stato realizzato all'interno di un corso pomeridiano scolastico dedicato all'intelligenza artificiale. Grazie a tutti per la collaborazione ❤️
Cocciadipollo/sentiment-analysis-test
Cocciadipollo
2025-04-16T12:51:27Z
405
7
[ "task_categories:text-classification", "task_ids:sentiment-analysis", "annotations_creators:expert-generated", "annotations_creators:crowdsourced", "language_creators:crowdsourced", "multilinguality:monolingual", "source_datasets:original", "language:it", "license:mit", "size_categories:n<1K", "format:parquet", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us", "school", "high school" ]
[ "text-classification" ]
2025-04-16T11:47:01Z
7
--- dataset_info: features: - name: text dtype: string - name: sentiment dtype: string splits: - name: train num_bytes: 28302.111747851002 num_examples: 279 - name: test num_bytes: 7100.888252148997 num_examples: 70 download_size: 23157 dataset_size: 35403.0 configs: - config_name: default data_files: - split: train path: data/train-* - split: test path: data/test-* annotations_creators: - expert-generated - crowdsourced language: - it language_creators: - crowdsourced license: - mit multilinguality: - monolingual pretty_name: A sentiment analisys database created in a school environment. size_categories: - n<1K source_datasets: - original tags: - school - high school task_categories: - text-classification task_ids: - sentiment-analysis --- # Progetto scolastico per l'analisi dei sentimenti Il dataset è stato creato con un questionario online in cui si chiedeva ad un pubblico di studenti, docenti, personale amministrativo, famiglie di rispondere ad alcune domande sul loro rapporto con la scuola. le annotazioni sono state effetuate coorelando le risposte testuali ad indicatori di gradimento. Il dataaset è stato realizzato all'interno di un corso pomeridiano scolastico dedicato all'IA. Grazie a tutti per la coollaborazione! ❤️
Merlinooooo/sentiment-analysis-test
Merlinooooo
2025-04-16T12:51:25Z
388
7
[ "task_categories:text-classification", "task_ids:sentiment-analysis", "annotations_creators:expert-generated", "annotations_creators:crowdsourced", "language_creators:crowdsourced", "multilinguality:monolingual", "source_datasets:original", "language:it", "license:mit", "size_categories:n<1K", "format:parquet", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "region:us", "school", "high-school" ]
[ "text-classification" ]
2025-04-16T11:47:06Z
7
--- dataset_info: features: - name: text dtype: string - name: sentiment dtype: string splits: - name: train num_bytes: 28302.111747851002 num_examples: 279 - name: test num_bytes: 7100.888252148997 num_examples: 70 download_size: 23157 dataset_size: 35403.0 configs: - config_name: default data_files: - split: train path: data/train-* - split: test path: data/test-* annotations_creators: - expert-generated - crowdsourced language: - it language_creators: - crowdsourced license: - mit multilinguality: - monolingual pretty_name: A sentiment analysis database created in a school environment size_categories: - n<1K source_datasets: - original tags: - school - high-school task_categories: - text-classification task_ids: - sentiment-analysis --- # Progetto scolastico per l'analisi dei sentimenti Il dataset è stato creato in un questionario online in cui si chiedeva ad un pubblico di studenti, docenti, personale amministrativo, famiglie di rispondere ad alcune domande sul loro rapporto con la scuola. Le annotazioni sono state effettuate correlando le risposte testuali ad indicatori di gradimento. il dataset è stato realizzato all'interno di un corso pomeridiano scolastico dedicato all'intelligenza artificiale Grazie a tutti per la collaborazione ❤️
CohereLabs/aya_dataset
CohereLabs
2025-04-15T08:51:55Z
1,341
302
[ "task_categories:other", "annotations_creators:crowdsourced", "annotations_creators:expert-generated", "language_creators:crowdsourced", "language_creators:expert-generated", "multilinguality:multilingual", "source_datasets:original", "language:amh", "language:arb", "language:ary", "language:ars", "language:acq", "language:arz", "language:apc", "language:ben", "language:ceb", "language:dan", "language:deu", "language:ell", "language:eng", "language:eus", "language:fil", "language:fin", "language:fra", "language:gle", "language:guj", "language:hat", "language:hau", "language:hin", "language:hun", "language:ibo", "language:ind", "language:ita", "language:jav", "language:jpn", "language:kan", "language:kir", "language:kor", "language:kur", "language:lit", "language:mal", "language:mar", "language:mlg", "language:msa", "language:mya", "language:nep", "language:nld", "language:nso", "language:nya", "language:pan", "language:pes", "language:pol", "language:por", "language:pus", "language:rus", "language:sin", "language:sna", "language:snd", "language:som", "language:spa", "language:sqi", "language:srp", "language:sun", "language:swa", "language:swe", "language:tam", "language:tel", "language:tha", "language:tur", "language:ukr", "language:urd", "language:vie", "language:wol", "language:xho", "language:yor", "language:zho", "language:zul", "license:apache-2.0", "size_categories:100K<n<1M", "format:parquet", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "arxiv:2402.06619", "region:us" ]
[ "other" ]
2024-01-31T21:40:16Z
null
--- annotations_creators: - crowdsourced - expert-generated language_creators: - crowdsourced - expert-generated language: - amh - arb - ary - ars - acq - arz - apc - ben - ceb - dan - deu - ell - eng - eus - fil - fin - fra - gle - guj - hat - hau - hin - hun - ibo - ind - ita - jav - jpn - kan - kir - kor - kur - lit - mal - mar - mlg - msa - mya - nep - nld - nso - nya - pan - pes - pol - por - pus - rus - sin - sna - snd - som - spa - sqi - srp - sun - swa - swe - tam - tel - tha - tur - ukr - urd - vie - wol - xho - yor - zho - zul license: apache-2.0 multilinguality: - multilingual size_categories: - 100K<n<1M source_datasets: - original task_categories: - other task_ids: [] pretty_name: Aya Dataset dataset_info: - config_name: default features: - name: inputs dtype: string - name: targets dtype: string - name: language dtype: string - name: language_code dtype: string - name: annotation_type dtype: string - name: user_id dtype: string splits: - name: test num_bytes: 1782208 num_examples: 1750 - name: train num_bytes: 254591851 num_examples: 202362 download_size: 275359572 dataset_size: 256374059 - config_name: demographics features: - name: user_id dtype: string - name: age_range sequence: int64 - name: gender dtype: string - name: country dtype: string - name: languages sequence: string - name: dialects sequence: string splits: - name: train num_bytes: 202127 num_examples: 1456 download_size: 113702 dataset_size: 202127 configs: - config_name: default data_files: - split: train path: data/train-* - split: test path: data/test-* - config_name: demographics data_files: - split: train path: demographics/train-* tags: [] --- ![Aya Header](https://huggingface.co/datasets/CohereLabs/aya_dataset/resolve/main/aya_header.png) # Dataset Summary The `Aya Dataset` is a multilingual instruction fine-tuning dataset curated by an open-science community via [Aya Annotation Platform](https://aya.for.ai/) from Cohere Labs. The dataset contains a total of 204k human-annotated prompt-completion pairs along with the demographics data of the annotators.<br> This dataset can be used to train, finetune, and evaluate multilingual LLMs. - **Curated by:** Contributors of [Aya Open Science Intiative](https://aya.for.ai/). - **Language(s):** 65 languages (71 including dialects & scripts). - **License:** [Apache 2.0](https://opensource.org/license/apache-2-0) - **Aya Datasets Family:** | Name | Explanation | |------|--------------| | [aya_dataset](https://huggingface.co/datasets/CohereLabs/aya_dataset) | Human-annotated multilingual instruction finetuning dataset, comprising over 204K instances across 65 languages. | | [aya_collection](https://huggingface.co/datasets/CohereLabs/aya_collection) | Created by applying instruction-style templates from fluent speakers to 44 datasets, including translations of 19 instruction-style datasets into 101 languages, providing 513M instances for various tasks.| | [aya_collection_language_split](https://huggingface.co/datasets/CohereLabs/aya_collection_language_split) | Aya Collection structured based on language level subsets. | | [aya_evaluation_suite](https://huggingface.co/datasets/CohereLabs/aya_evaluation_suite) | A diverse evaluation set for multilingual open-ended generation, featuring 250 culturally grounded prompts in 7 languages, 200 translated prompts in 24 languages, and human-edited versions selected for cross-cultural relevance from English Dolly in 6 languages.| | [aya_redteaming](https://huggingface.co/datasets/CohereLabs/aya_redteaming)| A red-teaming dataset consisting of harmful prompts in 8 languages across 9 different categories of harm with explicit labels for "global" and "local" harm.| # Dataset The `Aya Dataset` comprises of two types of data: 1. **Human Annotations:** Original annotations (brand new prompts and completions written by annotators) and re-annotations (human edits of automatically generated prompts and completions). 2. **Demographics Data:** Anonymized information for each annotator. ## Load with Datasets To load this dataset consisting of both prompt-completions and demographics data with `datasets`, you'll just need to install Datasets as `pip install datasets --upgrade` and then use the following code: ```python from datasets import load_dataset # Load the annotations dataset aya_dataset = load_dataset("CohereLabs/aya_dataset") # Load the demographics dataset aya_demographics = load_dataset("CohereLabs/aya_dataset", "demographics") ``` ## Data Fields ### Human Annotations (Default) The data fields are the same among all splits: - `inputs`: Prompt or input to the language model. - `targets`: Completion or output of the language model. - `language`: The language of the `inputs` and `targets`. - `language_code`: The ISO code for the language of the `inputs` and `targets`. - `annotation_type`: The value denoting whether `inputs` and `targets` are 'original_annotations' or 're-annotations'. - `user_id`: Unique identifier of the annotator who submitted the prompt-completion pair. ### Demographics Data The data fields are the same among all splits: - `user_id`: Unique identifier of the annotator who submitted the prompt-completion pair. - `age_range`: Age of the annotator. Ranges from 0 to 121. - `gender`: Gender of the annotator. The values are 'male', 'female', 'prefer not to say', 'non-binary' and 'others'. - `languages`: List of languages spoken by the annotator. - `dialects`: Dialects reported by the annotator. Some empty values may be represented as 'null'. ## Data Splits ### Human Annotations (Default) The following are the splits of the data: | Split | No. of instances | Language Coverage | |-------|------------------|-------------------| | train | 202,364 | All | | test | 1,750 | 7 ('Standard Arabic', 'Yoruba', 'Turkish', 'English', 'Simplified Chinese', 'Portuguese', 'Telugu')| ### Demographics Data The following are the splits of the data: | Split | No. of Instances | |-------|------------------| | train | 1,456 | ## Data Instances ### Human Annotations (Default) An example of `train` looks as follows: ```json { "inputs": "What cultural events or festivals add vibrancy to Colombo's calendar...", "targets": "Colombo's cultural calendar is adorned with diverse events and festivals that celebrate the city's rich tapestry of traditions...", "language": "English", "language_code": "eng", "annotation_type": "original-annotations", "user_id": "f0ff69570af705b75c5a0851883e..." } ``` ### Demographics Data An example of `train` looks as follows: ```json { "user_id": "f0ff69570af705b75c5a0851883e...", "age_range": [ 25, 35 ], "gender": "female", "languages": [ "English", "Hausa" ], "dialects": [ "Hausa" ] } ``` ## Statistics ### Annotation Types The following is the breakdown of original annotations and re-annotations in the final dataset. | Type of Annotation | Instances | |--------------------|-----------| | Original Annotations | 138,844 | | Re-Annotations | 65,270 | | Total | 204,114| ### Languages The dataset covers 65 languages: 28 high-resource, 12 mid-resource, and 31 low-resource languages. The following is details about the languages, dialects & scripts included in the dataset. <details> <summary> Languages Info </summary> | ISO Code | Language | Resources | |----------|----------|-----------| | `amh` | Amharic | Low | | `arb`, `ary`, `ars`, `acq`, `arz` & `apc` | Arabic (Standard, Moroccan, Najdi, Ta'izzi-Adeni, Egyptian & South Levantine) | High | | `ben` | Bengali | Mid | | `ceb` | Cebuano | Mid | | `dan` | Danish | Mid | | `deu` | German | High | | `ell` | Greek | Mid | | `eng` | English | High | | `eus` | Basque | High | | `fil` | Filipino | Mid | | `fin` | Finnish | Mid | | `fra` | French | High | | `gle` | Irish | Low | | `guj` | Gujarati | Low | | `hat` | Haitian Creole | Low | | `hau` | Hausa | Low | | `hin` | Hindi | High | | `hun` | Hungarian | High | | `ibo` | Igbo | Low | | `ind` | Indonesian | Mid | | `ita` | Italian | High | | `jav` | Javanese | Low | | `jpn` | Japanese | High | | `kan` | Kannada | Low | | `kir` | Kyrgyz | Low | | `kor` | Korean | Mid | | `kur` | Kurdish | Low | | `lit` | Lithuanian | Mid | | `mal` | Malayalam | Low | | `mar` | Marathi | Low | | `mlg` | Malagasy | Low | | `msa` | Malay | Mid | | `mya` | Burmese | Low | | `nep` | Nepali | Low | | `nld` | Dutch | High | | `nso` | Northern Sotho | Low | | `nya` | Chichewa | Low | | `pan` | Punjabi | Low | | `pes` | Persian | High | | `pol` | Polish | High | | `por` | Portuguese | High | | `pus` | Pashto | Low | | `rus` | Russian | High | | `sin` | Sinhala | Low | | `sna` | Shona | Low | | `snd` | Sindhi | Low | | `som` | Somali | Low | | `spa` | Spanish | High | | `sqi` | Albanian | Low | | `srp` | Serbian | High | | `sun` | Sundanese | Low | | `swa` | Swahili | Low | | `swe` | Swedish | High | | `tam` | Tamil | Mid | | `tel` | Telugu | Low | | `tha` | Thai | Mid | | `tur` | Turkish | High | | `ukr` | Ukrainian | Mid | | `urd` | Urdu | Mid | | `vie` | Vietnamese | High | | `wol` | Wolof | Low | | `xho` | Xhosa | Low | | `yor` | Yorùbá | Low | | `zho` | Chinese (Traditional & Simplified) | High | | `zul` | Zulu | Low | </details> <br> # Motivations & Intentions - **Curation Rationale:** The curation effort employed an open-science approach to create a diverse instruction-style dataset through annotators across the globe that ensures comprehensive representation across all languages. The success of the curation effort, led by volunteers across diverse backgrounds, was significantly influenced by their hope to meaningfully bring NLP advancements to their languages. # Known Limitations - **Language and dialect coverage:** The dataset covers a limited fraction of the world's linguistic diversity, with 93% of languages not represented, facing challenges in distinguishing between languages and dialects, lacking coverage for many regional dialects, and excluding programming languages. - **Uneven distribution of contributions:** The dataset contains contributions in annotation activities, with a 'long tail' of annotators making only one or two contributions, leading to potential dataset imbalances across languages and a lack of diversity within certain language annotations. - **Cultural and Personal Bias:** In the dataset, certain languages have limited representation due to a few dominant annotators, potentially leading to a narrow viewpoint and skewed distribution of content, particularly towards certain domains like news. - **Gendered Pronouns:** Many of the languages in the Aya Dataset only contain pronouns that are explicitly gendered (e.g., Arabic) or that lack gender-neutral third-person pronouns for gender-neutral reference (e.g. Estonian). - **Formality Distinctions:** The dataset encompasses languages with diverse formality distinctions, involving honorifics and situational choices in pronoun use, reflecting varying levels of standardization influenced by regional, cultural, and identity factors. - **Toxic or Offensive Speech:** The Aya Annotation Platform lacked specific flags for toxic speech, relying on human verification and peer review to mitigate offensive content, but there's no guarantee that all potentially offensive data points were removed during the annotation process. - **Accounting for mislabeled data:** The Aya Annotation Platform lacks re-labeling capabilities, leading to potential mislabeled data in the Aya Dataset, including instances of incorrect language assignments and non-compliance with instruction-style formatting. # Additional Information ## Provenance - **Methods Used:** Crowd-sourced through volunteer annotations, followed by a quality assessment phase in which samples from the dataset were checked. - **Methodology Details:** - *Source:* Original annotations and edits of opensource NLP datasets - *Platform:* [Aya Annotation Platform](https://aya.for.ai/) - *Dates of Collection:* May 2023 - Dec 2023 ## Dataset Version and Maintenance - **Maintenance Status:** Actively Maintained - **Version Details:** - *Current version:* 1.0 - *Last Update:* 02/2024 - *First Release:* 02/2024 - **Maintenance Plan:** Updates will be periodically made available based on volunteer contributions. ## Authorship - **Publishing Organization:** [Cohere Labs](https://cohere.com/research) - **Industry Type:** Not-for-profit - Tech - **Contact Details:** https://aya.for.ai/ ## Licensing Information This dataset can be used for any purpose, whether academic or commercial, under the terms of the [Apache 2.0](https://opensource.org/license/apache-2-0) License. ## Citation Information ```bibtex @misc{singh2024aya, title={Aya Dataset: An Open-Access Collection for Multilingual Instruction Tuning}, author={Shivalika Singh and Freddie Vargus and Daniel Dsouza and Börje F. Karlsson and Abinaya Mahendiran and Wei-Yin Ko and Herumb Shandilya and Jay Patel and Deividas Mataciunas and Laura OMahony and Mike Zhang and Ramith Hettiarachchi and Joseph Wilson and Marina Machado and Luisa Souza Moura and Dominik Krzemiński and Hakimeh Fadaei and Irem Ergün and Ifeoma Okoh and Aisha Alaagib and Oshan Mudannayake and Zaid Alyafeai and Vu Minh Chien and Sebastian Ruder and Surya Guthikonda and Emad A. Alghamdi and Sebastian Gehrmann and Niklas Muennighoff and Max Bartolo and Julia Kreutzer and Ahmet Üstün and Marzieh Fadaee and Sara Hooker}, year={2024}, eprint={2402.06619}, archivePrefix={arXiv}, primaryClass={cs.CL} } ```
CohereLabs/Global-MMLU
CohereLabs
2025-04-15T08:41:05Z
7,690
118
[ "language:en", "language:ar", "language:bn", "language:es", "language:fr", "language:hi", "language:ru", "language:de", "language:id", "language:it", "language:ja", "language:ko", "language:pt", "language:zh", "language:yo", "language:nl", "language:ro", "language:uk", "language:vi", "language:tr", "language:pl", "language:fa", "language:cs", "language:he", "language:el", "language:ms", "language:fil", "language:te", "language:si", "language:ne", "language:ky", "language:sv", "language:lt", "language:sr", "language:mg", "language:so", "language:ha", "language:am", "language:sn", "language:ig", "language:ny", "language:sw", "license:apache-2.0", "size_categories:100K<n<1M", "format:parquet", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "library:argilla", "arxiv:2412.03304", "region:us", "argilla" ]
[]
2024-12-01T22:45:59Z
null
--- dataset_info: - config_name: am features: - name: sample_id dtype: string - name: subject dtype: string - name: subject_category dtype: string - name: question dtype: string - name: option_a dtype: string - name: option_b dtype: string - name: option_c dtype: string - name: option_d dtype: string - name: answer dtype: string - name: required_knowledge dtype: string - name: time_sensitive dtype: string - name: reference dtype: string - name: culture dtype: string - name: region dtype: string - name: country dtype: string - name: cultural_sensitivity_label dtype: string - name: is_annotated dtype: bool splits: - name: dev num_bytes: 209505 num_examples: 285 - name: test num_bytes: 12085768 num_examples: 14042 download_size: 10260448 dataset_size: 12295273 - config_name: ar features: - name: sample_id dtype: string - name: subject dtype: string - name: subject_category dtype: string - name: question dtype: string - name: option_a dtype: string - name: option_b dtype: string - name: option_c dtype: string - name: option_d dtype: string - name: answer dtype: string - name: required_knowledge dtype: string - name: time_sensitive dtype: string - name: reference dtype: string - name: culture dtype: string - name: region dtype: string - name: country dtype: string - name: cultural_sensitivity_label dtype: string - name: is_annotated dtype: bool splits: - name: dev num_bytes: 202343 num_examples: 285 - name: test num_bytes: 11621977 num_examples: 14042 download_size: 9817049 dataset_size: 11824320 - config_name: bn features: - name: sample_id dtype: string - name: subject dtype: string - name: subject_category dtype: string - name: question dtype: string - name: option_a dtype: string - name: option_b dtype: string - name: option_c dtype: string - name: option_d dtype: string - name: answer dtype: string - name: required_knowledge dtype: string - name: time_sensitive dtype: string - name: reference dtype: string - name: culture dtype: string - name: region dtype: string - name: country dtype: string - name: cultural_sensitivity_label dtype: string - name: is_annotated dtype: bool splits: - name: dev num_bytes: 301875 num_examples: 285 - name: test num_bytes: 18061158 num_examples: 14042 download_size: 12524784 dataset_size: 18363033 - config_name: cs features: - name: sample_id dtype: string - name: subject dtype: string - name: subject_category dtype: string - name: question dtype: string - name: option_a dtype: string - name: option_b dtype: string - name: option_c dtype: string - name: option_d dtype: string - name: answer dtype: string - name: required_knowledge dtype: string - name: time_sensitive dtype: string - name: reference dtype: string - name: culture dtype: string - name: region dtype: string - name: country dtype: string - name: cultural_sensitivity_label dtype: string - name: is_annotated dtype: bool splits: - name: dev num_bytes: 149807 num_examples: 285 - name: test num_bytes: 8607308 num_examples: 14042 download_size: 8640151 dataset_size: 8757115 - config_name: de features: - name: sample_id dtype: string - name: subject dtype: string - name: subject_category dtype: string - name: question dtype: string - name: option_a dtype: string - name: option_b dtype: string - name: option_c dtype: string - name: option_d dtype: string - name: answer dtype: string - name: required_knowledge dtype: string - name: time_sensitive dtype: string - name: reference dtype: string - name: culture dtype: string - name: region dtype: string - name: country dtype: string - name: cultural_sensitivity_label dtype: string - name: is_annotated dtype: bool splits: - name: dev num_bytes: 162406 num_examples: 285 - name: test num_bytes: 9575360 num_examples: 14042 download_size: 9187953 dataset_size: 9737766 - config_name: el features: - name: sample_id dtype: string - name: subject dtype: string - name: subject_category dtype: string - name: question dtype: string - name: option_a dtype: string - name: option_b dtype: string - name: option_c dtype: string - name: option_d dtype: string - name: answer dtype: string - name: required_knowledge dtype: string - name: time_sensitive dtype: string - name: reference dtype: string - name: culture dtype: string - name: region dtype: string - name: country dtype: string - name: cultural_sensitivity_label dtype: string - name: is_annotated dtype: bool splits: - name: dev num_bytes: 254308 num_examples: 285 - name: test num_bytes: 14502137 num_examples: 14042 download_size: 12288940 dataset_size: 14756445 - config_name: en features: - name: sample_id dtype: string - name: subject dtype: string - name: subject_category dtype: string - name: question dtype: string - name: option_a dtype: string - name: option_b dtype: string - name: option_c dtype: string - name: option_d dtype: string - name: answer dtype: string - name: required_knowledge dtype: string - name: time_sensitive dtype: string - name: reference dtype: string - name: culture dtype: string - name: region dtype: string - name: country dtype: string - name: cultural_sensitivity_label dtype: string - name: is_annotated dtype: bool splits: - name: dev num_bytes: 146364 num_examples: 285 - name: test num_bytes: 8440632 num_examples: 14042 download_size: 7912429 dataset_size: 8586996 - config_name: es features: - name: sample_id dtype: string - name: subject dtype: string - name: subject_category dtype: string - name: question dtype: string - name: option_a dtype: string - name: option_b dtype: string - name: option_c dtype: string - name: option_d dtype: string - name: answer dtype: string - name: required_knowledge dtype: string - name: time_sensitive dtype: string - name: reference dtype: string - name: culture dtype: string - name: region dtype: string - name: country dtype: string - name: cultural_sensitivity_label dtype: string - name: is_annotated dtype: bool splits: - name: dev num_bytes: 160633 num_examples: 285 - name: test num_bytes: 9399724 num_examples: 14042 download_size: 8752720 dataset_size: 9560357 - config_name: fa features: - name: sample_id dtype: string - name: subject dtype: string - name: subject_category dtype: string - name: question dtype: string - name: option_a dtype: string - name: option_b dtype: string - name: option_c dtype: string - name: option_d dtype: string - name: answer dtype: string - name: required_knowledge dtype: string - name: time_sensitive dtype: string - name: reference dtype: string - name: culture dtype: string - name: region dtype: string - name: country dtype: string - name: cultural_sensitivity_label dtype: string - name: is_annotated dtype: bool splits: - name: dev num_bytes: 202609 num_examples: 285 - name: test num_bytes: 11611890 num_examples: 14042 download_size: 9564082 dataset_size: 11814499 - config_name: fil features: - name: sample_id dtype: string - name: subject dtype: string - name: subject_category dtype: string - name: question dtype: string - name: option_a dtype: string - name: option_b dtype: string - name: option_c dtype: string - name: option_d dtype: string - name: answer dtype: string - name: required_knowledge dtype: string - name: time_sensitive dtype: string - name: reference dtype: string - name: culture dtype: string - name: region dtype: string - name: country dtype: string - name: cultural_sensitivity_label dtype: string - name: is_annotated dtype: bool splits: - name: dev num_bytes: 165182 num_examples: 285 - name: test num_bytes: 9510179 num_examples: 14042 download_size: 8564879 dataset_size: 9675361 - config_name: fr features: - name: sample_id dtype: string - name: subject dtype: string - name: subject_category dtype: string - name: question dtype: string - name: option_a dtype: string - name: option_b dtype: string - name: option_c dtype: string - name: option_d dtype: string - name: answer dtype: string - name: required_knowledge dtype: string - name: time_sensitive dtype: string - name: reference dtype: string - name: culture dtype: string - name: region dtype: string - name: country dtype: string - name: cultural_sensitivity_label dtype: string - name: is_annotated dtype: bool splits: - name: dev num_bytes: 166173 num_examples: 285 - name: test num_bytes: 9858873 num_examples: 14042 download_size: 9202595 dataset_size: 10025046 - config_name: ha features: - name: sample_id dtype: string - name: subject dtype: string - name: subject_category dtype: string - name: question dtype: string - name: option_a dtype: string - name: option_b dtype: string - name: option_c dtype: string - name: option_d dtype: string - name: answer dtype: string - name: required_knowledge dtype: string - name: time_sensitive dtype: string - name: reference dtype: string - name: culture dtype: string - name: region dtype: string - name: country dtype: string - name: cultural_sensitivity_label dtype: string - name: is_annotated dtype: bool splits: - name: dev num_bytes: 147406 num_examples: 285 - name: test num_bytes: 8445707 num_examples: 14042 download_size: 7665529 dataset_size: 8593113 - config_name: he features: - name: sample_id dtype: string - name: subject dtype: string - name: subject_category dtype: string - name: question dtype: string - name: option_a dtype: string - name: option_b dtype: string - name: option_c dtype: string - name: option_d dtype: string - name: answer dtype: string - name: required_knowledge dtype: string - name: time_sensitive dtype: string - name: reference dtype: string - name: culture dtype: string - name: region dtype: string - name: country dtype: string - name: cultural_sensitivity_label dtype: string - name: is_annotated dtype: bool splits: - name: dev num_bytes: 178912 num_examples: 285 - name: test num_bytes: 10248592 num_examples: 14042 download_size: 8818618 dataset_size: 10427504 - config_name: hi features: - name: sample_id dtype: string - name: subject dtype: string - name: subject_category dtype: string - name: question dtype: string - name: option_a dtype: string - name: option_b dtype: string - name: option_c dtype: string - name: option_d dtype: string - name: answer dtype: string - name: required_knowledge dtype: string - name: time_sensitive dtype: string - name: reference dtype: string - name: culture dtype: string - name: region dtype: string - name: country dtype: string - name: cultural_sensitivity_label dtype: string - name: is_annotated dtype: bool splits: - name: dev num_bytes: 308254 num_examples: 285 - name: test num_bytes: 17970478 num_examples: 14042 download_size: 12407854 dataset_size: 18278732 - config_name: id features: - name: sample_id dtype: string - name: subject dtype: string - name: subject_category dtype: string - name: question dtype: string - name: option_a dtype: string - name: option_b dtype: string - name: option_c dtype: string - name: option_d dtype: string - name: answer dtype: string - name: required_knowledge dtype: string - name: time_sensitive dtype: string - name: reference dtype: string - name: culture dtype: string - name: region dtype: string - name: country dtype: string - name: cultural_sensitivity_label dtype: string - name: is_annotated dtype: bool splits: - name: dev num_bytes: 154692 num_examples: 285 - name: test num_bytes: 8886643 num_examples: 14042 download_size: 7793365 dataset_size: 9041335 - config_name: ig features: - name: sample_id dtype: string - name: subject dtype: string - name: subject_category dtype: string - name: question dtype: string - name: option_a dtype: string - name: option_b dtype: string - name: option_c dtype: string - name: option_d dtype: string - name: answer dtype: string - name: required_knowledge dtype: string - name: time_sensitive dtype: string - name: reference dtype: string - name: culture dtype: string - name: region dtype: string - name: country dtype: string - name: cultural_sensitivity_label dtype: string - name: is_annotated dtype: bool splits: - name: dev num_bytes: 157376 num_examples: 285 - name: test num_bytes: 9221405 num_examples: 14042 download_size: 7644102 dataset_size: 9378781 - config_name: it features: - name: sample_id dtype: string - name: subject dtype: string - name: subject_category dtype: string - name: question dtype: string - name: option_a dtype: string - name: option_b dtype: string - name: option_c dtype: string - name: option_d dtype: string - name: answer dtype: string - name: required_knowledge dtype: string - name: time_sensitive dtype: string - name: reference dtype: string - name: culture dtype: string - name: region dtype: string - name: country dtype: string - name: cultural_sensitivity_label dtype: string - name: is_annotated dtype: bool splits: - name: dev num_bytes: 157547 num_examples: 285 - name: test num_bytes: 9374481 num_examples: 14042 download_size: 8873034 dataset_size: 9532028 - config_name: ja features: - name: sample_id dtype: string - name: subject dtype: string - name: subject_category dtype: string - name: question dtype: string - name: option_a dtype: string - name: option_b dtype: string - name: option_c dtype: string - name: option_d dtype: string - name: answer dtype: string - name: required_knowledge dtype: string - name: time_sensitive dtype: string - name: reference dtype: string - name: culture dtype: string - name: region dtype: string - name: country dtype: string - name: cultural_sensitivity_label dtype: string - name: is_annotated dtype: bool splits: - name: dev num_bytes: 167646 num_examples: 285 - name: test num_bytes: 9830716 num_examples: 14042 download_size: 8826164 dataset_size: 9998362 - config_name: ko features: - name: sample_id dtype: string - name: subject dtype: string - name: subject_category dtype: string - name: question dtype: string - name: option_a dtype: string - name: option_b dtype: string - name: option_c dtype: string - name: option_d dtype: string - name: answer dtype: string - name: required_knowledge dtype: string - name: time_sensitive dtype: string - name: reference dtype: string - name: culture dtype: string - name: region dtype: string - name: country dtype: string - name: cultural_sensitivity_label dtype: string - name: is_annotated dtype: bool splits: - name: dev num_bytes: 160572 num_examples: 285 - name: test num_bytes: 9454859 num_examples: 14042 download_size: 8640457 dataset_size: 9615431 - config_name: ky features: - name: sample_id dtype: string - name: subject dtype: string - name: subject_category dtype: string - name: question dtype: string - name: option_a dtype: string - name: option_b dtype: string - name: option_c dtype: string - name: option_d dtype: string - name: answer dtype: string - name: required_knowledge dtype: string - name: time_sensitive dtype: string - name: reference dtype: string - name: culture dtype: string - name: region dtype: string - name: country dtype: string - name: cultural_sensitivity_label dtype: string - name: is_annotated dtype: bool splits: - name: dev num_bytes: 235001 num_examples: 285 - name: test num_bytes: 13483934 num_examples: 14042 download_size: 11148813 dataset_size: 13718935 - config_name: lt features: - name: sample_id dtype: string - name: subject dtype: string - name: subject_category dtype: string - name: question dtype: string - name: option_a dtype: string - name: option_b dtype: string - name: option_c dtype: string - name: option_d dtype: string - name: answer dtype: string - name: required_knowledge dtype: string - name: time_sensitive dtype: string - name: reference dtype: string - name: culture dtype: string - name: region dtype: string - name: country dtype: string - name: cultural_sensitivity_label dtype: string - name: is_annotated dtype: bool splits: - name: dev num_bytes: 148917 num_examples: 285 - name: test num_bytes: 8504949 num_examples: 14042 download_size: 8416467 dataset_size: 8653866 - config_name: mg features: - name: sample_id dtype: string - name: subject dtype: string - name: subject_category dtype: string - name: question dtype: string - name: option_a dtype: string - name: option_b dtype: string - name: option_c dtype: string - name: option_d dtype: string - name: answer dtype: string - name: required_knowledge dtype: string - name: time_sensitive dtype: string - name: reference dtype: string - name: culture dtype: string - name: region dtype: string - name: country dtype: string - name: cultural_sensitivity_label dtype: string - name: is_annotated dtype: bool splits: - name: dev num_bytes: 161992 num_examples: 285 - name: test num_bytes: 9337415 num_examples: 14042 download_size: 8011427 dataset_size: 9499407 - config_name: ms features: - name: sample_id dtype: string - name: subject dtype: string - name: subject_category dtype: string - name: question dtype: string - name: option_a dtype: string - name: option_b dtype: string - name: option_c dtype: string - name: option_d dtype: string - name: answer dtype: string - name: required_knowledge dtype: string - name: time_sensitive dtype: string - name: reference dtype: string - name: culture dtype: string - name: region dtype: string - name: country dtype: string - name: cultural_sensitivity_label dtype: string - name: is_annotated dtype: bool splits: - name: dev num_bytes: 152549 num_examples: 285 - name: test num_bytes: 8823844 num_examples: 14042 download_size: 7783581 dataset_size: 8976393 - config_name: ne features: - name: sample_id dtype: string - name: subject dtype: string - name: subject_category dtype: string - name: question dtype: string - name: option_a dtype: string - name: option_b dtype: string - name: option_c dtype: string - name: option_d dtype: string - name: answer dtype: string - name: required_knowledge dtype: string - name: time_sensitive dtype: string - name: reference dtype: string - name: culture dtype: string - name: region dtype: string - name: country dtype: string - name: cultural_sensitivity_label dtype: string - name: is_annotated dtype: bool splits: - name: dev num_bytes: 294790 num_examples: 285 - name: test num_bytes: 16972110 num_examples: 14042 download_size: 11895818 dataset_size: 17266900 - config_name: nl features: - name: sample_id dtype: string - name: subject dtype: string - name: subject_category dtype: string - name: question dtype: string - name: option_a dtype: string - name: option_b dtype: string - name: option_c dtype: string - name: option_d dtype: string - name: answer dtype: string - name: required_knowledge dtype: string - name: time_sensitive dtype: string - name: reference dtype: string - name: culture dtype: string - name: region dtype: string - name: country dtype: string - name: cultural_sensitivity_label dtype: string - name: is_annotated dtype: bool splits: - name: dev num_bytes: 158122 num_examples: 285 - name: test num_bytes: 9099176 num_examples: 14042 download_size: 8565959 dataset_size: 9257298 - config_name: ny features: - name: sample_id dtype: string - name: subject dtype: string - name: subject_category dtype: string - name: question dtype: string - name: option_a dtype: string - name: option_b dtype: string - name: option_c dtype: string - name: option_d dtype: string - name: answer dtype: string - name: required_knowledge dtype: string - name: time_sensitive dtype: string - name: reference dtype: string - name: culture dtype: string - name: region dtype: string - name: country dtype: string - name: cultural_sensitivity_label dtype: string - name: is_annotated dtype: bool splits: - name: dev num_bytes: 151315 num_examples: 285 - name: test num_bytes: 8686819 num_examples: 14042 download_size: 7822699 dataset_size: 8838134 - config_name: pl features: - name: sample_id dtype: string - name: subject dtype: string - name: subject_category dtype: string - name: question dtype: string - name: option_a dtype: string - name: option_b dtype: string - name: option_c dtype: string - name: option_d dtype: string - name: answer dtype: string - name: required_knowledge dtype: string - name: time_sensitive dtype: string - name: reference dtype: string - name: culture dtype: string - name: region dtype: string - name: country dtype: string - name: cultural_sensitivity_label dtype: string - name: is_annotated dtype: bool splits: - name: dev num_bytes: 157290 num_examples: 285 - name: test num_bytes: 8980730 num_examples: 14042 download_size: 8981270 dataset_size: 9138020 - config_name: pt features: - name: sample_id dtype: string - name: subject dtype: string - name: subject_category dtype: string - name: question dtype: string - name: option_a dtype: string - name: option_b dtype: string - name: option_c dtype: string - name: option_d dtype: string - name: answer dtype: string - name: required_knowledge dtype: string - name: time_sensitive dtype: string - name: reference dtype: string - name: culture dtype: string - name: region dtype: string - name: country dtype: string - name: cultural_sensitivity_label dtype: string - name: is_annotated dtype: bool splits: - name: dev num_bytes: 154592 num_examples: 285 - name: test num_bytes: 8983299 num_examples: 14042 download_size: 8517588 dataset_size: 9137891 - config_name: ro features: - name: sample_id dtype: string - name: subject dtype: string - name: subject_category dtype: string - name: question dtype: string - name: option_a dtype: string - name: option_b dtype: string - name: option_c dtype: string - name: option_d dtype: string - name: answer dtype: string - name: required_knowledge dtype: string - name: time_sensitive dtype: string - name: reference dtype: string - name: culture dtype: string - name: region dtype: string - name: country dtype: string - name: cultural_sensitivity_label dtype: string - name: is_annotated dtype: bool splits: - name: dev num_bytes: 158311 num_examples: 285 - name: test num_bytes: 9163189 num_examples: 14042 download_size: 8773232 dataset_size: 9321500 - config_name: ru features: - name: sample_id dtype: string - name: subject dtype: string - name: subject_category dtype: string - name: question dtype: string - name: option_a dtype: string - name: option_b dtype: string - name: option_c dtype: string - name: option_d dtype: string - name: answer dtype: string - name: required_knowledge dtype: string - name: time_sensitive dtype: string - name: reference dtype: string - name: culture dtype: string - name: region dtype: string - name: country dtype: string - name: cultural_sensitivity_label dtype: string - name: is_annotated dtype: bool splits: - name: dev num_bytes: 246059 num_examples: 285 - name: test num_bytes: 14059847 num_examples: 14042 download_size: 11904365 dataset_size: 14305906 - config_name: si features: - name: sample_id dtype: string - name: subject dtype: string - name: subject_category dtype: string - name: question dtype: string - name: option_a dtype: string - name: option_b dtype: string - name: option_c dtype: string - name: option_d dtype: string - name: answer dtype: string - name: required_knowledge dtype: string - name: time_sensitive dtype: string - name: reference dtype: string - name: culture dtype: string - name: region dtype: string - name: country dtype: string - name: cultural_sensitivity_label dtype: string - name: is_annotated dtype: bool splits: - name: dev num_bytes: 297843 num_examples: 285 - name: test num_bytes: 17374939 num_examples: 14042 download_size: 12790098 dataset_size: 17672782 - config_name: sn features: - name: sample_id dtype: string - name: subject dtype: string - name: subject_category dtype: string - name: question dtype: string - name: option_a dtype: string - name: option_b dtype: string - name: option_c dtype: string - name: option_d dtype: string - name: answer dtype: string - name: required_knowledge dtype: string - name: time_sensitive dtype: string - name: reference dtype: string - name: culture dtype: string - name: region dtype: string - name: country dtype: string - name: cultural_sensitivity_label dtype: string - name: is_annotated dtype: bool splits: - name: dev num_bytes: 147355 num_examples: 285 - name: test num_bytes: 8507368 num_examples: 14042 download_size: 7962672 dataset_size: 8654723 - config_name: so features: - name: sample_id dtype: string - name: subject dtype: string - name: subject_category dtype: string - name: question dtype: string - name: option_a dtype: string - name: option_b dtype: string - name: option_c dtype: string - name: option_d dtype: string - name: answer dtype: string - name: required_knowledge dtype: string - name: time_sensitive dtype: string - name: reference dtype: string - name: culture dtype: string - name: region dtype: string - name: country dtype: string - name: cultural_sensitivity_label dtype: string - name: is_annotated dtype: bool splits: - name: dev num_bytes: 156282 num_examples: 285 - name: test num_bytes: 9033243 num_examples: 14042 download_size: 8706693 dataset_size: 9189525 - config_name: sr features: - name: sample_id dtype: string - name: subject dtype: string - name: subject_category dtype: string - name: question dtype: string - name: option_a dtype: string - name: option_b dtype: string - name: option_c dtype: string - name: option_d dtype: string - name: answer dtype: string - name: required_knowledge dtype: string - name: time_sensitive dtype: string - name: reference dtype: string - name: culture dtype: string - name: region dtype: string - name: country dtype: string - name: cultural_sensitivity_label dtype: string - name: is_annotated dtype: bool splits: - name: dev num_bytes: 221580 num_examples: 285 - name: test num_bytes: 12695546 num_examples: 14042 download_size: 10748391 dataset_size: 12917126 - config_name: sv features: - name: sample_id dtype: string - name: subject dtype: string - name: subject_category dtype: string - name: question dtype: string - name: option_a dtype: string - name: option_b dtype: string - name: option_c dtype: string - name: option_d dtype: string - name: answer dtype: string - name: required_knowledge dtype: string - name: time_sensitive dtype: string - name: reference dtype: string - name: culture dtype: string - name: region dtype: string - name: country dtype: string - name: cultural_sensitivity_label dtype: string - name: is_annotated dtype: bool splits: - name: dev num_bytes: 147893 num_examples: 285 - name: test num_bytes: 8549708 num_examples: 14042 download_size: 8181997 dataset_size: 8697601 - config_name: sw features: - name: sample_id dtype: string - name: subject dtype: string - name: subject_category dtype: string - name: question dtype: string - name: option_a dtype: string - name: option_b dtype: string - name: option_c dtype: string - name: option_d dtype: string - name: answer dtype: string - name: required_knowledge dtype: string - name: time_sensitive dtype: string - name: reference dtype: string - name: culture dtype: string - name: region dtype: string - name: country dtype: string - name: cultural_sensitivity_label dtype: string - name: is_annotated dtype: bool splits: - name: dev num_bytes: 147069 num_examples: 285 - name: test num_bytes: 8653210 num_examples: 14042 download_size: 7932986 dataset_size: 8800279 - config_name: te features: - name: sample_id dtype: string - name: subject dtype: string - name: subject_category dtype: string - name: question dtype: string - name: option_a dtype: string - name: option_b dtype: string - name: option_c dtype: string - name: option_d dtype: string - name: answer dtype: string - name: required_knowledge dtype: string - name: time_sensitive dtype: string - name: reference dtype: string - name: culture dtype: string - name: region dtype: string - name: country dtype: string - name: cultural_sensitivity_label dtype: string - name: is_annotated dtype: bool splits: - name: dev num_bytes: 315724 num_examples: 285 - name: test num_bytes: 18170058 num_examples: 14042 download_size: 12631358 dataset_size: 18485782 - config_name: tr features: - name: sample_id dtype: string - name: subject dtype: string - name: subject_category dtype: string - name: question dtype: string - name: option_a dtype: string - name: option_b dtype: string - name: option_c dtype: string - name: option_d dtype: string - name: answer dtype: string - name: required_knowledge dtype: string - name: time_sensitive dtype: string - name: reference dtype: string - name: culture dtype: string - name: region dtype: string - name: country dtype: string - name: cultural_sensitivity_label dtype: string - name: is_annotated dtype: bool splits: - name: dev num_bytes: 153426 num_examples: 285 - name: test num_bytes: 8833244 num_examples: 14042 download_size: 8351339 dataset_size: 8986670 - config_name: uk features: - name: sample_id dtype: string - name: subject dtype: string - name: subject_category dtype: string - name: question dtype: string - name: option_a dtype: string - name: option_b dtype: string - name: option_c dtype: string - name: option_d dtype: string - name: answer dtype: string - name: required_knowledge dtype: string - name: time_sensitive dtype: string - name: reference dtype: string - name: culture dtype: string - name: region dtype: string - name: country dtype: string - name: cultural_sensitivity_label dtype: string - name: is_annotated dtype: bool splits: - name: dev num_bytes: 229888 num_examples: 285 - name: test num_bytes: 13233771 num_examples: 14042 download_size: 11347842 dataset_size: 13463659 - config_name: vi features: - name: sample_id dtype: string - name: subject dtype: string - name: subject_category dtype: string - name: question dtype: string - name: option_a dtype: string - name: option_b dtype: string - name: option_c dtype: string - name: option_d dtype: string - name: answer dtype: string - name: required_knowledge dtype: string - name: time_sensitive dtype: string - name: reference dtype: string - name: culture dtype: string - name: region dtype: string - name: country dtype: string - name: cultural_sensitivity_label dtype: string - name: is_annotated dtype: bool splits: - name: dev num_bytes: 185712 num_examples: 285 - name: test num_bytes: 10604332 num_examples: 14042 download_size: 8971266 dataset_size: 10790044 - config_name: yo features: - name: sample_id dtype: string - name: subject dtype: string - name: subject_category dtype: string - name: question dtype: string - name: option_a dtype: string - name: option_b dtype: string - name: option_c dtype: string - name: option_d dtype: string - name: answer dtype: string - name: required_knowledge dtype: string - name: time_sensitive dtype: string - name: reference dtype: string - name: culture dtype: string - name: region dtype: string - name: country dtype: string - name: cultural_sensitivity_label dtype: string - name: is_annotated dtype: bool splits: - name: dev num_bytes: 153810 num_examples: 285 - name: test num_bytes: 10694916 num_examples: 14042 download_size: 9303668 dataset_size: 10848726 - config_name: zh features: - name: sample_id dtype: string - name: subject dtype: string - name: subject_category dtype: string - name: question dtype: string - name: option_a dtype: string - name: option_b dtype: string - name: option_c dtype: string - name: option_d dtype: string - name: answer dtype: string - name: required_knowledge dtype: string - name: time_sensitive dtype: string - name: reference dtype: string - name: culture dtype: string - name: region dtype: string - name: country dtype: string - name: cultural_sensitivity_label dtype: string - name: is_annotated dtype: bool splits: - name: dev num_bytes: 127577 num_examples: 285 - name: test num_bytes: 7393764 num_examples: 14042 download_size: 7322261 dataset_size: 7521341 configs: - config_name: am data_files: - split: test path: am/test-* - split: dev path: am/dev-* - config_name: ar data_files: - split: test path: ar/test-* - split: dev path: ar/dev-* - config_name: bn data_files: - split: test path: bn/test-* - split: dev path: bn/dev-* - config_name: cs data_files: - split: test path: cs/test-* - split: dev path: cs/dev-* - config_name: de data_files: - split: test path: de/test-* - split: dev path: de/dev-* - config_name: el data_files: - split: test path: el/test-* - split: dev path: el/dev-* - config_name: en data_files: - split: test path: en/test-* - split: dev path: en/dev-* - config_name: es data_files: - split: test path: es/test-* - split: dev path: es/dev-* - config_name: fa data_files: - split: test path: fa/test-* - split: dev path: fa/dev-* - config_name: fil data_files: - split: test path: fil/test-* - split: dev path: fil/dev-* - config_name: fr data_files: - split: test path: fr/test-* - split: dev path: fr/dev-* - config_name: ha data_files: - split: test path: ha/test-* - split: dev path: ha/dev-* - config_name: he data_files: - split: test path: he/test-* - split: dev path: he/dev-* - config_name: hi data_files: - split: test path: hi/test-* - split: dev path: hi/dev-* - config_name: id data_files: - split: test path: id/test-* - split: dev path: id/dev-* - config_name: ig data_files: - split: test path: ig/test-* - split: dev path: ig/dev-* - config_name: it data_files: - split: test path: it/test-* - split: dev path: it/dev-* - config_name: ja data_files: - split: test path: ja/test-* - split: dev path: ja/dev-* - config_name: ko data_files: - split: test path: ko/test-* - split: dev path: ko/dev-* - config_name: ky data_files: - split: test path: ky/test-* - split: dev path: ky/dev-* - config_name: lt data_files: - split: test path: lt/test-* - split: dev path: lt/dev-* - config_name: mg data_files: - split: test path: mg/test-* - split: dev path: mg/dev-* - config_name: ms data_files: - split: test path: ms/test-* - split: dev path: ms/dev-* - config_name: ne data_files: - split: test path: ne/test-* - split: dev path: ne/dev-* - config_name: nl data_files: - split: test path: nl/test-* - split: dev path: nl/dev-* - config_name: ny data_files: - split: test path: ny/test-* - split: dev path: ny/dev-* - config_name: pl data_files: - split: test path: pl/test-* - split: dev path: pl/dev-* - config_name: pt data_files: - split: test path: pt/test-* - split: dev path: pt/dev-* - config_name: ro data_files: - split: test path: ro/test-* - split: dev path: ro/dev-* - config_name: ru data_files: - split: test path: ru/test-* - split: dev path: ru/dev-* - config_name: si data_files: - split: test path: si/test-* - split: dev path: si/dev-* - config_name: sn data_files: - split: test path: sn/test-* - split: dev path: sn/dev-* - config_name: so data_files: - split: test path: so/test-* - split: dev path: so/dev-* - config_name: sr data_files: - split: test path: sr/test-* - split: dev path: sr/dev-* - config_name: sv data_files: - split: test path: sv/test-* - split: dev path: sv/dev-* - config_name: sw data_files: - split: test path: sw/test-* - split: dev path: sw/dev-* - config_name: te data_files: - split: test path: te/test-* - split: dev path: te/dev-* - config_name: tr data_files: - split: test path: tr/test-* - split: dev path: tr/dev-* - config_name: uk data_files: - split: test path: uk/test-* - split: dev path: uk/dev-* - config_name: vi data_files: - split: test path: vi/test-* - split: dev path: vi/dev-* - config_name: yo data_files: - split: test path: yo/test-* - split: dev path: yo/dev-* - config_name: zh data_files: - split: test path: zh/test-* - split: dev path: zh/dev-* tags: - argilla license: apache-2.0 language: - en - ar - bn - es - fr - hi - ru - de - id - it - ja - ko - pt - zh - yo - nl - ro - uk - vi - tr - pl - fa - cs - he - el - ms - fil - te - si - ne - ky - sv - lt - sr - mg - so - ha - am - sn - ig - ny - sw --- ![GlobalMMLU Header](https://huggingface.co/datasets/CohereLabs/Global-MMLU/resolve/main/global_mmlu.jpg) # Dataset Summary [Global-MMLU](https://arxiv.org/abs/2412.03304) 🌍 is a multilingual evaluation set spanning 42 languages, including English. This dataset combines machine translations for [MMLU](https://huggingface.co/datasets/cais/mmlu) questions along with professional translations and crowd-sourced post-edits. It also includes cultural sensitivity annotations for a subset of the questions (2850 questions per language) and classifies them as *Culturally Sensitive* (CS) 🗽 or *Culturally Agnostic* (CA) ⚖️. These annotations were collected as part of an open science initiative led by Cohere Labs in collaboration with many external collaborators from both industry and academia. - **Curated by:** Professional annotators and contributors of [Cohere Labs Community](https://cohere.com/research). - **Language(s):** 42 languages. - **License:** [Apache 2.0](https://opensource.org/license/apache-2-0) **Note:** We also provide a "lite" version of Global MMLU called ["Global-MMLU-Lite"](https://huggingface.co/datasets/CohereLabs/Global-MMLU-Lite). This datatset is more balanced containing 200 samples each for CA and CS subsets for each language. And provides coverage for 15 languages with human translations. ### **Global-MMLU Dataset Family:** | Name | Explanation | |------|--------------| | [Global-MMLU](https://huggingface.co/datasets/CohereLabs/Global-MMLU) | Full Global-MMLU set with translations for all 14K samples including CS and CA subsets| | [Global-MMLU-Lite](https://huggingface.co/datasets/CohereLabs/Global-MMLU-Lite) | Lite version of Global-MMLU with human translated samples in 15 languages and containing 200 samples each for CS and CA subsets per language.| ## Load with Datasets To load this dataset with `datasets`, you'll first need to install it using `pip install datasets` and then use the following code: ```python from datasets import load_dataset # load HF dataset global_mmlu = load_dataset("CohereLabs/Global-MMLU", 'en') # can also be used as pandas dataframe global_mmlu.set_format("pandas") global_mmlu_test = global_mmlu['test'][:] global_mmlu_dev = global_mmlu['dev'][:] ``` <details> <summary> additional details </summary> The columns corresponding to annotations collected from our cultural bias study (i.e. 'required_knowledge', 'time_sensitive', 'reference', 'culture', 'region', 'country') contain a list of values representing annotations from different annotators. However, to avoid conversion issues to HF dataset, these columns are provided as string in the final dataset. You can convert these columns back to list of values for easier manipulation as follows: ```python import ast # convert string values to list global_mmlu_df['required_knowledge'] = global_mmlu_df['required_knowledge'].apply(lamda x: ast.literal_eval(x)) ``` </details> <br> ## Data Fields The data fields are the same among all splits. Brief description of each field is provided below. <details> <summary> data field description </summary> - `sample_id`: A unique identifier for the question. - `subject`: The main topic the question falls under. - `subject_category`: The high-level category the subject falls under i.e. STEM/Humanities/Social Sciences/Medical/Business/Other. - `question`: translated question from MMLU - `option_a`: one of the possible option choices - `option_b`: one of the possible option choices - `option_c`: one of the possible option choices - `option_d`: one of the possible option choices - `answer': the correct answer (A/B/C/D) - `required_knowledge`: annotator votes for knowledge needed to answer the question correctly. Possible values include: "cultural", "regional", "dialect" or "none" - `time_sensitive`: annotator votes indicating if the question's answer is time-dependent. Possible values include: Yes/No - `reference`: annotations for which part of the question contains cultural/regional/dialect references. The different items in the list are annotations from different annotators. - `culture`: annotations for which culture does the question belong to. The different items in the list correspond to annotations from different annotators. - `region`: Geographic region the question is relevant to. Each item in the list correspond to annotations from different annotators. - `country`: Specific country the question pertains to. Each item in the list correspond to annotations from different annotators. - `cultural_sensitivity_label`: Label to indicate if question is culturally sensitive (CS) or culturally agnostic (CA) based on annotator votes. - `is_annotated`: True/False flag to indicate if sample contains any annotations from our cultural bias study. </details> <br> ## Data Splits The following are the splits of the data: | Split | No. of instances | Language Coverage | |-------|------------------|-------------------| | test | 589,764 | 42 | | dev | 11,970 | 42 | ## Data Instances An example from `test` set looks as follows: ```json {'sample_id': 'world_religions/test/170' 'subject': 'world_religions', 'subject_category': 'Humanities', 'question': ' The numen of Augustus referred to which of the following characteristics?', 'option_a': 'Divine power', 'option_b': 'Sexual virility', 'option_c': 'Military acumen', 'option_d': 'Philosophical intellect', 'answer': 'A', 'required_knowledge': "['none', 'cultural', 'cultural', 'cultural']", 'time_sensitive': "['No', 'No', 'No', 'No']", 'reference': "['-', '-', {'end': 22, 'label': 'Cultural', 'score': None, 'start': 5}, {'end': 22, 'label': 'Cultural', 'score': None, 'start': 5}]", 'culture': "['Western Culture', 'Western Culture', 'Western Culture']", 'region': "['North America', 'Europe']", 'country': "['Italy']", 'cultural_sensitivity_label': 'CS', 'is_annotated': True, } ``` ## Statistics ### Annotation Types The following is the breakdown of CS🗽, CA⚖️ and MA📝 samples in the final dataset. | Type of Annotation | Instances per language | No. of languages | Total instances |--------------------|------------------------|------------------|----------------| | Culturally Sensitive 🗽 | 792 | 42 | 33,264 | | Culturally Agnostic ⚖️ | 2058 |42 | 86,436 | | MMLU Annotated 📝| 2850 |42 | 119,700 | ### Languages The dataset covers 42 languages: 20 high-resource, 9 mid-resource, and 13 low-resource languages. The following is details about the languages included in the dataset. <details> <summary> Languages Info </summary> | ISO Code | Language | Resources | |----------|----------|-----------| | `am` | Amharic | Low | | `ar` | Arabic (Standard)| High | | `bn` | Bengali | Mid | | `de` | German | High | | `el` | Greek | Mid | | `en` | English | High | | `fil` | Filipino | Mid | | `fr` | French | High | | `ha` | Hausa | Low | | `he` | Hebrew | Mid | | `hi` | Hindi | High | | `ig` | Igbo | Low | | `id` | Indonesian | Mid | | `it` | Italian | High | | `ja` | Japanese | High | | `ky` | Kyrgyz | Low | | `ko` | Korean | Mid | | `lt` | Lithuanian | Mid | | `mg` | Malagasy | Low | | `ms` | Malay | Mid | | `ne` | Nepali | Low | | `nl` | Dutch | High | | `ny` | Chichewa | Low | | `fa` | Persian | High | | `pl` | Polish | High | | `pt` | Portuguese | High | | `ru` | Russian | High | | `si` | Sinhala | Low | | `sn` | Shona | Low | | `so` | Somali | Low | | `es` | Spanish | High | | `sr` | Serbian | High | | `sw` | Swahili | Low | | `sv` | Swedish | High | | `te` | Telugu | Low | | `tr` | Turkish | High | | `uk` | Ukrainian | Mid | | `vi` | Vietnamese | High | | `yo` | Yorùbá | Low | | `zh` | Chinese (Simplified) | High | </details> <br> # Known Limitations A brief overview of limitations of this dataset is provided below. <details> <summary> show limitations </summary> - **Language and dialect coverage:** Global-MMLU focusses on 42 languages. However, this is still only a tiny fraction of the world’s linguistic diversity. Future work is needed to continue to improve evaluations beyond these 42 languages and take into account how technology serves different dialects. - **Uneven distribution of contributions:** The dataset contains translation post-edits from community volunteers, with a 'long tail' of volunteers making only one or two contributions. Similarly, there is a huge gap between languages with the highest number of contributions and ones with the lowest number of contributions. - **Toxic or offensive speech:** Our annotation process did not focus on flagging for toxic,harmful, or offensive speech, so it is possible that Global-MMLU contains some data that could be considered harmful. We believe this is of relatively low risk because of the nature of the original MMLU and the focus on examination material. - **Region Category Assignment:** For the annotation of geographically sensitive questions, we classified regions into six geographic regions (Africa, Asia, Europe, North America, Oceania,and South America). However, based upon discussions we would going forward recommend switching to the taxonomy proposed by the World Bank which is more granular and includes separate designations for Central America and Sub-Saharan Africa. - **Identifying cultural sensitivity does not guarantee cultural inclusion:** While Global-MMLU highlights important limitations in current datasets by identifying gaps in non-Western cultural representation. Future work must prioritize the integration of diverse culturally grounded knowledge to achieve true inclusivity and fairness in multilingual AI evaluation. </details> <br> # Additional Information ## Provenance - **Methods Used:** Professional annotations as well as crowd-sourced through volunteer annotations. - **Methodology Details:** We collected cultural bias annotations as well as post-edits of translations for different mmlu questions. - [Cultural Sensitivity Annotation Platform](https://huggingface.co/spaces/CohereLabs/MMLU-evaluation) - [Translation Quality Annotation Platform](https://huggingface.co/spaces/CohereLabs/review-mmlu-translations) - Dates of Collection: May 2024 - Aug 2024 ## Dataset Version and Maintenance - **Maintenance Status:** Actively Maintained - **Version Details:** - *Current version:* 1.0 - *Last Update:* 12/2024 - *First Release:* 12/2024 ## Authorship - **Publishing Organization:** [Cohere Labs](https://cohere.com/research) - **Industry Type:** Not-for-profit - Tech ## Licensing Information This dataset can be used for any purpose, under the terms of the [Apache 2.0](https://opensource.org/license/apache-2-0) License. ## Additional Details For any additional details, please check our paper, [Global MMLU: Understanding and Addressing Cultural and Linguistic Biases in Multilingual Evaluation](https://arxiv.org/abs/2412.03304). ## Citation Information ```bibtex @misc{singh2024globalmmluunderstandingaddressing, title={Global MMLU: Understanding and Addressing Cultural and Linguistic Biases in Multilingual Evaluation}, author={Shivalika Singh and Angelika Romanou and Clémentine Fourrier and David I. Adelani and Jian Gang Ngui and Daniel Vila-Suero and Peerat Limkonchotiwat and Kelly Marchisio and Wei Qi Leong and Yosephine Susanto and Raymond Ng and Shayne Longpre and Wei-Yin Ko and Madeline Smith and Antoine Bosselut and Alice Oh and Andre F. T. Martins and Leshem Choshen and Daphne Ippolito and Enzo Ferrante and Marzieh Fadaee and Beyza Ermis and Sara Hooker}, year={2024}, eprint={2412.03304}, archivePrefix={arXiv}, primaryClass={cs.CL}, url={https://arxiv.org/abs/2412.03304}, } ```
lmarena-ai/search-arena-v1-7k
lmarena-ai
2025-04-14T16:01:06Z
875
14
[ "size_categories:1K<n<10K", "format:parquet", "modality:tabular", "modality:text", "library:datasets", "library:pandas", "library:mlcroissant", "library:polars", "arxiv:2403.04132", "region:us" ]
[]
2025-04-14T06:28:29Z
2
--- size_categories: - 100K<n<1M configs: - config_name: default data_files: - split: test path: data/search-arena-* --- ## Overview This dataset contains 7k leaderboard conversation votes collected from [Search Arena](https://lmarena.ai/?search) between March 18, 2025 and April 13, 2025. All entries have been redacted for PII and sensitive user information to ensure privacy. Each data point includes: - Two model responses (`messages_a` and `messages_b`) - The human vote result - A timestamp - Full system metadata, LLM + web search trace, and post-processed metadata for controlled experiments (`conv_meta`) To reproduce the leaderboard results and analyses: - Check out the [Colab notebook](https://colab.research.google.com/drive/1h7rR7rhePBPuIfaWsVNlW87kv3DLibPS?usp=sharing) for a step-by-step walkthrough. - A companion [blog post](https://blog.lmarena.ai/blog/2025/search-arena) will provide deeper insights and commentary on the dataset and evaluation process. ## License User prompts are licensed under CC-BY-4.0, and model outputs are governed by the terms of use set by the respective model providers. ## Citation ```bibtex @misc{searcharena2025, title = {Introducing the Search Arena: Evaluating Search-Enabled AI}, url = {https://blog.lmarena.ai/blog/2025/search-arena/}, author = {Mihran Miroyan*, Tsung-Han Wu*, Logan Kenneth King, Tianle Li, Anastasios N. Angelopoulos, Wei-Lin Chiang, Narges Norouzi, Joseph E. Gonzalez}, month = {April}, year = {2025} } @misc{chiang2024chatbot, title={Chatbot Arena: An Open Platform for Evaluating LLMs by Human Preference}, author={Wei-Lin Chiang and Lianmin Zheng and Ying Sheng and Anastasios Nikolas Angelopoulos and Tianle Li and Dacheng Li and Hao Zhang and Banghua Zhu and Michael Jordan and Joseph E. Gonzalez and Ion Stoica}, year={2024}, eprint={2403.04132}, archivePrefix={arXiv}, primaryClass={cs.AI} } ```
ashynf/OpenFWI
ashynf
2025-04-14T02:05:50Z
5,502
4
[ "license:cc-by-nc-sa-4.0", "size_categories:100B<n<1T", "region:us" ]
[]
2025-04-11T21:42:02Z
2
--- license: cc-by-nc-sa-4.0 size_categories: - 100B<n<1T --- **OpenFWI: Large-scale Multi-structural Benchmark Datasets for Full Waveform Inversion** OpenFWI is the first collection of large-scale, multi-structural benchmark datasets for machine learning driven seismic Full Wave Inversion (FWI). It includes eleven 2D datasets and one 3D dataset, in four subsurface structure families. Here, We provide three of them (Vel, Fault, and Style). For both 2D and 3D Kimberlina datasets, please refer to [OpenFWI](https://smileunc.github.io/projects/openfwi) for more details. **Resources** [Paper](https://proceedings.neurips.cc/paper_files/paper/2022/file/27d3ef263c7cb8d542c4f9815a49b69b-Paper-Datasets_and_Benchmarks.pdf) - Discover the technical details and baseline methods. [Github Repo](https://github.com/lanl/OpenFWI) - Pytorch Implementation of OpenFWI Benchmarks. [Tutorial](https://www.kaggle.com/competitions/waveform-inversion/data) - A simple example of using the data and our baseline models. [OpenFWI Website](https://openfwi-lanl.github.io) - Explore more resources on the official website of OpenFWI. **License** This dataset is licensed under the [Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License](https://creativecommons.org/licenses/by-nc-sa/4.0/) (CC BY-NC-SA 4.0). **Citation** If you find the data useful, please cite: ```bibtex @article{deng2022openfwi, title={OpenFWI: Large-scale multi-structural benchmark datasets for full waveform inversion}, author={Deng, Chengyuan and Feng, Shihang and Wang, Hanchen and Zhang, Xitong and Jin, Peng and Feng, Yinan and Zeng, Qili and Chen, Yinpeng and Lin, Youzuo}, journal={Advances in Neural Information Processing Systems}, volume={35}, pages={6007--6020}, year={2022} } ```
pietrolesci/pythia-deduped-stats-raw
pietrolesci
2025-04-09T19:26:24Z
21,014
0
[ "language:en", "size_categories:10M<n<100M", "format:parquet", "library:datasets", "library:dask", "library:mlcroissant", "library:polars", "arxiv:2406.04327", "region:us" ]
[]
2024-02-12T01:19:05Z
null
--- language: - en --- This dataset has been created as an artefact of the paper [Causal Estimation of Memorisation Profiles (Lesci et al., 2024)](https://arxiv.org/abs/2406.04327). More info about this dataset in the related collection [Memorisation-Profiles](https://huggingface.co/collections/pietrolesci/memorisation-profiles-6619604c4594c878cd9d451f). Collection of data statistics computed using the intermediate checkpoints (step0, step1000, ..., step143k) of all Pythia deduped versions. This folder contains the model evaluations (or "stats") for each model size included in the study. This is the "raw" version where we have stats at the token level. We gathered these statistics "just in case" since the inference process was expensive. However, we provide the sequence-level statistics in the `pietrolesci/pythia-deduped-stats` dataset.
openbmb/DCAD-2000
openbmb
2025-04-04T09:51:21Z
20,678
6
[ "multilinguality:multilingual", "language:eng", "language:rus", "language:cmn", "language:deu", "language:spa", "language:fra", "language:jpn", "language:ita", "language:por", "language:pol", "language:nld", "language:ind", "language:tur", "language:vie", "language:fas", "language:kor", "language:swe", "language:hun", "language:ell", "language:ukr", "language:tha", "language:aze", "language:arb", "language:slv", "language:cat", "language:fin", "language:hbs", "language:ces", "language:fil", "language:guj", "language:nob", "language:mal", "language:bul", "language:kan", "language:hin", "language:tam", "language:ara", "language:kaz", "language:sqi", "language:pan", "language:est", "language:srp", "language:heb", "language:isl", "language:mlt", "language:mkd", "language:bos", "language:lit", "language:kat", "language:hrv", "language:ben", "language:glg", "language:ceb", "language:ron", "language:msa", "language:hye", "language:tgk", "language:mar", "language:bel", "language:nep", "language:urd", "language:slk", "language:mon", "language:swa", "language:dan", "language:eus", "language:azj", "language:sin", "language:lat", "language:als", "language:tel", "language:afr", "language:ekk", "language:ltz", "language:zsm", "language:som", "language:cym", "language:und", "language:kir", "language:nor", "language:uzb", "language:mya", "language:epo", "language:hau", "language:ary", "language:gle", "language:lvs", "language:nno", "language:ory", "language:amh", "language:tat", "language:khm", "language:arz", "language:pus", "language:lav", "language:war", "language:mlg", "language:uzn", "language:hat", "language:zul", "language:kur", "language:tgl", "language:div", "language:fry", "language:sna", "language:fao", "language:sun", "language:snd", "language:cos", "language:bak", "language:ckb", "language:asm", "language:ast", "language:ibo", "language:mri", "language:ars", "language:anp", "language:khk", "language:lao", "language:ori", "language:yid", "language:hmn", "language:jav", "language:nya", "language:tir", "language:uig", "language:wln", "language:yor", "language:swh", "language:gla", "language:xho", "language:bre", "language:sot", "language:che", "language:nan", "language:bew", "language:glk", "language:smo", "language:orm", "language:zho", "language:haw", "language:pnb", "language:oci", "language:chv", "language:kin", "language:pcm", "language:kaa", "language:lug", "language:pbt", "language:gsw", "language:min", "language:tuk", "language:lim", "language:kmr", "language:roh", "language:vec", "language:san", "language:gmh", "language:nds", "language:sco", "language:azb", "language:tsn", "language:sah", "language:pri", "language:pap", "language:plt", "language:lmo", "language:bod", "language:yue", "language:bar", "language:hif", "language:arg", "language:pms", "language:nrm", "language:vol", "language:aeb", "language:ido", "language:pam", "language:kal", "language:lus", "language:new", "language:sme", "language:nso", "language:lld", "language:ina", "language:scn", "language:hyw", "language:abk", "language:bpy", "language:lin", "language:oss", "language:que", "language:szl", "language:sdh", "language:mww", "language:mhr", "language:rmy", "language:diq", "language:srd", "language:ydd", "language:run", "language:hsb", "language:wol", "language:kiu", "language:ilo", "language:cak", "language:acm", "language:grc", "language:xmf", "language:vls", "language:tso", "language:crh", "language:nav", "language:mwl", "language:grn", "language:mui", "language:mvf", "language:udm", "language:mzn", "language:bcl", "language:ban", "language:afb", "language:cnh", "language:cbk", "language:tyv", "language:ful", "language:zea", "language:frp", "language:frr", "language:pfl", "language:lij", "language:apc", "language:mai", "language:pon", "language:eml", "language:uzs", "language:nde", "language:hil", "language:mam", "language:gom", "language:dag", "language:nap", "language:bjn", "language:hbo", "language:tet", "language:inh", "language:ven", "language:gaz", "language:tpi", "language:ajp", "language:sat", "language:wuu", "language:bam", "language:vep", "language:glv", "language:tzo", "language:bik", "language:acr", "language:twi", "language:myv", "language:ewe", "language:poh", "language:ile", "language:rue", "language:ace", "language:kha", "language:ksh", "language:hui", "language:gor", "language:krc", "language:lfn", "language:tls", "language:fur", "language:gag", "language:chk", "language:hmo", "language:kek", "language:fon", "language:ssw", "language:quh", "language:meu", "language:csb", "language:ksd", "language:ext", "language:rop", "language:aln", "language:kab", "language:shn", "language:tbz", "language:knv", "language:bqc", "language:npi", "language:rug", "language:kom", "language:kpg", "language:kbd", "language:dzo", "language:wal", "language:zlm", "language:mzh", "language:hac", "language:tum", "language:sgs", "language:fij", "language:tuc", "language:jbo", "language:iba", "language:pcd", "language:kjh", "language:ixl", "language:ndo", "language:suz", "language:mfe", "language:bxr", "language:ava", "language:bho", "language:brh", "language:ctd", "language:aym", "language:ady", "language:nbl", "language:hus", "language:gug", "language:bat", "language:hne", "language:srn", "language:ctu", "language:ton", "language:bug", "language:aoj", "language:jam", "language:lez", "language:xav", "language:luo", "language:ltg", "language:bih", "language:mrj", "language:avk", "language:kos", "language:kbp", "language:alt", "language:cfm", "language:goh", "language:lad", "language:cor", "language:quc", "language:kpv", "language:aka", "language:hak", "language:rmn", "language:pls", "language:sag", "language:rcf", "language:gym", "language:kik", "language:pag", "language:arn", "language:ngu", "language:dsb", "language:tca", "language:mni", "language:naq", "language:pdc", "language:mps", "language:olo", "language:fro", "language:fit", "language:cdo", "language:kac", "language:mdf", "language:mad", "language:cab", "language:srm", "language:toj", "language:zai", "language:guc", "language:smn", "language:ote", "language:nch", "language:prs", "language:bis", "language:cuk", "language:bzj", "language:map", "language:djk", "language:ncj", "language:mco", "language:shp", "language:ang", "language:seh", "language:ikk", "language:qvi", "language:yom", "language:top", "language:awa", "language:ike", "language:mau", "language:kum", "language:mos", "language:tly", "language:ayr", "language:tdt", "language:yap", "language:evn", "language:tah", "language:xal", "language:szy", "language:quw", "language:stq", "language:tlh", "language:otq", "language:nov", "language:skr", "language:yua", "language:raw", "language:fiu", "language:enm", "language:bbc", "language:non", "language:ber", "language:kea", "language:bua", "language:kua", "language:zza", "language:csy", "language:mnw", "language:dtp", "language:nah", "language:yao", "language:nog", "language:abt", "language:meo", "language:ame", "language:mbf", "language:chm", "language:kas", "language:crs", "language:tcy", "language:lbe", "language:cni", "language:rom", "language:zom", "language:blk", "language:efi", "language:ada", "language:koi", "language:bbj", "language:trv", "language:ach", "language:mgh", "language:gcr", "language:krl", "language:iso", "language:lki", "language:atj", "language:cmo", "language:chu", "language:gur", "language:iku", "language:sma", "language:lrc", "language:gos", "language:dty", "language:nzi", "language:btx", "language:brx", "language:bts", "language:nyu", "language:ckt", "language:guw", "language:pck", "language:quy", "language:gcf", "language:ksw", "language:nia", "language:bal", "language:nhe", "language:hwc", "language:mup", "language:dyu", "language:ami", "language:tay", "language:gaa", "language:loz", "language:abq", "language:pnt", "language:ium", "language:abs", "language:cmr", "language:shi", "language:tzm", "language:gil", "language:ewo", "language:kon", "language:sms", "language:ape", "language:tzh", "language:bci", "language:acd", "language:tab", "language:fuv", "language:swc", "language:tcz", "language:sgc", "language:aaz", "language:tvl", "language:zha", "language:pib", "language:pli", "language:bru", "language:quz", "language:cha", "language:aeu", "language:nqo", "language:mer", "language:acf", "language:rmc", "language:msi", "language:smj", "language:kng", "language:dhv", "language:doi", "language:zap", "language:hla", "language:cac", "language:pdt", "language:mbt", "language:taq", "language:bgp", "language:tyz", "language:aau", "language:mah", "language:nak", "language:kmb", "language:mjw", "language:fuf", "language:bba", "language:amu", "language:alz", "language:cop", "language:rar", "language:emp", "language:arc", "language:nyn", "language:rut", "language:lub", "language:tzj", "language:fat", "language:lzh", "language:aak", "language:nnb", "language:lhu", "language:bjv", "language:bum", "language:maq", "language:roa", "language:bvr", "language:kcg", "language:tok", "language:yrk", "language:khs", "language:rup", "language:qub", "language:jac", "language:hmr", "language:sid", "language:chr", "language:amp", "language:qug", "language:mas", "language:fkv", "language:bem", "language:cav", "language:her", "language:kwn", "language:ify", "language:gpe", "language:zpa", "language:agu", "language:ppk", "language:tiv", "language:ahk", "language:pis", "language:hns", "language:sus", "language:bon", "language:sja", "language:kri", "language:dik", "language:wrk", "language:got", "language:dar", "language:gui", "language:mak", "language:jiv", "language:kjb", "language:maz", "language:bsn", "language:dov", "language:cok", "language:rwo", "language:mag", "language:hnj", "language:krj", "language:teo", "language:enq", "language:aom", "language:ffm", "language:din", "language:ubu", "language:pwn", "language:gum", "language:sny", "language:twu", "language:med", "language:acn", "language:rme", "language:cbs", "language:abx", "language:ndc", "language:ibb", "language:pih", "language:zty", "language:rad", "language:aby", "language:zne", "language:cui", "language:lua", "language:yby", "language:bno", "language:niu", "language:agd", "language:nij", "language:agg", "language:yre", "language:fip", "language:qve", "language:tsg", "language:skg", "language:adj", "language:mfq", "language:agx", "language:kmg", "language:umb", "language:cnk", "language:nhw", "language:maa", "language:kwy", "language:jra", "language:syr", "language:inb", "language:pau", "language:xsm", "language:vap", "language:alp", "language:des", "language:syl", "language:mkn", "language:gan", "language:qwh", "language:jvn", "language:bib", "language:mpx", "language:bin", "language:klv", "language:xon", "language:jbu", "language:bdh", "language:knj", "language:ptu", "language:mmn", "language:mwq", "language:sop", "language:cgc", "language:dop", "language:mgr", "language:rmo", "language:tbc", "language:kyq", "language:zpu", "language:kkc", "language:acu", "language:ata", "language:mif", "language:tsz", "language:mbi", "language:imo", "language:fuh", "language:zyp", "language:apr", "language:att", "language:bus", "language:kqp", "language:koo", "language:bas", "language:urh", "language:bbr", "language:msy", "language:mdy", "language:nas", "language:lee", "language:mcu", "language:dig", "language:izr", "language:nfr", "language:hto", "language:mzw", "language:qxh", "language:chz", "language:guh", "language:ded", "language:bvz", "language:kkj", "language:gux", "language:ikt", "language:bnp", "language:bgt", "language:quf", "language:dwr", "language:nod", "language:yal", "language:bmr", "language:ljp", "language:xla", "language:zyb", "language:gof", "language:niv", "language:nrf", "language:npy", "language:ura", "language:dob", "language:gfk", "language:gdr", "language:sdc", "language:pjt", "language:spp", "language:irk", "language:cpa", "language:knc", "language:gvl", "language:agr", "language:njo", "language:xrb", "language:kao", "language:zpz", "language:bgr", "language:rnd", "language:ntu", "language:aai", "language:gai", "language:dgi", "language:kez", "language:dad", "language:kmh", "language:aii", "language:ifk", "language:aji", "language:noa", "language:bhl", "language:ztq", "language:gde", "language:dnj", "language:men", "language:miq", "language:isd", "language:lgg", "language:ncx", "language:gub", "language:itv", "language:tac", "language:cce", "language:kwi", "language:nhu", "language:sml", "language:pah", "language:sas", "language:cwt", "language:spy", "language:enb", "language:fai", "language:cas", "language:sue", "language:yss", "language:ozm", "language:bku", "language:biv", "language:akp", "language:dts", "language:kdr", "language:fal", "language:mmo", "language:byr", "language:cbv", "language:cbu", "language:eza", "language:kgp", "language:sda", "language:gbo", "language:mfi", "language:bdd", "language:boj", "language:blh", "language:ibg", "language:agn", "language:vmy", "language:sua", "language:gul", "language:rap", "language:bmh", "language:sxn", "language:dyi", "language:tdx", "language:car", "language:heh", "language:dgc", "language:sgb", "language:bpr", "language:gbi", "language:avu", "language:udu", "language:yli", "language:cjs", "language:gnn", "language:nhi", "language:urk", "language:kpy", "language:qup", "language:qvh", "language:pir", "language:tsc", "language:thk", "language:grt", "language:kqn", "language:hay", "language:mqb", "language:tbg", "language:bss", "language:trn", "language:jae", "language:myy", "language:ipk", "language:bao", "language:qvz", "language:dua", "language:mwv", "language:nct", "language:kyc", "language:tuo", "language:tee", "language:mop", "language:mxt", "language:qvm", "language:gkn", "language:lam", "language:ntr", "language:sbe", "language:dyo", "language:lex", "language:smk", "language:coe", "language:kpe", "language:xbi", "language:atb", "language:tcf", "language:bfo", "language:cbc", "language:swg", "language:cnt", "language:izz", "language:knf", "language:mpp", "language:kki", "language:nho", "language:kbr", "language:bmu", "language:hig", "language:qxr", "language:orv", "language:pma", "language:zpl", "language:kpr", "language:sig", "language:mxp", "language:lef", "language:mlp", "language:soq", "language:cre", "language:aey", "language:qvs", "language:sur", "language:suk", "language:cax", "language:alq", "language:chw", "language:dww", "language:gog", "language:viv", "language:qvw", "language:ted", "language:yuj", "language:ziw", "language:kto", "language:mcq", "language:sim", "language:lue", "language:cya", "language:knk", "language:tpt", "language:taj", "language:blz", "language:mil", "language:kde", "language:kpz", "language:hag", "language:vun", "language:kus", "language:lid", "language:tos", "language:bfd", "language:chd", "language:cle", "language:yon", "language:bgs", "language:lsm", "language:msb", "language:ota", "language:mxq", "language:myw", "language:aso", "language:esu", "language:kdi", "language:bhp", "language:mfz", "language:sgw", "language:xsi", "language:gun", "language:kxc", "language:mux", "language:otw", "language:cjp", "language:nyo", "language:zia", "language:cco", "language:ojb", "language:ktu", "language:cpy", "language:txu", "language:bjr", "language:bud", "language:ken", "language:bzh", "language:ipi", "language:crx", "language:cko", "language:gnd", "language:kix", "language:mmx", "language:esk", "language:kud", "language:toh", "language:byx", "language:mtp", "language:kyz", "language:lmp", "language:tiy", "language:ccp", "language:sbl", "language:tcs", "language:usp", "language:buk", "language:hvn", "language:amm", "language:kzj", "language:gng", "language:ter", "language:kij", "language:ngl", "language:pab", "language:pad", "language:kca", "language:meq", "language:niq", "language:dah", "language:lia", "language:kss", "language:bch", "language:khz", "language:nop", "language:yle", "language:ain", "language:pem", "language:yml", "language:gaw", "language:kze", "language:tby", "language:cso", "language:ifb", "language:rki", "language:aly", "language:kmu", "language:mhl", "language:kmo", "language:mva", "language:opm", "language:rtm", "language:xtn", "language:dgr", "language:ksr", "language:snp", "language:tvk", "language:zpv", "language:hrx", "language:kwj", "language:muh", "language:kak", "language:kpw", "language:akb", "language:ybb", "language:kjs", "language:wuv", "language:laj", "language:ogo", "language:nim", "language:czt", "language:aoz", "language:rro", "language:yuw", "language:heg", "language:mrw", "language:zac", "language:kpf", "language:leu", "language:nwi", "language:awx", "language:gso", "language:keo", "language:tte", "language:ceg", "language:nhy", "language:bhw", "language:gnw", "language:nin", "language:xmv", "language:crm", "language:mxv", "language:ubr", "language:gld", "language:bzi", "language:ann", "language:lac", "language:mej", "language:zpc", "language:dje", "language:caq", "language:guk", "language:wos", "language:msm", "language:zaw", "language:cap", "language:mox", "language:zab", "language:mca", "language:wrs", "language:hae", "language:nss", "language:rav", "language:uvh", "language:dga", "language:ppo", "language:arq", "language:bim", "language:lwg", "language:kms", "language:kne", "language:tkl", "language:caf", "language:zat", "language:aba", "language:chf", "language:lew", "language:npl", "language:uvl", "language:rmq", "language:mek", "language:snc", "language:nsn", "language:amn", "language:abz", "language:cek", "language:ong", "language:txq", "language:bjp", "language:gvf", "language:crk", "language:snf", "language:nhg", "language:kqw", "language:mic", "language:mie", "language:nus", "language:plu", "language:idu", "language:sll", "language:thl", "language:bef", "language:kwd", "language:mqy", "language:caa", "language:aia", "language:wer", "language:agm", "language:mbl", "language:mbs", "language:adh", "language:hub", "language:mav", "language:nab", "language:bqp", "language:guo", "language:auy", "language:big", "language:nii", "language:row", "language:kim", "language:mti", "language:otm", "language:prg", "language:icr", "language:mcp", "language:tbo", "language:chy", "language:oji", "language:kpx", "language:lbb", "language:pao", "language:xuo", "language:zgh", "language:wls", "language:aca", "language:apy", "language:wnc", "language:ssg", "language:eve", "language:mee", "language:yut", "language:kxm", "language:mur", "language:mcd", "language:cto", "language:mqj", "language:usa", "language:emi", "language:liv", "language:guz", "language:tgp", "language:ncl", "language:azg", "language:kck", "language:apb", "language:kbm", "language:sgz", "language:wsk", "language:gbm", "language:csw", "language:ctp", "language:eri", "language:vid", "language:bsq", "language:dak", "language:tif", "language:xsr", "language:mwp", "language:roo", "language:wbp", "language:gdn", "language:tgo", "language:zpm", "language:mih", "language:qvc", "language:tod", "language:abn", "language:cao", "language:kbc", "language:kog", "language:moh", "language:leh", "language:nsu", "language:okv", "language:tih", "language:zos", "language:llg", "language:gmv", "language:kyf", "language:mhx", "language:mvp", "language:naf", "language:ino", "language:dng", "language:mnk", "language:oke", "language:omw", "language:tnk", "language:xtd", "language:agt", "language:cbr", "language:eko", "language:nhk", "language:hot", "language:iou", "language:mpm", "language:myx", "language:ndh", "language:pbb", "language:rng", "language:cri", "language:bvd", "language:kmz", "language:cta", "language:blw", "language:clu", "language:cof", "language:kje", "language:nmf", "language:kbh", "language:qxn", "language:ria", "language:wed", "language:zar", "language:huv", "language:lsi", "language:msk", "language:nbq", "language:ati", "language:lww", "language:yva", "language:amr", "language:lmk", "language:tim", "language:bbb", "language:bzd", "language:djr", "language:kwf", "language:nuy", "language:poi", "language:ptp", "language:atd", "language:bps", "language:lcm", "language:mbh", "language:nuj", "language:ayp", "language:cwe", "language:geb", "language:aoi", "language:mto", "language:cbi", "language:gvn", "language:kqf", "language:rkb", "language:xed", "language:zas", "language:apz", "language:cpb", "language:ese", "language:gdg", "language:iws", "language:apu", "language:cag", "language:kdc", "language:nou", "language:too", "language:xmm", "language:beq", "language:hnn", "language:myk", "language:sby", "language:arl", "language:ncu", "language:rml", "language:gam", "language:kvn", "language:mbc", "language:mfy", "language:urt", "language:luc", "language:mio", "language:snn", "language:tar", "language:tku", "language:upv", "language:kjg", "language:btd", "language:cut", "language:mwn", "language:tue", "language:adz", "language:kyg", "language:rel", "language:tew", "language:ttc", "language:nkf", "language:yrb", "language:fub", "language:nsm", "language:nyf", "language:bkd", "language:cpc", "language:lif", "language:mwc", "language:apw", "language:azz", "language:zpo", "language:faa", "language:ood", "language:toc", "language:wmt", "language:ajg", "language:nko", "language:ssd", "language:sxb", "language:zxx", "language:not", "language:amk", "language:hch", "language:kqc", "language:aui", "language:llb", "language:avt", "language:kaq", "language:tna", "language:zpt", "language:bwd", "language:stn", "language:kgk", "language:mck", "language:bhg", "language:ifu", "language:kkl", "language:zao", "language:huu", "language:sil", "language:trq", "language:tuf", "language:wnu", "language:amf", "language:fan", "language:zdj", "language:bco", "language:kdl", "language:lln", "language:mit", "language:yaq", "language:ake", "language:ige", "language:mna", "language:nhx", "language:njn", "language:wba", "language:zaa", "language:syb", "language:qxo", "language:rgu", "language:wiu", "language:bla", "language:etr", "language:miy", "language:mjc", "language:nif", "language:pov", "language:qxl", "language:sey", "language:bxh", "language:ddg", "language:mks", "language:nki", "language:pot", "language:prf", "language:hra", "language:law", "language:plw", "language:zpi", "language:atg", "language:daf", "language:drg", "language:dru", "language:ngp", "language:boa", "language:cbt", "language:dhm", "language:ikw", "language:kqe", "language:lbj", "language:loq", "language:xsb", "language:yad", "language:kyu", "language:mcb", "language:tks", "language:cnl", "language:far", "language:miz", "language:maj", "language:mgm", "language:mib", "language:mza", "language:nhr", "language:smt", "language:ish", "language:mlu", "language:yka", "language:anv", "language:ign", "language:mir", "language:tav", "language:gyr", "language:klt", "language:apn", "language:wbm", "language:cjk", "language:dln", "language:mcf", "language:twx", "language:ldi", "language:poy", "language:yaa", "language:zae", "language:bdq", "language:con", "language:dnw", "language:gah", "language:lcp", "language:mbj", "language:xog", "language:yan", "language:sdq", "language:sgh", "language:gwi", "language:lbk", "language:ses", "language:ycn", "language:kau", "language:hix", "language:mig", "language:mlh", "language:aer", "language:amx", "language:auc", "language:cly", "language:dsh", "language:gnb", "language:neb", "language:ttt", "language:jic", "language:xtm", "language:tll", "language:apt", "language:kgr", "language:pio", "language:pmf", "language:akh", "language:box", "language:cub", "language:cux", "language:mrq", "language:bsp", "language:for", "language:kny", "language:uri", "language:zsr", "language:pce", "language:cme", "language:cot", "language:kxw", "language:omb", "language:srq", "language:vag", "language:cpu", "language:kmy", "language:ksc", "language:lai", "language:nnp", "language:ruf", "language:spm", "language:tro", "language:ttj", "language:waj", "language:raj", "language:ian", "language:ivb", "language:kmm", "language:kzf", "language:maw", "language:tuv", "language:whk", "language:aph", "language:anm", "language:mny", "language:trp", "language:are", "language:esi", "language:mhi", "language:obo", "language:bwu", "language:lwo", "language:mrg", "language:old", "language:zav", "language:snk", "language:daa", "language:fuq", "language:lud", "language:nlg", "language:sck", "language:tbl", "language:khb", "language:tpu", "language:fue", "language:tmd", "language:whg", "language:blt", "language:gqr", "language:kub", "language:ots", "language:tpp", "language:var", "language:hbb", "language:krr", "language:alj", "language:ebk", "language:goa", "language:hav", "language:kmk", "language:kqo", "language:muy", "language:njb", "language:snw", "language:wib", "language:tkr", "language:ksp", "language:lip", "language:mog", "language:nla", "language:pss", "language:zad", "language:awi", "language:frd", "language:kei", "language:kmd", "language:mim", "language:adi", "language:any", "language:arp", "language:biu", "language:bvc", "language:ghs", "language:ttq", "language:vmk", "language:zpj", "language:crl", "language:dtb", "language:dug", "language:kam", "language:nmo", "language:qva", "language:rhg", "language:taw", "language:tui", "language:uth", "language:wap", "language:zpg", "language:brb", "language:bqj", "language:cjv", "language:cnw", "language:kgf", "language:mfh", "language:mns", "language:nnw", "language:ntp", "language:oto", "language:dks", "language:mpt", "language:nmw", "language:sba", "language:sfw", "language:shu", "language:ury", "language:yrl", "language:xkg", "language:iqw", "language:kpj", "language:nfa", "language:nph", "language:nre", "language:wsg", "language:xsu", "language:nut", "language:bmk", "language:cdf", "language:enl", "language:hop", "language:ife", "language:lnd", "language:mrv", "language:nzm", "language:tem", "language:gwc", "language:udi", "language:ajz", "language:kru", "language:qvo", "language:sld", "language:yup", "language:kwu", "language:tdg", "language:ade", "language:bcw", "language:bgz", "language:cul", "language:fmu", "language:gvc", "language:ktj", "language:lis", "language:mev", "language:msc", "language:mta", "language:nbc", "language:ncq", "language:way", "language:wlv", "language:wlx", "language:jmx", "language:wni", "language:aim", "language:byv", "language:cgg", "language:fud", "language:jmc", "language:ker", "language:kia", "language:lom", "language:tnc", "language:bwx", "language:gjn", "language:guq", "language:lap", "language:mip", "language:mkl", "language:mtg", "language:mzk", "language:njm", "language:nse", "language:pne", "language:pww", "language:zam", "language:zca", "language:the", "language:dis", "language:lgl", "language:mbd", "language:mda", "language:mhy", "language:nst", "language:pkb", "language:pmq", "language:poe", "language:swk", "language:trs", "language:vot", "language:yam", "language:zpq", "language:atq", "language:bav", "language:cho", "language:chq", "language:cjo", "language:due", "language:duo", "language:etu", "language:ifa", "language:kby", "language:khq", "language:kno", "language:lem", "language:mcn", "language:mkz", "language:mnb", "language:myb", "language:niy", "language:pxm", "language:rub", "language:sju", "language:tmc", "language:tpm", "language:wob", "language:nxa", "language:sdk", "language:bhz", "language:bth", "language:crj", "language:eka", "language:khy", "language:led", "language:met", "language:mgo", "language:mus", "language:mzm", "language:naw", "language:sri", "language:tbk", "language:xbr", "language:tnl", "language:tnn", "language:awb", "language:bsc", "language:iri", "language:kpq", "language:ktm", "language:loe", "language:lun", "language:mfg", "language:moa", "language:ndj", "language:otn", "language:sjo", "language:wat", "language:bkm", "language:tuz", "language:ybi", "language:gbr", "language:gej", "language:giz", "language:gna", "language:gwr", "language:kex", "language:kle", "language:knx", "language:ktz", "language:kvj", "language:lob", "language:mhw", "language:mph", "language:pmx", "language:qus", "language:tcc", "language:azn", "language:bra", "language:gal", "language:nlv", "language:stk", "language:yux", "language:dip", "language:ivv", "language:kdj", "language:kff", "language:mnf", "language:moc", "language:mpg", "language:mua", "language:nma", "language:rim", "language:rjs", "language:rnl", "language:ssx", "language:tig", "language:urw", "language:idt", "language:mrn", "language:oki", "language:bbk", "language:bea", "language:bkq", "language:btt", "language:cou", "language:ddn", "language:did", "language:diu", "language:ilb", "language:ksj", "language:lhi", "language:ndi", "language:ngb", "language:njz", "language:nmz", "language:npo", "language:nwb", "language:pbi", "language:sbd", "language:soe", "language:tpw", "language:twb", "language:bfn", "language:boz", "language:lwl", "language:psp", "language:sgd", "language:ymp", "language:bkl", "language:bkv", "language:csk", "language:enx", "language:ess", "language:gud", "language:guu", "language:kdh", "language:krx", "language:kzn", "language:maf", "language:max", "language:mgc", "language:mvn", "language:mwm", "language:nbe", "language:nlc", "language:nxd", "language:pse", "language:sbs", "language:tlb", "language:wwa", "language:yim", "language:bkx", "language:bze", "language:kbq", "language:mry", "language:tkd", "language:ybh", "language:abi", "language:ayo", "language:bex", "language:bom", "language:bov", "language:dhg", "language:jaa", "language:kqs", "language:kqy", "language:ksb", "language:ldn", "language:log", "language:lol", "language:mfk", "language:mtj", "language:ndz", "language:ngc", "language:nio", "language:nri", "language:nuz", "language:pps", "language:shk", "language:tlj", "language:vut", "language:xnn", "language:bcc", "language:cuh", "language:hro", "language:nco", "language:nxl", "language:pex", "language:tnp", "language:vif", "language:aha", "language:avn", "language:bmv", "language:chj", "language:crt", "language:dos", "language:eto", "language:gya", "language:hlt", "language:jun", "language:ktb", "language:lgm", "language:mnx", "language:mor", "language:mug", "language:nng", "language:nnh", "language:nsa", "language:nyk", "language:otd", "language:oym", "language:phm", "language:prq", "language:saj", "language:szb", "language:tap", "language:tnr", "language:tzl", "language:wew", "language:yas", "language:yuz", "language:baw", "language:bwt", "language:bxa", "language:fli", "language:kqr", "language:lan", "language:lgr", "language:lhm", "language:mlk", "language:mve", "language:odk", "language:pea", "language:saq", "language:sea", "language:sps", "language:tio", "language:yin", "language:adl", "language:agw", "language:ald", "language:asg", "language:bnj", "language:bwi", "language:bwq", "language:dbq", "language:dow", "language:fad", "language:hoc", "language:kql", "language:ksf", "language:lea", "language:mse", "language:mzl", "language:nhd", "language:nnl", "language:nwx", "language:pbc", "language:pny", "language:tul", "language:aaa", "language:abc", "language:agq", "language:ags", "language:azo", "language:bag", "language:bax", "language:bce", "language:bfm", "language:bgf", "language:bhs", "language:bkc", "language:bkh", "language:bob", "language:bqm", "language:bri", "language:brv", "language:buo", "language:bya", "language:chp", "language:cim", "language:clo", "language:cuv", "language:dmg", "language:dtr", "language:eee", "language:ekm", "language:enc", "language:gbj", "language:gou", "language:hao", "language:hna", "language:hre", "language:isu", "language:jgo", "language:kbx", "language:kvt", "language:kxp", "language:lfa", "language:lkb", "language:lns", "language:loh", "language:lts", "language:luy", "language:mdr", "language:mfj", "language:mgg", "language:mgq", "language:mle", "language:mlw", "language:mmu", "language:mne", "language:muv", "language:mxu", "language:nge", "language:ngn", "language:njy", "language:nsk", "language:nwe", "language:nza", "language:pcg", "language:pdu", "language:pnz", "language:pwg", "language:qaa", "language:snl", "language:sox", "language:ssn", "language:syw", "language:tbj", "language:tdb", "language:thy", "language:tnt", "language:tom", "language:tpl", "language:tsb", "language:tvs", "language:udg", "language:unr", "language:wbr", "language:wms", "language:wnk", "language:wtk", "language:xmd", "language:xmg", "language:xty", "language:yav", "language:yea", "language:zuh", "license:other", "size_categories:1B<n<10B", "format:json", "modality:tabular", "modality:text", "library:datasets", "library:dask", "library:mlcroissant", "arxiv:2502.11546", "region:us", "multilingual" ]
[]
2025-03-25T04:55:24Z
null
--- license: other pretty_name: dcad-2000 license_name: license license_link: LICENSE configs: - config_name: eng_Latn data_files: - split: train path: eng_Latn/*_keep.jsonl - config_name: rus_Cyrl data_files: - split: train path: rus_Cyrl/*_keep.jsonl - config_name: cmn_Hani data_files: - split: train path: cmn_Hani/*_keep.jsonl - config_name: deu_Latn data_files: - split: train path: deu_Latn/*_keep.jsonl - config_name: spa_Latn data_files: - split: train path: spa_Latn/*_keep.jsonl - config_name: fra_Latn data_files: - split: train path: fra_Latn/*_keep.jsonl - config_name: jpn_Jpan data_files: - split: train path: jpn_Jpan/*_keep.jsonl - config_name: ita_Latn data_files: - split: train path: ita_Latn/*_keep.jsonl - config_name: por_Latn data_files: - split: train path: por_Latn/*_keep.jsonl - config_name: pol_Latn data_files: - split: train path: pol_Latn/*_keep.jsonl - config_name: nld_Latn data_files: - split: train path: nld_Latn/*_keep.jsonl - config_name: ind_Latn data_files: - split: train path: ind_Latn/*_keep.jsonl - config_name: tur_Latn data_files: - split: train path: tur_Latn/*_keep.jsonl - config_name: vie_Latn data_files: - split: train path: vie_Latn/*_keep.jsonl - config_name: fas_Arab data_files: - split: train path: fas_Arab/*_keep.jsonl - config_name: kor_Hang data_files: - split: train path: kor_Hang/*_keep.jsonl - config_name: swe_Latn data_files: - split: train path: swe_Latn/*_keep.jsonl - config_name: hun_Latn data_files: - split: train path: hun_Latn/*_keep.jsonl - config_name: ell_Grek data_files: - split: train path: ell_Grek/*_keep.jsonl - config_name: ukr_Cyrl data_files: - split: train path: ukr_Cyrl/*_keep.jsonl - config_name: tha_Thai data_files: - split: train path: tha_Thai/*_keep.jsonl - config_name: aze_Latn data_files: - split: train path: aze_Latn/*_keep.jsonl - config_name: arb_Arab data_files: - split: train path: arb_Arab/*_keep.jsonl - config_name: slv_Latn data_files: - split: train path: slv_Latn/*_keep.jsonl - config_name: cat_Latn data_files: - split: train path: cat_Latn/*_keep.jsonl - config_name: fin_Latn data_files: - split: train path: fin_Latn/*_keep.jsonl - config_name: hbs_Latn data_files: - split: train path: hbs_Latn/*_keep.jsonl - config_name: ces_Latn data_files: - split: train path: ces_Latn/*_keep.jsonl - config_name: fil_Latn data_files: - split: train path: fil_Latn/*_keep.jsonl - config_name: guj_Gujr data_files: - split: train path: guj_Gujr/*_keep.jsonl - config_name: nob_Latn data_files: - split: train path: nob_Latn/*_keep.jsonl - config_name: mal_Mlym data_files: - split: train path: mal_Mlym/*_keep.jsonl - config_name: bul_Cyrl data_files: - split: train path: bul_Cyrl/*_keep.jsonl - config_name: kan_Knda data_files: - split: train path: kan_Knda/*_keep.jsonl - config_name: hin_Deva data_files: - split: train path: hin_Deva/*_keep.jsonl - config_name: tam_Taml data_files: - split: train path: tam_Taml/*_keep.jsonl - config_name: ara_Arab data_files: - split: train path: ara_Arab/*_keep.jsonl - config_name: kaz_Cyrl data_files: - split: train path: kaz_Cyrl/*_keep.jsonl - config_name: sqi_Latn data_files: - split: train path: sqi_Latn/*_keep.jsonl - config_name: pan_Guru data_files: - split: train path: pan_Guru/*_keep.jsonl - config_name: est_Latn data_files: - split: train path: est_Latn/*_keep.jsonl - config_name: srp_Cyrl data_files: - split: train path: srp_Cyrl/*_keep.jsonl - config_name: heb_Hebr data_files: - split: train path: heb_Hebr/*_keep.jsonl - config_name: isl_Latn data_files: - split: train path: isl_Latn/*_keep.jsonl - config_name: mlt_Latn data_files: - split: train path: mlt_Latn/*_keep.jsonl - config_name: mkd_Cyrl data_files: - split: train path: mkd_Cyrl/*_keep.jsonl - config_name: bos_Latn data_files: - split: train path: bos_Latn/*_keep.jsonl - config_name: lit_Latn data_files: - split: train path: lit_Latn/*_keep.jsonl - config_name: kat_Geor data_files: - split: train path: kat_Geor/*_keep.jsonl - config_name: hrv_Latn data_files: - split: train path: hrv_Latn/*_keep.jsonl - config_name: ben_Beng data_files: - split: train path: ben_Beng/*_keep.jsonl - config_name: glg_Latn data_files: - split: train path: glg_Latn/*_keep.jsonl - config_name: ceb_Latn data_files: - split: train path: ceb_Latn/*_keep.jsonl - config_name: ron_Latn data_files: - split: train path: ron_Latn/*_keep.jsonl - config_name: msa_Latn data_files: - split: train path: msa_Latn/*_keep.jsonl - config_name: hye_Armn data_files: - split: train path: hye_Armn/*_keep.jsonl - config_name: tgk_Cyrl data_files: - split: train path: tgk_Cyrl/*_keep.jsonl - config_name: mar_Deva data_files: - split: train path: mar_Deva/*_keep.jsonl - config_name: bel_Cyrl data_files: - split: train path: bel_Cyrl/*_keep.jsonl - config_name: nep_Deva data_files: - split: train path: nep_Deva/*_keep.jsonl - config_name: urd_Arab data_files: - split: train path: urd_Arab/*_keep.jsonl - config_name: slk_Latn data_files: - split: train path: slk_Latn/*_keep.jsonl - config_name: mon_Cyrl data_files: - split: train path: mon_Cyrl/*_keep.jsonl - config_name: swa_Latn data_files: - split: train path: swa_Latn/*_keep.jsonl - config_name: dan_Latn data_files: - split: train path: dan_Latn/*_keep.jsonl - config_name: eus_Latn data_files: - split: train path: eus_Latn/*_keep.jsonl - config_name: azj_Latn data_files: - split: train path: azj_Latn/*_keep.jsonl - config_name: sin_Sinh data_files: - split: train path: sin_Sinh/*_keep.jsonl - config_name: lat_Latn data_files: - split: train path: lat_Latn/*_keep.jsonl - config_name: als_Latn data_files: - split: train path: als_Latn/*_keep.jsonl - config_name: tel_Telu data_files: - split: train path: tel_Telu/*_keep.jsonl - config_name: afr_Latn data_files: - split: train path: afr_Latn/*_keep.jsonl - config_name: ekk_Latn data_files: - split: train path: ekk_Latn/*_keep.jsonl - config_name: ltz_Latn data_files: - split: train path: ltz_Latn/*_keep.jsonl - config_name: zsm_Latn data_files: - split: train path: zsm_Latn/*_keep.jsonl - config_name: som_Latn data_files: - split: train path: som_Latn/*_keep.jsonl - config_name: cym_Latn data_files: - split: train path: cym_Latn/*_keep.jsonl - config_name: und_Kana data_files: - split: train path: und_Kana/*_keep.jsonl - config_name: kir_Cyrl data_files: - split: train path: kir_Cyrl/*_keep.jsonl - config_name: nor_Latn data_files: - split: train path: nor_Latn/*_keep.jsonl - config_name: uzb_Latn data_files: - split: train path: uzb_Latn/*_keep.jsonl - config_name: mya_Mymr data_files: - split: train path: mya_Mymr/*_keep.jsonl - config_name: epo_Latn data_files: - split: train path: epo_Latn/*_keep.jsonl - config_name: hau_Latn data_files: - split: train path: hau_Latn/*_keep.jsonl - config_name: ary_Arab data_files: - split: train path: ary_Arab/*_keep.jsonl - config_name: gle_Latn data_files: - split: train path: gle_Latn/*_keep.jsonl - config_name: lvs_Latn data_files: - split: train path: lvs_Latn/*_keep.jsonl - config_name: nno_Latn data_files: - split: train path: nno_Latn/*_keep.jsonl - config_name: und_Bamu data_files: - split: train path: und_Bamu/*_keep.jsonl - config_name: ory_Orya data_files: - split: train path: ory_Orya/*_keep.jsonl - config_name: amh_Ethi data_files: - split: train path: amh_Ethi/*_keep.jsonl - config_name: tat_Cyrl data_files: - split: train path: tat_Cyrl/*_keep.jsonl - config_name: khm_Khmr data_files: - split: train path: khm_Khmr/*_keep.jsonl - config_name: und_Copt data_files: - split: train path: und_Copt/*_keep.jsonl - config_name: und_Tang data_files: - split: train path: und_Tang/*_keep.jsonl - config_name: und_Xsux data_files: - split: train path: und_Xsux/*_keep.jsonl - config_name: arz_Arab data_files: - split: train path: arz_Arab/*_keep.jsonl - config_name: pus_Arab data_files: - split: train path: pus_Arab/*_keep.jsonl - config_name: lav_Latn data_files: - split: train path: lav_Latn/*_keep.jsonl - config_name: hbs_Cyrl data_files: - split: train path: hbs_Cyrl/*_keep.jsonl - config_name: und_Yiii data_files: - split: train path: und_Yiii/*_keep.jsonl - config_name: war_Latn data_files: - split: train path: war_Latn/*_keep.jsonl - config_name: multi_Latn data_files: - split: train path: multi_Latn/*_keep.jsonl - config_name: und_Hira data_files: - split: train path: und_Hira/*_keep.jsonl - config_name: mlg_Latn data_files: - split: train path: mlg_Latn/*_keep.jsonl - config_name: uzn_Cyrl data_files: - split: train path: uzn_Cyrl/*_keep.jsonl - config_name: hat_Latn data_files: - split: train path: hat_Latn/*_keep.jsonl - config_name: zul_Latn data_files: - split: train path: zul_Latn/*_keep.jsonl - config_name: kur_Latn data_files: - split: train path: kur_Latn/*_keep.jsonl - config_name: tgl_Latn data_files: - split: train path: tgl_Latn/*_keep.jsonl - config_name: uzb_Cyrl data_files: - split: train path: uzb_Cyrl/*_keep.jsonl - config_name: div_Thaa data_files: - split: train path: div_Thaa/*_keep.jsonl - config_name: und_Laoo data_files: - split: train path: und_Laoo/*_keep.jsonl - config_name: fry_Latn data_files: - split: train path: fry_Latn/*_keep.jsonl - config_name: sna_Latn data_files: - split: train path: sna_Latn/*_keep.jsonl - config_name: und_Cyrl data_files: - split: train path: und_Cyrl/*_keep.jsonl - config_name: fao_Latn data_files: - split: train path: fao_Latn/*_keep.jsonl - config_name: sun_Latn data_files: - split: train path: sun_Latn/*_keep.jsonl - config_name: und_Kits data_files: - split: train path: und_Kits/*_keep.jsonl - config_name: und_Hluw data_files: - split: train path: und_Hluw/*_keep.jsonl - config_name: snd_Arab data_files: - split: train path: snd_Arab/*_keep.jsonl - config_name: cos_Latn data_files: - split: train path: cos_Latn/*_keep.jsonl - config_name: bak_Cyrl data_files: - split: train path: bak_Cyrl/*_keep.jsonl - config_name: ckb_Arab data_files: - split: train path: ckb_Arab/*_keep.jsonl - config_name: asm_Beng data_files: - split: train path: asm_Beng/*_keep.jsonl - config_name: und_Grek data_files: - split: train path: und_Grek/*_keep.jsonl - config_name: jpn_Japn data_files: - split: train path: jpn_Japn/*_keep.jsonl - config_name: ast_Latn data_files: - split: train path: ast_Latn/*_keep.jsonl - config_name: ibo_Latn data_files: - split: train path: ibo_Latn/*_keep.jsonl - config_name: mri_Latn data_files: - split: train path: mri_Latn/*_keep.jsonl - config_name: und_Lina data_files: - split: train path: und_Lina/*_keep.jsonl - config_name: ars_Arab data_files: - split: train path: ars_Arab/*_keep.jsonl - config_name: und_Shrd data_files: - split: train path: und_Shrd/*_keep.jsonl - config_name: anp_Deva data_files: - split: train path: anp_Deva/*_keep.jsonl - config_name: khk_Cyrl data_files: - split: train path: khk_Cyrl/*_keep.jsonl - config_name: und_Samr data_files: - split: train path: und_Samr/*_keep.jsonl - config_name: lao_Laoo data_files: - split: train path: lao_Laoo/*_keep.jsonl - config_name: ori_Orya data_files: - split: train path: ori_Orya/*_keep.jsonl - config_name: und_Cans data_files: - split: train path: und_Cans/*_keep.jsonl - config_name: yid_Hebr data_files: - split: train path: yid_Hebr/*_keep.jsonl - config_name: hmn_Latn data_files: - split: train path: hmn_Latn/*_keep.jsonl - config_name: jav_Latn data_files: - split: train path: jav_Latn/*_keep.jsonl - config_name: nya_Latn data_files: - split: train path: nya_Latn/*_keep.jsonl - config_name: und_Mend data_files: - split: train path: und_Mend/*_keep.jsonl - config_name: und_Adlm data_files: - split: train path: und_Adlm/*_keep.jsonl - config_name: und_Egyp data_files: - split: train path: und_Egyp/*_keep.jsonl - config_name: tir_Ethi data_files: - split: train path: tir_Ethi/*_keep.jsonl - config_name: uig_Arab data_files: - split: train path: uig_Arab/*_keep.jsonl - config_name: wln_Latn data_files: - split: train path: wln_Latn/*_keep.jsonl - config_name: und_Syrc data_files: - split: train path: und_Syrc/*_keep.jsonl - config_name: yor_Latn data_files: - split: train path: yor_Latn/*_keep.jsonl - config_name: swh_Latn data_files: - split: train path: swh_Latn/*_keep.jsonl - config_name: gla_Latn data_files: - split: train path: gla_Latn/*_keep.jsonl - config_name: xho_Latn data_files: - split: train path: xho_Latn/*_keep.jsonl - config_name: uzn_Latn data_files: - split: train path: uzn_Latn/*_keep.jsonl - config_name: bre_Latn data_files: - split: train path: bre_Latn/*_keep.jsonl - config_name: sot_Latn data_files: - split: train path: sot_Latn/*_keep.jsonl - config_name: che_Cyrl data_files: - split: train path: che_Cyrl/*_keep.jsonl - config_name: nan_Latn data_files: - split: train path: nan_Latn/*_keep.jsonl - config_name: tel_Latn data_files: - split: train path: tel_Latn/*_keep.jsonl - config_name: bew_Latn data_files: - split: train path: bew_Latn/*_keep.jsonl - config_name: glk_Arab data_files: - split: train path: glk_Arab/*_keep.jsonl - config_name: smo_Latn data_files: - split: train path: smo_Latn/*_keep.jsonl - config_name: orm_Latn data_files: - split: train path: orm_Latn/*_keep.jsonl - config_name: zho_Hani data_files: - split: train path: zho_Hani/*_keep.jsonl - config_name: haw_Latn data_files: - split: train path: haw_Latn/*_keep.jsonl - config_name: pnb_Arab data_files: - split: train path: pnb_Arab/*_keep.jsonl - config_name: und_Linb data_files: - split: train path: und_Linb/*_keep.jsonl - config_name: oci_Latn data_files: - split: train path: oci_Latn/*_keep.jsonl - config_name: chv_Cyrl data_files: - split: train path: chv_Cyrl/*_keep.jsonl - config_name: kin_Latn data_files: - split: train path: kin_Latn/*_keep.jsonl - config_name: und_Brai data_files: - split: train path: und_Brai/*_keep.jsonl - config_name: srp_Latn data_files: - split: train path: srp_Latn/*_keep.jsonl - config_name: und_Hung data_files: - split: train path: und_Hung/*_keep.jsonl - config_name: und_Sgnw data_files: - split: train path: und_Sgnw/*_keep.jsonl - config_name: pcm_Latn data_files: - split: train path: pcm_Latn/*_keep.jsonl - config_name: kaa_Cyrl data_files: - split: train path: kaa_Cyrl/*_keep.jsonl - config_name: lug_Latn data_files: - split: train path: lug_Latn/*_keep.jsonl - config_name: pbt_Arab data_files: - split: train path: pbt_Arab/*_keep.jsonl - config_name: aze_Arab data_files: - split: train path: aze_Arab/*_keep.jsonl - config_name: gsw_Latn data_files: - split: train path: gsw_Latn/*_keep.jsonl - config_name: min_Latn data_files: - split: train path: min_Latn/*_keep.jsonl - config_name: tuk_Latn data_files: - split: train path: tuk_Latn/*_keep.jsonl - config_name: lim_Latn data_files: - split: train path: lim_Latn/*_keep.jsonl - config_name: kmr_Latn data_files: - split: train path: kmr_Latn/*_keep.jsonl - config_name: und_Nshu data_files: - split: train path: und_Nshu/*_keep.jsonl - config_name: roh_Latn data_files: - split: train path: roh_Latn/*_keep.jsonl - config_name: und_Bali data_files: - split: train path: und_Bali/*_keep.jsonl - config_name: und_Lana data_files: - split: train path: und_Lana/*_keep.jsonl - config_name: vec_Latn data_files: - split: train path: vec_Latn/*_keep.jsonl - config_name: san_Deva data_files: - split: train path: san_Deva/*_keep.jsonl - config_name: und_Modi data_files: - split: train path: und_Modi/*_keep.jsonl - config_name: gmh_Latn data_files: - split: train path: gmh_Latn/*_keep.jsonl - config_name: nds_Latn data_files: - split: train path: nds_Latn/*_keep.jsonl - config_name: sco_Latn data_files: - split: train path: sco_Latn/*_keep.jsonl - config_name: und_Mong data_files: - split: train path: und_Mong/*_keep.jsonl - config_name: und_Ethi data_files: - split: train path: und_Ethi/*_keep.jsonl - config_name: azb_Arab data_files: - split: train path: azb_Arab/*_keep.jsonl - config_name: tsn_Latn data_files: - split: train path: tsn_Latn/*_keep.jsonl - config_name: rus_Latn data_files: - split: train path: rus_Latn/*_keep.jsonl - config_name: und_Hebr data_files: - split: train path: und_Hebr/*_keep.jsonl - config_name: mon_Latn data_files: - split: train path: mon_Latn/*_keep.jsonl - config_name: und_Saur data_files: - split: train path: und_Saur/*_keep.jsonl - config_name: tgk_Latn data_files: - split: train path: tgk_Latn/*_keep.jsonl - config_name: sah_Cyrl data_files: - split: train path: sah_Cyrl/*_keep.jsonl - config_name: pri_Latn data_files: - split: train path: pri_Latn/*_keep.jsonl - config_name: pap_Latn data_files: - split: train path: pap_Latn/*_keep.jsonl - config_name: plt_Latn data_files: - split: train path: plt_Latn/*_keep.jsonl - config_name: lmo_Latn data_files: - split: train path: lmo_Latn/*_keep.jsonl - config_name: bod_Tibt data_files: - split: train path: bod_Tibt/*_keep.jsonl - config_name: und_Vaii data_files: - split: train path: und_Vaii/*_keep.jsonl - config_name: yue_Hani data_files: - split: train path: yue_Hani/*_keep.jsonl - config_name: und_Dupl data_files: - split: train path: und_Dupl/*_keep.jsonl - config_name: und_Glag data_files: - split: train path: und_Glag/*_keep.jsonl - config_name: und_Thai data_files: - split: train path: und_Thai/*_keep.jsonl - config_name: bar_Latn data_files: - split: train path: bar_Latn/*_keep.jsonl - config_name: und_Thaa data_files: - split: train path: und_Thaa/*_keep.jsonl - config_name: hif_Latn data_files: - split: train path: hif_Latn/*_keep.jsonl - config_name: und_Runr data_files: - split: train path: und_Runr/*_keep.jsonl - config_name: arg_Latn data_files: - split: train path: arg_Latn/*_keep.jsonl - config_name: pms_Latn data_files: - split: train path: pms_Latn/*_keep.jsonl - config_name: kat_Latn data_files: - split: train path: kat_Latn/*_keep.jsonl - config_name: nrm_Latn data_files: - split: train path: nrm_Latn/*_keep.jsonl - config_name: vol_Latn data_files: - split: train path: vol_Latn/*_keep.jsonl - config_name: aeb_Arab data_files: - split: train path: aeb_Arab/*_keep.jsonl - config_name: und_Khmr data_files: - split: train path: und_Khmr/*_keep.jsonl - config_name: und_Geor data_files: - split: train path: und_Geor/*_keep.jsonl - config_name: ido_Latn data_files: - split: train path: ido_Latn/*_keep.jsonl - config_name: pam_Latn data_files: - split: train path: pam_Latn/*_keep.jsonl - config_name: kal_Latn data_files: - split: train path: kal_Latn/*_keep.jsonl - config_name: und_Dsrt data_files: - split: train path: und_Dsrt/*_keep.jsonl - config_name: und_Tibt data_files: - split: train path: und_Tibt/*_keep.jsonl - config_name: und_Mymr data_files: - split: train path: und_Mymr/*_keep.jsonl - config_name: lus_Latn data_files: - split: train path: lus_Latn/*_keep.jsonl - config_name: und_Armn data_files: - split: train path: und_Armn/*_keep.jsonl - config_name: und_Mroo data_files: - split: train path: und_Mroo/*_keep.jsonl - config_name: und_Bopo data_files: - split: train path: und_Bopo/*_keep.jsonl - config_name: new_Deva data_files: - split: train path: new_Deva/*_keep.jsonl - config_name: sme_Latn data_files: - split: train path: sme_Latn/*_keep.jsonl - config_name: und_Khar data_files: - split: train path: und_Khar/*_keep.jsonl - config_name: und_Mtei data_files: - split: train path: und_Mtei/*_keep.jsonl - config_name: nso_Latn data_files: - split: train path: nso_Latn/*_keep.jsonl - config_name: lld_Latn data_files: - split: train path: lld_Latn/*_keep.jsonl - config_name: ina_Latn data_files: - split: train path: ina_Latn/*_keep.jsonl - config_name: scn_Latn data_files: - split: train path: scn_Latn/*_keep.jsonl - config_name: und_Deva data_files: - split: train path: und_Deva/*_keep.jsonl - config_name: und_Brah data_files: - split: train path: und_Brah/*_keep.jsonl - config_name: und_Bhks data_files: - split: train path: und_Bhks/*_keep.jsonl - config_name: hyw_Armn data_files: - split: train path: hyw_Armn/*_keep.jsonl - config_name: abk_Cyrl data_files: - split: train path: abk_Cyrl/*_keep.jsonl - config_name: bew_Cyrl data_files: - split: train path: bew_Cyrl/*_keep.jsonl - config_name: und_Telu data_files: - split: train path: und_Telu/*_keep.jsonl - config_name: und_Merc data_files: - split: train path: und_Merc/*_keep.jsonl - config_name: bpy_Beng data_files: - split: train path: bpy_Beng/*_keep.jsonl - config_name: tgk_Arab data_files: - split: train path: tgk_Arab/*_keep.jsonl - config_name: lin_Latn data_files: - split: train path: lin_Latn/*_keep.jsonl - config_name: oss_Cyrl data_files: - split: train path: oss_Cyrl/*_keep.jsonl - config_name: que_Latn data_files: - split: train path: que_Latn/*_keep.jsonl - config_name: szl_Latn data_files: - split: train path: szl_Latn/*_keep.jsonl - config_name: sdh_Arab data_files: - split: train path: sdh_Arab/*_keep.jsonl - config_name: mww_Latn data_files: - split: train path: mww_Latn/*_keep.jsonl - config_name: mhr_Cyrl data_files: - split: train path: mhr_Cyrl/*_keep.jsonl - config_name: und_Hmnp data_files: - split: train path: und_Hmnp/*_keep.jsonl - config_name: und_Orya data_files: - split: train path: und_Orya/*_keep.jsonl - config_name: und_Kali data_files: - split: train path: und_Kali/*_keep.jsonl - config_name: rmy_Latn data_files: - split: train path: rmy_Latn/*_keep.jsonl - config_name: diq_Latn data_files: - split: train path: diq_Latn/*_keep.jsonl - config_name: srd_Latn data_files: - split: train path: srd_Latn/*_keep.jsonl - config_name: und_Plrd data_files: - split: train path: und_Plrd/*_keep.jsonl - config_name: und_Phag data_files: - split: train path: und_Phag/*_keep.jsonl - config_name: ydd_Hebr data_files: - split: train path: ydd_Hebr/*_keep.jsonl - config_name: und_Hmng data_files: - split: train path: und_Hmng/*_keep.jsonl - config_name: run_Latn data_files: - split: train path: run_Latn/*_keep.jsonl - config_name: und_Nkoo data_files: - split: train path: und_Nkoo/*_keep.jsonl - config_name: hsb_Latn data_files: - split: train path: hsb_Latn/*_keep.jsonl - config_name: und_Lisu data_files: - split: train path: und_Lisu/*_keep.jsonl - config_name: wol_Latn data_files: - split: train path: wol_Latn/*_keep.jsonl - config_name: und_Gran data_files: - split: train path: und_Gran/*_keep.jsonl - config_name: urd_Latn data_files: - split: train path: urd_Latn/*_keep.jsonl - config_name: und_Cher data_files: - split: train path: und_Cher/*_keep.jsonl - config_name: kiu_Latn data_files: - split: train path: kiu_Latn/*_keep.jsonl - config_name: ilo_Latn data_files: - split: train path: ilo_Latn/*_keep.jsonl - config_name: und_Taml data_files: - split: train path: und_Taml/*_keep.jsonl - config_name: cak_Latn data_files: - split: train path: cak_Latn/*_keep.jsonl - config_name: und_Gonm data_files: - split: train path: und_Gonm/*_keep.jsonl - config_name: acm_Arab data_files: - split: train path: acm_Arab/*_keep.jsonl - config_name: und_Tnsa data_files: - split: train path: und_Tnsa/*_keep.jsonl - config_name: und_Cprt data_files: - split: train path: und_Cprt/*_keep.jsonl - config_name: grc_Grek data_files: - split: train path: grc_Grek/*_keep.jsonl - config_name: xmf_Geor data_files: - split: train path: xmf_Geor/*_keep.jsonl - config_name: vls_Latn data_files: - split: train path: vls_Latn/*_keep.jsonl - config_name: und_Gujr data_files: - split: train path: und_Gujr/*_keep.jsonl - config_name: und_Cari data_files: - split: train path: und_Cari/*_keep.jsonl - config_name: tso_Latn data_files: - split: train path: tso_Latn/*_keep.jsonl - config_name: crh_Latn data_files: - split: train path: crh_Latn/*_keep.jsonl - config_name: nav_Latn data_files: - split: train path: nav_Latn/*_keep.jsonl - config_name: mwl_Latn data_files: - split: train path: mwl_Latn/*_keep.jsonl - config_name: und_Diak data_files: - split: train path: und_Diak/*_keep.jsonl - config_name: tat_Latn data_files: - split: train path: tat_Latn/*_keep.jsonl - config_name: grn_Latn data_files: - split: train path: grn_Latn/*_keep.jsonl - config_name: mui_Latn data_files: - split: train path: mui_Latn/*_keep.jsonl - config_name: hin_Latn data_files: - split: train path: hin_Latn/*_keep.jsonl - config_name: mvf_Mong data_files: - split: train path: mvf_Mong/*_keep.jsonl - config_name: udm_Cyrl data_files: - split: train path: udm_Cyrl/*_keep.jsonl - config_name: mzn_Arab data_files: - split: train path: mzn_Arab/*_keep.jsonl - config_name: und_Marc data_files: - split: train path: und_Marc/*_keep.jsonl - config_name: und_Talu data_files: - split: train path: und_Talu/*_keep.jsonl - config_name: und_Vith data_files: - split: train path: und_Vith/*_keep.jsonl - config_name: bcl_Latn data_files: - split: train path: bcl_Latn/*_keep.jsonl - config_name: und_Nagm data_files: - split: train path: und_Nagm/*_keep.jsonl - config_name: und_Mani data_files: - split: train path: und_Mani/*_keep.jsonl - config_name: ban_Latn data_files: - split: train path: ban_Latn/*_keep.jsonl - config_name: und_Mlym data_files: - split: train path: und_Mlym/*_keep.jsonl - config_name: afb_Arab data_files: - split: train path: afb_Arab/*_keep.jsonl - config_name: und_Java data_files: - split: train path: und_Java/*_keep.jsonl - config_name: cnh_Latn data_files: - split: train path: cnh_Latn/*_keep.jsonl - config_name: und_Ahom data_files: - split: train path: und_Ahom/*_keep.jsonl - config_name: cbk_Latn data_files: - split: train path: cbk_Latn/*_keep.jsonl - config_name: kaa_Latn data_files: - split: train path: kaa_Latn/*_keep.jsonl - config_name: tyv_Cyrl data_files: - split: train path: tyv_Cyrl/*_keep.jsonl - config_name: kur_Arab data_files: - split: train path: kur_Arab/*_keep.jsonl - config_name: ful_Latn data_files: - split: train path: ful_Latn/*_keep.jsonl - config_name: und_Beng data_files: - split: train path: und_Beng/*_keep.jsonl - config_name: zea_Latn data_files: - split: train path: zea_Latn/*_keep.jsonl - config_name: frp_Latn data_files: - split: train path: frp_Latn/*_keep.jsonl - config_name: frr_Latn data_files: - split: train path: frr_Latn/*_keep.jsonl - config_name: pfl_Latn data_files: - split: train path: pfl_Latn/*_keep.jsonl - config_name: lij_Latn data_files: - split: train path: lij_Latn/*_keep.jsonl - config_name: apc_Arab data_files: - split: train path: apc_Arab/*_keep.jsonl - config_name: mai_Deva data_files: - split: train path: mai_Deva/*_keep.jsonl - config_name: und_Wara data_files: - split: train path: und_Wara/*_keep.jsonl - config_name: und_Palm data_files: - split: train path: und_Palm/*_keep.jsonl - config_name: pon_Latn data_files: - split: train path: pon_Latn/*_keep.jsonl - config_name: bak_Latn data_files: - split: train path: bak_Latn/*_keep.jsonl - config_name: eml_Latn data_files: - split: train path: eml_Latn/*_keep.jsonl - config_name: uzs_Arab data_files: - split: train path: uzs_Arab/*_keep.jsonl - config_name: nde_Latn data_files: - split: train path: nde_Latn/*_keep.jsonl - config_name: hil_Latn data_files: - split: train path: hil_Latn/*_keep.jsonl - config_name: mam_Latn data_files: - split: train path: mam_Latn/*_keep.jsonl - config_name: gom_Latn data_files: - split: train path: gom_Latn/*_keep.jsonl - config_name: dag_Latn data_files: - split: train path: dag_Latn/*_keep.jsonl - config_name: nap_Latn data_files: - split: train path: nap_Latn/*_keep.jsonl - config_name: bjn_Latn data_files: - split: train path: bjn_Latn/*_keep.jsonl - config_name: und_Khoj data_files: - split: train path: und_Khoj/*_keep.jsonl - config_name: hbo_Hebr data_files: - split: train path: hbo_Hebr/*_keep.jsonl - config_name: tet_Latn data_files: - split: train path: tet_Latn/*_keep.jsonl - config_name: inh_Cyrl data_files: - split: train path: inh_Cyrl/*_keep.jsonl - config_name: ven_Latn data_files: - split: train path: ven_Latn/*_keep.jsonl - config_name: gaz_Latn data_files: - split: train path: gaz_Latn/*_keep.jsonl - config_name: tpi_Latn data_files: - split: train path: tpi_Latn/*_keep.jsonl - config_name: und_Sidd data_files: - split: train path: und_Sidd/*_keep.jsonl - config_name: ajp_Arab data_files: - split: train path: ajp_Arab/*_keep.jsonl - config_name: sat_Olck data_files: - split: train path: sat_Olck/*_keep.jsonl - config_name: wuu_Hani data_files: - split: train path: wuu_Hani/*_keep.jsonl - config_name: bam_Latn data_files: - split: train path: bam_Latn/*_keep.jsonl - config_name: vep_Latn data_files: - split: train path: vep_Latn/*_keep.jsonl - config_name: glv_Latn data_files: - split: train path: glv_Latn/*_keep.jsonl - config_name: tzo_Latn data_files: - split: train path: tzo_Latn/*_keep.jsonl - config_name: bik_Latn data_files: - split: train path: bik_Latn/*_keep.jsonl - config_name: und_Rohg data_files: - split: train path: und_Rohg/*_keep.jsonl - config_name: acr_Latn data_files: - split: train path: acr_Latn/*_keep.jsonl - config_name: twi_Latn data_files: - split: train path: twi_Latn/*_keep.jsonl - config_name: myv_Cyrl data_files: - split: train path: myv_Cyrl/*_keep.jsonl - config_name: ewe_Latn data_files: - split: train path: ewe_Latn/*_keep.jsonl - config_name: poh_Latn data_files: - split: train path: poh_Latn/*_keep.jsonl - config_name: und_Avst data_files: - split: train path: und_Avst/*_keep.jsonl - config_name: ile_Latn data_files: - split: train path: ile_Latn/*_keep.jsonl - config_name: rue_Cyrl data_files: - split: train path: rue_Cyrl/*_keep.jsonl - config_name: und_Ougr data_files: - split: train path: und_Ougr/*_keep.jsonl - config_name: und_Wcho data_files: - split: train path: und_Wcho/*_keep.jsonl - config_name: ace_Latn data_files: - split: train path: ace_Latn/*_keep.jsonl - config_name: kha_Latn data_files: - split: train path: kha_Latn/*_keep.jsonl - config_name: und_Yezi data_files: - split: train path: und_Yezi/*_keep.jsonl - config_name: und_Ital data_files: - split: train path: und_Ital/*_keep.jsonl - config_name: und_Kthi data_files: - split: train path: und_Kthi/*_keep.jsonl - config_name: und_Takr data_files: - split: train path: und_Takr/*_keep.jsonl - config_name: ksh_Latn data_files: - split: train path: ksh_Latn/*_keep.jsonl - config_name: hui_Latn data_files: - split: train path: hui_Latn/*_keep.jsonl - config_name: gor_Latn data_files: - split: train path: gor_Latn/*_keep.jsonl - config_name: krc_Cyrl data_files: - split: train path: krc_Cyrl/*_keep.jsonl - config_name: lfn_Latn data_files: - split: train path: lfn_Latn/*_keep.jsonl - config_name: tls_Latn data_files: - split: train path: tls_Latn/*_keep.jsonl - config_name: fur_Latn data_files: - split: train path: fur_Latn/*_keep.jsonl - config_name: und_Tavt data_files: - split: train path: und_Tavt/*_keep.jsonl - config_name: gag_Latn data_files: - split: train path: gag_Latn/*_keep.jsonl - config_name: chk_Latn data_files: - split: train path: chk_Latn/*_keep.jsonl - config_name: hmo_Latn data_files: - split: train path: hmo_Latn/*_keep.jsonl - config_name: kek_Latn data_files: - split: train path: kek_Latn/*_keep.jsonl - config_name: fon_Latn data_files: - split: train path: fon_Latn/*_keep.jsonl - config_name: ssw_Latn data_files: - split: train path: ssw_Latn/*_keep.jsonl - config_name: quh_Latn data_files: - split: train path: quh_Latn/*_keep.jsonl - config_name: meu_Latn data_files: - split: train path: meu_Latn/*_keep.jsonl - config_name: csb_Latn data_files: - split: train path: csb_Latn/*_keep.jsonl - config_name: ksd_Latn data_files: - split: train path: ksd_Latn/*_keep.jsonl - config_name: ext_Latn data_files: - split: train path: ext_Latn/*_keep.jsonl - config_name: rop_Latn data_files: - split: train path: rop_Latn/*_keep.jsonl - config_name: und_Tfng data_files: - split: train path: und_Tfng/*_keep.jsonl - config_name: aln_Latn data_files: - split: train path: aln_Latn/*_keep.jsonl - config_name: kab_Latn data_files: - split: train path: kab_Latn/*_keep.jsonl - config_name: shn_Mymr data_files: - split: train path: shn_Mymr/*_keep.jsonl - config_name: tbz_Latn data_files: - split: train path: tbz_Latn/*_keep.jsonl - config_name: knv_Latn data_files: - split: train path: knv_Latn/*_keep.jsonl - config_name: bqc_Latn data_files: - split: train path: bqc_Latn/*_keep.jsonl - config_name: und_Elba data_files: - split: train path: und_Elba/*_keep.jsonl - config_name: und_Zanb data_files: - split: train path: und_Zanb/*_keep.jsonl - config_name: npi_Deva data_files: - split: train path: npi_Deva/*_keep.jsonl - config_name: rug_Latn data_files: - split: train path: rug_Latn/*_keep.jsonl - config_name: und_Tale data_files: - split: train path: und_Tale/*_keep.jsonl - config_name: kom_Cyrl data_files: - split: train path: kom_Cyrl/*_keep.jsonl - config_name: kpg_Latn data_files: - split: train path: kpg_Latn/*_keep.jsonl - config_name: kbd_Cyrl data_files: - split: train path: kbd_Cyrl/*_keep.jsonl - config_name: dzo_Tibt data_files: - split: train path: dzo_Tibt/*_keep.jsonl - config_name: wal_Latn data_files: - split: train path: wal_Latn/*_keep.jsonl - config_name: zlm_Latn data_files: - split: train path: zlm_Latn/*_keep.jsonl - config_name: und_Soyo data_files: - split: train path: und_Soyo/*_keep.jsonl - config_name: mzh_Latn data_files: - split: train path: mzh_Latn/*_keep.jsonl - config_name: pan_Latn data_files: - split: train path: pan_Latn/*_keep.jsonl - config_name: hac_Arab data_files: - split: train path: hac_Arab/*_keep.jsonl - config_name: und_Sogo data_files: - split: train path: und_Sogo/*_keep.jsonl - config_name: tum_Latn data_files: - split: train path: tum_Latn/*_keep.jsonl - config_name: sgs_Latn data_files: - split: train path: sgs_Latn/*_keep.jsonl - config_name: fij_Latn data_files: - split: train path: fij_Latn/*_keep.jsonl - config_name: tuc_Latn data_files: - split: train path: tuc_Latn/*_keep.jsonl - config_name: und_Dogr data_files: - split: train path: und_Dogr/*_keep.jsonl - config_name: und_Kawi data_files: - split: train path: und_Kawi/*_keep.jsonl - config_name: jbo_Latn data_files: - split: train path: jbo_Latn/*_keep.jsonl - config_name: iba_Latn data_files: - split: train path: iba_Latn/*_keep.jsonl - config_name: pcd_Latn data_files: - split: train path: pcd_Latn/*_keep.jsonl - config_name: kjh_Cyrl data_files: - split: train path: kjh_Cyrl/*_keep.jsonl - config_name: ixl_Latn data_files: - split: train path: ixl_Latn/*_keep.jsonl - config_name: und_Phli data_files: - split: train path: und_Phli/*_keep.jsonl - config_name: und_Cham data_files: - split: train path: und_Cham/*_keep.jsonl - config_name: und_Guru data_files: - split: train path: und_Guru/*_keep.jsonl - config_name: ndo_Latn data_files: - split: train path: ndo_Latn/*_keep.jsonl - config_name: suz_Deva data_files: - split: train path: suz_Deva/*_keep.jsonl - config_name: und_Nbat data_files: - split: train path: und_Nbat/*_keep.jsonl - config_name: und_Nand data_files: - split: train path: und_Nand/*_keep.jsonl - config_name: mfe_Latn data_files: - split: train path: mfe_Latn/*_keep.jsonl - config_name: bxr_Cyrl data_files: - split: train path: bxr_Cyrl/*_keep.jsonl - config_name: ava_Cyrl data_files: - split: train path: ava_Cyrl/*_keep.jsonl - config_name: und_Osma data_files: - split: train path: und_Osma/*_keep.jsonl - config_name: bho_Deva data_files: - split: train path: bho_Deva/*_keep.jsonl - config_name: brh_Arab data_files: - split: train path: brh_Arab/*_keep.jsonl - config_name: und_Sind data_files: - split: train path: und_Sind/*_keep.jsonl - config_name: ctd_Latn data_files: - split: train path: ctd_Latn/*_keep.jsonl - config_name: aym_Latn data_files: - split: train path: aym_Latn/*_keep.jsonl - config_name: ady_Cyrl data_files: - split: train path: ady_Cyrl/*_keep.jsonl - config_name: nbl_Latn data_files: - split: train path: nbl_Latn/*_keep.jsonl - config_name: gom_Deva data_files: - split: train path: gom_Deva/*_keep.jsonl - config_name: hus_Latn data_files: - split: train path: hus_Latn/*_keep.jsonl - config_name: gug_Latn data_files: - split: train path: gug_Latn/*_keep.jsonl - config_name: und_Pauc data_files: - split: train path: und_Pauc/*_keep.jsonl - config_name: und_Sogd data_files: - split: train path: und_Sogd/*_keep.jsonl - config_name: bat_Latn data_files: - split: train path: bat_Latn/*_keep.jsonl - config_name: guj_Deva data_files: - split: train path: guj_Deva/*_keep.jsonl - config_name: hne_Deva data_files: - split: train path: hne_Deva/*_keep.jsonl - config_name: srn_Latn data_files: - split: train path: srn_Latn/*_keep.jsonl - config_name: ctu_Latn data_files: - split: train path: ctu_Latn/*_keep.jsonl - config_name: san_Latn data_files: - split: train path: san_Latn/*_keep.jsonl - config_name: ton_Latn data_files: - split: train path: ton_Latn/*_keep.jsonl - config_name: bug_Latn data_files: - split: train path: bug_Latn/*_keep.jsonl - config_name: aoj_Latn data_files: - split: train path: aoj_Latn/*_keep.jsonl - config_name: jam_Latn data_files: - split: train path: jam_Latn/*_keep.jsonl - config_name: und_Sylo data_files: - split: train path: und_Sylo/*_keep.jsonl - config_name: lez_Cyrl data_files: - split: train path: lez_Cyrl/*_keep.jsonl - config_name: xav_Latn data_files: - split: train path: xav_Latn/*_keep.jsonl - config_name: luo_Latn data_files: - split: train path: luo_Latn/*_keep.jsonl - config_name: ltg_Latn data_files: - split: train path: ltg_Latn/*_keep.jsonl - config_name: bih_Deva data_files: - split: train path: bih_Deva/*_keep.jsonl - config_name: uig_Cyrl data_files: - split: train path: uig_Cyrl/*_keep.jsonl - config_name: mrj_Cyrl data_files: - split: train path: mrj_Cyrl/*_keep.jsonl - config_name: avk_Latn data_files: - split: train path: avk_Latn/*_keep.jsonl - config_name: kos_Latn data_files: - split: train path: kos_Latn/*_keep.jsonl - config_name: kbp_Latn data_files: - split: train path: kbp_Latn/*_keep.jsonl - config_name: alt_Cyrl data_files: - split: train path: alt_Cyrl/*_keep.jsonl - config_name: cfm_Latn data_files: - split: train path: cfm_Latn/*_keep.jsonl - config_name: goh_Latn data_files: - split: train path: goh_Latn/*_keep.jsonl - config_name: lad_Latn data_files: - split: train path: lad_Latn/*_keep.jsonl - config_name: und_Rjng data_files: - split: train path: und_Rjng/*_keep.jsonl - config_name: cor_Latn data_files: - split: train path: cor_Latn/*_keep.jsonl - config_name: quc_Latn data_files: - split: train path: quc_Latn/*_keep.jsonl - config_name: und_Goth data_files: - split: train path: und_Goth/*_keep.jsonl - config_name: kpv_Cyrl data_files: - split: train path: kpv_Cyrl/*_keep.jsonl - config_name: und_Chrs data_files: - split: train path: und_Chrs/*_keep.jsonl - config_name: aka_Latn data_files: - split: train path: aka_Latn/*_keep.jsonl - config_name: und_Phlp data_files: - split: train path: und_Phlp/*_keep.jsonl - config_name: hak_Latn data_files: - split: train path: hak_Latn/*_keep.jsonl - config_name: rmn_Latn data_files: - split: train path: rmn_Latn/*_keep.jsonl - config_name: pls_Latn data_files: - split: train path: pls_Latn/*_keep.jsonl - config_name: sag_Latn data_files: - split: train path: sag_Latn/*_keep.jsonl - config_name: min_Arab data_files: - split: train path: min_Arab/*_keep.jsonl - config_name: und_Tglg data_files: - split: train path: und_Tglg/*_keep.jsonl - config_name: rcf_Latn data_files: - split: train path: rcf_Latn/*_keep.jsonl - config_name: som_Arab data_files: - split: train path: som_Arab/*_keep.jsonl - config_name: gym_Latn data_files: - split: train path: gym_Latn/*_keep.jsonl - config_name: und_Mand data_files: - split: train path: und_Mand/*_keep.jsonl - config_name: kik_Latn data_files: - split: train path: kik_Latn/*_keep.jsonl - config_name: pag_Latn data_files: - split: train path: pag_Latn/*_keep.jsonl - config_name: arn_Latn data_files: - split: train path: arn_Latn/*_keep.jsonl - config_name: und_Shaw data_files: - split: train path: und_Shaw/*_keep.jsonl - config_name: ngu_Latn data_files: - split: train path: ngu_Latn/*_keep.jsonl - config_name: dsb_Latn data_files: - split: train path: dsb_Latn/*_keep.jsonl - config_name: tca_Latn data_files: - split: train path: tca_Latn/*_keep.jsonl - config_name: mni_Mtei data_files: - split: train path: mni_Mtei/*_keep.jsonl - config_name: naq_Latn data_files: - split: train path: naq_Latn/*_keep.jsonl - config_name: pdc_Latn data_files: - split: train path: pdc_Latn/*_keep.jsonl - config_name: mps_Latn data_files: - split: train path: mps_Latn/*_keep.jsonl - config_name: und_Hatr data_files: - split: train path: und_Hatr/*_keep.jsonl - config_name: olo_Latn data_files: - split: train path: olo_Latn/*_keep.jsonl - config_name: fro_Latn data_files: - split: train path: fro_Latn/*_keep.jsonl - config_name: fit_Latn data_files: - split: train path: fit_Latn/*_keep.jsonl - config_name: crh_Cyrl data_files: - split: train path: crh_Cyrl/*_keep.jsonl - config_name: cdo_Latn data_files: - split: train path: cdo_Latn/*_keep.jsonl - config_name: kac_Latn data_files: - split: train path: kac_Latn/*_keep.jsonl - config_name: mdf_Cyrl data_files: - split: train path: mdf_Cyrl/*_keep.jsonl - config_name: mad_Latn data_files: - split: train path: mad_Latn/*_keep.jsonl - config_name: cab_Latn data_files: - split: train path: cab_Latn/*_keep.jsonl - config_name: srm_Latn data_files: - split: train path: srm_Latn/*_keep.jsonl - config_name: toj_Latn data_files: - split: train path: toj_Latn/*_keep.jsonl - config_name: zai_Latn data_files: - split: train path: zai_Latn/*_keep.jsonl - config_name: guc_Latn data_files: - split: train path: guc_Latn/*_keep.jsonl - config_name: smn_Latn data_files: - split: train path: smn_Latn/*_keep.jsonl - config_name: ote_Latn data_files: - split: train path: ote_Latn/*_keep.jsonl - config_name: nch_Latn data_files: - split: train path: nch_Latn/*_keep.jsonl - config_name: und_Bugi data_files: - split: train path: und_Bugi/*_keep.jsonl - config_name: prs_Arab data_files: - split: train path: prs_Arab/*_keep.jsonl - config_name: tuk_Arab data_files: - split: train path: tuk_Arab/*_keep.jsonl - config_name: bis_Latn data_files: - split: train path: bis_Latn/*_keep.jsonl - config_name: cuk_Latn data_files: - split: train path: cuk_Latn/*_keep.jsonl - config_name: bzj_Latn data_files: - split: train path: bzj_Latn/*_keep.jsonl - config_name: map_Latn data_files: - split: train path: map_Latn/*_keep.jsonl - config_name: djk_Latn data_files: - split: train path: djk_Latn/*_keep.jsonl - config_name: ncj_Latn data_files: - split: train path: ncj_Latn/*_keep.jsonl - config_name: und_Olck data_files: - split: train path: und_Olck/*_keep.jsonl - config_name: mco_Latn data_files: - split: train path: mco_Latn/*_keep.jsonl - config_name: shp_Latn data_files: - split: train path: shp_Latn/*_keep.jsonl - config_name: und_Tagb data_files: - split: train path: und_Tagb/*_keep.jsonl - config_name: ang_Latn data_files: - split: train path: ang_Latn/*_keep.jsonl - config_name: seh_Latn data_files: - split: train path: seh_Latn/*_keep.jsonl - config_name: ikk_Latn data_files: - split: train path: ikk_Latn/*_keep.jsonl - config_name: qvi_Latn data_files: - split: train path: qvi_Latn/*_keep.jsonl - config_name: yom_Latn data_files: - split: train path: yom_Latn/*_keep.jsonl - config_name: top_Latn data_files: - split: train path: top_Latn/*_keep.jsonl - config_name: awa_Deva data_files: - split: train path: awa_Deva/*_keep.jsonl - config_name: ike_Cans data_files: - split: train path: ike_Cans/*_keep.jsonl - config_name: mau_Latn data_files: - split: train path: mau_Latn/*_keep.jsonl - config_name: kum_Cyrl data_files: - split: train path: kum_Cyrl/*_keep.jsonl - config_name: mos_Latn data_files: - split: train path: mos_Latn/*_keep.jsonl - config_name: tly_Latn data_files: - split: train path: tly_Latn/*_keep.jsonl - config_name: und_Prti data_files: - split: train path: und_Prti/*_keep.jsonl - config_name: ayr_Latn data_files: - split: train path: ayr_Latn/*_keep.jsonl - config_name: tdt_Latn data_files: - split: train path: tdt_Latn/*_keep.jsonl - config_name: yap_Latn data_files: - split: train path: yap_Latn/*_keep.jsonl - config_name: evn_Cyrl data_files: - split: train path: evn_Cyrl/*_keep.jsonl - config_name: tah_Latn data_files: - split: train path: tah_Latn/*_keep.jsonl - config_name: xal_Cyrl data_files: - split: train path: xal_Cyrl/*_keep.jsonl - config_name: szy_Latn data_files: - split: train path: szy_Latn/*_keep.jsonl - config_name: quw_Latn data_files: - split: train path: quw_Latn/*_keep.jsonl - config_name: stq_Latn data_files: - split: train path: stq_Latn/*_keep.jsonl - config_name: tlh_Latn data_files: - split: train path: tlh_Latn/*_keep.jsonl - config_name: otq_Latn data_files: - split: train path: otq_Latn/*_keep.jsonl - config_name: und_Narb data_files: - split: train path: und_Narb/*_keep.jsonl - config_name: nov_Latn data_files: - split: train path: nov_Latn/*_keep.jsonl - config_name: skr_Arab data_files: - split: train path: skr_Arab/*_keep.jsonl - config_name: yua_Latn data_files: - split: train path: yua_Latn/*_keep.jsonl - config_name: raw_Latn data_files: - split: train path: raw_Latn/*_keep.jsonl - config_name: und_Sarb data_files: - split: train path: und_Sarb/*_keep.jsonl - config_name: fiu_Latn data_files: - split: train path: fiu_Latn/*_keep.jsonl - config_name: enm_Latn data_files: - split: train path: enm_Latn/*_keep.jsonl - config_name: bbc_Latn data_files: - split: train path: bbc_Latn/*_keep.jsonl - config_name: und_Lydi data_files: - split: train path: und_Lydi/*_keep.jsonl - config_name: non_Latn data_files: - split: train path: non_Latn/*_keep.jsonl - config_name: und_Ugar data_files: - split: train path: und_Ugar/*_keep.jsonl - config_name: ber_Latn data_files: - split: train path: ber_Latn/*_keep.jsonl - config_name: kea_Latn data_files: - split: train path: kea_Latn/*_keep.jsonl - config_name: uig_Latn data_files: - split: train path: uig_Latn/*_keep.jsonl - config_name: bua_Cyrl data_files: - split: train path: bua_Cyrl/*_keep.jsonl - config_name: kua_Latn data_files: - split: train path: kua_Latn/*_keep.jsonl - config_name: zza_Latn data_files: - split: train path: zza_Latn/*_keep.jsonl - config_name: csy_Latn data_files: - split: train path: csy_Latn/*_keep.jsonl - config_name: mnw_Mymr data_files: - split: train path: mnw_Mymr/*_keep.jsonl - config_name: dtp_Latn data_files: - split: train path: dtp_Latn/*_keep.jsonl - config_name: nah_Latn data_files: - split: train path: nah_Latn/*_keep.jsonl - config_name: yao_Latn data_files: - split: train path: yao_Latn/*_keep.jsonl - config_name: azj_Cyrl data_files: - split: train path: azj_Cyrl/*_keep.jsonl - config_name: hau_Arab data_files: - split: train path: hau_Arab/*_keep.jsonl - config_name: nog_Cyrl data_files: - split: train path: nog_Cyrl/*_keep.jsonl - config_name: abt_Latn data_files: - split: train path: abt_Latn/*_keep.jsonl - config_name: meo_Latn data_files: - split: train path: meo_Latn/*_keep.jsonl - config_name: ame_Latn data_files: - split: train path: ame_Latn/*_keep.jsonl - config_name: mbf_Latn data_files: - split: train path: mbf_Latn/*_keep.jsonl - config_name: chm_Cyrl data_files: - split: train path: chm_Cyrl/*_keep.jsonl - config_name: arb_Latn data_files: - split: train path: arb_Latn/*_keep.jsonl - config_name: kas_Arab data_files: - split: train path: kas_Arab/*_keep.jsonl - config_name: tam_Latn data_files: - split: train path: tam_Latn/*_keep.jsonl - config_name: crs_Latn data_files: - split: train path: crs_Latn/*_keep.jsonl - config_name: tcy_Knda data_files: - split: train path: tcy_Knda/*_keep.jsonl - config_name: lbe_Cyrl data_files: - split: train path: lbe_Cyrl/*_keep.jsonl - config_name: cni_Latn data_files: - split: train path: cni_Latn/*_keep.jsonl - config_name: ben_Latn data_files: - split: train path: ben_Latn/*_keep.jsonl - config_name: rom_Latn data_files: - split: train path: rom_Latn/*_keep.jsonl - config_name: zom_Latn data_files: - split: train path: zom_Latn/*_keep.jsonl - config_name: blk_Mymr data_files: - split: train path: blk_Mymr/*_keep.jsonl - config_name: efi_Latn data_files: - split: train path: efi_Latn/*_keep.jsonl - config_name: ada_Latn data_files: - split: train path: ada_Latn/*_keep.jsonl - config_name: und_Buhd data_files: - split: train path: und_Buhd/*_keep.jsonl - config_name: koi_Cyrl data_files: - split: train path: koi_Cyrl/*_keep.jsonl - config_name: und_Perm data_files: - split: train path: und_Perm/*_keep.jsonl - config_name: bbj_Latn data_files: - split: train path: bbj_Latn/*_keep.jsonl - config_name: trv_Latn data_files: - split: train path: trv_Latn/*_keep.jsonl - config_name: ach_Latn data_files: - split: train path: ach_Latn/*_keep.jsonl - config_name: mgh_Latn data_files: - split: train path: mgh_Latn/*_keep.jsonl - config_name: kas_Latn data_files: - split: train path: kas_Latn/*_keep.jsonl - config_name: gcr_Latn data_files: - split: train path: gcr_Latn/*_keep.jsonl - config_name: krl_Latn data_files: - split: train path: krl_Latn/*_keep.jsonl - config_name: iso_Latn data_files: - split: train path: iso_Latn/*_keep.jsonl - config_name: lki_Arab data_files: - split: train path: lki_Arab/*_keep.jsonl - config_name: atj_Latn data_files: - split: train path: atj_Latn/*_keep.jsonl - config_name: cmo_Latn data_files: - split: train path: cmo_Latn/*_keep.jsonl - config_name: chu_Cyrl data_files: - split: train path: chu_Cyrl/*_keep.jsonl - config_name: gur_Latn data_files: - split: train path: gur_Latn/*_keep.jsonl - config_name: mni_Beng data_files: - split: train path: mni_Beng/*_keep.jsonl - config_name: iku_Cans data_files: - split: train path: iku_Cans/*_keep.jsonl - config_name: sma_Latn data_files: - split: train path: sma_Latn/*_keep.jsonl - config_name: lrc_Arab data_files: - split: train path: lrc_Arab/*_keep.jsonl - config_name: gos_Latn data_files: - split: train path: gos_Latn/*_keep.jsonl - config_name: dty_Deva data_files: - split: train path: dty_Deva/*_keep.jsonl - config_name: nzi_Latn data_files: - split: train path: nzi_Latn/*_keep.jsonl - config_name: btx_Latn data_files: - split: train path: btx_Latn/*_keep.jsonl - config_name: brx_Deva data_files: - split: train path: brx_Deva/*_keep.jsonl - config_name: bts_Latn data_files: - split: train path: bts_Latn/*_keep.jsonl - config_name: nyu_Latn data_files: - split: train path: nyu_Latn/*_keep.jsonl - config_name: ckt_Latn data_files: - split: train path: ckt_Latn/*_keep.jsonl - config_name: guw_Latn data_files: - split: train path: guw_Latn/*_keep.jsonl - config_name: pck_Latn data_files: - split: train path: pck_Latn/*_keep.jsonl - config_name: quy_Latn data_files: - split: train path: quy_Latn/*_keep.jsonl - config_name: gcf_Latn data_files: - split: train path: gcf_Latn/*_keep.jsonl - config_name: ksw_Mymr data_files: - split: train path: ksw_Mymr/*_keep.jsonl - config_name: nia_Latn data_files: - split: train path: nia_Latn/*_keep.jsonl - config_name: bal_Arab data_files: - split: train path: bal_Arab/*_keep.jsonl - config_name: nhe_Latn data_files: - split: train path: nhe_Latn/*_keep.jsonl - config_name: hwc_Latn data_files: - split: train path: hwc_Latn/*_keep.jsonl - config_name: mup_Deva data_files: - split: train path: mup_Deva/*_keep.jsonl - config_name: dyu_Latn data_files: - split: train path: dyu_Latn/*_keep.jsonl - config_name: ami_Latn data_files: - split: train path: ami_Latn/*_keep.jsonl - config_name: tay_Latn data_files: - split: train path: tay_Latn/*_keep.jsonl - config_name: gaa_Latn data_files: - split: train path: gaa_Latn/*_keep.jsonl - config_name: loz_Latn data_files: - split: train path: loz_Latn/*_keep.jsonl - config_name: abq_Cyrl data_files: - split: train path: abq_Cyrl/*_keep.jsonl - config_name: pnt_Grek data_files: - split: train path: pnt_Grek/*_keep.jsonl - config_name: snd_Latn data_files: - split: train path: snd_Latn/*_keep.jsonl - config_name: ium_Latn data_files: - split: train path: ium_Latn/*_keep.jsonl - config_name: abs_Latn data_files: - split: train path: abs_Latn/*_keep.jsonl - config_name: cmr_Latn data_files: - split: train path: cmr_Latn/*_keep.jsonl - config_name: shi_Latn data_files: - split: train path: shi_Latn/*_keep.jsonl - config_name: tzm_Tfng data_files: - split: train path: tzm_Tfng/*_keep.jsonl - config_name: gil_Latn data_files: - split: train path: gil_Latn/*_keep.jsonl - config_name: ewo_Latn data_files: - split: train path: ewo_Latn/*_keep.jsonl - config_name: kon_Latn data_files: - split: train path: kon_Latn/*_keep.jsonl - config_name: sms_Latn data_files: - split: train path: sms_Latn/*_keep.jsonl - config_name: ape_Latn data_files: - split: train path: ape_Latn/*_keep.jsonl - config_name: tzh_Latn data_files: - split: train path: tzh_Latn/*_keep.jsonl - config_name: bci_Latn data_files: - split: train path: bci_Latn/*_keep.jsonl - config_name: acd_Latn data_files: - split: train path: acd_Latn/*_keep.jsonl - config_name: tab_Cyrl data_files: - split: train path: tab_Cyrl/*_keep.jsonl - config_name: fuv_Latn data_files: - split: train path: fuv_Latn/*_keep.jsonl - config_name: swc_Latn data_files: - split: train path: swc_Latn/*_keep.jsonl - config_name: tcz_Latn data_files: - split: train path: tcz_Latn/*_keep.jsonl - config_name: und_Elym data_files: - split: train path: und_Elym/*_keep.jsonl - config_name: sgc_Latn data_files: - split: train path: sgc_Latn/*_keep.jsonl - config_name: aaz_Latn data_files: - split: train path: aaz_Latn/*_keep.jsonl - config_name: tly_Arab data_files: - split: train path: tly_Arab/*_keep.jsonl - config_name: tvl_Latn data_files: - split: train path: tvl_Latn/*_keep.jsonl - config_name: zha_Latn data_files: - split: train path: zha_Latn/*_keep.jsonl - config_name: pib_Latn data_files: - split: train path: pib_Latn/*_keep.jsonl - config_name: pli_Deva data_files: - split: train path: pli_Deva/*_keep.jsonl - config_name: bru_Latn data_files: - split: train path: bru_Latn/*_keep.jsonl - config_name: quz_Latn data_files: - split: train path: quz_Latn/*_keep.jsonl - config_name: cha_Latn data_files: - split: train path: cha_Latn/*_keep.jsonl - config_name: aeu_Latn data_files: - split: train path: aeu_Latn/*_keep.jsonl - config_name: bjn_Arab data_files: - split: train path: bjn_Arab/*_keep.jsonl - config_name: nqo_Nkoo data_files: - split: train path: nqo_Nkoo/*_keep.jsonl - config_name: mer_Latn data_files: - split: train path: mer_Latn/*_keep.jsonl - config_name: acf_Latn data_files: - split: train path: acf_Latn/*_keep.jsonl - config_name: rmc_Latn data_files: - split: train path: rmc_Latn/*_keep.jsonl - config_name: msi_Latn data_files: - split: train path: msi_Latn/*_keep.jsonl - config_name: smj_Latn data_files: - split: train path: smj_Latn/*_keep.jsonl - config_name: kng_Latn data_files: - split: train path: kng_Latn/*_keep.jsonl - config_name: dhv_Latn data_files: - split: train path: dhv_Latn/*_keep.jsonl - config_name: doi_Deva data_files: - split: train path: doi_Deva/*_keep.jsonl - config_name: zap_Latn data_files: - split: train path: zap_Latn/*_keep.jsonl - config_name: hla_Latn data_files: - split: train path: hla_Latn/*_keep.jsonl - config_name: cac_Latn data_files: - split: train path: cac_Latn/*_keep.jsonl - config_name: pdt_Latn data_files: - split: train path: pdt_Latn/*_keep.jsonl - config_name: mbt_Latn data_files: - split: train path: mbt_Latn/*_keep.jsonl - config_name: taq_Latn data_files: - split: train path: taq_Latn/*_keep.jsonl - config_name: bgp_Latn data_files: - split: train path: bgp_Latn/*_keep.jsonl - config_name: tyz_Latn data_files: - split: train path: tyz_Latn/*_keep.jsonl - config_name: aau_Latn data_files: - split: train path: aau_Latn/*_keep.jsonl - config_name: mah_Latn data_files: - split: train path: mah_Latn/*_keep.jsonl - config_name: nak_Latn data_files: - split: train path: nak_Latn/*_keep.jsonl - config_name: kmb_Latn data_files: - split: train path: kmb_Latn/*_keep.jsonl - config_name: mjw_Latn data_files: - split: train path: mjw_Latn/*_keep.jsonl - config_name: fuf_Latn data_files: - split: train path: fuf_Latn/*_keep.jsonl - config_name: bba_Latn data_files: - split: train path: bba_Latn/*_keep.jsonl - config_name: ron_Cyrl data_files: - split: train path: ron_Cyrl/*_keep.jsonl - config_name: amu_Latn data_files: - split: train path: amu_Latn/*_keep.jsonl - config_name: alz_Latn data_files: - split: train path: alz_Latn/*_keep.jsonl - config_name: cop_Copt data_files: - split: train path: cop_Copt/*_keep.jsonl - config_name: rar_Latn data_files: - split: train path: rar_Latn/*_keep.jsonl - config_name: emp_Latn data_files: - split: train path: emp_Latn/*_keep.jsonl - config_name: arc_Syrc data_files: - split: train path: arc_Syrc/*_keep.jsonl - config_name: nyn_Latn data_files: - split: train path: nyn_Latn/*_keep.jsonl - config_name: rut_Cyrl data_files: - split: train path: rut_Cyrl/*_keep.jsonl - config_name: lub_Latn data_files: - split: train path: lub_Latn/*_keep.jsonl - config_name: tzj_Latn data_files: - split: train path: tzj_Latn/*_keep.jsonl - config_name: fat_Latn data_files: - split: train path: fat_Latn/*_keep.jsonl - config_name: lzh_Hani data_files: - split: train path: lzh_Hani/*_keep.jsonl - config_name: aak_Latn data_files: - split: train path: aak_Latn/*_keep.jsonl - config_name: nnb_Latn data_files: - split: train path: nnb_Latn/*_keep.jsonl - config_name: lhu_Latn data_files: - split: train path: lhu_Latn/*_keep.jsonl - config_name: bjv_Latn data_files: - split: train path: bjv_Latn/*_keep.jsonl - config_name: bum_Latn data_files: - split: train path: bum_Latn/*_keep.jsonl - config_name: maq_Latn data_files: - split: train path: maq_Latn/*_keep.jsonl - config_name: roa_Latn data_files: - split: train path: roa_Latn/*_keep.jsonl - config_name: bvr_Latn data_files: - split: train path: bvr_Latn/*_keep.jsonl - config_name: kcg_Latn data_files: - split: train path: kcg_Latn/*_keep.jsonl - config_name: tok_Latn data_files: - split: train path: tok_Latn/*_keep.jsonl - config_name: yrk_Cyrl data_files: - split: train path: yrk_Cyrl/*_keep.jsonl - config_name: khs_Latn data_files: - split: train path: khs_Latn/*_keep.jsonl - config_name: rup_Latn data_files: - split: train path: rup_Latn/*_keep.jsonl - config_name: tuk_Cyrl data_files: - split: train path: tuk_Cyrl/*_keep.jsonl - config_name: qub_Latn data_files: - split: train path: qub_Latn/*_keep.jsonl - config_name: jac_Latn data_files: - split: train path: jac_Latn/*_keep.jsonl - config_name: hmr_Latn data_files: - split: train path: hmr_Latn/*_keep.jsonl - config_name: sid_Latn data_files: - split: train path: sid_Latn/*_keep.jsonl - config_name: chr_Cher data_files: - split: train path: chr_Cher/*_keep.jsonl - config_name: amp_Latn data_files: - split: train path: amp_Latn/*_keep.jsonl - config_name: qug_Latn data_files: - split: train path: qug_Latn/*_keep.jsonl - config_name: mas_Latn data_files: - split: train path: mas_Latn/*_keep.jsonl - config_name: fkv_Latn data_files: - split: train path: fkv_Latn/*_keep.jsonl - config_name: mal_Latn data_files: - split: train path: mal_Latn/*_keep.jsonl - config_name: bem_Latn data_files: - split: train path: bem_Latn/*_keep.jsonl - config_name: cav_Latn data_files: - split: train path: cav_Latn/*_keep.jsonl - config_name: her_Latn data_files: - split: train path: her_Latn/*_keep.jsonl - config_name: kwn_Latn data_files: - split: train path: kwn_Latn/*_keep.jsonl - config_name: ify_Latn data_files: - split: train path: ify_Latn/*_keep.jsonl - config_name: asm_Latn data_files: - split: train path: asm_Latn/*_keep.jsonl - config_name: gpe_Latn data_files: - split: train path: gpe_Latn/*_keep.jsonl - config_name: zpa_Latn data_files: - split: train path: zpa_Latn/*_keep.jsonl - config_name: agu_Latn data_files: - split: train path: agu_Latn/*_keep.jsonl - config_name: ppk_Latn data_files: - split: train path: ppk_Latn/*_keep.jsonl - config_name: tiv_Latn data_files: - split: train path: tiv_Latn/*_keep.jsonl - config_name: npi_Latn data_files: - split: train path: npi_Latn/*_keep.jsonl - config_name: ahk_Latn data_files: - split: train path: ahk_Latn/*_keep.jsonl - config_name: pis_Latn data_files: - split: train path: pis_Latn/*_keep.jsonl - config_name: hns_Latn data_files: - split: train path: hns_Latn/*_keep.jsonl - config_name: sus_Latn data_files: - split: train path: sus_Latn/*_keep.jsonl - config_name: hak_Hani data_files: - split: train path: hak_Hani/*_keep.jsonl - config_name: bon_Latn data_files: - split: train path: bon_Latn/*_keep.jsonl - config_name: sja_Latn data_files: - split: train path: sja_Latn/*_keep.jsonl - config_name: kas_Deva data_files: - split: train path: kas_Deva/*_keep.jsonl - config_name: mar_Latn data_files: - split: train path: mar_Latn/*_keep.jsonl - config_name: kri_Latn data_files: - split: train path: kri_Latn/*_keep.jsonl - config_name: dik_Latn data_files: - split: train path: dik_Latn/*_keep.jsonl - config_name: wrk_Latn data_files: - split: train path: wrk_Latn/*_keep.jsonl - config_name: got_Goth data_files: - split: train path: got_Goth/*_keep.jsonl - config_name: dar_Cyrl data_files: - split: train path: dar_Cyrl/*_keep.jsonl - config_name: gui_Latn data_files: - split: train path: gui_Latn/*_keep.jsonl - config_name: mak_Latn data_files: - split: train path: mak_Latn/*_keep.jsonl - config_name: jiv_Latn data_files: - split: train path: jiv_Latn/*_keep.jsonl - config_name: kjb_Latn data_files: - split: train path: kjb_Latn/*_keep.jsonl - config_name: maz_Latn data_files: - split: train path: maz_Latn/*_keep.jsonl - config_name: bsn_Latn data_files: - split: train path: bsn_Latn/*_keep.jsonl - config_name: dov_Latn data_files: - split: train path: dov_Latn/*_keep.jsonl - config_name: cok_Latn data_files: - split: train path: cok_Latn/*_keep.jsonl - config_name: rwo_Latn data_files: - split: train path: rwo_Latn/*_keep.jsonl - config_name: mag_Deva data_files: - split: train path: mag_Deva/*_keep.jsonl - config_name: hnj_Latn data_files: - split: train path: hnj_Latn/*_keep.jsonl - config_name: krj_Latn data_files: - split: train path: krj_Latn/*_keep.jsonl - config_name: teo_Latn data_files: - split: train path: teo_Latn/*_keep.jsonl - config_name: enq_Latn data_files: - split: train path: enq_Latn/*_keep.jsonl - config_name: aom_Latn data_files: - split: train path: aom_Latn/*_keep.jsonl - config_name: ffm_Latn data_files: - split: train path: ffm_Latn/*_keep.jsonl - config_name: din_Latn data_files: - split: train path: din_Latn/*_keep.jsonl - config_name: ubu_Latn data_files: - split: train path: ubu_Latn/*_keep.jsonl - config_name: pwn_Latn data_files: - split: train path: pwn_Latn/*_keep.jsonl - config_name: gum_Latn data_files: - split: train path: gum_Latn/*_keep.jsonl - config_name: sny_Latn data_files: - split: train path: sny_Latn/*_keep.jsonl - config_name: twu_Latn data_files: - split: train path: twu_Latn/*_keep.jsonl - config_name: med_Latn data_files: - split: train path: med_Latn/*_keep.jsonl - config_name: acn_Latn data_files: - split: train path: acn_Latn/*_keep.jsonl - config_name: rme_Latn data_files: - split: train path: rme_Latn/*_keep.jsonl - config_name: cbs_Latn data_files: - split: train path: cbs_Latn/*_keep.jsonl - config_name: abx_Latn data_files: - split: train path: abx_Latn/*_keep.jsonl - config_name: ndc_Latn data_files: - split: train path: ndc_Latn/*_keep.jsonl - config_name: ibb_Latn data_files: - split: train path: ibb_Latn/*_keep.jsonl - config_name: pih_Latn data_files: - split: train path: pih_Latn/*_keep.jsonl - config_name: zty_Latn data_files: - split: train path: zty_Latn/*_keep.jsonl - config_name: rad_Latn data_files: - split: train path: rad_Latn/*_keep.jsonl - config_name: aby_Latn data_files: - split: train path: aby_Latn/*_keep.jsonl - config_name: zne_Latn data_files: - split: train path: zne_Latn/*_keep.jsonl - config_name: cui_Latn data_files: - split: train path: cui_Latn/*_keep.jsonl - config_name: lua_Latn data_files: - split: train path: lua_Latn/*_keep.jsonl - config_name: yby_Latn data_files: - split: train path: yby_Latn/*_keep.jsonl - config_name: bno_Latn data_files: - split: train path: bno_Latn/*_keep.jsonl - config_name: niu_Latn data_files: - split: train path: niu_Latn/*_keep.jsonl - config_name: agd_Latn data_files: - split: train path: agd_Latn/*_keep.jsonl - config_name: nij_Latn data_files: - split: train path: nij_Latn/*_keep.jsonl - config_name: agg_Latn data_files: - split: train path: agg_Latn/*_keep.jsonl - config_name: yre_Latn data_files: - split: train path: yre_Latn/*_keep.jsonl - config_name: fip_Latn data_files: - split: train path: fip_Latn/*_keep.jsonl - config_name: qve_Latn data_files: - split: train path: qve_Latn/*_keep.jsonl - config_name: tsg_Latn data_files: - split: train path: tsg_Latn/*_keep.jsonl - config_name: skg_Latn data_files: - split: train path: skg_Latn/*_keep.jsonl - config_name: adj_Latn data_files: - split: train path: adj_Latn/*_keep.jsonl - config_name: mfq_Latn data_files: - split: train path: mfq_Latn/*_keep.jsonl - config_name: agx_Cyrl data_files: - split: train path: agx_Cyrl/*_keep.jsonl - config_name: kmg_Latn data_files: - split: train path: kmg_Latn/*_keep.jsonl - config_name: umb_Latn data_files: - split: train path: umb_Latn/*_keep.jsonl - config_name: cnk_Latn data_files: - split: train path: cnk_Latn/*_keep.jsonl - config_name: nhw_Latn data_files: - split: train path: nhw_Latn/*_keep.jsonl - config_name: maa_Latn data_files: - split: train path: maa_Latn/*_keep.jsonl - config_name: kwy_Latn data_files: - split: train path: kwy_Latn/*_keep.jsonl - config_name: jra_Latn data_files: - split: train path: jra_Latn/*_keep.jsonl - config_name: syr_Syrc data_files: - split: train path: syr_Syrc/*_keep.jsonl - config_name: inb_Latn data_files: - split: train path: inb_Latn/*_keep.jsonl - config_name: pau_Latn data_files: - split: train path: pau_Latn/*_keep.jsonl - config_name: xsm_Latn data_files: - split: train path: xsm_Latn/*_keep.jsonl - config_name: vap_Latn data_files: - split: train path: vap_Latn/*_keep.jsonl - config_name: alp_Latn data_files: - split: train path: alp_Latn/*_keep.jsonl - config_name: des_Latn data_files: - split: train path: des_Latn/*_keep.jsonl - config_name: syl_Latn data_files: - split: train path: syl_Latn/*_keep.jsonl - config_name: kmr_Cyrl data_files: - split: train path: kmr_Cyrl/*_keep.jsonl - config_name: mkn_Latn data_files: - split: train path: mkn_Latn/*_keep.jsonl - config_name: gan_Hani data_files: - split: train path: gan_Hani/*_keep.jsonl - config_name: qwh_Latn data_files: - split: train path: qwh_Latn/*_keep.jsonl - config_name: jvn_Latn data_files: - split: train path: jvn_Latn/*_keep.jsonl - config_name: bib_Latn data_files: - split: train path: bib_Latn/*_keep.jsonl - config_name: ckt_Cyrl data_files: - split: train path: ckt_Cyrl/*_keep.jsonl - config_name: mpx_Latn data_files: - split: train path: mpx_Latn/*_keep.jsonl - config_name: bin_Latn data_files: - split: train path: bin_Latn/*_keep.jsonl - config_name: klv_Latn data_files: - split: train path: klv_Latn/*_keep.jsonl - config_name: xon_Latn data_files: - split: train path: xon_Latn/*_keep.jsonl - config_name: jbu_Latn data_files: - split: train path: jbu_Latn/*_keep.jsonl - config_name: bdh_Latn data_files: - split: train path: bdh_Latn/*_keep.jsonl - config_name: knj_Latn data_files: - split: train path: knj_Latn/*_keep.jsonl - config_name: ptu_Latn data_files: - split: train path: ptu_Latn/*_keep.jsonl - config_name: mmn_Latn data_files: - split: train path: mmn_Latn/*_keep.jsonl - config_name: mwq_Latn data_files: - split: train path: mwq_Latn/*_keep.jsonl - config_name: sop_Latn data_files: - split: train path: sop_Latn/*_keep.jsonl - config_name: cgc_Latn data_files: - split: train path: cgc_Latn/*_keep.jsonl - config_name: rmy_Cyrl data_files: - split: train path: rmy_Cyrl/*_keep.jsonl - config_name: dop_Latn data_files: - split: train path: dop_Latn/*_keep.jsonl - config_name: mgr_Latn data_files: - split: train path: mgr_Latn/*_keep.jsonl - config_name: rmo_Latn data_files: - split: train path: rmo_Latn/*_keep.jsonl - config_name: tbc_Latn data_files: - split: train path: tbc_Latn/*_keep.jsonl - config_name: kyq_Latn data_files: - split: train path: kyq_Latn/*_keep.jsonl - config_name: zpu_Latn data_files: - split: train path: zpu_Latn/*_keep.jsonl - config_name: kkc_Latn data_files: - split: train path: kkc_Latn/*_keep.jsonl - config_name: acu_Latn data_files: - split: train path: acu_Latn/*_keep.jsonl - config_name: ata_Latn data_files: - split: train path: ata_Latn/*_keep.jsonl - config_name: gag_Cyrl data_files: - split: train path: gag_Cyrl/*_keep.jsonl - config_name: kan_Latn data_files: - split: train path: kan_Latn/*_keep.jsonl - config_name: mif_Latn data_files: - split: train path: mif_Latn/*_keep.jsonl - config_name: sus_Arab data_files: - split: train path: sus_Arab/*_keep.jsonl - config_name: tsz_Latn data_files: - split: train path: tsz_Latn/*_keep.jsonl - config_name: mbi_Latn data_files: - split: train path: mbi_Latn/*_keep.jsonl - config_name: imo_Latn data_files: - split: train path: imo_Latn/*_keep.jsonl - config_name: fuh_Latn data_files: - split: train path: fuh_Latn/*_keep.jsonl - config_name: zyp_Latn data_files: - split: train path: zyp_Latn/*_keep.jsonl - config_name: apr_Latn data_files: - split: train path: apr_Latn/*_keep.jsonl - config_name: att_Latn data_files: - split: train path: att_Latn/*_keep.jsonl - config_name: bus_Latn data_files: - split: train path: bus_Latn/*_keep.jsonl - config_name: kqp_Latn data_files: - split: train path: kqp_Latn/*_keep.jsonl - config_name: koo_Latn data_files: - split: train path: koo_Latn/*_keep.jsonl - config_name: bas_Latn data_files: - split: train path: bas_Latn/*_keep.jsonl - config_name: urh_Latn data_files: - split: train path: urh_Latn/*_keep.jsonl - config_name: bbr_Latn data_files: - split: train path: bbr_Latn/*_keep.jsonl - config_name: msy_Latn data_files: - split: train path: msy_Latn/*_keep.jsonl - config_name: mdy_Ethi data_files: - split: train path: mdy_Ethi/*_keep.jsonl - config_name: nas_Latn data_files: - split: train path: nas_Latn/*_keep.jsonl - config_name: lee_Latn data_files: - split: train path: lee_Latn/*_keep.jsonl - config_name: mcu_Latn data_files: - split: train path: mcu_Latn/*_keep.jsonl - config_name: mrj_Latn data_files: - split: train path: mrj_Latn/*_keep.jsonl - config_name: dig_Latn data_files: - split: train path: dig_Latn/*_keep.jsonl - config_name: izr_Latn data_files: - split: train path: izr_Latn/*_keep.jsonl - config_name: nfr_Latn data_files: - split: train path: nfr_Latn/*_keep.jsonl - config_name: hto_Latn data_files: - split: train path: hto_Latn/*_keep.jsonl - config_name: mzw_Latn data_files: - split: train path: mzw_Latn/*_keep.jsonl - config_name: qxh_Latn data_files: - split: train path: qxh_Latn/*_keep.jsonl - config_name: chz_Latn data_files: - split: train path: chz_Latn/*_keep.jsonl - config_name: guh_Latn data_files: - split: train path: guh_Latn/*_keep.jsonl - config_name: ded_Latn data_files: - split: train path: ded_Latn/*_keep.jsonl - config_name: bvz_Latn data_files: - split: train path: bvz_Latn/*_keep.jsonl - config_name: kkj_Latn data_files: - split: train path: kkj_Latn/*_keep.jsonl - config_name: gux_Latn data_files: - split: train path: gux_Latn/*_keep.jsonl - config_name: ikt_Latn data_files: - split: train path: ikt_Latn/*_keep.jsonl - config_name: bnp_Latn data_files: - split: train path: bnp_Latn/*_keep.jsonl - config_name: bgt_Latn data_files: - split: train path: bgt_Latn/*_keep.jsonl - config_name: quf_Latn data_files: - split: train path: quf_Latn/*_keep.jsonl - config_name: dwr_Latn data_files: - split: train path: dwr_Latn/*_keep.jsonl - config_name: nod_Thai data_files: - split: train path: nod_Thai/*_keep.jsonl - config_name: yal_Latn data_files: - split: train path: yal_Latn/*_keep.jsonl - config_name: bmr_Latn data_files: - split: train path: bmr_Latn/*_keep.jsonl - config_name: ljp_Latn data_files: - split: train path: ljp_Latn/*_keep.jsonl - config_name: xla_Latn data_files: - split: train path: xla_Latn/*_keep.jsonl - config_name: zyb_Latn data_files: - split: train path: zyb_Latn/*_keep.jsonl - config_name: gof_Latn data_files: - split: train path: gof_Latn/*_keep.jsonl - config_name: niv_Cyrl data_files: - split: train path: niv_Cyrl/*_keep.jsonl - config_name: nrf_Latn data_files: - split: train path: nrf_Latn/*_keep.jsonl - config_name: npy_Latn data_files: - split: train path: npy_Latn/*_keep.jsonl - config_name: ura_Latn data_files: - split: train path: ura_Latn/*_keep.jsonl - config_name: dob_Latn data_files: - split: train path: dob_Latn/*_keep.jsonl - config_name: gfk_Latn data_files: - split: train path: gfk_Latn/*_keep.jsonl - config_name: gdr_Latn data_files: - split: train path: gdr_Latn/*_keep.jsonl - config_name: sdc_Latn data_files: - split: train path: sdc_Latn/*_keep.jsonl - config_name: pjt_Latn data_files: - split: train path: pjt_Latn/*_keep.jsonl - config_name: spp_Latn data_files: - split: train path: spp_Latn/*_keep.jsonl - config_name: irk_Latn data_files: - split: train path: irk_Latn/*_keep.jsonl - config_name: cpa_Latn data_files: - split: train path: cpa_Latn/*_keep.jsonl - config_name: knc_Latn data_files: - split: train path: knc_Latn/*_keep.jsonl - config_name: gvl_Latn data_files: - split: train path: gvl_Latn/*_keep.jsonl - config_name: agr_Latn data_files: - split: train path: agr_Latn/*_keep.jsonl - config_name: njo_Latn data_files: - split: train path: njo_Latn/*_keep.jsonl - config_name: xrb_Latn data_files: - split: train path: xrb_Latn/*_keep.jsonl - config_name: kao_Latn data_files: - split: train path: kao_Latn/*_keep.jsonl - config_name: zpz_Latn data_files: - split: train path: zpz_Latn/*_keep.jsonl - config_name: bgr_Latn data_files: - split: train path: bgr_Latn/*_keep.jsonl - config_name: rnd_Latn data_files: - split: train path: rnd_Latn/*_keep.jsonl - config_name: ntu_Latn data_files: - split: train path: ntu_Latn/*_keep.jsonl - config_name: aai_Latn data_files: - split: train path: aai_Latn/*_keep.jsonl - config_name: gai_Latn data_files: - split: train path: gai_Latn/*_keep.jsonl - config_name: dgi_Latn data_files: - split: train path: dgi_Latn/*_keep.jsonl - config_name: kez_Latn data_files: - split: train path: kez_Latn/*_keep.jsonl - config_name: dad_Latn data_files: - split: train path: dad_Latn/*_keep.jsonl - config_name: kmh_Latn data_files: - split: train path: kmh_Latn/*_keep.jsonl - config_name: aii_Syrc data_files: - split: train path: aii_Syrc/*_keep.jsonl - config_name: ifk_Latn data_files: - split: train path: ifk_Latn/*_keep.jsonl - config_name: aji_Latn data_files: - split: train path: aji_Latn/*_keep.jsonl - config_name: noa_Latn data_files: - split: train path: noa_Latn/*_keep.jsonl - config_name: bhl_Latn data_files: - split: train path: bhl_Latn/*_keep.jsonl - config_name: ztq_Latn data_files: - split: train path: ztq_Latn/*_keep.jsonl - config_name: rmn_Cyrl data_files: - split: train path: rmn_Cyrl/*_keep.jsonl - config_name: gde_Latn data_files: - split: train path: gde_Latn/*_keep.jsonl - config_name: dnj_Latn data_files: - split: train path: dnj_Latn/*_keep.jsonl - config_name: men_Latn data_files: - split: train path: men_Latn/*_keep.jsonl - config_name: miq_Latn data_files: - split: train path: miq_Latn/*_keep.jsonl - config_name: mdf_Latn data_files: - split: train path: mdf_Latn/*_keep.jsonl - config_name: isd_Latn data_files: - split: train path: isd_Latn/*_keep.jsonl - config_name: lgg_Latn data_files: - split: train path: lgg_Latn/*_keep.jsonl - config_name: ncx_Latn data_files: - split: train path: ncx_Latn/*_keep.jsonl - config_name: gub_Latn data_files: - split: train path: gub_Latn/*_keep.jsonl - config_name: itv_Latn data_files: - split: train path: itv_Latn/*_keep.jsonl - config_name: tac_Latn data_files: - split: train path: tac_Latn/*_keep.jsonl - config_name: cce_Latn data_files: - split: train path: cce_Latn/*_keep.jsonl - config_name: kwi_Latn data_files: - split: train path: kwi_Latn/*_keep.jsonl - config_name: nhu_Latn data_files: - split: train path: nhu_Latn/*_keep.jsonl - config_name: snd_Deva data_files: - split: train path: snd_Deva/*_keep.jsonl - config_name: sml_Latn data_files: - split: train path: sml_Latn/*_keep.jsonl - config_name: pah_Latn data_files: - split: train path: pah_Latn/*_keep.jsonl - config_name: sas_Latn data_files: - split: train path: sas_Latn/*_keep.jsonl - config_name: cwt_Latn data_files: - split: train path: cwt_Latn/*_keep.jsonl - config_name: spy_Latn data_files: - split: train path: spy_Latn/*_keep.jsonl - config_name: enb_Latn data_files: - split: train path: enb_Latn/*_keep.jsonl - config_name: fai_Latn data_files: - split: train path: fai_Latn/*_keep.jsonl - config_name: cas_Latn data_files: - split: train path: cas_Latn/*_keep.jsonl - config_name: sue_Latn data_files: - split: train path: sue_Latn/*_keep.jsonl - config_name: yss_Latn data_files: - split: train path: yss_Latn/*_keep.jsonl - config_name: ozm_Latn data_files: - split: train path: ozm_Latn/*_keep.jsonl - config_name: bku_Latn data_files: - split: train path: bku_Latn/*_keep.jsonl - config_name: biv_Latn data_files: - split: train path: biv_Latn/*_keep.jsonl - config_name: akp_Latn data_files: - split: train path: akp_Latn/*_keep.jsonl - config_name: dts_Latn data_files: - split: train path: dts_Latn/*_keep.jsonl - config_name: kdr_Latn data_files: - split: train path: kdr_Latn/*_keep.jsonl - config_name: fal_Latn data_files: - split: train path: fal_Latn/*_keep.jsonl - config_name: mmo_Latn data_files: - split: train path: mmo_Latn/*_keep.jsonl - config_name: byr_Latn data_files: - split: train path: byr_Latn/*_keep.jsonl - config_name: cbv_Latn data_files: - split: train path: cbv_Latn/*_keep.jsonl - config_name: cbu_Latn data_files: - split: train path: cbu_Latn/*_keep.jsonl - config_name: eza_Latn data_files: - split: train path: eza_Latn/*_keep.jsonl - config_name: kgp_Latn data_files: - split: train path: kgp_Latn/*_keep.jsonl - config_name: sda_Latn data_files: - split: train path: sda_Latn/*_keep.jsonl - config_name: gbo_Latn data_files: - split: train path: gbo_Latn/*_keep.jsonl - config_name: mfi_Latn data_files: - split: train path: mfi_Latn/*_keep.jsonl - config_name: bdd_Latn data_files: - split: train path: bdd_Latn/*_keep.jsonl - config_name: boj_Latn data_files: - split: train path: boj_Latn/*_keep.jsonl - config_name: blh_Latn data_files: - split: train path: blh_Latn/*_keep.jsonl - config_name: ibg_Latn data_files: - split: train path: ibg_Latn/*_keep.jsonl - config_name: agn_Latn data_files: - split: train path: agn_Latn/*_keep.jsonl - config_name: vmy_Latn data_files: - split: train path: vmy_Latn/*_keep.jsonl - config_name: sua_Latn data_files: - split: train path: sua_Latn/*_keep.jsonl - config_name: gul_Latn data_files: - split: train path: gul_Latn/*_keep.jsonl - config_name: rap_Latn data_files: - split: train path: rap_Latn/*_keep.jsonl - config_name: bmh_Latn data_files: - split: train path: bmh_Latn/*_keep.jsonl - config_name: sxn_Latn data_files: - split: train path: sxn_Latn/*_keep.jsonl - config_name: dyi_Latn data_files: - split: train path: dyi_Latn/*_keep.jsonl - config_name: tdx_Latn data_files: - split: train path: tdx_Latn/*_keep.jsonl - config_name: car_Latn data_files: - split: train path: car_Latn/*_keep.jsonl - config_name: heh_Latn data_files: - split: train path: heh_Latn/*_keep.jsonl - config_name: dgc_Latn data_files: - split: train path: dgc_Latn/*_keep.jsonl - config_name: sgb_Latn data_files: - split: train path: sgb_Latn/*_keep.jsonl - config_name: bpr_Latn data_files: - split: train path: bpr_Latn/*_keep.jsonl - config_name: gbi_Latn data_files: - split: train path: gbi_Latn/*_keep.jsonl - config_name: avu_Latn data_files: - split: train path: avu_Latn/*_keep.jsonl - config_name: udu_Latn data_files: - split: train path: udu_Latn/*_keep.jsonl - config_name: yli_Latn data_files: - split: train path: yli_Latn/*_keep.jsonl - config_name: cjs_Cyrl data_files: - split: train path: cjs_Cyrl/*_keep.jsonl - config_name: gnn_Latn data_files: - split: train path: gnn_Latn/*_keep.jsonl - config_name: nhi_Latn data_files: - split: train path: nhi_Latn/*_keep.jsonl - config_name: urk_Thai data_files: - split: train path: urk_Thai/*_keep.jsonl - config_name: kpy_Cyrl data_files: - split: train path: kpy_Cyrl/*_keep.jsonl - config_name: qup_Latn data_files: - split: train path: qup_Latn/*_keep.jsonl - config_name: qvh_Latn data_files: - split: train path: qvh_Latn/*_keep.jsonl - config_name: pir_Latn data_files: - split: train path: pir_Latn/*_keep.jsonl - config_name: tsc_Latn data_files: - split: train path: tsc_Latn/*_keep.jsonl - config_name: bxr_Latn data_files: - split: train path: bxr_Latn/*_keep.jsonl - config_name: thk_Latn data_files: - split: train path: thk_Latn/*_keep.jsonl - config_name: grt_Beng data_files: - split: train path: grt_Beng/*_keep.jsonl - config_name: kqn_Latn data_files: - split: train path: kqn_Latn/*_keep.jsonl - config_name: hay_Latn data_files: - split: train path: hay_Latn/*_keep.jsonl - config_name: mqb_Latn data_files: - split: train path: mqb_Latn/*_keep.jsonl - config_name: tbg_Latn data_files: - split: train path: tbg_Latn/*_keep.jsonl - config_name: bss_Latn data_files: - split: train path: bss_Latn/*_keep.jsonl - config_name: trn_Latn data_files: - split: train path: trn_Latn/*_keep.jsonl - config_name: jae_Latn data_files: - split: train path: jae_Latn/*_keep.jsonl - config_name: myy_Latn data_files: - split: train path: myy_Latn/*_keep.jsonl - config_name: ipk_Latn data_files: - split: train path: ipk_Latn/*_keep.jsonl - config_name: bao_Latn data_files: - split: train path: bao_Latn/*_keep.jsonl - config_name: qvz_Latn data_files: - split: train path: qvz_Latn/*_keep.jsonl - config_name: dua_Latn data_files: - split: train path: dua_Latn/*_keep.jsonl - config_name: mwv_Latn data_files: - split: train path: mwv_Latn/*_keep.jsonl - config_name: nct_Latn data_files: - split: train path: nct_Latn/*_keep.jsonl - config_name: kyc_Latn data_files: - split: train path: kyc_Latn/*_keep.jsonl - config_name: tuo_Latn data_files: - split: train path: tuo_Latn/*_keep.jsonl - config_name: tee_Latn data_files: - split: train path: tee_Latn/*_keep.jsonl - config_name: mop_Latn data_files: - split: train path: mop_Latn/*_keep.jsonl - config_name: mxt_Latn data_files: - split: train path: mxt_Latn/*_keep.jsonl - config_name: qvm_Latn data_files: - split: train path: qvm_Latn/*_keep.jsonl - config_name: gkn_Latn data_files: - split: train path: gkn_Latn/*_keep.jsonl - config_name: lam_Latn data_files: - split: train path: lam_Latn/*_keep.jsonl - config_name: ntr_Latn data_files: - split: train path: ntr_Latn/*_keep.jsonl - config_name: sbe_Latn data_files: - split: train path: sbe_Latn/*_keep.jsonl - config_name: dyo_Latn data_files: - split: train path: dyo_Latn/*_keep.jsonl - config_name: lex_Latn data_files: - split: train path: lex_Latn/*_keep.jsonl - config_name: smk_Latn data_files: - split: train path: smk_Latn/*_keep.jsonl - config_name: coe_Latn data_files: - split: train path: coe_Latn/*_keep.jsonl - config_name: kpe_Latn data_files: - split: train path: kpe_Latn/*_keep.jsonl - config_name: mni_Latn data_files: - split: train path: mni_Latn/*_keep.jsonl - config_name: xbi_Latn data_files: - split: train path: xbi_Latn/*_keep.jsonl - config_name: atb_Latn data_files: - split: train path: atb_Latn/*_keep.jsonl - config_name: tcf_Latn data_files: - split: train path: tcf_Latn/*_keep.jsonl - config_name: bfo_Latn data_files: - split: train path: bfo_Latn/*_keep.jsonl - config_name: cbc_Latn data_files: - split: train path: cbc_Latn/*_keep.jsonl - config_name: swg_Latn data_files: - split: train path: swg_Latn/*_keep.jsonl - config_name: niv_Latn data_files: - split: train path: niv_Latn/*_keep.jsonl - config_name: cnt_Latn data_files: - split: train path: cnt_Latn/*_keep.jsonl - config_name: izz_Latn data_files: - split: train path: izz_Latn/*_keep.jsonl - config_name: knf_Latn data_files: - split: train path: knf_Latn/*_keep.jsonl - config_name: mpp_Latn data_files: - split: train path: mpp_Latn/*_keep.jsonl - config_name: kki_Latn data_files: - split: train path: kki_Latn/*_keep.jsonl - config_name: nho_Latn data_files: - split: train path: nho_Latn/*_keep.jsonl - config_name: sat_Latn data_files: - split: train path: sat_Latn/*_keep.jsonl - config_name: kbr_Latn data_files: - split: train path: kbr_Latn/*_keep.jsonl - config_name: bmu_Latn data_files: - split: train path: bmu_Latn/*_keep.jsonl - config_name: hig_Latn data_files: - split: train path: hig_Latn/*_keep.jsonl - config_name: qxr_Latn data_files: - split: train path: qxr_Latn/*_keep.jsonl - config_name: orv_Cyrl data_files: - split: train path: orv_Cyrl/*_keep.jsonl - config_name: pma_Latn data_files: - split: train path: pma_Latn/*_keep.jsonl - config_name: zpl_Latn data_files: - split: train path: zpl_Latn/*_keep.jsonl - config_name: kpr_Latn data_files: - split: train path: kpr_Latn/*_keep.jsonl - config_name: sig_Latn data_files: - split: train path: sig_Latn/*_keep.jsonl - config_name: ory_Latn data_files: - split: train path: ory_Latn/*_keep.jsonl - config_name: mxp_Latn data_files: - split: train path: mxp_Latn/*_keep.jsonl - config_name: lef_Latn data_files: - split: train path: lef_Latn/*_keep.jsonl - config_name: mlp_Latn data_files: - split: train path: mlp_Latn/*_keep.jsonl - config_name: soq_Latn data_files: - split: train path: soq_Latn/*_keep.jsonl - config_name: cre_Latn data_files: - split: train path: cre_Latn/*_keep.jsonl - config_name: aey_Latn data_files: - split: train path: aey_Latn/*_keep.jsonl - config_name: qvs_Latn data_files: - split: train path: qvs_Latn/*_keep.jsonl - config_name: sur_Latn data_files: - split: train path: sur_Latn/*_keep.jsonl - config_name: knc_Arab data_files: - split: train path: knc_Arab/*_keep.jsonl - config_name: suk_Latn data_files: - split: train path: suk_Latn/*_keep.jsonl - config_name: cax_Latn data_files: - split: train path: cax_Latn/*_keep.jsonl - config_name: alq_Latn data_files: - split: train path: alq_Latn/*_keep.jsonl - config_name: chw_Latn data_files: - split: train path: chw_Latn/*_keep.jsonl - config_name: dww_Latn data_files: - split: train path: dww_Latn/*_keep.jsonl - config_name: gog_Latn data_files: - split: train path: gog_Latn/*_keep.jsonl - config_name: viv_Latn data_files: - split: train path: viv_Latn/*_keep.jsonl - config_name: qvw_Latn data_files: - split: train path: qvw_Latn/*_keep.jsonl - config_name: ted_Latn data_files: - split: train path: ted_Latn/*_keep.jsonl - config_name: yuj_Latn data_files: - split: train path: yuj_Latn/*_keep.jsonl - config_name: ziw_Latn data_files: - split: train path: ziw_Latn/*_keep.jsonl - config_name: kto_Latn data_files: - split: train path: kto_Latn/*_keep.jsonl - config_name: mcq_Latn data_files: - split: train path: mcq_Latn/*_keep.jsonl - config_name: sim_Latn data_files: - split: train path: sim_Latn/*_keep.jsonl - config_name: koi_Latn data_files: - split: train path: koi_Latn/*_keep.jsonl - config_name: lue_Latn data_files: - split: train path: lue_Latn/*_keep.jsonl - config_name: cya_Latn data_files: - split: train path: cya_Latn/*_keep.jsonl - config_name: knk_Latn data_files: - split: train path: knk_Latn/*_keep.jsonl - config_name: tpt_Latn data_files: - split: train path: tpt_Latn/*_keep.jsonl - config_name: taj_Deva data_files: - split: train path: taj_Deva/*_keep.jsonl - config_name: blz_Latn data_files: - split: train path: blz_Latn/*_keep.jsonl - config_name: mil_Latn data_files: - split: train path: mil_Latn/*_keep.jsonl - config_name: kde_Latn data_files: - split: train path: kde_Latn/*_keep.jsonl - config_name: kpz_Latn data_files: - split: train path: kpz_Latn/*_keep.jsonl - config_name: hag_Latn data_files: - split: train path: hag_Latn/*_keep.jsonl - config_name: vun_Latn data_files: - split: train path: vun_Latn/*_keep.jsonl - config_name: kus_Latn data_files: - split: train path: kus_Latn/*_keep.jsonl - config_name: lid_Latn data_files: - split: train path: lid_Latn/*_keep.jsonl - config_name: tos_Latn data_files: - split: train path: tos_Latn/*_keep.jsonl - config_name: bfd_Latn data_files: - split: train path: bfd_Latn/*_keep.jsonl - config_name: chd_Latn data_files: - split: train path: chd_Latn/*_keep.jsonl - config_name: cle_Latn data_files: - split: train path: cle_Latn/*_keep.jsonl - config_name: yon_Latn data_files: - split: train path: yon_Latn/*_keep.jsonl - config_name: bgs_Latn data_files: - split: train path: bgs_Latn/*_keep.jsonl - config_name: lsm_Latn data_files: - split: train path: lsm_Latn/*_keep.jsonl - config_name: msb_Latn data_files: - split: train path: msb_Latn/*_keep.jsonl - config_name: ota_Arab data_files: - split: train path: ota_Arab/*_keep.jsonl - config_name: guj_Latn data_files: - split: train path: guj_Latn/*_keep.jsonl - config_name: mxq_Latn data_files: - split: train path: mxq_Latn/*_keep.jsonl - config_name: lfn_Cyrl data_files: - split: train path: lfn_Cyrl/*_keep.jsonl - config_name: myw_Latn data_files: - split: train path: myw_Latn/*_keep.jsonl - config_name: aso_Latn data_files: - split: train path: aso_Latn/*_keep.jsonl - config_name: esu_Latn data_files: - split: train path: esu_Latn/*_keep.jsonl - config_name: kdi_Latn data_files: - split: train path: kdi_Latn/*_keep.jsonl - config_name: bhp_Latn data_files: - split: train path: bhp_Latn/*_keep.jsonl - config_name: mfz_Latn data_files: - split: train path: mfz_Latn/*_keep.jsonl - config_name: sgw_Ethi data_files: - split: train path: sgw_Ethi/*_keep.jsonl - config_name: xsi_Latn data_files: - split: train path: xsi_Latn/*_keep.jsonl - config_name: gun_Latn data_files: - split: train path: gun_Latn/*_keep.jsonl - config_name: kxc_Ethi data_files: - split: train path: kxc_Ethi/*_keep.jsonl - config_name: mux_Latn data_files: - split: train path: mux_Latn/*_keep.jsonl - config_name: otw_Latn data_files: - split: train path: otw_Latn/*_keep.jsonl - config_name: cjp_Latn data_files: - split: train path: cjp_Latn/*_keep.jsonl - config_name: nyo_Latn data_files: - split: train path: nyo_Latn/*_keep.jsonl - config_name: zia_Latn data_files: - split: train path: zia_Latn/*_keep.jsonl - config_name: cco_Latn data_files: - split: train path: cco_Latn/*_keep.jsonl - config_name: ojb_Cans data_files: - split: train path: ojb_Cans/*_keep.jsonl - config_name: ktu_Latn data_files: - split: train path: ktu_Latn/*_keep.jsonl - config_name: cpy_Latn data_files: - split: train path: cpy_Latn/*_keep.jsonl - config_name: txu_Latn data_files: - split: train path: txu_Latn/*_keep.jsonl - config_name: bjr_Latn data_files: - split: train path: bjr_Latn/*_keep.jsonl - config_name: bud_Latn data_files: - split: train path: bud_Latn/*_keep.jsonl - config_name: ken_Latn data_files: - split: train path: ken_Latn/*_keep.jsonl - config_name: bzh_Latn data_files: - split: train path: bzh_Latn/*_keep.jsonl - config_name: ipi_Latn data_files: - split: train path: ipi_Latn/*_keep.jsonl - config_name: crx_Latn data_files: - split: train path: crx_Latn/*_keep.jsonl - config_name: cko_Latn data_files: - split: train path: cko_Latn/*_keep.jsonl - config_name: gnd_Latn data_files: - split: train path: gnd_Latn/*_keep.jsonl - config_name: kix_Latn data_files: - split: train path: kix_Latn/*_keep.jsonl - config_name: mmx_Latn data_files: - split: train path: mmx_Latn/*_keep.jsonl - config_name: esk_Latn data_files: - split: train path: esk_Latn/*_keep.jsonl - config_name: kud_Latn data_files: - split: train path: kud_Latn/*_keep.jsonl - config_name: toh_Latn data_files: - split: train path: toh_Latn/*_keep.jsonl - config_name: byx_Latn data_files: - split: train path: byx_Latn/*_keep.jsonl - config_name: mtp_Latn data_files: - split: train path: mtp_Latn/*_keep.jsonl - config_name: kpv_Latn data_files: - split: train path: kpv_Latn/*_keep.jsonl - config_name: kyz_Latn data_files: - split: train path: kyz_Latn/*_keep.jsonl - config_name: lmp_Latn data_files: - split: train path: lmp_Latn/*_keep.jsonl - config_name: tiy_Latn data_files: - split: train path: tiy_Latn/*_keep.jsonl - config_name: ccp_Latn data_files: - split: train path: ccp_Latn/*_keep.jsonl - config_name: sbl_Latn data_files: - split: train path: sbl_Latn/*_keep.jsonl - config_name: tcs_Latn data_files: - split: train path: tcs_Latn/*_keep.jsonl - config_name: usp_Latn data_files: - split: train path: usp_Latn/*_keep.jsonl - config_name: buk_Latn data_files: - split: train path: buk_Latn/*_keep.jsonl - config_name: hvn_Latn data_files: - split: train path: hvn_Latn/*_keep.jsonl - config_name: amm_Latn data_files: - split: train path: amm_Latn/*_keep.jsonl - config_name: kzj_Latn data_files: - split: train path: kzj_Latn/*_keep.jsonl - config_name: gng_Latn data_files: - split: train path: gng_Latn/*_keep.jsonl - config_name: ter_Latn data_files: - split: train path: ter_Latn/*_keep.jsonl - config_name: kij_Latn data_files: - split: train path: kij_Latn/*_keep.jsonl - config_name: ngl_Latn data_files: - split: train path: ngl_Latn/*_keep.jsonl - config_name: pab_Latn data_files: - split: train path: pab_Latn/*_keep.jsonl - config_name: pad_Latn data_files: - split: train path: pad_Latn/*_keep.jsonl - config_name: kca_Cyrl data_files: - split: train path: kca_Cyrl/*_keep.jsonl - config_name: meq_Latn data_files: - split: train path: meq_Latn/*_keep.jsonl - config_name: kas_Cyrl data_files: - split: train path: kas_Cyrl/*_keep.jsonl - config_name: niq_Latn data_files: - split: train path: niq_Latn/*_keep.jsonl - config_name: dah_Latn data_files: - split: train path: dah_Latn/*_keep.jsonl - config_name: lia_Latn data_files: - split: train path: lia_Latn/*_keep.jsonl - config_name: kss_Latn data_files: - split: train path: kss_Latn/*_keep.jsonl - config_name: bch_Latn data_files: - split: train path: bch_Latn/*_keep.jsonl - config_name: khz_Latn data_files: - split: train path: khz_Latn/*_keep.jsonl - config_name: nop_Latn data_files: - split: train path: nop_Latn/*_keep.jsonl - config_name: yle_Latn data_files: - split: train path: yle_Latn/*_keep.jsonl - config_name: ain_Latn data_files: - split: train path: ain_Latn/*_keep.jsonl - config_name: pem_Latn data_files: - split: train path: pem_Latn/*_keep.jsonl - config_name: yml_Latn data_files: - split: train path: yml_Latn/*_keep.jsonl - config_name: taq_Tfng data_files: - split: train path: taq_Tfng/*_keep.jsonl - config_name: gaw_Latn data_files: - split: train path: gaw_Latn/*_keep.jsonl - config_name: kze_Latn data_files: - split: train path: kze_Latn/*_keep.jsonl - config_name: tby_Latn data_files: - split: train path: tby_Latn/*_keep.jsonl - config_name: cso_Latn data_files: - split: train path: cso_Latn/*_keep.jsonl - config_name: ifb_Latn data_files: - split: train path: ifb_Latn/*_keep.jsonl - config_name: rki_Mymr data_files: - split: train path: rki_Mymr/*_keep.jsonl - config_name: aly_Latn data_files: - split: train path: aly_Latn/*_keep.jsonl - config_name: kmu_Latn data_files: - split: train path: kmu_Latn/*_keep.jsonl - config_name: mhl_Latn data_files: - split: train path: mhl_Latn/*_keep.jsonl - config_name: kmo_Latn data_files: - split: train path: kmo_Latn/*_keep.jsonl - config_name: mva_Latn data_files: - split: train path: mva_Latn/*_keep.jsonl - config_name: opm_Latn data_files: - split: train path: opm_Latn/*_keep.jsonl - config_name: rtm_Latn data_files: - split: train path: rtm_Latn/*_keep.jsonl - config_name: xtn_Latn data_files: - split: train path: xtn_Latn/*_keep.jsonl - config_name: dgr_Latn data_files: - split: train path: dgr_Latn/*_keep.jsonl - config_name: ksr_Latn data_files: - split: train path: ksr_Latn/*_keep.jsonl - config_name: snp_Latn data_files: - split: train path: snp_Latn/*_keep.jsonl - config_name: tvk_Latn data_files: - split: train path: tvk_Latn/*_keep.jsonl - config_name: zpv_Latn data_files: - split: train path: zpv_Latn/*_keep.jsonl - config_name: hrx_Latn data_files: - split: train path: hrx_Latn/*_keep.jsonl - config_name: kwj_Latn data_files: - split: train path: kwj_Latn/*_keep.jsonl - config_name: muh_Latn data_files: - split: train path: muh_Latn/*_keep.jsonl - config_name: kak_Latn data_files: - split: train path: kak_Latn/*_keep.jsonl - config_name: kpw_Latn data_files: - split: train path: kpw_Latn/*_keep.jsonl - config_name: akb_Latn data_files: - split: train path: akb_Latn/*_keep.jsonl - config_name: ybb_Latn data_files: - split: train path: ybb_Latn/*_keep.jsonl - config_name: kjs_Latn data_files: - split: train path: kjs_Latn/*_keep.jsonl - config_name: wuv_Latn data_files: - split: train path: wuv_Latn/*_keep.jsonl - config_name: laj_Latn data_files: - split: train path: laj_Latn/*_keep.jsonl - config_name: ogo_Latn data_files: - split: train path: ogo_Latn/*_keep.jsonl - config_name: nim_Latn data_files: - split: train path: nim_Latn/*_keep.jsonl - config_name: czt_Latn data_files: - split: train path: czt_Latn/*_keep.jsonl - config_name: aoz_Latn data_files: - split: train path: aoz_Latn/*_keep.jsonl - config_name: rro_Latn data_files: - split: train path: rro_Latn/*_keep.jsonl - config_name: yuw_Latn data_files: - split: train path: yuw_Latn/*_keep.jsonl - config_name: heg_Latn data_files: - split: train path: heg_Latn/*_keep.jsonl - config_name: mrw_Latn data_files: - split: train path: mrw_Latn/*_keep.jsonl - config_name: zac_Latn data_files: - split: train path: zac_Latn/*_keep.jsonl - config_name: kpf_Latn data_files: - split: train path: kpf_Latn/*_keep.jsonl - config_name: leu_Latn data_files: - split: train path: leu_Latn/*_keep.jsonl - config_name: nwi_Latn data_files: - split: train path: nwi_Latn/*_keep.jsonl - config_name: awx_Latn data_files: - split: train path: awx_Latn/*_keep.jsonl - config_name: gso_Latn data_files: - split: train path: gso_Latn/*_keep.jsonl - config_name: keo_Latn data_files: - split: train path: keo_Latn/*_keep.jsonl - config_name: tte_Latn data_files: - split: train path: tte_Latn/*_keep.jsonl - config_name: ceg_Latn data_files: - split: train path: ceg_Latn/*_keep.jsonl - config_name: nhy_Latn data_files: - split: train path: nhy_Latn/*_keep.jsonl - config_name: bhw_Latn data_files: - split: train path: bhw_Latn/*_keep.jsonl - config_name: gnw_Latn data_files: - split: train path: gnw_Latn/*_keep.jsonl - config_name: nin_Latn data_files: - split: train path: nin_Latn/*_keep.jsonl - config_name: xmv_Latn data_files: - split: train path: xmv_Latn/*_keep.jsonl - config_name: crm_Cans data_files: - split: train path: crm_Cans/*_keep.jsonl - config_name: mxv_Latn data_files: - split: train path: mxv_Latn/*_keep.jsonl - config_name: ubr_Latn data_files: - split: train path: ubr_Latn/*_keep.jsonl - config_name: gld_Cyrl data_files: - split: train path: gld_Cyrl/*_keep.jsonl - config_name: bzi_Thai data_files: - split: train path: bzi_Thai/*_keep.jsonl - config_name: ann_Latn data_files: - split: train path: ann_Latn/*_keep.jsonl - config_name: lac_Latn data_files: - split: train path: lac_Latn/*_keep.jsonl - config_name: mej_Latn data_files: - split: train path: mej_Latn/*_keep.jsonl - config_name: zpc_Latn data_files: - split: train path: zpc_Latn/*_keep.jsonl - config_name: dje_Latn data_files: - split: train path: dje_Latn/*_keep.jsonl - config_name: caq_Latn data_files: - split: train path: caq_Latn/*_keep.jsonl - config_name: guk_Ethi data_files: - split: train path: guk_Ethi/*_keep.jsonl - config_name: wos_Latn data_files: - split: train path: wos_Latn/*_keep.jsonl - config_name: msm_Latn data_files: - split: train path: msm_Latn/*_keep.jsonl - config_name: zaw_Latn data_files: - split: train path: zaw_Latn/*_keep.jsonl - config_name: cap_Latn data_files: - split: train path: cap_Latn/*_keep.jsonl - config_name: mox_Latn data_files: - split: train path: mox_Latn/*_keep.jsonl - config_name: zab_Latn data_files: - split: train path: zab_Latn/*_keep.jsonl - config_name: mca_Latn data_files: - split: train path: mca_Latn/*_keep.jsonl - config_name: wrs_Latn data_files: - split: train path: wrs_Latn/*_keep.jsonl - config_name: hae_Latn data_files: - split: train path: hae_Latn/*_keep.jsonl - config_name: nss_Latn data_files: - split: train path: nss_Latn/*_keep.jsonl - config_name: rav_Deva data_files: - split: train path: rav_Deva/*_keep.jsonl - config_name: uvh_Latn data_files: - split: train path: uvh_Latn/*_keep.jsonl - config_name: dga_Latn data_files: - split: train path: dga_Latn/*_keep.jsonl - config_name: ppo_Latn data_files: - split: train path: ppo_Latn/*_keep.jsonl - config_name: nog_Latn data_files: - split: train path: nog_Latn/*_keep.jsonl - config_name: arq_Arab data_files: - split: train path: arq_Arab/*_keep.jsonl - config_name: bim_Latn data_files: - split: train path: bim_Latn/*_keep.jsonl - config_name: lwg_Latn data_files: - split: train path: lwg_Latn/*_keep.jsonl - config_name: kms_Latn data_files: - split: train path: kms_Latn/*_keep.jsonl - config_name: kne_Latn data_files: - split: train path: kne_Latn/*_keep.jsonl - config_name: tkl_Latn data_files: - split: train path: tkl_Latn/*_keep.jsonl - config_name: ady_Latn data_files: - split: train path: ady_Latn/*_keep.jsonl - config_name: caf_Latn data_files: - split: train path: caf_Latn/*_keep.jsonl - config_name: zat_Latn data_files: - split: train path: zat_Latn/*_keep.jsonl - config_name: aba_Latn data_files: - split: train path: aba_Latn/*_keep.jsonl - config_name: chf_Latn data_files: - split: train path: chf_Latn/*_keep.jsonl - config_name: lew_Latn data_files: - split: train path: lew_Latn/*_keep.jsonl - config_name: npl_Latn data_files: - split: train path: npl_Latn/*_keep.jsonl - config_name: uvl_Latn data_files: - split: train path: uvl_Latn/*_keep.jsonl - config_name: cdo_Hani data_files: - split: train path: cdo_Hani/*_keep.jsonl - config_name: rmq_Latn data_files: - split: train path: rmq_Latn/*_keep.jsonl - config_name: mek_Latn data_files: - split: train path: mek_Latn/*_keep.jsonl - config_name: snc_Latn data_files: - split: train path: snc_Latn/*_keep.jsonl - config_name: nsn_Latn data_files: - split: train path: nsn_Latn/*_keep.jsonl - config_name: amn_Latn data_files: - split: train path: amn_Latn/*_keep.jsonl - config_name: abz_Latn data_files: - split: train path: abz_Latn/*_keep.jsonl - config_name: cek_Latn data_files: - split: train path: cek_Latn/*_keep.jsonl - config_name: ong_Latn data_files: - split: train path: ong_Latn/*_keep.jsonl - config_name: txq_Latn data_files: - split: train path: txq_Latn/*_keep.jsonl - config_name: bjp_Latn data_files: - split: train path: bjp_Latn/*_keep.jsonl - config_name: gvf_Latn data_files: - split: train path: gvf_Latn/*_keep.jsonl - config_name: crk_Cans data_files: - split: train path: crk_Cans/*_keep.jsonl - config_name: snf_Latn data_files: - split: train path: snf_Latn/*_keep.jsonl - config_name: nhg_Latn data_files: - split: train path: nhg_Latn/*_keep.jsonl - config_name: kqw_Latn data_files: - split: train path: kqw_Latn/*_keep.jsonl - config_name: mic_Latn data_files: - split: train path: mic_Latn/*_keep.jsonl - config_name: mie_Latn data_files: - split: train path: mie_Latn/*_keep.jsonl - config_name: nus_Latn data_files: - split: train path: nus_Latn/*_keep.jsonl - config_name: plu_Latn data_files: - split: train path: plu_Latn/*_keep.jsonl - config_name: idu_Latn data_files: - split: train path: idu_Latn/*_keep.jsonl - config_name: sll_Latn data_files: - split: train path: sll_Latn/*_keep.jsonl - config_name: thl_Deva data_files: - split: train path: thl_Deva/*_keep.jsonl - config_name: bef_Latn data_files: - split: train path: bef_Latn/*_keep.jsonl - config_name: kwd_Latn data_files: - split: train path: kwd_Latn/*_keep.jsonl - config_name: mqy_Latn data_files: - split: train path: mqy_Latn/*_keep.jsonl - config_name: caa_Latn data_files: - split: train path: caa_Latn/*_keep.jsonl - config_name: aia_Latn data_files: - split: train path: aia_Latn/*_keep.jsonl - config_name: wer_Latn data_files: - split: train path: wer_Latn/*_keep.jsonl - config_name: agm_Latn data_files: - split: train path: agm_Latn/*_keep.jsonl - config_name: mbl_Latn data_files: - split: train path: mbl_Latn/*_keep.jsonl - config_name: mbs_Latn data_files: - split: train path: mbs_Latn/*_keep.jsonl - config_name: adh_Latn data_files: - split: train path: adh_Latn/*_keep.jsonl - config_name: hub_Latn data_files: - split: train path: hub_Latn/*_keep.jsonl - config_name: mav_Latn data_files: - split: train path: mav_Latn/*_keep.jsonl - config_name: nab_Latn data_files: - split: train path: nab_Latn/*_keep.jsonl - config_name: bqp_Latn data_files: - split: train path: bqp_Latn/*_keep.jsonl - config_name: guo_Latn data_files: - split: train path: guo_Latn/*_keep.jsonl - config_name: auy_Latn data_files: - split: train path: auy_Latn/*_keep.jsonl - config_name: big_Latn data_files: - split: train path: big_Latn/*_keep.jsonl - config_name: nii_Latn data_files: - split: train path: nii_Latn/*_keep.jsonl - config_name: row_Latn data_files: - split: train path: row_Latn/*_keep.jsonl - config_name: kim_Cyrl data_files: - split: train path: kim_Cyrl/*_keep.jsonl - config_name: mti_Latn data_files: - split: train path: mti_Latn/*_keep.jsonl - config_name: otm_Latn data_files: - split: train path: otm_Latn/*_keep.jsonl - config_name: prg_Latn data_files: - split: train path: prg_Latn/*_keep.jsonl - config_name: icr_Latn data_files: - split: train path: icr_Latn/*_keep.jsonl - config_name: mcp_Latn data_files: - split: train path: mcp_Latn/*_keep.jsonl - config_name: tbo_Latn data_files: - split: train path: tbo_Latn/*_keep.jsonl - config_name: chy_Latn data_files: - split: train path: chy_Latn/*_keep.jsonl - config_name: oji_Latn data_files: - split: train path: oji_Latn/*_keep.jsonl - config_name: kpx_Latn data_files: - split: train path: kpx_Latn/*_keep.jsonl - config_name: lbb_Latn data_files: - split: train path: lbb_Latn/*_keep.jsonl - config_name: pao_Latn data_files: - split: train path: pao_Latn/*_keep.jsonl - config_name: xuo_Latn data_files: - split: train path: xuo_Latn/*_keep.jsonl - config_name: zgh_Tfng data_files: - split: train path: zgh_Tfng/*_keep.jsonl - config_name: wls_Latn data_files: - split: train path: wls_Latn/*_keep.jsonl - config_name: aca_Latn data_files: - split: train path: aca_Latn/*_keep.jsonl - config_name: apy_Latn data_files: - split: train path: apy_Latn/*_keep.jsonl - config_name: wnc_Latn data_files: - split: train path: wnc_Latn/*_keep.jsonl - config_name: ssg_Latn data_files: - split: train path: ssg_Latn/*_keep.jsonl - config_name: eve_Cyrl data_files: - split: train path: eve_Cyrl/*_keep.jsonl - config_name: mee_Latn data_files: - split: train path: mee_Latn/*_keep.jsonl - config_name: yut_Latn data_files: - split: train path: yut_Latn/*_keep.jsonl - config_name: chr_Latn data_files: - split: train path: chr_Latn/*_keep.jsonl - config_name: kxm_Thai data_files: - split: train path: kxm_Thai/*_keep.jsonl - config_name: mur_Latn data_files: - split: train path: mur_Latn/*_keep.jsonl - config_name: lad_Hebr data_files: - split: train path: lad_Hebr/*_keep.jsonl - config_name: mcd_Latn data_files: - split: train path: mcd_Latn/*_keep.jsonl - config_name: cto_Latn data_files: - split: train path: cto_Latn/*_keep.jsonl - config_name: mqj_Latn data_files: - split: train path: mqj_Latn/*_keep.jsonl - config_name: usa_Latn data_files: - split: train path: usa_Latn/*_keep.jsonl - config_name: emi_Latn data_files: - split: train path: emi_Latn/*_keep.jsonl - config_name: liv_Latn data_files: - split: train path: liv_Latn/*_keep.jsonl - config_name: guz_Latn data_files: - split: train path: guz_Latn/*_keep.jsonl - config_name: tgp_Latn data_files: - split: train path: tgp_Latn/*_keep.jsonl - config_name: ncl_Latn data_files: - split: train path: ncl_Latn/*_keep.jsonl - config_name: azg_Latn data_files: - split: train path: azg_Latn/*_keep.jsonl - config_name: kck_Latn data_files: - split: train path: kck_Latn/*_keep.jsonl - config_name: apb_Latn data_files: - split: train path: apb_Latn/*_keep.jsonl - config_name: kbm_Latn data_files: - split: train path: kbm_Latn/*_keep.jsonl - config_name: sgz_Latn data_files: - split: train path: sgz_Latn/*_keep.jsonl - config_name: wsk_Latn data_files: - split: train path: wsk_Latn/*_keep.jsonl - config_name: gbm_Deva data_files: - split: train path: gbm_Deva/*_keep.jsonl - config_name: csw_Latn data_files: - split: train path: csw_Latn/*_keep.jsonl - config_name: ctp_Latn data_files: - split: train path: ctp_Latn/*_keep.jsonl - config_name: eri_Latn data_files: - split: train path: eri_Latn/*_keep.jsonl - config_name: vid_Latn data_files: - split: train path: vid_Latn/*_keep.jsonl - config_name: bsq_Latn data_files: - split: train path: bsq_Latn/*_keep.jsonl - config_name: dak_Latn data_files: - split: train path: dak_Latn/*_keep.jsonl - config_name: tif_Latn data_files: - split: train path: tif_Latn/*_keep.jsonl - config_name: xsr_Deva data_files: - split: train path: xsr_Deva/*_keep.jsonl - config_name: mwp_Latn data_files: - split: train path: mwp_Latn/*_keep.jsonl - config_name: roo_Latn data_files: - split: train path: roo_Latn/*_keep.jsonl - config_name: wbp_Latn data_files: - split: train path: wbp_Latn/*_keep.jsonl - config_name: gdn_Latn data_files: - split: train path: gdn_Latn/*_keep.jsonl - config_name: tgo_Latn data_files: - split: train path: tgo_Latn/*_keep.jsonl - config_name: zpm_Latn data_files: - split: train path: zpm_Latn/*_keep.jsonl - config_name: mih_Latn data_files: - split: train path: mih_Latn/*_keep.jsonl - config_name: qvc_Latn data_files: - split: train path: qvc_Latn/*_keep.jsonl - config_name: tod_Latn data_files: - split: train path: tod_Latn/*_keep.jsonl - config_name: abn_Latn data_files: - split: train path: abn_Latn/*_keep.jsonl - config_name: cao_Latn data_files: - split: train path: cao_Latn/*_keep.jsonl - config_name: kbc_Latn data_files: - split: train path: kbc_Latn/*_keep.jsonl - config_name: kog_Latn data_files: - split: train path: kog_Latn/*_keep.jsonl - config_name: moh_Latn data_files: - split: train path: moh_Latn/*_keep.jsonl - config_name: leh_Latn data_files: - split: train path: leh_Latn/*_keep.jsonl - config_name: nsu_Latn data_files: - split: train path: nsu_Latn/*_keep.jsonl - config_name: okv_Latn data_files: - split: train path: okv_Latn/*_keep.jsonl - config_name: tih_Latn data_files: - split: train path: tih_Latn/*_keep.jsonl - config_name: zos_Latn data_files: - split: train path: zos_Latn/*_keep.jsonl - config_name: llg_Latn data_files: - split: train path: llg_Latn/*_keep.jsonl - config_name: gmv_Ethi data_files: - split: train path: gmv_Ethi/*_keep.jsonl - config_name: kyf_Latn data_files: - split: train path: kyf_Latn/*_keep.jsonl - config_name: mhx_Latn data_files: - split: train path: mhx_Latn/*_keep.jsonl - config_name: mvp_Latn data_files: - split: train path: mvp_Latn/*_keep.jsonl - config_name: naf_Latn data_files: - split: train path: naf_Latn/*_keep.jsonl - config_name: ino_Latn data_files: - split: train path: ino_Latn/*_keep.jsonl - config_name: dng_Cyrl data_files: - split: train path: dng_Cyrl/*_keep.jsonl - config_name: mnk_Latn data_files: - split: train path: mnk_Latn/*_keep.jsonl - config_name: oke_Latn data_files: - split: train path: oke_Latn/*_keep.jsonl - config_name: omw_Latn data_files: - split: train path: omw_Latn/*_keep.jsonl - config_name: tnk_Latn data_files: - split: train path: tnk_Latn/*_keep.jsonl - config_name: xtd_Latn data_files: - split: train path: xtd_Latn/*_keep.jsonl - config_name: agt_Latn data_files: - split: train path: agt_Latn/*_keep.jsonl - config_name: cbr_Latn data_files: - split: train path: cbr_Latn/*_keep.jsonl - config_name: eko_Latn data_files: - split: train path: eko_Latn/*_keep.jsonl - config_name: nhk_Latn data_files: - split: train path: nhk_Latn/*_keep.jsonl - config_name: hot_Latn data_files: - split: train path: hot_Latn/*_keep.jsonl - config_name: iou_Latn data_files: - split: train path: iou_Latn/*_keep.jsonl - config_name: mpm_Latn data_files: - split: train path: mpm_Latn/*_keep.jsonl - config_name: myx_Latn data_files: - split: train path: myx_Latn/*_keep.jsonl - config_name: ndh_Latn data_files: - split: train path: ndh_Latn/*_keep.jsonl - config_name: pbb_Latn data_files: - split: train path: pbb_Latn/*_keep.jsonl - config_name: rng_Latn data_files: - split: train path: rng_Latn/*_keep.jsonl - config_name: cri_Latn data_files: - split: train path: cri_Latn/*_keep.jsonl - config_name: bvd_Latn data_files: - split: train path: bvd_Latn/*_keep.jsonl - config_name: kmz_Latn data_files: - split: train path: kmz_Latn/*_keep.jsonl - config_name: cta_Latn data_files: - split: train path: cta_Latn/*_keep.jsonl - config_name: blw_Latn data_files: - split: train path: blw_Latn/*_keep.jsonl - config_name: clu_Latn data_files: - split: train path: clu_Latn/*_keep.jsonl - config_name: cof_Latn data_files: - split: train path: cof_Latn/*_keep.jsonl - config_name: kje_Latn data_files: - split: train path: kje_Latn/*_keep.jsonl - config_name: nmf_Latn data_files: - split: train path: nmf_Latn/*_keep.jsonl - config_name: kbh_Latn data_files: - split: train path: kbh_Latn/*_keep.jsonl - config_name: qxn_Latn data_files: - split: train path: qxn_Latn/*_keep.jsonl - config_name: ria_Latn data_files: - split: train path: ria_Latn/*_keep.jsonl - config_name: wed_Latn data_files: - split: train path: wed_Latn/*_keep.jsonl - config_name: zar_Latn data_files: - split: train path: zar_Latn/*_keep.jsonl - config_name: huv_Latn data_files: - split: train path: huv_Latn/*_keep.jsonl - config_name: lsi_Latn data_files: - split: train path: lsi_Latn/*_keep.jsonl - config_name: msk_Latn data_files: - split: train path: msk_Latn/*_keep.jsonl - config_name: nbq_Latn data_files: - split: train path: nbq_Latn/*_keep.jsonl - config_name: ati_Latn data_files: - split: train path: ati_Latn/*_keep.jsonl - config_name: lww_Latn data_files: - split: train path: lww_Latn/*_keep.jsonl - config_name: yva_Latn data_files: - split: train path: yva_Latn/*_keep.jsonl - config_name: amr_Latn data_files: - split: train path: amr_Latn/*_keep.jsonl - config_name: lmk_Latn data_files: - split: train path: lmk_Latn/*_keep.jsonl - config_name: tim_Latn data_files: - split: train path: tim_Latn/*_keep.jsonl - config_name: bbb_Latn data_files: - split: train path: bbb_Latn/*_keep.jsonl - config_name: bzd_Latn data_files: - split: train path: bzd_Latn/*_keep.jsonl - config_name: djr_Latn data_files: - split: train path: djr_Latn/*_keep.jsonl - config_name: kwf_Latn data_files: - split: train path: kwf_Latn/*_keep.jsonl - config_name: nuy_Latn data_files: - split: train path: nuy_Latn/*_keep.jsonl - config_name: poi_Latn data_files: - split: train path: poi_Latn/*_keep.jsonl - config_name: ptp_Latn data_files: - split: train path: ptp_Latn/*_keep.jsonl - config_name: atd_Latn data_files: - split: train path: atd_Latn/*_keep.jsonl - config_name: bps_Latn data_files: - split: train path: bps_Latn/*_keep.jsonl - config_name: lcm_Latn data_files: - split: train path: lcm_Latn/*_keep.jsonl - config_name: mbh_Latn data_files: - split: train path: mbh_Latn/*_keep.jsonl - config_name: nuj_Latn data_files: - split: train path: nuj_Latn/*_keep.jsonl - config_name: ayp_Arab data_files: - split: train path: ayp_Arab/*_keep.jsonl - config_name: cwe_Latn data_files: - split: train path: cwe_Latn/*_keep.jsonl - config_name: geb_Latn data_files: - split: train path: geb_Latn/*_keep.jsonl - config_name: aoi_Latn data_files: - split: train path: aoi_Latn/*_keep.jsonl - config_name: mto_Latn data_files: - split: train path: mto_Latn/*_keep.jsonl - config_name: cbi_Latn data_files: - split: train path: cbi_Latn/*_keep.jsonl - config_name: gvn_Latn data_files: - split: train path: gvn_Latn/*_keep.jsonl - config_name: kqf_Latn data_files: - split: train path: kqf_Latn/*_keep.jsonl - config_name: rkb_Latn data_files: - split: train path: rkb_Latn/*_keep.jsonl - config_name: xed_Latn data_files: - split: train path: xed_Latn/*_keep.jsonl - config_name: zas_Latn data_files: - split: train path: zas_Latn/*_keep.jsonl - config_name: apz_Latn data_files: - split: train path: apz_Latn/*_keep.jsonl - config_name: cpb_Latn data_files: - split: train path: cpb_Latn/*_keep.jsonl - config_name: ese_Latn data_files: - split: train path: ese_Latn/*_keep.jsonl - config_name: gdg_Latn data_files: - split: train path: gdg_Latn/*_keep.jsonl - config_name: iws_Latn data_files: - split: train path: iws_Latn/*_keep.jsonl - config_name: apu_Latn data_files: - split: train path: apu_Latn/*_keep.jsonl - config_name: cag_Latn data_files: - split: train path: cag_Latn/*_keep.jsonl - config_name: kdc_Latn data_files: - split: train path: kdc_Latn/*_keep.jsonl - config_name: nou_Latn data_files: - split: train path: nou_Latn/*_keep.jsonl - config_name: too_Latn data_files: - split: train path: too_Latn/*_keep.jsonl - config_name: xmm_Latn data_files: - split: train path: xmm_Latn/*_keep.jsonl - config_name: beq_Latn data_files: - split: train path: beq_Latn/*_keep.jsonl - config_name: hnn_Latn data_files: - split: train path: hnn_Latn/*_keep.jsonl - config_name: myk_Latn data_files: - split: train path: myk_Latn/*_keep.jsonl - config_name: sby_Latn data_files: - split: train path: sby_Latn/*_keep.jsonl - config_name: arl_Latn data_files: - split: train path: arl_Latn/*_keep.jsonl - config_name: ncu_Latn data_files: - split: train path: ncu_Latn/*_keep.jsonl - config_name: rml_Latn data_files: - split: train path: rml_Latn/*_keep.jsonl - config_name: gam_Latn data_files: - split: train path: gam_Latn/*_keep.jsonl - config_name: kvn_Latn data_files: - split: train path: kvn_Latn/*_keep.jsonl - config_name: mbc_Latn data_files: - split: train path: mbc_Latn/*_keep.jsonl - config_name: mfy_Latn data_files: - split: train path: mfy_Latn/*_keep.jsonl - config_name: urt_Latn data_files: - split: train path: urt_Latn/*_keep.jsonl - config_name: luc_Latn data_files: - split: train path: luc_Latn/*_keep.jsonl - config_name: mio_Latn data_files: - split: train path: mio_Latn/*_keep.jsonl - config_name: snn_Latn data_files: - split: train path: snn_Latn/*_keep.jsonl - config_name: tar_Latn data_files: - split: train path: tar_Latn/*_keep.jsonl - config_name: tku_Latn data_files: - split: train path: tku_Latn/*_keep.jsonl - config_name: upv_Latn data_files: - split: train path: upv_Latn/*_keep.jsonl - config_name: kjg_Laoo data_files: - split: train path: kjg_Laoo/*_keep.jsonl - config_name: btd_Latn data_files: - split: train path: btd_Latn/*_keep.jsonl - config_name: cut_Latn data_files: - split: train path: cut_Latn/*_keep.jsonl - config_name: mwn_Latn data_files: - split: train path: mwn_Latn/*_keep.jsonl - config_name: tue_Latn data_files: - split: train path: tue_Latn/*_keep.jsonl - config_name: adz_Latn data_files: - split: train path: adz_Latn/*_keep.jsonl - config_name: kyg_Latn data_files: - split: train path: kyg_Latn/*_keep.jsonl - config_name: rel_Latn data_files: - split: train path: rel_Latn/*_keep.jsonl - config_name: tew_Latn data_files: - split: train path: tew_Latn/*_keep.jsonl - config_name: ttc_Latn data_files: - split: train path: ttc_Latn/*_keep.jsonl - config_name: nkf_Latn data_files: - split: train path: nkf_Latn/*_keep.jsonl - config_name: yrb_Latn data_files: - split: train path: yrb_Latn/*_keep.jsonl - config_name: fub_Latn data_files: - split: train path: fub_Latn/*_keep.jsonl - config_name: nsm_Latn data_files: - split: train path: nsm_Latn/*_keep.jsonl - config_name: nyf_Latn data_files: - split: train path: nyf_Latn/*_keep.jsonl - config_name: ava_Latn data_files: - split: train path: ava_Latn/*_keep.jsonl - config_name: bkd_Latn data_files: - split: train path: bkd_Latn/*_keep.jsonl - config_name: cpc_Latn data_files: - split: train path: cpc_Latn/*_keep.jsonl - config_name: lif_Limb data_files: - split: train path: lif_Limb/*_keep.jsonl - config_name: mwc_Latn data_files: - split: train path: mwc_Latn/*_keep.jsonl - config_name: gan_Latn data_files: - split: train path: gan_Latn/*_keep.jsonl - config_name: apw_Latn data_files: - split: train path: apw_Latn/*_keep.jsonl - config_name: azz_Latn data_files: - split: train path: azz_Latn/*_keep.jsonl - config_name: zpo_Latn data_files: - split: train path: zpo_Latn/*_keep.jsonl - config_name: ace_Arab data_files: - split: train path: ace_Arab/*_keep.jsonl - config_name: faa_Latn data_files: - split: train path: faa_Latn/*_keep.jsonl - config_name: ood_Latn data_files: - split: train path: ood_Latn/*_keep.jsonl - config_name: toc_Latn data_files: - split: train path: toc_Latn/*_keep.jsonl - config_name: wmt_Latn data_files: - split: train path: wmt_Latn/*_keep.jsonl - config_name: ajg_Latn data_files: - split: train path: ajg_Latn/*_keep.jsonl - config_name: nko_Latn data_files: - split: train path: nko_Latn/*_keep.jsonl - config_name: ssd_Latn data_files: - split: train path: ssd_Latn/*_keep.jsonl - config_name: sxb_Latn data_files: - split: train path: sxb_Latn/*_keep.jsonl - config_name: zxx_Latn data_files: - split: train path: zxx_Latn/*_keep.jsonl - config_name: not_Latn data_files: - split: train path: not_Latn/*_keep.jsonl - config_name: amk_Latn data_files: - split: train path: amk_Latn/*_keep.jsonl - config_name: hch_Latn data_files: - split: train path: hch_Latn/*_keep.jsonl - config_name: kqc_Latn data_files: - split: train path: kqc_Latn/*_keep.jsonl - config_name: aui_Latn data_files: - split: train path: aui_Latn/*_keep.jsonl - config_name: llb_Latn data_files: - split: train path: llb_Latn/*_keep.jsonl - config_name: avt_Latn data_files: - split: train path: avt_Latn/*_keep.jsonl - config_name: kaq_Latn data_files: - split: train path: kaq_Latn/*_keep.jsonl - config_name: tna_Latn data_files: - split: train path: tna_Latn/*_keep.jsonl - config_name: zpt_Latn data_files: - split: train path: zpt_Latn/*_keep.jsonl - config_name: mnw_Latn data_files: - split: train path: mnw_Latn/*_keep.jsonl - config_name: bwd_Latn data_files: - split: train path: bwd_Latn/*_keep.jsonl - config_name: lif_Deva data_files: - split: train path: lif_Deva/*_keep.jsonl - config_name: stn_Latn data_files: - split: train path: stn_Latn/*_keep.jsonl - config_name: crk_Latn data_files: - split: train path: crk_Latn/*_keep.jsonl - config_name: kgk_Latn data_files: - split: train path: kgk_Latn/*_keep.jsonl - config_name: mck_Latn data_files: - split: train path: mck_Latn/*_keep.jsonl - config_name: bhg_Latn data_files: - split: train path: bhg_Latn/*_keep.jsonl - config_name: ifu_Latn data_files: - split: train path: ifu_Latn/*_keep.jsonl - config_name: kkl_Latn data_files: - split: train path: kkl_Latn/*_keep.jsonl - config_name: zao_Latn data_files: - split: train path: zao_Latn/*_keep.jsonl - config_name: huu_Latn data_files: - split: train path: huu_Latn/*_keep.jsonl - config_name: sil_Latn data_files: - split: train path: sil_Latn/*_keep.jsonl - config_name: trq_Latn data_files: - split: train path: trq_Latn/*_keep.jsonl - config_name: tuf_Latn data_files: - split: train path: tuf_Latn/*_keep.jsonl - config_name: wnu_Latn data_files: - split: train path: wnu_Latn/*_keep.jsonl - config_name: amf_Latn data_files: - split: train path: amf_Latn/*_keep.jsonl - config_name: fan_Latn data_files: - split: train path: fan_Latn/*_keep.jsonl - config_name: zdj_Latn data_files: - split: train path: zdj_Latn/*_keep.jsonl - config_name: bco_Latn data_files: - split: train path: bco_Latn/*_keep.jsonl - config_name: kdl_Latn data_files: - split: train path: kdl_Latn/*_keep.jsonl - config_name: lln_Latn data_files: - split: train path: lln_Latn/*_keep.jsonl - config_name: mit_Latn data_files: - split: train path: mit_Latn/*_keep.jsonl - config_name: yaq_Latn data_files: - split: train path: yaq_Latn/*_keep.jsonl - config_name: ake_Latn data_files: - split: train path: ake_Latn/*_keep.jsonl - config_name: ige_Latn data_files: - split: train path: ige_Latn/*_keep.jsonl - config_name: mna_Latn data_files: - split: train path: mna_Latn/*_keep.jsonl - config_name: nhx_Latn data_files: - split: train path: nhx_Latn/*_keep.jsonl - config_name: njn_Latn data_files: - split: train path: njn_Latn/*_keep.jsonl - config_name: wba_Latn data_files: - split: train path: wba_Latn/*_keep.jsonl - config_name: zaa_Latn data_files: - split: train path: zaa_Latn/*_keep.jsonl - config_name: lez_Latn data_files: - split: train path: lez_Latn/*_keep.jsonl - config_name: syb_Latn data_files: - split: train path: syb_Latn/*_keep.jsonl - config_name: qxo_Latn data_files: - split: train path: qxo_Latn/*_keep.jsonl - config_name: rgu_Latn data_files: - split: train path: rgu_Latn/*_keep.jsonl - config_name: wiu_Latn data_files: - split: train path: wiu_Latn/*_keep.jsonl - config_name: bla_Latn data_files: - split: train path: bla_Latn/*_keep.jsonl - config_name: etr_Latn data_files: - split: train path: etr_Latn/*_keep.jsonl - config_name: miy_Latn data_files: - split: train path: miy_Latn/*_keep.jsonl - config_name: mjc_Latn data_files: - split: train path: mjc_Latn/*_keep.jsonl - config_name: nif_Latn data_files: - split: train path: nif_Latn/*_keep.jsonl - config_name: pov_Latn data_files: - split: train path: pov_Latn/*_keep.jsonl - config_name: qxl_Latn data_files: - split: train path: qxl_Latn/*_keep.jsonl - config_name: sey_Latn data_files: - split: train path: sey_Latn/*_keep.jsonl - config_name: bxh_Latn data_files: - split: train path: bxh_Latn/*_keep.jsonl - config_name: ddg_Latn data_files: - split: train path: ddg_Latn/*_keep.jsonl - config_name: mks_Latn data_files: - split: train path: mks_Latn/*_keep.jsonl - config_name: nki_Latn data_files: - split: train path: nki_Latn/*_keep.jsonl - config_name: pot_Latn data_files: - split: train path: pot_Latn/*_keep.jsonl - config_name: prf_Latn data_files: - split: train path: prf_Latn/*_keep.jsonl - config_name: hra_Latn data_files: - split: train path: hra_Latn/*_keep.jsonl - config_name: law_Latn data_files: - split: train path: law_Latn/*_keep.jsonl - config_name: plw_Latn data_files: - split: train path: plw_Latn/*_keep.jsonl - config_name: zpi_Latn data_files: - split: train path: zpi_Latn/*_keep.jsonl - config_name: atg_Latn data_files: - split: train path: atg_Latn/*_keep.jsonl - config_name: daf_Latn data_files: - split: train path: daf_Latn/*_keep.jsonl - config_name: drg_Latn data_files: - split: train path: drg_Latn/*_keep.jsonl - config_name: dru_Latn data_files: - split: train path: dru_Latn/*_keep.jsonl - config_name: ngp_Latn data_files: - split: train path: ngp_Latn/*_keep.jsonl - config_name: boa_Latn data_files: - split: train path: boa_Latn/*_keep.jsonl - config_name: cbt_Latn data_files: - split: train path: cbt_Latn/*_keep.jsonl - config_name: dhm_Latn data_files: - split: train path: dhm_Latn/*_keep.jsonl - config_name: ikw_Latn data_files: - split: train path: ikw_Latn/*_keep.jsonl - config_name: kqe_Latn data_files: - split: train path: kqe_Latn/*_keep.jsonl - config_name: lbj_Tibt data_files: - split: train path: lbj_Tibt/*_keep.jsonl - config_name: loq_Latn data_files: - split: train path: loq_Latn/*_keep.jsonl - config_name: xsb_Latn data_files: - split: train path: xsb_Latn/*_keep.jsonl - config_name: yad_Latn data_files: - split: train path: yad_Latn/*_keep.jsonl - config_name: kyu_Latn data_files: - split: train path: kyu_Latn/*_keep.jsonl - config_name: mcb_Latn data_files: - split: train path: mcb_Latn/*_keep.jsonl - config_name: tks_Arab data_files: - split: train path: tks_Arab/*_keep.jsonl - config_name: cnl_Latn data_files: - split: train path: cnl_Latn/*_keep.jsonl - config_name: far_Latn data_files: - split: train path: far_Latn/*_keep.jsonl - config_name: miz_Latn data_files: - split: train path: miz_Latn/*_keep.jsonl - config_name: maj_Latn data_files: - split: train path: maj_Latn/*_keep.jsonl - config_name: mgm_Latn data_files: - split: train path: mgm_Latn/*_keep.jsonl - config_name: mib_Latn data_files: - split: train path: mib_Latn/*_keep.jsonl - config_name: mza_Latn data_files: - split: train path: mza_Latn/*_keep.jsonl - config_name: nhr_Latn data_files: - split: train path: nhr_Latn/*_keep.jsonl - config_name: smt_Latn data_files: - split: train path: smt_Latn/*_keep.jsonl - config_name: ish_Latn data_files: - split: train path: ish_Latn/*_keep.jsonl - config_name: mlu_Latn data_files: - split: train path: mlu_Latn/*_keep.jsonl - config_name: yka_Latn data_files: - split: train path: yka_Latn/*_keep.jsonl - config_name: anv_Latn data_files: - split: train path: anv_Latn/*_keep.jsonl - config_name: ign_Latn data_files: - split: train path: ign_Latn/*_keep.jsonl - config_name: mir_Latn data_files: - split: train path: mir_Latn/*_keep.jsonl - config_name: tav_Latn data_files: - split: train path: tav_Latn/*_keep.jsonl - config_name: gyr_Latn data_files: - split: train path: gyr_Latn/*_keep.jsonl - config_name: klt_Latn data_files: - split: train path: klt_Latn/*_keep.jsonl - config_name: apn_Latn data_files: - split: train path: apn_Latn/*_keep.jsonl - config_name: wbm_Latn data_files: - split: train path: wbm_Latn/*_keep.jsonl - config_name: cjk_Latn data_files: - split: train path: cjk_Latn/*_keep.jsonl - config_name: dln_Latn data_files: - split: train path: dln_Latn/*_keep.jsonl - config_name: mcf_Latn data_files: - split: train path: mcf_Latn/*_keep.jsonl - config_name: twx_Latn data_files: - split: train path: twx_Latn/*_keep.jsonl - config_name: ldi_Latn data_files: - split: train path: ldi_Latn/*_keep.jsonl - config_name: poy_Latn data_files: - split: train path: poy_Latn/*_keep.jsonl - config_name: yaa_Latn data_files: - split: train path: yaa_Latn/*_keep.jsonl - config_name: zae_Latn data_files: - split: train path: zae_Latn/*_keep.jsonl - config_name: bdq_Latn data_files: - split: train path: bdq_Latn/*_keep.jsonl - config_name: con_Latn data_files: - split: train path: con_Latn/*_keep.jsonl - config_name: dnw_Latn data_files: - split: train path: dnw_Latn/*_keep.jsonl - config_name: gah_Latn data_files: - split: train path: gah_Latn/*_keep.jsonl - config_name: lcp_Thai data_files: - split: train path: lcp_Thai/*_keep.jsonl - config_name: mbj_Latn data_files: - split: train path: mbj_Latn/*_keep.jsonl - config_name: xog_Latn data_files: - split: train path: xog_Latn/*_keep.jsonl - config_name: yan_Latn data_files: - split: train path: yan_Latn/*_keep.jsonl - config_name: sdq_Latn data_files: - split: train path: sdq_Latn/*_keep.jsonl - config_name: sgh_Cyrl data_files: - split: train path: sgh_Cyrl/*_keep.jsonl - config_name: gwi_Latn data_files: - split: train path: gwi_Latn/*_keep.jsonl - config_name: lbk_Latn data_files: - split: train path: lbk_Latn/*_keep.jsonl - config_name: ses_Latn data_files: - split: train path: ses_Latn/*_keep.jsonl - config_name: ycn_Latn data_files: - split: train path: ycn_Latn/*_keep.jsonl - config_name: kau_Latn data_files: - split: train path: kau_Latn/*_keep.jsonl - config_name: hix_Latn data_files: - split: train path: hix_Latn/*_keep.jsonl - config_name: mig_Latn data_files: - split: train path: mig_Latn/*_keep.jsonl - config_name: mlh_Latn data_files: - split: train path: mlh_Latn/*_keep.jsonl - config_name: aer_Latn data_files: - split: train path: aer_Latn/*_keep.jsonl - config_name: amx_Latn data_files: - split: train path: amx_Latn/*_keep.jsonl - config_name: auc_Latn data_files: - split: train path: auc_Latn/*_keep.jsonl - config_name: cly_Latn data_files: - split: train path: cly_Latn/*_keep.jsonl - config_name: dsh_Latn data_files: - split: train path: dsh_Latn/*_keep.jsonl - config_name: gnb_Latn data_files: - split: train path: gnb_Latn/*_keep.jsonl - config_name: neb_Latn data_files: - split: train path: neb_Latn/*_keep.jsonl - config_name: ttt_Cyrl data_files: - split: train path: ttt_Cyrl/*_keep.jsonl - config_name: jic_Latn data_files: - split: train path: jic_Latn/*_keep.jsonl - config_name: xtm_Latn data_files: - split: train path: xtm_Latn/*_keep.jsonl - config_name: tll_Latn data_files: - split: train path: tll_Latn/*_keep.jsonl - config_name: apt_Latn data_files: - split: train path: apt_Latn/*_keep.jsonl - config_name: kgr_Latn data_files: - split: train path: kgr_Latn/*_keep.jsonl - config_name: pio_Latn data_files: - split: train path: pio_Latn/*_keep.jsonl - config_name: pmf_Latn data_files: - split: train path: pmf_Latn/*_keep.jsonl - config_name: yrk_Latn data_files: - split: train path: yrk_Latn/*_keep.jsonl - config_name: akh_Latn data_files: - split: train path: akh_Latn/*_keep.jsonl - config_name: box_Latn data_files: - split: train path: box_Latn/*_keep.jsonl - config_name: cub_Latn data_files: - split: train path: cub_Latn/*_keep.jsonl - config_name: cux_Latn data_files: - split: train path: cux_Latn/*_keep.jsonl - config_name: mrq_Latn data_files: - split: train path: mrq_Latn/*_keep.jsonl - config_name: inh_Latn data_files: - split: train path: inh_Latn/*_keep.jsonl - config_name: bsp_Latn data_files: - split: train path: bsp_Latn/*_keep.jsonl - config_name: for_Latn data_files: - split: train path: for_Latn/*_keep.jsonl - config_name: kny_Latn data_files: - split: train path: kny_Latn/*_keep.jsonl - config_name: uri_Latn data_files: - split: train path: uri_Latn/*_keep.jsonl - config_name: zsr_Latn data_files: - split: train path: zsr_Latn/*_keep.jsonl - config_name: pce_Mymr data_files: - split: train path: pce_Mymr/*_keep.jsonl - config_name: cme_Latn data_files: - split: train path: cme_Latn/*_keep.jsonl - config_name: cot_Latn data_files: - split: train path: cot_Latn/*_keep.jsonl - config_name: kxw_Latn data_files: - split: train path: kxw_Latn/*_keep.jsonl - config_name: omb_Latn data_files: - split: train path: omb_Latn/*_keep.jsonl - config_name: srq_Latn data_files: - split: train path: srq_Latn/*_keep.jsonl - config_name: vag_Latn data_files: - split: train path: vag_Latn/*_keep.jsonl - config_name: cpu_Latn data_files: - split: train path: cpu_Latn/*_keep.jsonl - config_name: kmy_Latn data_files: - split: train path: kmy_Latn/*_keep.jsonl - config_name: ksc_Latn data_files: - split: train path: ksc_Latn/*_keep.jsonl - config_name: lai_Latn data_files: - split: train path: lai_Latn/*_keep.jsonl - config_name: nnp_Latn data_files: - split: train path: nnp_Latn/*_keep.jsonl - config_name: ruf_Latn data_files: - split: train path: ruf_Latn/*_keep.jsonl - config_name: spm_Latn data_files: - split: train path: spm_Latn/*_keep.jsonl - config_name: tro_Latn data_files: - split: train path: tro_Latn/*_keep.jsonl - config_name: ttj_Latn data_files: - split: train path: ttj_Latn/*_keep.jsonl - config_name: waj_Latn data_files: - split: train path: waj_Latn/*_keep.jsonl - config_name: raj_Deva data_files: - split: train path: raj_Deva/*_keep.jsonl - config_name: ian_Latn data_files: - split: train path: ian_Latn/*_keep.jsonl - config_name: ivb_Latn data_files: - split: train path: ivb_Latn/*_keep.jsonl - config_name: kmm_Latn data_files: - split: train path: kmm_Latn/*_keep.jsonl - config_name: kzf_Latn data_files: - split: train path: kzf_Latn/*_keep.jsonl - config_name: maw_Latn data_files: - split: train path: maw_Latn/*_keep.jsonl - config_name: tuv_Latn data_files: - split: train path: tuv_Latn/*_keep.jsonl - config_name: whk_Latn data_files: - split: train path: whk_Latn/*_keep.jsonl - config_name: zsm_Arab data_files: - split: train path: zsm_Arab/*_keep.jsonl - config_name: aph_Deva data_files: - split: train path: aph_Deva/*_keep.jsonl - config_name: anm_Latn data_files: - split: train path: anm_Latn/*_keep.jsonl - config_name: mny_Latn data_files: - split: train path: mny_Latn/*_keep.jsonl - config_name: bug_Bugi data_files: - split: train path: bug_Bugi/*_keep.jsonl - config_name: trp_Latn data_files: - split: train path: trp_Latn/*_keep.jsonl - config_name: are_Latn data_files: - split: train path: are_Latn/*_keep.jsonl - config_name: esi_Latn data_files: - split: train path: esi_Latn/*_keep.jsonl - config_name: mhi_Latn data_files: - split: train path: mhi_Latn/*_keep.jsonl - config_name: obo_Latn data_files: - split: train path: obo_Latn/*_keep.jsonl - config_name: bwu_Latn data_files: - split: train path: bwu_Latn/*_keep.jsonl - config_name: lwo_Latn data_files: - split: train path: lwo_Latn/*_keep.jsonl - config_name: mrg_Latn data_files: - split: train path: mrg_Latn/*_keep.jsonl - config_name: old_Latn data_files: - split: train path: old_Latn/*_keep.jsonl - config_name: zav_Latn data_files: - split: train path: zav_Latn/*_keep.jsonl - config_name: snk_Latn data_files: - split: train path: snk_Latn/*_keep.jsonl - config_name: daa_Latn data_files: - split: train path: daa_Latn/*_keep.jsonl - config_name: fuq_Latn data_files: - split: train path: fuq_Latn/*_keep.jsonl - config_name: lud_Latn data_files: - split: train path: lud_Latn/*_keep.jsonl - config_name: nlg_Latn data_files: - split: train path: nlg_Latn/*_keep.jsonl - config_name: sck_Deva data_files: - split: train path: sck_Deva/*_keep.jsonl - config_name: tbl_Latn data_files: - split: train path: tbl_Latn/*_keep.jsonl - config_name: khb_Talu data_files: - split: train path: khb_Talu/*_keep.jsonl - config_name: tpu_Khmr data_files: - split: train path: tpu_Khmr/*_keep.jsonl - config_name: fue_Latn data_files: - split: train path: fue_Latn/*_keep.jsonl - config_name: tmd_Latn data_files: - split: train path: tmd_Latn/*_keep.jsonl - config_name: whg_Latn data_files: - split: train path: whg_Latn/*_keep.jsonl - config_name: blt_Latn data_files: - split: train path: blt_Latn/*_keep.jsonl - config_name: gqr_Latn data_files: - split: train path: gqr_Latn/*_keep.jsonl - config_name: kub_Latn data_files: - split: train path: kub_Latn/*_keep.jsonl - config_name: ots_Latn data_files: - split: train path: ots_Latn/*_keep.jsonl - config_name: tpp_Latn data_files: - split: train path: tpp_Latn/*_keep.jsonl - config_name: var_Latn data_files: - split: train path: var_Latn/*_keep.jsonl - config_name: hbb_Latn data_files: - split: train path: hbb_Latn/*_keep.jsonl - config_name: krr_Khmr data_files: - split: train path: krr_Khmr/*_keep.jsonl - config_name: zxx_Zyyy data_files: - split: train path: zxx_Zyyy/*_keep.jsonl - config_name: alj_Latn data_files: - split: train path: alj_Latn/*_keep.jsonl - config_name: ebk_Latn data_files: - split: train path: ebk_Latn/*_keep.jsonl - config_name: goa_Latn data_files: - split: train path: goa_Latn/*_keep.jsonl - config_name: hav_Latn data_files: - split: train path: hav_Latn/*_keep.jsonl - config_name: kmk_Latn data_files: - split: train path: kmk_Latn/*_keep.jsonl - config_name: kqo_Latn data_files: - split: train path: kqo_Latn/*_keep.jsonl - config_name: muy_Latn data_files: - split: train path: muy_Latn/*_keep.jsonl - config_name: njb_Latn data_files: - split: train path: njb_Latn/*_keep.jsonl - config_name: snw_Latn data_files: - split: train path: snw_Latn/*_keep.jsonl - config_name: wib_Latn data_files: - split: train path: wib_Latn/*_keep.jsonl - config_name: shn_Latn data_files: - split: train path: shn_Latn/*_keep.jsonl - config_name: tkr_Latn data_files: - split: train path: tkr_Latn/*_keep.jsonl - config_name: ksp_Latn data_files: - split: train path: ksp_Latn/*_keep.jsonl - config_name: lip_Latn data_files: - split: train path: lip_Latn/*_keep.jsonl - config_name: mog_Latn data_files: - split: train path: mog_Latn/*_keep.jsonl - config_name: nla_Latn data_files: - split: train path: nla_Latn/*_keep.jsonl - config_name: pss_Latn data_files: - split: train path: pss_Latn/*_keep.jsonl - config_name: zad_Latn data_files: - split: train path: zad_Latn/*_keep.jsonl - config_name: awi_Latn data_files: - split: train path: awi_Latn/*_keep.jsonl - config_name: frd_Latn data_files: - split: train path: frd_Latn/*_keep.jsonl - config_name: kei_Latn data_files: - split: train path: kei_Latn/*_keep.jsonl - config_name: kmd_Latn data_files: - split: train path: kmd_Latn/*_keep.jsonl - config_name: mim_Latn data_files: - split: train path: mim_Latn/*_keep.jsonl - config_name: got_Latn data_files: - split: train path: got_Latn/*_keep.jsonl - config_name: lbe_Latn data_files: - split: train path: lbe_Latn/*_keep.jsonl - config_name: adi_Latn data_files: - split: train path: adi_Latn/*_keep.jsonl - config_name: any_Latn data_files: - split: train path: any_Latn/*_keep.jsonl - config_name: arp_Latn data_files: - split: train path: arp_Latn/*_keep.jsonl - config_name: biu_Latn data_files: - split: train path: biu_Latn/*_keep.jsonl - config_name: bvc_Latn data_files: - split: train path: bvc_Latn/*_keep.jsonl - config_name: ghs_Latn data_files: - split: train path: ghs_Latn/*_keep.jsonl - config_name: ttq_Latn data_files: - split: train path: ttq_Latn/*_keep.jsonl - config_name: vmk_Latn data_files: - split: train path: vmk_Latn/*_keep.jsonl - config_name: zpj_Latn data_files: - split: train path: zpj_Latn/*_keep.jsonl - config_name: crl_Cans data_files: - split: train path: crl_Cans/*_keep.jsonl - config_name: dtb_Latn data_files: - split: train path: dtb_Latn/*_keep.jsonl - config_name: dug_Latn data_files: - split: train path: dug_Latn/*_keep.jsonl - config_name: kam_Latn data_files: - split: train path: kam_Latn/*_keep.jsonl - config_name: nmo_Latn data_files: - split: train path: nmo_Latn/*_keep.jsonl - config_name: qva_Latn data_files: - split: train path: qva_Latn/*_keep.jsonl - config_name: rhg_Latn data_files: - split: train path: rhg_Latn/*_keep.jsonl - config_name: taw_Latn data_files: - split: train path: taw_Latn/*_keep.jsonl - config_name: tui_Latn data_files: - split: train path: tui_Latn/*_keep.jsonl - config_name: uth_Latn data_files: - split: train path: uth_Latn/*_keep.jsonl - config_name: wap_Latn data_files: - split: train path: wap_Latn/*_keep.jsonl - config_name: zpg_Latn data_files: - split: train path: zpg_Latn/*_keep.jsonl - config_name: brb_Khmr data_files: - split: train path: brb_Khmr/*_keep.jsonl - config_name: dty_Latn data_files: - split: train path: dty_Latn/*_keep.jsonl - config_name: syr_Latn data_files: - split: train path: syr_Latn/*_keep.jsonl - config_name: bih_Arab data_files: - split: train path: bih_Arab/*_keep.jsonl - config_name: bqj_Latn data_files: - split: train path: bqj_Latn/*_keep.jsonl - config_name: cjv_Latn data_files: - split: train path: cjv_Latn/*_keep.jsonl - config_name: cnw_Latn data_files: - split: train path: cnw_Latn/*_keep.jsonl - config_name: kgf_Latn data_files: - split: train path: kgf_Latn/*_keep.jsonl - config_name: mfh_Latn data_files: - split: train path: mfh_Latn/*_keep.jsonl - config_name: mns_Cyrl data_files: - split: train path: mns_Cyrl/*_keep.jsonl - config_name: nnw_Latn data_files: - split: train path: nnw_Latn/*_keep.jsonl - config_name: ntp_Latn data_files: - split: train path: ntp_Latn/*_keep.jsonl - config_name: oto_Latn data_files: - split: train path: oto_Latn/*_keep.jsonl - config_name: dks_Latn data_files: - split: train path: dks_Latn/*_keep.jsonl - config_name: mpt_Latn data_files: - split: train path: mpt_Latn/*_keep.jsonl - config_name: nmw_Latn data_files: - split: train path: nmw_Latn/*_keep.jsonl - config_name: sba_Latn data_files: - split: train path: sba_Latn/*_keep.jsonl - config_name: sfw_Latn data_files: - split: train path: sfw_Latn/*_keep.jsonl - config_name: shu_Arab data_files: - split: train path: shu_Arab/*_keep.jsonl - config_name: ury_Latn data_files: - split: train path: ury_Latn/*_keep.jsonl - config_name: yrl_Latn data_files: - split: train path: yrl_Latn/*_keep.jsonl - config_name: arc_Latn data_files: - split: train path: arc_Latn/*_keep.jsonl - config_name: cmo_Khmr data_files: - split: train path: cmo_Khmr/*_keep.jsonl - config_name: xkg_Latn data_files: - split: train path: xkg_Latn/*_keep.jsonl - config_name: iqw_Latn data_files: - split: train path: iqw_Latn/*_keep.jsonl - config_name: kpj_Latn data_files: - split: train path: kpj_Latn/*_keep.jsonl - config_name: nfa_Latn data_files: - split: train path: nfa_Latn/*_keep.jsonl - config_name: nph_Latn data_files: - split: train path: nph_Latn/*_keep.jsonl - config_name: nre_Latn data_files: - split: train path: nre_Latn/*_keep.jsonl - config_name: wsg_Telu data_files: - split: train path: wsg_Telu/*_keep.jsonl - config_name: xsu_Latn data_files: - split: train path: xsu_Latn/*_keep.jsonl - config_name: nut_Latn data_files: - split: train path: nut_Latn/*_keep.jsonl - config_name: bmk_Latn data_files: - split: train path: bmk_Latn/*_keep.jsonl - config_name: cdf_Latn data_files: - split: train path: cdf_Latn/*_keep.jsonl - config_name: enl_Latn data_files: - split: train path: enl_Latn/*_keep.jsonl - config_name: fuv_Arab data_files: - split: train path: fuv_Arab/*_keep.jsonl - config_name: hop_Latn data_files: - split: train path: hop_Latn/*_keep.jsonl - config_name: ife_Latn data_files: - split: train path: ife_Latn/*_keep.jsonl - config_name: lnd_Latn data_files: - split: train path: lnd_Latn/*_keep.jsonl - config_name: mrv_Latn data_files: - split: train path: mrv_Latn/*_keep.jsonl - config_name: nzm_Latn data_files: - split: train path: nzm_Latn/*_keep.jsonl - config_name: tem_Latn data_files: - split: train path: tem_Latn/*_keep.jsonl - config_name: dwr_Ethi data_files: - split: train path: dwr_Ethi/*_keep.jsonl - config_name: gwc_Arab data_files: - split: train path: gwc_Arab/*_keep.jsonl - config_name: udi_Cyrl data_files: - split: train path: udi_Cyrl/*_keep.jsonl - config_name: zha_Hani data_files: - split: train path: zha_Hani/*_keep.jsonl - config_name: ajz_Latn data_files: - split: train path: ajz_Latn/*_keep.jsonl - config_name: kru_Deva data_files: - split: train path: kru_Deva/*_keep.jsonl - config_name: qvo_Latn data_files: - split: train path: qvo_Latn/*_keep.jsonl - config_name: sld_Latn data_files: - split: train path: sld_Latn/*_keep.jsonl - config_name: yup_Latn data_files: - split: train path: yup_Latn/*_keep.jsonl - config_name: anp_Latn data_files: - split: train path: anp_Latn/*_keep.jsonl - config_name: kwu_Latn data_files: - split: train path: kwu_Latn/*_keep.jsonl - config_name: szy_Hani data_files: - split: train path: szy_Hani/*_keep.jsonl - config_name: tdg_Deva data_files: - split: train path: tdg_Deva/*_keep.jsonl - config_name: ade_Latn data_files: - split: train path: ade_Latn/*_keep.jsonl - config_name: bcw_Latn data_files: - split: train path: bcw_Latn/*_keep.jsonl - config_name: bgz_Latn data_files: - split: train path: bgz_Latn/*_keep.jsonl - config_name: cul_Latn data_files: - split: train path: cul_Latn/*_keep.jsonl - config_name: fmu_Deva data_files: - split: train path: fmu_Deva/*_keep.jsonl - config_name: gvc_Latn data_files: - split: train path: gvc_Latn/*_keep.jsonl - config_name: ktj_Latn data_files: - split: train path: ktj_Latn/*_keep.jsonl - config_name: lis_Lisu data_files: - split: train path: lis_Lisu/*_keep.jsonl - config_name: mev_Latn data_files: - split: train path: mev_Latn/*_keep.jsonl - config_name: msc_Latn data_files: - split: train path: msc_Latn/*_keep.jsonl - config_name: mta_Latn data_files: - split: train path: mta_Latn/*_keep.jsonl - config_name: nbc_Latn data_files: - split: train path: nbc_Latn/*_keep.jsonl - config_name: ncq_Laoo data_files: - split: train path: ncq_Laoo/*_keep.jsonl - config_name: way_Latn data_files: - split: train path: way_Latn/*_keep.jsonl - config_name: wlv_Latn data_files: - split: train path: wlv_Latn/*_keep.jsonl - config_name: wlx_Latn data_files: - split: train path: wlx_Latn/*_keep.jsonl - config_name: jmx_Latn data_files: - split: train path: jmx_Latn/*_keep.jsonl - config_name: pli_Latn data_files: - split: train path: pli_Latn/*_keep.jsonl - config_name: tkr_Cyrl data_files: - split: train path: tkr_Cyrl/*_keep.jsonl - config_name: wni_Latn data_files: - split: train path: wni_Latn/*_keep.jsonl - config_name: aim_Latn data_files: - split: train path: aim_Latn/*_keep.jsonl - config_name: byv_Latn data_files: - split: train path: byv_Latn/*_keep.jsonl - config_name: cgg_Latn data_files: - split: train path: cgg_Latn/*_keep.jsonl - config_name: fud_Latn data_files: - split: train path: fud_Latn/*_keep.jsonl - config_name: jmc_Latn data_files: - split: train path: jmc_Latn/*_keep.jsonl - config_name: ker_Latn data_files: - split: train path: ker_Latn/*_keep.jsonl - config_name: kia_Latn data_files: - split: train path: kia_Latn/*_keep.jsonl - config_name: lom_Latn data_files: - split: train path: lom_Latn/*_keep.jsonl - config_name: ojb_Latn data_files: - split: train path: ojb_Latn/*_keep.jsonl - config_name: tnc_Latn data_files: - split: train path: tnc_Latn/*_keep.jsonl - config_name: bwx_Latn data_files: - split: train path: bwx_Latn/*_keep.jsonl - config_name: cre_Cans data_files: - split: train path: cre_Cans/*_keep.jsonl - config_name: skr_Latn data_files: - split: train path: skr_Latn/*_keep.jsonl - config_name: xal_Latn data_files: - split: train path: xal_Latn/*_keep.jsonl - config_name: gjn_Latn data_files: - split: train path: gjn_Latn/*_keep.jsonl - config_name: guq_Latn data_files: - split: train path: guq_Latn/*_keep.jsonl - config_name: lap_Latn data_files: - split: train path: lap_Latn/*_keep.jsonl - config_name: mip_Latn data_files: - split: train path: mip_Latn/*_keep.jsonl - config_name: mkl_Latn data_files: - split: train path: mkl_Latn/*_keep.jsonl - config_name: mtg_Latn data_files: - split: train path: mtg_Latn/*_keep.jsonl - config_name: mzk_Latn data_files: - split: train path: mzk_Latn/*_keep.jsonl - config_name: njm_Latn data_files: - split: train path: njm_Latn/*_keep.jsonl - config_name: nse_Latn data_files: - split: train path: nse_Latn/*_keep.jsonl - config_name: pne_Latn data_files: - split: train path: pne_Latn/*_keep.jsonl - config_name: pww_Thai data_files: - split: train path: pww_Thai/*_keep.jsonl - config_name: zam_Latn data_files: - split: train path: zam_Latn/*_keep.jsonl - config_name: zca_Latn data_files: - split: train path: zca_Latn/*_keep.jsonl - config_name: the_Deva data_files: - split: train path: the_Deva/*_keep.jsonl - config_name: dis_Latn data_files: - split: train path: dis_Latn/*_keep.jsonl - config_name: lgl_Latn data_files: - split: train path: lgl_Latn/*_keep.jsonl - config_name: mbd_Latn data_files: - split: train path: mbd_Latn/*_keep.jsonl - config_name: mda_Latn data_files: - split: train path: mda_Latn/*_keep.jsonl - config_name: mhy_Latn data_files: - split: train path: mhy_Latn/*_keep.jsonl - config_name: nst_Latn data_files: - split: train path: nst_Latn/*_keep.jsonl - config_name: pkb_Latn data_files: - split: train path: pkb_Latn/*_keep.jsonl - config_name: pmq_Latn data_files: - split: train path: pmq_Latn/*_keep.jsonl - config_name: poe_Latn data_files: - split: train path: poe_Latn/*_keep.jsonl - config_name: swk_Latn data_files: - split: train path: swk_Latn/*_keep.jsonl - config_name: trs_Latn data_files: - split: train path: trs_Latn/*_keep.jsonl - config_name: vot_Latn data_files: - split: train path: vot_Latn/*_keep.jsonl - config_name: yam_Latn data_files: - split: train path: yam_Latn/*_keep.jsonl - config_name: zpq_Latn data_files: - split: train path: zpq_Latn/*_keep.jsonl - config_name: atq_Latn data_files: - split: train path: atq_Latn/*_keep.jsonl - config_name: bav_Latn data_files: - split: train path: bav_Latn/*_keep.jsonl - config_name: cho_Latn data_files: - split: train path: cho_Latn/*_keep.jsonl - config_name: chq_Latn data_files: - split: train path: chq_Latn/*_keep.jsonl - config_name: cjo_Latn data_files: - split: train path: cjo_Latn/*_keep.jsonl - config_name: due_Latn data_files: - split: train path: due_Latn/*_keep.jsonl - config_name: duo_Latn data_files: - split: train path: duo_Latn/*_keep.jsonl - config_name: etu_Latn data_files: - split: train path: etu_Latn/*_keep.jsonl - config_name: ifa_Latn data_files: - split: train path: ifa_Latn/*_keep.jsonl - config_name: kby_Latn data_files: - split: train path: kby_Latn/*_keep.jsonl - config_name: khq_Latn data_files: - split: train path: khq_Latn/*_keep.jsonl - config_name: kno_Latn data_files: - split: train path: kno_Latn/*_keep.jsonl - config_name: lem_Latn data_files: - split: train path: lem_Latn/*_keep.jsonl - config_name: mcn_Latn data_files: - split: train path: mcn_Latn/*_keep.jsonl - config_name: mkz_Latn data_files: - split: train path: mkz_Latn/*_keep.jsonl - config_name: mnb_Latn data_files: - split: train path: mnb_Latn/*_keep.jsonl - config_name: myb_Latn data_files: - split: train path: myb_Latn/*_keep.jsonl - config_name: niy_Latn data_files: - split: train path: niy_Latn/*_keep.jsonl - config_name: pxm_Latn data_files: - split: train path: pxm_Latn/*_keep.jsonl - config_name: rub_Latn data_files: - split: train path: rub_Latn/*_keep.jsonl - config_name: sju_Latn data_files: - split: train path: sju_Latn/*_keep.jsonl - config_name: tmc_Latn data_files: - split: train path: tmc_Latn/*_keep.jsonl - config_name: tpm_Latn data_files: - split: train path: tpm_Latn/*_keep.jsonl - config_name: wob_Latn data_files: - split: train path: wob_Latn/*_keep.jsonl - config_name: nxa_Latn data_files: - split: train path: nxa_Latn/*_keep.jsonl - config_name: pnt_Latn data_files: - split: train path: pnt_Latn/*_keep.jsonl - config_name: sdk_Latn data_files: - split: train path: sdk_Latn/*_keep.jsonl - config_name: tcy_Latn data_files: - split: train path: tcy_Latn/*_keep.jsonl - config_name: bhz_Latn data_files: - split: train path: bhz_Latn/*_keep.jsonl - config_name: bth_Latn data_files: - split: train path: bth_Latn/*_keep.jsonl - config_name: crj_Cans data_files: - split: train path: crj_Cans/*_keep.jsonl - config_name: eka_Latn data_files: - split: train path: eka_Latn/*_keep.jsonl - config_name: khy_Latn data_files: - split: train path: khy_Latn/*_keep.jsonl - config_name: led_Latn data_files: - split: train path: led_Latn/*_keep.jsonl - config_name: met_Latn data_files: - split: train path: met_Latn/*_keep.jsonl - config_name: mgo_Latn data_files: - split: train path: mgo_Latn/*_keep.jsonl - config_name: mus_Latn data_files: - split: train path: mus_Latn/*_keep.jsonl - config_name: mzm_Latn data_files: - split: train path: mzm_Latn/*_keep.jsonl - config_name: naw_Latn data_files: - split: train path: naw_Latn/*_keep.jsonl - config_name: sri_Latn data_files: - split: train path: sri_Latn/*_keep.jsonl - config_name: tbk_Latn data_files: - split: train path: tbk_Latn/*_keep.jsonl - config_name: xbr_Latn data_files: - split: train path: xbr_Latn/*_keep.jsonl - config_name: awa_Latn data_files: - split: train path: awa_Latn/*_keep.jsonl - config_name: tnl_Latn data_files: - split: train path: tnl_Latn/*_keep.jsonl - config_name: tnn_Latn data_files: - split: train path: tnn_Latn/*_keep.jsonl - config_name: awb_Latn data_files: - split: train path: awb_Latn/*_keep.jsonl - config_name: bsc_Latn data_files: - split: train path: bsc_Latn/*_keep.jsonl - config_name: iri_Latn data_files: - split: train path: iri_Latn/*_keep.jsonl - config_name: kpq_Latn data_files: - split: train path: kpq_Latn/*_keep.jsonl - config_name: ktm_Latn data_files: - split: train path: ktm_Latn/*_keep.jsonl - config_name: loe_Latn data_files: - split: train path: loe_Latn/*_keep.jsonl - config_name: lun_Latn data_files: - split: train path: lun_Latn/*_keep.jsonl - config_name: mfg_Latn data_files: - split: train path: mfg_Latn/*_keep.jsonl - config_name: moa_Latn data_files: - split: train path: moa_Latn/*_keep.jsonl - config_name: ndj_Latn data_files: - split: train path: ndj_Latn/*_keep.jsonl - config_name: otn_Latn data_files: - split: train path: otn_Latn/*_keep.jsonl - config_name: sjo_Mong data_files: - split: train path: sjo_Mong/*_keep.jsonl - config_name: wat_Latn data_files: - split: train path: wat_Latn/*_keep.jsonl - config_name: bkm_Latn data_files: - split: train path: bkm_Latn/*_keep.jsonl - config_name: tuz_Latn data_files: - split: train path: tuz_Latn/*_keep.jsonl - config_name: ybi_Deva data_files: - split: train path: ybi_Deva/*_keep.jsonl - config_name: gbr_Latn data_files: - split: train path: gbr_Latn/*_keep.jsonl - config_name: gej_Latn data_files: - split: train path: gej_Latn/*_keep.jsonl - config_name: giz_Latn data_files: - split: train path: giz_Latn/*_keep.jsonl - config_name: gna_Latn data_files: - split: train path: gna_Latn/*_keep.jsonl - config_name: gwr_Latn data_files: - split: train path: gwr_Latn/*_keep.jsonl - config_name: kex_Deva data_files: - split: train path: kex_Deva/*_keep.jsonl - config_name: kle_Deva data_files: - split: train path: kle_Deva/*_keep.jsonl - config_name: knx_Latn data_files: - split: train path: knx_Latn/*_keep.jsonl - config_name: ktz_Latn data_files: - split: train path: ktz_Latn/*_keep.jsonl - config_name: kvj_Latn data_files: - split: train path: kvj_Latn/*_keep.jsonl - config_name: lob_Latn data_files: - split: train path: lob_Latn/*_keep.jsonl - config_name: mhw_Latn data_files: - split: train path: mhw_Latn/*_keep.jsonl - config_name: mph_Latn data_files: - split: train path: mph_Latn/*_keep.jsonl - config_name: pmx_Latn data_files: - split: train path: pmx_Latn/*_keep.jsonl - config_name: qus_Latn data_files: - split: train path: qus_Latn/*_keep.jsonl - config_name: tcc_Latn data_files: - split: train path: tcc_Latn/*_keep.jsonl - config_name: ang_Runr data_files: - split: train path: ang_Runr/*_keep.jsonl - config_name: azn_Latn data_files: - split: train path: azn_Latn/*_keep.jsonl - config_name: bra_Deva data_files: - split: train path: bra_Deva/*_keep.jsonl - config_name: gal_Latn data_files: - split: train path: gal_Latn/*_keep.jsonl - config_name: jra_Khmr data_files: - split: train path: jra_Khmr/*_keep.jsonl - config_name: nlv_Latn data_files: - split: train path: nlv_Latn/*_keep.jsonl - config_name: rki_Latn data_files: - split: train path: rki_Latn/*_keep.jsonl - config_name: stk_Latn data_files: - split: train path: stk_Latn/*_keep.jsonl - config_name: yux_Cyrl data_files: - split: train path: yux_Cyrl/*_keep.jsonl - config_name: dip_Latn data_files: - split: train path: dip_Latn/*_keep.jsonl - config_name: ivv_Latn data_files: - split: train path: ivv_Latn/*_keep.jsonl - config_name: kdj_Latn data_files: - split: train path: kdj_Latn/*_keep.jsonl - config_name: kff_Telu data_files: - split: train path: kff_Telu/*_keep.jsonl - config_name: kyu_Kali data_files: - split: train path: kyu_Kali/*_keep.jsonl - config_name: mnf_Latn data_files: - split: train path: mnf_Latn/*_keep.jsonl - config_name: moc_Latn data_files: - split: train path: moc_Latn/*_keep.jsonl - config_name: mpg_Latn data_files: - split: train path: mpg_Latn/*_keep.jsonl - config_name: mua_Latn data_files: - split: train path: mua_Latn/*_keep.jsonl - config_name: nma_Latn data_files: - split: train path: nma_Latn/*_keep.jsonl - config_name: rim_Latn data_files: - split: train path: rim_Latn/*_keep.jsonl - config_name: rjs_Deva data_files: - split: train path: rjs_Deva/*_keep.jsonl - config_name: rnl_Latn data_files: - split: train path: rnl_Latn/*_keep.jsonl - config_name: ssx_Latn data_files: - split: train path: ssx_Latn/*_keep.jsonl - config_name: tig_Ethi data_files: - split: train path: tig_Ethi/*_keep.jsonl - config_name: urw_Latn data_files: - split: train path: urw_Latn/*_keep.jsonl - config_name: idt_Latn data_files: - split: train path: idt_Latn/*_keep.jsonl - config_name: mrn_Latn data_files: - split: train path: mrn_Latn/*_keep.jsonl - config_name: oki_Latn data_files: - split: train path: oki_Latn/*_keep.jsonl - config_name: bbk_Latn data_files: - split: train path: bbk_Latn/*_keep.jsonl - config_name: bea_Latn data_files: - split: train path: bea_Latn/*_keep.jsonl - config_name: bkq_Latn data_files: - split: train path: bkq_Latn/*_keep.jsonl - config_name: btt_Latn data_files: - split: train path: btt_Latn/*_keep.jsonl - config_name: cou_Latn data_files: - split: train path: cou_Latn/*_keep.jsonl - config_name: ddn_Latn data_files: - split: train path: ddn_Latn/*_keep.jsonl - config_name: did_Latn data_files: - split: train path: did_Latn/*_keep.jsonl - config_name: diu_Latn data_files: - split: train path: diu_Latn/*_keep.jsonl - config_name: ilb_Latn data_files: - split: train path: ilb_Latn/*_keep.jsonl - config_name: ksj_Latn data_files: - split: train path: ksj_Latn/*_keep.jsonl - config_name: lhi_Latn data_files: - split: train path: lhi_Latn/*_keep.jsonl - config_name: ndi_Latn data_files: - split: train path: ndi_Latn/*_keep.jsonl - config_name: ngb_Latn data_files: - split: train path: ngb_Latn/*_keep.jsonl - config_name: njz_Latn data_files: - split: train path: njz_Latn/*_keep.jsonl - config_name: nmz_Latn data_files: - split: train path: nmz_Latn/*_keep.jsonl - config_name: npo_Latn data_files: - split: train path: npo_Latn/*_keep.jsonl - config_name: nwb_Latn data_files: - split: train path: nwb_Latn/*_keep.jsonl - config_name: pbi_Latn data_files: - split: train path: pbi_Latn/*_keep.jsonl - config_name: sbd_Latn data_files: - split: train path: sbd_Latn/*_keep.jsonl - config_name: soe_Latn data_files: - split: train path: soe_Latn/*_keep.jsonl - config_name: tpw_Latn data_files: - split: train path: tpw_Latn/*_keep.jsonl - config_name: twb_Latn data_files: - split: train path: twb_Latn/*_keep.jsonl - config_name: bfn_Latn data_files: - split: train path: bfn_Latn/*_keep.jsonl - config_name: boz_Latn data_files: - split: train path: boz_Latn/*_keep.jsonl - config_name: lwl_Thai data_files: - split: train path: lwl_Thai/*_keep.jsonl - config_name: psp_Latn data_files: - split: train path: psp_Latn/*_keep.jsonl - config_name: sgd_Latn data_files: - split: train path: sgd_Latn/*_keep.jsonl - config_name: ymp_Latn data_files: - split: train path: ymp_Latn/*_keep.jsonl - config_name: bkl_Latn data_files: - split: train path: bkl_Latn/*_keep.jsonl - config_name: bkv_Latn data_files: - split: train path: bkv_Latn/*_keep.jsonl - config_name: csk_Latn data_files: - split: train path: csk_Latn/*_keep.jsonl - config_name: enx_Latn data_files: - split: train path: enx_Latn/*_keep.jsonl - config_name: ess_Latn data_files: - split: train path: ess_Latn/*_keep.jsonl - config_name: gof_Ethi data_files: - split: train path: gof_Ethi/*_keep.jsonl - config_name: gud_Latn data_files: - split: train path: gud_Latn/*_keep.jsonl - config_name: guu_Latn data_files: - split: train path: guu_Latn/*_keep.jsonl - config_name: kdh_Latn data_files: - split: train path: kdh_Latn/*_keep.jsonl - config_name: krx_Latn data_files: - split: train path: krx_Latn/*_keep.jsonl - config_name: kzn_Latn data_files: - split: train path: kzn_Latn/*_keep.jsonl - config_name: maf_Latn data_files: - split: train path: maf_Latn/*_keep.jsonl - config_name: max_Latn data_files: - split: train path: max_Latn/*_keep.jsonl - config_name: mgc_Latn data_files: - split: train path: mgc_Latn/*_keep.jsonl - config_name: mvn_Latn data_files: - split: train path: mvn_Latn/*_keep.jsonl - config_name: mwm_Latn data_files: - split: train path: mwm_Latn/*_keep.jsonl - config_name: nbe_Latn data_files: - split: train path: nbe_Latn/*_keep.jsonl - config_name: nlc_Latn data_files: - split: train path: nlc_Latn/*_keep.jsonl - config_name: nxd_Latn data_files: - split: train path: nxd_Latn/*_keep.jsonl - config_name: pse_Latn data_files: - split: train path: pse_Latn/*_keep.jsonl - config_name: sbs_Latn data_files: - split: train path: sbs_Latn/*_keep.jsonl - config_name: tlb_Latn data_files: - split: train path: tlb_Latn/*_keep.jsonl - config_name: wwa_Latn data_files: - split: train path: wwa_Latn/*_keep.jsonl - config_name: yim_Latn data_files: - split: train path: yim_Latn/*_keep.jsonl - config_name: bkx_Latn data_files: - split: train path: bkx_Latn/*_keep.jsonl - config_name: bze_Latn data_files: - split: train path: bze_Latn/*_keep.jsonl - config_name: chu_Latn data_files: - split: train path: chu_Latn/*_keep.jsonl - config_name: kbq_Latn data_files: - split: train path: kbq_Latn/*_keep.jsonl - config_name: kpy_Armn data_files: - split: train path: kpy_Armn/*_keep.jsonl - config_name: mry_Latn data_files: - split: train path: mry_Latn/*_keep.jsonl - config_name: tkd_Latn data_files: - split: train path: tkd_Latn/*_keep.jsonl - config_name: ybh_Deva data_files: - split: train path: ybh_Deva/*_keep.jsonl - config_name: abi_Latn data_files: - split: train path: abi_Latn/*_keep.jsonl - config_name: ayo_Latn data_files: - split: train path: ayo_Latn/*_keep.jsonl - config_name: bex_Latn data_files: - split: train path: bex_Latn/*_keep.jsonl - config_name: bom_Latn data_files: - split: train path: bom_Latn/*_keep.jsonl - config_name: bov_Latn data_files: - split: train path: bov_Latn/*_keep.jsonl - config_name: dhg_Latn data_files: - split: train path: dhg_Latn/*_keep.jsonl - config_name: jaa_Latn data_files: - split: train path: jaa_Latn/*_keep.jsonl - config_name: kqs_Latn data_files: - split: train path: kqs_Latn/*_keep.jsonl - config_name: kqy_Ethi data_files: - split: train path: kqy_Ethi/*_keep.jsonl - config_name: ksb_Latn data_files: - split: train path: ksb_Latn/*_keep.jsonl - config_name: ldn_Latn data_files: - split: train path: ldn_Latn/*_keep.jsonl - config_name: log_Latn data_files: - split: train path: log_Latn/*_keep.jsonl - config_name: lol_Latn data_files: - split: train path: lol_Latn/*_keep.jsonl - config_name: mfk_Latn data_files: - split: train path: mfk_Latn/*_keep.jsonl - config_name: mtj_Latn data_files: - split: train path: mtj_Latn/*_keep.jsonl - config_name: ndz_Latn data_files: - split: train path: ndz_Latn/*_keep.jsonl - config_name: ngc_Latn data_files: - split: train path: ngc_Latn/*_keep.jsonl - config_name: nio_Cyrl data_files: - split: train path: nio_Cyrl/*_keep.jsonl - config_name: nri_Latn data_files: - split: train path: nri_Latn/*_keep.jsonl - config_name: nuz_Latn data_files: - split: train path: nuz_Latn/*_keep.jsonl - config_name: pps_Latn data_files: - split: train path: pps_Latn/*_keep.jsonl - config_name: shk_Latn data_files: - split: train path: shk_Latn/*_keep.jsonl - config_name: tlj_Latn data_files: - split: train path: tlj_Latn/*_keep.jsonl - config_name: vut_Latn data_files: - split: train path: vut_Latn/*_keep.jsonl - config_name: xnn_Latn data_files: - split: train path: xnn_Latn/*_keep.jsonl - config_name: bcc_Arab data_files: - split: train path: bcc_Arab/*_keep.jsonl - config_name: cuh_Latn data_files: - split: train path: cuh_Latn/*_keep.jsonl - config_name: emp_Cyrl data_files: - split: train path: emp_Cyrl/*_keep.jsonl - config_name: hro_Latn data_files: - split: train path: hro_Latn/*_keep.jsonl - config_name: nco_Latn data_files: - split: train path: nco_Latn/*_keep.jsonl - config_name: nxl_Latn data_files: - split: train path: nxl_Latn/*_keep.jsonl - config_name: pex_Latn data_files: - split: train path: pex_Latn/*_keep.jsonl - config_name: tnp_Latn data_files: - split: train path: tnp_Latn/*_keep.jsonl - config_name: trv_Hani data_files: - split: train path: trv_Hani/*_keep.jsonl - config_name: vif_Latn data_files: - split: train path: vif_Latn/*_keep.jsonl - config_name: aha_Latn data_files: - split: train path: aha_Latn/*_keep.jsonl - config_name: avn_Latn data_files: - split: train path: avn_Latn/*_keep.jsonl - config_name: bmv_Latn data_files: - split: train path: bmv_Latn/*_keep.jsonl - config_name: chj_Latn data_files: - split: train path: chj_Latn/*_keep.jsonl - config_name: crt_Latn data_files: - split: train path: crt_Latn/*_keep.jsonl - config_name: dos_Latn data_files: - split: train path: dos_Latn/*_keep.jsonl - config_name: eto_Latn data_files: - split: train path: eto_Latn/*_keep.jsonl - config_name: gya_Latn data_files: - split: train path: gya_Latn/*_keep.jsonl - config_name: hlt_Latn data_files: - split: train path: hlt_Latn/*_keep.jsonl - config_name: jun_Orya data_files: - split: train path: jun_Orya/*_keep.jsonl - config_name: ktb_Ethi data_files: - split: train path: ktb_Ethi/*_keep.jsonl - config_name: lgm_Latn data_files: - split: train path: lgm_Latn/*_keep.jsonl - config_name: mnx_Latn data_files: - split: train path: mnx_Latn/*_keep.jsonl - config_name: mor_Latn data_files: - split: train path: mor_Latn/*_keep.jsonl - config_name: mug_Latn data_files: - split: train path: mug_Latn/*_keep.jsonl - config_name: nng_Latn data_files: - split: train path: nng_Latn/*_keep.jsonl - config_name: nnh_Latn data_files: - split: train path: nnh_Latn/*_keep.jsonl - config_name: nsa_Latn data_files: - split: train path: nsa_Latn/*_keep.jsonl - config_name: nyk_Latn data_files: - split: train path: nyk_Latn/*_keep.jsonl - config_name: otd_Latn data_files: - split: train path: otd_Latn/*_keep.jsonl - config_name: oym_Latn data_files: - split: train path: oym_Latn/*_keep.jsonl - config_name: phm_Latn data_files: - split: train path: phm_Latn/*_keep.jsonl - config_name: prq_Latn data_files: - split: train path: prq_Latn/*_keep.jsonl - config_name: saj_Latn data_files: - split: train path: saj_Latn/*_keep.jsonl - config_name: szb_Latn data_files: - split: train path: szb_Latn/*_keep.jsonl - config_name: tap_Latn data_files: - split: train path: tap_Latn/*_keep.jsonl - config_name: tnr_Latn data_files: - split: train path: tnr_Latn/*_keep.jsonl - config_name: tzl_Latn data_files: - split: train path: tzl_Latn/*_keep.jsonl - config_name: wew_Latn data_files: - split: train path: wew_Latn/*_keep.jsonl - config_name: yas_Latn data_files: - split: train path: yas_Latn/*_keep.jsonl - config_name: yuz_Latn data_files: - split: train path: yuz_Latn/*_keep.jsonl - config_name: avk_Cyrl data_files: - split: train path: avk_Cyrl/*_keep.jsonl - config_name: bat_Cyrl data_files: - split: train path: bat_Cyrl/*_keep.jsonl - config_name: baw_Latn data_files: - split: train path: baw_Latn/*_keep.jsonl - config_name: bwt_Latn data_files: - split: train path: bwt_Latn/*_keep.jsonl - config_name: bxa_Latn data_files: - split: train path: bxa_Latn/*_keep.jsonl - config_name: cjs_Latn data_files: - split: train path: cjs_Latn/*_keep.jsonl - config_name: csw_Cans data_files: - split: train path: csw_Cans/*_keep.jsonl - config_name: fli_Latn data_files: - split: train path: fli_Latn/*_keep.jsonl - config_name: kqr_Latn data_files: - split: train path: kqr_Latn/*_keep.jsonl - config_name: lan_Latn data_files: - split: train path: lan_Latn/*_keep.jsonl - config_name: lgr_Latn data_files: - split: train path: lgr_Latn/*_keep.jsonl - config_name: lhm_Deva data_files: - split: train path: lhm_Deva/*_keep.jsonl - config_name: mlk_Latn data_files: - split: train path: mlk_Latn/*_keep.jsonl - config_name: mve_Arab data_files: - split: train path: mve_Arab/*_keep.jsonl - config_name: nqo_Latn data_files: - split: train path: nqo_Latn/*_keep.jsonl - config_name: odk_Arab data_files: - split: train path: odk_Arab/*_keep.jsonl - config_name: olo_Cyrl data_files: - split: train path: olo_Cyrl/*_keep.jsonl - config_name: otq_Arab data_files: - split: train path: otq_Arab/*_keep.jsonl - config_name: pea_Latn data_files: - split: train path: pea_Latn/*_keep.jsonl - config_name: saq_Latn data_files: - split: train path: saq_Latn/*_keep.jsonl - config_name: sea_Latn data_files: - split: train path: sea_Latn/*_keep.jsonl - config_name: sps_Latn data_files: - split: train path: sps_Latn/*_keep.jsonl - config_name: tio_Latn data_files: - split: train path: tio_Latn/*_keep.jsonl - config_name: yin_Latn data_files: - split: train path: yin_Latn/*_keep.jsonl - config_name: adl_Latn data_files: - split: train path: adl_Latn/*_keep.jsonl - config_name: agw_Latn data_files: - split: train path: agw_Latn/*_keep.jsonl - config_name: ald_Latn data_files: - split: train path: ald_Latn/*_keep.jsonl - config_name: asg_Latn data_files: - split: train path: asg_Latn/*_keep.jsonl - config_name: bnj_Latn data_files: - split: train path: bnj_Latn/*_keep.jsonl - config_name: brx_Latn data_files: - split: train path: brx_Latn/*_keep.jsonl - config_name: bwi_Latn data_files: - split: train path: bwi_Latn/*_keep.jsonl - config_name: bwq_Latn data_files: - split: train path: bwq_Latn/*_keep.jsonl - config_name: dbq_Latn data_files: - split: train path: dbq_Latn/*_keep.jsonl - config_name: dow_Latn data_files: - split: train path: dow_Latn/*_keep.jsonl - config_name: fad_Latn data_files: - split: train path: fad_Latn/*_keep.jsonl - config_name: hoc_Latn data_files: - split: train path: hoc_Latn/*_keep.jsonl - config_name: kql_Latn data_files: - split: train path: kql_Latn/*_keep.jsonl - config_name: ksf_Latn data_files: - split: train path: ksf_Latn/*_keep.jsonl - config_name: kyu_Mymr data_files: - split: train path: kyu_Mymr/*_keep.jsonl - config_name: lea_Latn data_files: - split: train path: lea_Latn/*_keep.jsonl - config_name: mse_Latn data_files: - split: train path: mse_Latn/*_keep.jsonl - config_name: mzl_Latn data_files: - split: train path: mzl_Latn/*_keep.jsonl - config_name: nhd_Latn data_files: - split: train path: nhd_Latn/*_keep.jsonl - config_name: nnl_Latn data_files: - split: train path: nnl_Latn/*_keep.jsonl - config_name: nwx_Deva data_files: - split: train path: nwx_Deva/*_keep.jsonl - config_name: pbc_Latn data_files: - split: train path: pbc_Latn/*_keep.jsonl - config_name: pny_Latn data_files: - split: train path: pny_Latn/*_keep.jsonl - config_name: tul_Latn data_files: - split: train path: tul_Latn/*_keep.jsonl - config_name: aaa_Latn data_files: - split: train path: aaa_Latn/*_keep.jsonl - config_name: abc_Latn data_files: - split: train path: abc_Latn/*_keep.jsonl - config_name: agq_Latn data_files: - split: train path: agq_Latn/*_keep.jsonl - config_name: ags_Latn data_files: - split: train path: ags_Latn/*_keep.jsonl - config_name: azo_Latn data_files: - split: train path: azo_Latn/*_keep.jsonl - config_name: bag_Latn data_files: - split: train path: bag_Latn/*_keep.jsonl - config_name: bax_Latn data_files: - split: train path: bax_Latn/*_keep.jsonl - config_name: bce_Latn data_files: - split: train path: bce_Latn/*_keep.jsonl - config_name: bfm_Latn data_files: - split: train path: bfm_Latn/*_keep.jsonl - config_name: bgf_Latn data_files: - split: train path: bgf_Latn/*_keep.jsonl - config_name: bhs_Latn data_files: - split: train path: bhs_Latn/*_keep.jsonl - config_name: bkc_Latn data_files: - split: train path: bkc_Latn/*_keep.jsonl - config_name: bkh_Latn data_files: - split: train path: bkh_Latn/*_keep.jsonl - config_name: blk_Latn data_files: - split: train path: blk_Latn/*_keep.jsonl - config_name: bob_Latn data_files: - split: train path: bob_Latn/*_keep.jsonl - config_name: bqm_Latn data_files: - split: train path: bqm_Latn/*_keep.jsonl - config_name: bri_Latn data_files: - split: train path: bri_Latn/*_keep.jsonl - config_name: brv_Laoo data_files: - split: train path: brv_Laoo/*_keep.jsonl - config_name: buo_Latn data_files: - split: train path: buo_Latn/*_keep.jsonl - config_name: bya_Latn data_files: - split: train path: bya_Latn/*_keep.jsonl - config_name: chp_Cans data_files: - split: train path: chp_Cans/*_keep.jsonl - config_name: cim_Latn data_files: - split: train path: cim_Latn/*_keep.jsonl - config_name: clo_Latn data_files: - split: train path: clo_Latn/*_keep.jsonl - config_name: cuv_Latn data_files: - split: train path: cuv_Latn/*_keep.jsonl - config_name: dag_Arab data_files: - split: train path: dag_Arab/*_keep.jsonl - config_name: dmg_Latn data_files: - split: train path: dmg_Latn/*_keep.jsonl - config_name: dtr_Latn data_files: - split: train path: dtr_Latn/*_keep.jsonl - config_name: eee_Thai data_files: - split: train path: eee_Thai/*_keep.jsonl - config_name: ekm_Latn data_files: - split: train path: ekm_Latn/*_keep.jsonl - config_name: enc_Latn data_files: - split: train path: enc_Latn/*_keep.jsonl - config_name: fiu_Cyrl data_files: - split: train path: fiu_Cyrl/*_keep.jsonl - config_name: gbj_Orya data_files: - split: train path: gbj_Orya/*_keep.jsonl - config_name: gou_Latn data_files: - split: train path: gou_Latn/*_keep.jsonl - config_name: hao_Latn data_files: - split: train path: hao_Latn/*_keep.jsonl - config_name: hna_Latn data_files: - split: train path: hna_Latn/*_keep.jsonl - config_name: hre_Latn data_files: - split: train path: hre_Latn/*_keep.jsonl - config_name: isu_Latn data_files: - split: train path: isu_Latn/*_keep.jsonl - config_name: jgo_Latn data_files: - split: train path: jgo_Latn/*_keep.jsonl - config_name: kbx_Latn data_files: - split: train path: kbx_Latn/*_keep.jsonl - config_name: kvt_Latn data_files: - split: train path: kvt_Latn/*_keep.jsonl - config_name: kxp_Arab data_files: - split: train path: kxp_Arab/*_keep.jsonl - config_name: lfa_Latn data_files: - split: train path: lfa_Latn/*_keep.jsonl - config_name: lkb_Latn data_files: - split: train path: lkb_Latn/*_keep.jsonl - config_name: lns_Latn data_files: - split: train path: lns_Latn/*_keep.jsonl - config_name: loh_Latn data_files: - split: train path: loh_Latn/*_keep.jsonl - config_name: lrc_Latn data_files: - split: train path: lrc_Latn/*_keep.jsonl - config_name: lts_Latn data_files: - split: train path: lts_Latn/*_keep.jsonl - config_name: luy_Latn data_files: - split: train path: luy_Latn/*_keep.jsonl - config_name: mdr_Latn data_files: - split: train path: mdr_Latn/*_keep.jsonl - config_name: mfj_Latn data_files: - split: train path: mfj_Latn/*_keep.jsonl - config_name: mgg_Latn data_files: - split: train path: mgg_Latn/*_keep.jsonl - config_name: mgq_Latn data_files: - split: train path: mgq_Latn/*_keep.jsonl - config_name: mle_Latn data_files: - split: train path: mle_Latn/*_keep.jsonl - config_name: mlw_Latn data_files: - split: train path: mlw_Latn/*_keep.jsonl - config_name: mmu_Latn data_files: - split: train path: mmu_Latn/*_keep.jsonl - config_name: mne_Latn data_files: - split: train path: mne_Latn/*_keep.jsonl - config_name: muv_Mlym data_files: - split: train path: muv_Mlym/*_keep.jsonl - config_name: mxu_Latn data_files: - split: train path: mxu_Latn/*_keep.jsonl - config_name: nge_Latn data_files: - split: train path: nge_Latn/*_keep.jsonl - config_name: ngn_Latn data_files: - split: train path: ngn_Latn/*_keep.jsonl - config_name: njy_Latn data_files: - split: train path: njy_Latn/*_keep.jsonl - config_name: nsk_Cans data_files: - split: train path: nsk_Cans/*_keep.jsonl - config_name: nsk_Latn data_files: - split: train path: nsk_Latn/*_keep.jsonl - config_name: nwe_Latn data_files: - split: train path: nwe_Latn/*_keep.jsonl - config_name: nza_Latn data_files: - split: train path: nza_Latn/*_keep.jsonl - config_name: pcg_Mlym data_files: - split: train path: pcg_Mlym/*_keep.jsonl - config_name: pdu_Latn data_files: - split: train path: pdu_Latn/*_keep.jsonl - config_name: pnz_Latn data_files: - split: train path: pnz_Latn/*_keep.jsonl - config_name: pwg_Latn data_files: - split: train path: pwg_Latn/*_keep.jsonl - config_name: qaa_Latn data_files: - split: train path: qaa_Latn/*_keep.jsonl - config_name: shi_Arab data_files: - split: train path: shi_Arab/*_keep.jsonl - config_name: smn_Cyrl data_files: - split: train path: smn_Cyrl/*_keep.jsonl - config_name: snk_Arab data_files: - split: train path: snk_Arab/*_keep.jsonl - config_name: snl_Latn data_files: - split: train path: snl_Latn/*_keep.jsonl - config_name: sox_Latn data_files: - split: train path: sox_Latn/*_keep.jsonl - config_name: ssn_Latn data_files: - split: train path: ssn_Latn/*_keep.jsonl - config_name: syw_Deva data_files: - split: train path: syw_Deva/*_keep.jsonl - config_name: tay_Hani data_files: - split: train path: tay_Hani/*_keep.jsonl - config_name: tbj_Latn data_files: - split: train path: tbj_Latn/*_keep.jsonl - config_name: tdb_Deva data_files: - split: train path: tdb_Deva/*_keep.jsonl - config_name: thy_Thai data_files: - split: train path: thy_Thai/*_keep.jsonl - config_name: tnt_Latn data_files: - split: train path: tnt_Latn/*_keep.jsonl - config_name: tom_Latn data_files: - split: train path: tom_Latn/*_keep.jsonl - config_name: tpl_Latn data_files: - split: train path: tpl_Latn/*_keep.jsonl - config_name: tsb_Latn data_files: - split: train path: tsb_Latn/*_keep.jsonl - config_name: tvs_Latn data_files: - split: train path: tvs_Latn/*_keep.jsonl - config_name: udg_Mlym data_files: - split: train path: udg_Mlym/*_keep.jsonl - config_name: unr_Orya data_files: - split: train path: unr_Orya/*_keep.jsonl - config_name: wbr_Deva data_files: - split: train path: wbr_Deva/*_keep.jsonl - config_name: wms_Latn data_files: - split: train path: wms_Latn/*_keep.jsonl - config_name: wnk_Latn data_files: - split: train path: wnk_Latn/*_keep.jsonl - config_name: wtk_Latn data_files: - split: train path: wtk_Latn/*_keep.jsonl - config_name: xmd_Latn data_files: - split: train path: xmd_Latn/*_keep.jsonl - config_name: xmg_Latn data_files: - split: train path: xmg_Latn/*_keep.jsonl - config_name: xty_Latn data_files: - split: train path: xty_Latn/*_keep.jsonl - config_name: yav_Latn data_files: - split: train path: yav_Latn/*_keep.jsonl - config_name: yea_Mlym data_files: - split: train path: yea_Mlym/*_keep.jsonl - config_name: zuh_Latn data_files: - split: train path: zuh_Latn/*_keep.jsonl language: - eng - rus - cmn - deu - spa - fra - jpn - ita - por - pol - nld - ind - tur - vie - fas - kor - swe - hun - ell - ukr - tha - aze - arb - slv - cat - fin - hbs - ces - fil - guj - nob - mal - bul - kan - hin - tam - ara - kaz - sqi - pan - est - srp - heb - isl - mlt - mkd - bos - lit - kat - hrv - ben - glg - ceb - ron - msa - hye - tgk - mar - bel - nep - urd - slk - mon - swa - dan - eus - azj - sin - lat - als - tel - afr - ekk - ltz - zsm - som - cym - und - kir - nor - uzb - mya - epo - hau - ary - gle - lvs - nno - und - ory - amh - tat - khm - und - und - und - arz - pus - lav - hbs - und - war - und - mlg - uzn - hat - zul - kur - tgl - uzb - div - und - fry - sna - und - fao - sun - und - und - snd - cos - bak - ckb - asm - und - jpn - ast - ibo - mri - und - ars - und - anp - khk - und - lao - ori - und - yid - hmn - jav - nya - und - und - und - tir - uig - wln - und - yor - swh - gla - xho - uzn - bre - sot - che - nan - tel - bew - glk - smo - orm - zho - haw - pnb - und - oci - chv - kin - und - srp - und - und - pcm - kaa - lug - pbt - aze - gsw - min - tuk - lim - kmr - und - roh - und - und - vec - san - und - gmh - nds - sco - und - und - azb - tsn - rus - und - mon - und - tgk - sah - pri - pap - plt - lmo - bod - und - yue - und - und - und - bar - und - hif - und - arg - pms - kat - nrm - vol - aeb - und - und - ido - pam - kal - und - und - und - lus - und - und - und - new - sme - und - und - nso - lld - ina - scn - und - und - und - hyw - abk - bew - und - und - bpy - tgk - lin - oss - que - szl - sdh - mww - mhr - und - und - und - rmy - diq - srd - und - und - ydd - und - run - und - hsb - und - wol - und - urd - und - kiu - ilo - und - cak - und - acm - und - und - grc - xmf - vls - und - und - tso - crh - nav - mwl - und - tat - grn - mui - hin - mvf - udm - mzn - und - und - und - bcl - und - und - ban - und - afb - und - cnh - und - cbk - kaa - tyv - kur - ful - und - zea - frp - frr - pfl - lij - apc - mai - und - und - pon - bak - eml - uzs - nde - hil - mam - gom - dag - nap - bjn - und - hbo - tet - inh - ven - gaz - tpi - und - ajp - sat - wuu - bam - vep - glv - tzo - bik - und - acr - twi - myv - ewe - poh - und - ile - rue - und - und - ace - kha - und - und - und - und - ksh - hui - gor - krc - lfn - tls - fur - und - gag - chk - hmo - kek - fon - ssw - quh - meu - csb - ksd - ext - rop - und - aln - kab - shn - tbz - knv - bqc - und - und - npi - rug - und - kom - kpg - kbd - dzo - wal - zlm - und - mzh - pan - hac - und - tum - sgs - fij - tuc - und - und - jbo - iba - pcd - kjh - ixl - und - und - und - ndo - suz - und - und - mfe - bxr - ava - und - bho - brh - und - ctd - aym - ady - nbl - gom - hus - gug - und - und - bat - guj - hne - srn - ctu - san - ton - bug - aoj - jam - und - lez - xav - luo - ltg - bih - uig - mrj - avk - kos - kbp - alt - cfm - goh - lad - und - cor - quc - und - kpv - und - aka - und - hak - rmn - pls - sag - min - und - rcf - som - gym - und - kik - pag - arn - und - ngu - dsb - tca - mni - naq - pdc - mps - und - olo - fro - fit - crh - cdo - kac - mdf - mad - cab - srm - toj - zai - guc - smn - ote - nch - und - prs - tuk - bis - cuk - bzj - map - djk - ncj - und - mco - shp - und - ang - seh - ikk - qvi - yom - top - awa - ike - mau - kum - mos - tly - und - ayr - tdt - yap - evn - tah - xal - szy - quw - stq - tlh - otq - und - nov - skr - yua - raw - und - fiu - enm - bbc - und - non - und - ber - kea - uig - bua - kua - zza - csy - mnw - dtp - nah - yao - azj - hau - nog - abt - meo - ame - mbf - chm - arb - kas - tam - crs - tcy - lbe - cni - ben - rom - zom - blk - efi - ada - und - koi - und - bbj - trv - ach - mgh - kas - gcr - krl - iso - lki - atj - cmo - chu - gur - mni - iku - sma - lrc - gos - dty - nzi - btx - brx - bts - nyu - ckt - guw - pck - quy - gcf - ksw - nia - bal - nhe - hwc - mup - dyu - ami - tay - gaa - loz - abq - pnt - snd - ium - abs - cmr - shi - tzm - gil - ewo - kon - sms - ape - tzh - bci - acd - tab - fuv - swc - tcz - und - sgc - aaz - tly - tvl - zha - pib - pli - bru - quz - cha - aeu - bjn - nqo - mer - acf - rmc - msi - smj - kng - dhv - doi - zap - hla - cac - pdt - mbt - taq - bgp - tyz - aau - mah - nak - kmb - mjw - fuf - bba - ron - amu - alz - cop - rar - emp - arc - nyn - rut - lub - tzj - fat - lzh - aak - nnb - lhu - bjv - bum - maq - roa - bvr - kcg - tok - yrk - khs - rup - tuk - qub - jac - hmr - sid - chr - amp - qug - mas - fkv - mal - bem - cav - her - kwn - ify - asm - gpe - zpa - agu - ppk - tiv - npi - ahk - pis - hns - sus - hak - bon - sja - kas - mar - kri - dik - wrk - got - dar - gui - mak - jiv - kjb - maz - bsn - dov - cok - rwo - mag - hnj - krj - teo - enq - aom - ffm - din - ubu - pwn - gum - sny - twu - med - acn - rme - cbs - abx - ndc - ibb - pih - zty - rad - aby - zne - cui - lua - yby - bno - niu - agd - nij - agg - yre - fip - qve - tsg - skg - adj - mfq - agx - kmg - umb - cnk - nhw - maa - kwy - jra - syr - inb - pau - xsm - vap - alp - des - syl - kmr - mkn - gan - qwh - jvn - bib - ckt - mpx - bin - klv - xon - jbu - bdh - knj - ptu - mmn - mwq - sop - cgc - rmy - dop - mgr - rmo - tbc - kyq - zpu - kkc - acu - ata - gag - kan - mif - sus - tsz - mbi - imo - fuh - zyp - apr - att - bus - kqp - koo - bas - urh - bbr - msy - mdy - nas - lee - mcu - mrj - dig - izr - nfr - hto - mzw - qxh - chz - guh - ded - bvz - kkj - gux - ikt - bnp - bgt - quf - dwr - nod - yal - bmr - ljp - xla - zyb - gof - niv - nrf - npy - ura - dob - gfk - gdr - sdc - pjt - spp - irk - cpa - knc - gvl - agr - njo - xrb - kao - zpz - bgr - rnd - ntu - aai - gai - dgi - kez - dad - kmh - aii - ifk - aji - noa - bhl - ztq - rmn - gde - dnj - men - miq - mdf - isd - lgg - ncx - gub - itv - tac - cce - kwi - nhu - snd - sml - pah - sas - cwt - spy - enb - fai - cas - sue - yss - ozm - bku - biv - akp - dts - kdr - fal - mmo - byr - cbv - cbu - eza - kgp - sda - gbo - mfi - bdd - boj - blh - ibg - agn - vmy - sua - gul - rap - bmh - sxn - dyi - tdx - car - heh - dgc - sgb - bpr - gbi - avu - udu - yli - cjs - gnn - nhi - urk - kpy - qup - qvh - pir - tsc - bxr - thk - grt - kqn - hay - mqb - tbg - bss - trn - jae - myy - ipk - bao - qvz - dua - mwv - nct - kyc - tuo - tee - mop - mxt - qvm - gkn - lam - ntr - sbe - dyo - lex - smk - coe - kpe - mni - xbi - atb - tcf - bfo - cbc - swg - niv - cnt - izz - knf - mpp - kki - nho - sat - kbr - bmu - hig - qxr - orv - pma - zpl - kpr - sig - ory - mxp - lef - mlp - soq - cre - aey - qvs - sur - knc - suk - cax - alq - chw - dww - gog - viv - qvw - ted - yuj - ziw - kto - mcq - sim - koi - lue - cya - knk - tpt - taj - blz - mil - kde - kpz - hag - vun - kus - lid - tos - bfd - chd - cle - yon - bgs - lsm - msb - ota - guj - mxq - lfn - myw - aso - esu - kdi - bhp - mfz - sgw - xsi - gun - kxc - mux - otw - cjp - nyo - zia - cco - ojb - ktu - cpy - txu - bjr - bud - ken - bzh - ipi - crx - cko - gnd - kix - mmx - esk - kud - toh - byx - mtp - kpv - kyz - lmp - tiy - ccp - sbl - tcs - usp - buk - hvn - amm - kzj - gng - ter - kij - ngl - pab - pad - kca - meq - kas - niq - dah - lia - kss - bch - khz - nop - yle - ain - pem - yml - taq - gaw - kze - tby - cso - ifb - rki - aly - kmu - mhl - kmo - mva - opm - rtm - xtn - dgr - ksr - snp - tvk - zpv - hrx - kwj - muh - kak - kpw - akb - ybb - kjs - wuv - laj - ogo - nim - czt - aoz - rro - yuw - heg - mrw - zac - kpf - leu - nwi - awx - gso - keo - tte - ceg - nhy - bhw - gnw - nin - xmv - crm - mxv - ubr - gld - bzi - ann - lac - mej - zpc - dje - caq - guk - wos - msm - zaw - cap - mox - zab - mca - wrs - hae - nss - rav - uvh - dga - ppo - nog - arq - bim - lwg - kms - kne - tkl - ady - caf - zat - aba - chf - lew - npl - uvl - cdo - rmq - mek - snc - nsn - amn - abz - cek - ong - txq - bjp - gvf - crk - snf - nhg - kqw - mic - mie - nus - plu - idu - sll - thl - bef - kwd - mqy - caa - aia - wer - agm - mbl - mbs - adh - hub - mav - nab - bqp - guo - auy - big - nii - row - kim - mti - otm - prg - icr - mcp - tbo - chy - oji - kpx - lbb - pao - xuo - zgh - wls - aca - apy - wnc - ssg - eve - mee - yut - chr - kxm - mur - lad - mcd - cto - mqj - usa - emi - liv - guz - tgp - ncl - azg - kck - apb - kbm - sgz - wsk - gbm - csw - ctp - eri - vid - bsq - dak - tif - xsr - mwp - roo - wbp - gdn - tgo - zpm - mih - qvc - tod - abn - cao - kbc - kog - moh - leh - nsu - okv - tih - zos - llg - gmv - kyf - mhx - mvp - naf - ino - dng - mnk - oke - omw - tnk - xtd - agt - cbr - eko - nhk - hot - iou - mpm - myx - ndh - pbb - rng - cri - bvd - kmz - cta - blw - clu - cof - kje - nmf - kbh - qxn - ria - wed - zar - huv - lsi - msk - nbq - ati - lww - yva - amr - lmk - tim - bbb - bzd - djr - kwf - nuy - poi - ptp - atd - bps - lcm - mbh - nuj - ayp - cwe - geb - aoi - mto - cbi - gvn - kqf - rkb - xed - zas - apz - cpb - ese - gdg - iws - apu - cag - kdc - nou - too - xmm - beq - hnn - myk - sby - arl - ncu - rml - gam - kvn - mbc - mfy - urt - luc - mio - snn - tar - tku - upv - kjg - btd - cut - mwn - tue - adz - kyg - rel - tew - ttc - nkf - yrb - fub - nsm - nyf - ava - bkd - cpc - lif - mwc - gan - apw - azz - zpo - ace - faa - ood - toc - wmt - ajg - nko - ssd - sxb - zxx - not - amk - hch - kqc - aui - llb - avt - kaq - tna - zpt - mnw - bwd - lif - stn - crk - kgk - mck - bhg - ifu - kkl - zao - huu - sil - trq - tuf - wnu - amf - fan - zdj - bco - kdl - lln - mit - yaq - ake - ige - mna - nhx - njn - wba - zaa - lez - syb - qxo - rgu - wiu - bla - etr - miy - mjc - nif - pov - qxl - sey - bxh - ddg - mks - nki - pot - prf - hra - law - plw - zpi - atg - daf - drg - dru - ngp - boa - cbt - dhm - ikw - kqe - lbj - loq - xsb - yad - kyu - mcb - tks - cnl - far - miz - maj - mgm - mib - mza - nhr - smt - ish - mlu - yka - anv - ign - mir - tav - gyr - klt - apn - wbm - cjk - dln - mcf - twx - ldi - poy - yaa - zae - bdq - con - dnw - gah - lcp - mbj - xog - yan - sdq - sgh - gwi - lbk - ses - ycn - kau - hix - mig - mlh - aer - amx - auc - cly - dsh - gnb - neb - ttt - jic - xtm - tll - apt - kgr - pio - pmf - yrk - akh - box - cub - cux - mrq - inh - bsp - for - kny - uri - zsr - pce - cme - cot - kxw - omb - srq - vag - cpu - kmy - ksc - lai - nnp - ruf - spm - tro - ttj - waj - raj - ian - ivb - kmm - kzf - maw - tuv - whk - zsm - aph - anm - mny - bug - trp - are - esi - mhi - obo - bwu - lwo - mrg - old - zav - snk - daa - fuq - lud - nlg - sck - tbl - khb - tpu - fue - tmd - whg - blt - gqr - kub - ots - tpp - var - hbb - krr - zxx - alj - ebk - goa - hav - kmk - kqo - muy - njb - snw - wib - shn - tkr - ksp - lip - mog - nla - pss - zad - awi - frd - kei - kmd - mim - got - lbe - adi - any - arp - biu - bvc - ghs - ttq - vmk - zpj - crl - dtb - dug - kam - nmo - qva - rhg - taw - tui - uth - wap - zpg - brb - dty - syr - bih - bqj - cjv - cnw - kgf - mfh - mns - nnw - ntp - oto - dks - mpt - nmw - sba - sfw - shu - ury - yrl - arc - cmo - xkg - iqw - kpj - nfa - nph - nre - wsg - xsu - nut - bmk - cdf - enl - fuv - hop - ife - lnd - mrv - nzm - tem - dwr - gwc - udi - zha - ajz - kru - qvo - sld - yup - anp - kwu - szy - tdg - ade - bcw - bgz - cul - fmu - gvc - ktj - lis - mev - msc - mta - nbc - ncq - way - wlv - wlx - jmx - pli - tkr - wni - aim - byv - cgg - fud - jmc - ker - kia - lom - ojb - tnc - bwx - cre - skr - xal - gjn - guq - lap - mip - mkl - mtg - mzk - njm - nse - pne - pww - zam - zca - the - dis - lgl - mbd - mda - mhy - nst - pkb - pmq - poe - swk - trs - vot - yam - zpq - atq - bav - cho - chq - cjo - due - duo - etu - ifa - kby - khq - kno - lem - mcn - mkz - mnb - myb - niy - pxm - rub - sju - tmc - tpm - wob - nxa - pnt - sdk - tcy - bhz - bth - crj - eka - khy - led - met - mgo - mus - mzm - naw - sri - tbk - xbr - awa - tnl - tnn - awb - bsc - iri - kpq - ktm - loe - lun - mfg - moa - ndj - otn - sjo - wat - bkm - tuz - ybi - gbr - gej - giz - gna - gwr - kex - kle - knx - ktz - kvj - lob - mhw - mph - pmx - qus - tcc - ang - azn - bra - gal - jra - nlv - rki - stk - yux - dip - ivv - kdj - kff - kyu - mnf - moc - mpg - mua - nma - rim - rjs - rnl - ssx - tig - urw - idt - mrn - oki - bbk - bea - bkq - btt - cou - ddn - did - diu - ilb - ksj - lhi - ndi - ngb - njz - nmz - npo - nwb - pbi - sbd - soe - tpw - twb - bfn - boz - lwl - psp - sgd - ymp - bkl - bkv - csk - enx - ess - gof - gud - guu - kdh - krx - kzn - maf - max - mgc - mvn - mwm - nbe - nlc - nxd - pse - sbs - tlb - wwa - yim - bkx - bze - chu - kbq - kpy - mry - tkd - ybh - abi - ayo - bex - bom - bov - dhg - jaa - kqs - kqy - ksb - ldn - log - lol - mfk - mtj - ndz - ngc - nio - nri - nuz - pps - shk - tlj - vut - xnn - bcc - cuh - emp - hro - nco - nxl - pex - tnp - trv - vif - aha - avn - bmv - chj - crt - dos - eto - gya - hlt - jun - ktb - lgm - mnx - mor - mug - nng - nnh - nsa - nyk - otd - oym - phm - prq - saj - szb - tap - tnr - tzl - wew - yas - yuz - avk - bat - baw - bwt - bxa - cjs - csw - fli - kqr - lan - lgr - lhm - mlk - mve - nqo - odk - olo - otq - pea - saq - sea - sps - tio - yin - adl - agw - ald - asg - bnj - brx - bwi - bwq - dbq - dow - fad - hoc - kql - ksf - kyu - lea - mse - mzl - nhd - nnl - nwx - pbc - pny - tul - aaa - abc - agq - ags - azo - bag - bax - bce - bfm - bgf - bhs - bkc - bkh - blk - bob - bqm - bri - brv - buo - bya - chp - cim - clo - cuv - dag - dmg - dtr - eee - ekm - enc - fiu - gbj - gou - hao - hna - hre - isu - jgo - kbx - kvt - kxp - lfa - lkb - lns - loh - lrc - lts - luy - mdr - mfj - mgg - mgq - mle - mlw - mmu - mne - muv - mxu - nge - ngn - njy - nsk - nsk - nwe - nza - pcg - pdu - pnz - pwg - qaa - shi - smn - snk - snl - sox - ssn - syw - tay - tbj - tdb - thy - tnt - tom - tpl - tsb - tvs - udg - unr - wbr - wms - wnk - wtk - xmd - xmg - xty - yav - yea - zuh multilinguality: - multilingual pinned: true tags: - multilingual --- ## Dataset Summary DCAD-2000 is a large-scale multilingual corpus built using newly extracted Common Crawl data (CC-MAIN-2024-46) and existing multilingual datasets. It includes over 2,282 languages, 46.72TB of data, and 8.63 billion documents, spanning 155 highand medium-resource languages and 159 writing scripts. We propose reframing data cleaning as an anomaly detection task. This dynamic filtering approach significantly enhances data quality by identifying and removing noisy or anomalous content. + Paper: [A Multilingual Dataset across 2000+ Languages with Data Cleaning as Anomaly Detection](https://www.arxiv.org/abs/2502.11546) + Github: [https://github.com/yl-shen/DCAD-2000](https://github.com/yl-shen/DCAD-2000) + Dataset (HuggingFace): [openbmb/DCAD-2000](https://huggingface.co/datasets/openbmb/DCAD-2000) ## Dataset Overview Comparison of multilingual datasets constructed from Common Crawl (CC) and our constructed DCAD-2000, focusing on the latest CC version used, the total number of languages supported, distribution across resource categories (high, medium, low, very low), and training readiness. The CC version marked with **bold** indicates an inferred version due to the lack of explicit specification in the original paper. The ``Training-Ready'' column indicates whether the dataset is ready for training LLMs without requiring further data cleaning. | **Dataset** | **CC Version** | **#Langs (total)** | **#Langs (high)** | **#Langs (medium)** | **#Langs (low)** | **#Langs (very low)** | **Training-Ready** | |---------------------------|--------------------|--------------------|-------------------|---------------------|------------------|-----------------------|--------------------| | mC4 (Raffel et al., 2020) | CC-MAIN-2020-34 | 101 | 0 | 43 | 52 | 6 | ✘ | | OSCAR 23.01 (Abadji et al., 2022) | CC-MAIN-2022-49 | 153 | 6 | 42 | 25 | 80 | ✘ | | Glot500 (Imani et al., 2023) | **CC-MAIN-2020-34** | 511 | 0 | 108 | 79 | 324 | ✘ | | CulturaX (Nguyen et al., 2024) | **CC-MAIN-2022-49** | 167 | 11 | 47 | 27 | 82 | ✘ | | Madlad-400 (Kudugunta et al., 2024) | CC-MAIN-2022-33 | 419 | 7 | 46 | 39 | 327 | ✘ | | MaLA (Ji et al., 2024) | **CC-MAIN-2022-49** | 939 | 1 | 125 | 78 | 735 | ✘ | | Glotcc (Kargaran et al., 2024) | CC-MAIN-2023-50 | 1331 | 0 | 10 | 52 | 1269 | ✘ | | HPLT-v1.2 (de Gilbert et al., 2024) | **CC-MAIN-2022-40** | 191 | 12 | 53 | 38 | 88 | ✘ | | Fineweb-2 (Penedo et al., 2024) | CC-MAIN-2024-18 | 1915 | 10 | 62 | 49 | 1794 | ✘ | | **DCAD-2000** | CC-MAIN-2024-46 | 2282 | 13 | 142 | 124 | 2003 | ✓ | ## Dataset Creation + **Data Collection:** DCAD-2000 integrates data from four main sources: MaLA, Fineweb, Fineweb-2, and newly extracted Common Crawl data. + **Data Cleaning as Anomaly Detection:** Traditional data cleaning methods rely on fixed thresholds for document-level features, making them less adaptable to the diversity of multilingual data. To address this, we propose a novel framework that formulates data cleaning as an anomaly detection task, which involves the feature extraction and anomaly detection. - **Feature Extraction:** For each document, we consider the following eight features: (1) Number of Words; (2) Character Repetition Ratio; (3) Word Repetition Ratio; (4) Special Characters Ratio; (5) Stop- words Ratio; (6) Flagged Words Ratio; (7) Language Identification (LID) Score; (8) Perplexity Score. - **Anomaly Detection:** We evaluate several classical anomaly detection algorithms including (1) Isolation Forest; (2) One Class SVM; (3) Local Outlier Factor and (4) K-Means. - Visualization<br> ![ad_overview](https://github.com/yl-shen/DCAD-2000/raw/master/images/ad_visual.png) ## Data Statistics + [Geographical Distribution](https://github.com/yl-shen/DCAD-2000/blob/master/statistics/geo_group.tsv) + [Script Distribution](https://github.com/yl-shen/DCAD-2000/blob/master/statistics/script_group_detail.tsv) + [Statistical Details](https://github.com/yl-shen/DCAD-2000/blob/master/statistics/statistic_detail.tsv) + [language list](https://github.com/yl-shen/DCAD-2000/blob/master/statistics/statistic_lang_num.tsv) + [language mapping](https://github.com/yl-shen/DCAD-2000/blob/master/lang_mapping/all_mapping.csv) ## Usage (Dataset) ``` from datasets import load_dataset data = load_dataset("openbmb/DCAD-2000") ``` You can also specifiy the language you wanted ``` from datasets import load_dataset data = load_dataset("openbmb/DCAD-2000", name="eng_Latn", split="train") ``` ## Citation Information ``` @article{shen2025dcad, title={DCAD-2000: A Multilingual Dataset across 2000+ Languages with Data Cleaning as Anomaly Detection}, author={Shen, Yingli and Lai, Wen and Wang, Shuo and Zhang, Xueren and Luo, Kangyang and Fraser, Alexander and Sun, Maosong}, journal={arXiv preprint arXiv:2502.11546}, year={2025} } ``` ## Acknowledgements We introduce DCAD-2000, a large- scale multilingual dataset designed to address the increasing demand for high-quality and diverse training data for multilingual LLMs. This work is done by researchers at [Tsinghua NLP group](https://nlp.csai.tsinghua.edu.cn) in collaboration with partners from [TUM](https://wenlai-lavine.github.io/) and [Modelbest Inc.](https://huggingface.co/openbmb) ## Contact Information Yingli Shen ([email protected]) Wen Lai ([email protected])