|
--- |
|
dataset_info: |
|
features: |
|
- name: doc_id |
|
dtype: string |
|
- name: type |
|
dtype: string |
|
- name: text |
|
dtype: string |
|
splits: |
|
- name: train |
|
num_bytes: 25324509618 |
|
num_examples: 806930 |
|
download_size: 9419131940 |
|
dataset_size: 25324509618 |
|
configs: |
|
- config_name: default |
|
data_files: |
|
- split: train |
|
path: data/train-* |
|
license: cc-by-4.0 |
|
task_categories: |
|
- text-generation |
|
language: |
|
- hi |
|
- en |
|
pretty_name: long_context |
|
size_categories: |
|
- 100K<n<1M |
|
--- |
|
|
|
# Dataset |
|
|
|
This dataset was filtered from AI4BHarat dataset [sangraha](https://huggingface.co/datasets/ai4bharat/sangraha),which is the largest high-quality, cleaned Indic language pretraining data containing 251B tokens summed up over 22 languages, extracted from curated sources, existing multilingual corpora and large scale translations. |
|
|
|
This dataset contains only Hindi as of now |
|
|
|
# Information |
|
* First this dataset is mainly for long context training |
|
* The minimum len is `6000` and maximum len is `3754718` |
|
|
|
# Getting started |
|
|
|
For downloading the entire dataset: |
|
```python |
|
from datasets import load_dataset |
|
dataset = load_dataset("damerajee/long_context_hindi") |
|
``` |
|
|
|
If dataset is too big you can simply stream: |
|
```python |
|
from datasets import load_dataset |
|
|
|
dataset = load_dataset("damerajee/long_context_hindi",split='train',streaming=True) |
|
``` |
|
```python |
|
dataset.take(2) |
|
``` |