{"description": "Wikipedia dataset containing cleaned articles of all languages.\nThe datasets are built from the Wikipedia dump\n(https://dumps.wikimedia.org/) with one split per language. Each example\ncontains the content of one full Wikipedia article with cleaning to strip\nmarkdown and unwanted sections (references, etc.).\n", "citation": "@ONLINE {wikidump,\n author = \"Wikimedia Foundation\",\n title = \"Wikimedia Downloads\",\n url = \"https://dumps.wikimedia.org\"\n}\n", "homepage": "https://dumps.wikimedia.org", "license": "", "features": {"title": {"dtype": "string", "id": null, "_type": "Value"}, "text": {"dtype": "string", "id": null, "_type": "Value"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "wikipedia", "config_name": "20200501.en", "version": {"version_str": "1.0.0", "description": "New split API (https://tensorflow.org/datasets/splits)", "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 18330235071, "num_examples": 6078422, "dataset_name": "wikipedia"}}, "download_checksums": {}, "download_size": 18241319609, "post_processing_size": null, "dataset_size": 18330235071, "size_in_bytes": 36571554680} |