Commit
·
c6abeb0
1
Parent(s):
9cc038e
Upload dataset_infos.json with huggingface_hub
Browse files- dataset_infos.json +1 -1
dataset_infos.json
CHANGED
|
@@ -1 +1 @@
|
|
| 1 |
-
{"carlosejimenez--wikipedia-20220301.en-0.005-validation": {"description": "Wikipedia dataset containing cleaned articles of all languages.\nThe datasets are built from the Wikipedia dump\n(https://dumps.wikimedia.org/) with one split per language. Each example\ncontains the content of one full Wikipedia article with cleaning to strip\nmarkdown and unwanted sections (references, etc.).\n", "citation": "@ONLINE {wikidump,\n author = {Wikimedia Foundation},\n title = {Wikimedia Downloads},\n url = {https://dumps.wikimedia.org}\n}\n", "homepage": "https://dumps.wikimedia.org", "license": "", "features": {"id": {"dtype": "string", "id": null, "_type": "Value"}, "url": {"dtype": "string", "id": null, "_type": "Value"}, "title": {"dtype": "string", "id": null, "_type": "Value"}, "text": {"dtype": "string", "id": null, "_type": "Value"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name":
|
|
|
|
| 1 |
+
{"carlosejimenez--wikipedia-20220301.en-0.005-validation": {"description": "Wikipedia dataset containing cleaned articles of all languages.\nThe datasets are built from the Wikipedia dump\n(https://dumps.wikimedia.org/) with one split per language. Each example\ncontains the content of one full Wikipedia article with cleaning to strip\nmarkdown and unwanted sections (references, etc.).\n", "citation": "@ONLINE {wikidump,\n author = {Wikimedia Foundation},\n title = {Wikimedia Downloads},\n url = {https://dumps.wikimedia.org}\n}\n", "homepage": "https://dumps.wikimedia.org", "license": "", "features": {"id": {"dtype": "string", "id": null, "_type": "Value"}, "url": {"dtype": "string", "id": null, "_type": "Value"}, "title": {"dtype": "string", "id": null, "_type": "Value"}, "text": {"dtype": "string", "id": null, "_type": "Value"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": null, "config_name": null, "version": null, "splits": {"train": {"name": "train", "num_bytes": 20173799762.046375, "num_examples": 6426377, "dataset_name": "wikipedia-20220301.en-0.005-validation"}, "validation": {"name": "validation", "num_bytes": 101374773.95362327, "num_examples": 32293, "dataset_name": "wikipedia-20220301.en-0.005-validation"}}, "download_checksums": null, "download_size": 11628131978, "post_processing_size": null, "dataset_size": 20275174536.0, "size_in_bytes": 31903306514.0}}
|