Datasets:
License:
File size: 2,497 Bytes
7725656 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 |
'''
Script on Generating Wikipedia Data that are dumped into https://dumps.wikimedia.org/
More info can be read on https://huggingface.co/datasets/wikipedia
-------------------
Check here to see available indexed data: https://dumps.wikimedia.org/backup-index.html
Also check here to see language meta from its code: https://meta.wikimedia.org/wiki/List_of_Wikipedias
'''
import os
import logging
import argparse
import pandas as pd
from datasets import load_dataset
def set_logger():
# Set up the logger
logging.basicConfig(
level=logging.INFO, # Set the desired logging level (DEBUG, INFO, WARNING, ERROR, CRITICAL)
format='%(asctime)s [%(levelname)s]: %(message)s', # Customize the log message format
datefmt='%Y-%m-%d %H:%M:%S' # Customize the date/time format
)
# Create a file handler to write logs into a file
file_handler = logging.FileHandler('app.log')
# Set the log level for the file handler
file_handler.setLevel(logging.INFO)
# Create a formatter for the file handler (customize the log format for the file)
file_formatter = logging.Formatter('%(asctime)s [%(levelname)s]: %(message)s', datefmt='%Y-%m-%d %H:%M:%S')
file_handler.setFormatter(file_formatter)
logger = logging.getLogger("Wiki Dataset Generation")
logger.addHandler(file_handler)
return logger
#only executed if called directly
if __name__ == "__main__":
parser = argparse.ArgumentParser()
parser.add_argument("--lang-id", help="Lang ID from Wikipedia Data to extract")
parser.add_argument("--date-ver", help="Date of Wikipedia Data (YYYYMMDD) generation to extract")
parser.add_argument("--save-dir-path", help="""Relative dir path of saved Wikipedia CSV data
to the `extract_raw_wiki_data.py` script dir""",
default=os.path.dirname(os.path.abspath(__file__)))
args = parser.parse_args()
dset_name = "wikipedia"
logger = set_logger()
logger.info("Parsing arguments...")
lang_id = args.lang_id
date_ver = args.date_ver
save_dir = args.save_dir_path
logger.info("Loading the dataset from Wikipedia...")
df = load_dataset(dset_name, language=lang_id, date=date_ver, beam_runner='DirectRunner', split="train").to_pandas()
logger.info("Loading done!")
logger.info(f"#Data collected: {df.shape[0]}")
logger.info("Saving dataset raw form...")
df.to_csv(f"{save_dir}/wiki_{lang_id}_{date_ver}_raw_dataset.csv", index=False)
|