| import json |
|
|
| import datasets |
| import pandas as pd |
| from huggingface_hub.file_download import hf_hub_url |
| from collections import OrderedDict |
|
|
| try: |
| import lzma as xz |
| except ImportError: |
| import pylzma as xz |
|
|
| datasets.logging.set_verbosity_info() |
| logger = datasets.logging.get_logger(__name__) |
|
|
| _DESCRIPTION ="""\ |
| |
| """ |
|
|
| _HOMEPAGE = "" |
|
|
| _LICENSE = "" |
|
|
| _CITATION = "" |
|
|
| _URL = { |
| 'data/' |
| } |
| _LANGUAGES = [ |
| "de", "fr", "it", "swiss", "en" |
| ] |
| _SUBSETS = [ |
| "_sherlock", "_sfu", "_bioscope", "_dalloux", "" |
| ] |
|
|
| _BUILDS = ['de', 'fr', 'it', 'swiss', 'fr_dalloux', 'fr_all', 'en_bioscope', 'en_sherlock', 'en_sfu', 'en_all', 'all_all'] |
|
|
|
|
|
|
| class MultiLegalNegConfig(datasets.BuilderConfig): |
| |
| def __init__(self, name:str, **kwargs): |
| super( MultiLegalNegConfig, self).__init__(**kwargs) |
| self.name = name |
| self.language = name.split("_")[0] |
| self.subset = f'_{name.split("_")[1]}' if len(name.split("_"))==2 else "" |
|
|
| class MultiLegalNeg(datasets.GeneratorBasedBuilder): |
|
|
| BUILDER_CONFIG_CLASS = MultiLegalNegConfig |
| |
| BUILDER_CONFIGS = [ |
| MultiLegalNegConfig(f"{build}") for build in _BUILDS |
| ] |
|
|
| def _info(self): |
| features = datasets.Features( |
| { |
| "text": datasets.Value("string"), |
| "spans": [ |
| { |
| "start": datasets.Value("int64"), |
| "end": datasets.Value("int64"), |
| "token_start": datasets.Value("int64"), |
| "token_end": datasets.Value("int64"), |
| "label": datasets.Value("string") |
| } |
| ], |
| "tokens": [ |
| { |
| "text": datasets.Value("string"), |
| "start": datasets.Value("int64"), |
| "end": datasets.Value("int64"), |
| "id": datasets.Value("int64"), |
| "ws": datasets.Value("bool") |
| } |
| ] |
| } |
| ) |
| return datasets.DatasetInfo( |
| description=_DESCRIPTION, |
| features = features, |
| homepage = _HOMEPAGE, |
| citation=_CITATION |
| ) |
| |
| def _split_generators(self, dl_manager): |
| languages = _LANGUAGES if self.config.language == "all" else [self.config.language] |
| subsets = _SUBSETS if self.config.subset == "_all" else [self.config.subset] |
|
|
| split_generators = [] |
| for split in [datasets.Split.TRAIN, datasets.Split.TEST, datasets.Split.VALIDATION]: |
| filepaths = [] |
| for language in languages: |
| for subset in subsets: |
| try: |
| filepaths.append(dl_manager.download((f'data/{split}/{language}{subset}_{split}.jsonl.xz'))) |
| except: |
| break |
| split_generators.append(datasets.SplitGenerator(name=split, gen_kwargs={'filepaths': filepaths})) |
|
|
| return split_generators |
|
|
| def _generate_examples(self, filepaths): |
| id_ = 0 |
| for filepath in filepaths: |
| if filepath: |
| logger.info("Generating examples from = %s", filepath) |
| try: |
| with xz.open(open(filepath,'rb'), 'rt', encoding='utf-8') as f: |
| json_list = list(f) |
| |
| for json_str in json_list: |
| example = json.loads(json_str) |
| if example is not None and isinstance(example, dict): |
| yield id_, example |
| id_ +=1 |
|
|
| except Exception: |
| logger.exception("Error while processing file %s", filepath) |