Datasets:
Tasks:
Text Generation
Modalities:
Text
Sub-tasks:
language-modeling
Languages:
Japanese
Size:
10K - 100K
ArXiv:
Tags:
question-generation
License:
update
Browse files
generate_reference_files.py
CHANGED
@@ -6,9 +6,14 @@ os.makedirs('./reference_files', exist_ok=True)
|
|
6 |
|
7 |
|
8 |
for split in ['validation', 'test']:
|
9 |
-
dataset = load_dataset('asahi417/qg_jaquad', split=split)
|
10 |
for data in ['question', 'answer', 'sentence', 'paragraph']:
|
11 |
with open('./reference_files/{}-{}.txt'.format(data, split), 'w') as f:
|
12 |
-
|
|
|
|
|
|
|
|
|
13 |
length = [len(open(i).read().split('\n')) for i in glob(f'reference_files/*{split}.txt')]
|
14 |
assert len(list(set(length))) == 1, length
|
|
|
|
6 |
|
7 |
|
8 |
for split in ['validation', 'test']:
|
9 |
+
dataset = load_dataset('asahi417/qg_jaquad', split=split, download_mode='force_redownload')
|
10 |
for data in ['question', 'answer', 'sentence', 'paragraph']:
|
11 |
with open('./reference_files/{}-{}.txt'.format(data, split), 'w') as f:
|
12 |
+
if data == 'paragraph':
|
13 |
+
tmp_data = dataset['paragraph_id']
|
14 |
+
else:
|
15 |
+
tmp_data = dataset[data]
|
16 |
+
f.write('\n'.join([i.replace('\n', '') for i in tmp_data]))
|
17 |
length = [len(open(i).read().split('\n')) for i in glob(f'reference_files/*{split}.txt')]
|
18 |
assert len(list(set(length))) == 1, length
|
19 |
+
assert length[0] == len(dataset), f"{length[0]} != {len(dataset)}"
|