Commit
·
28c621e
1
Parent(s):
6e86234
added information on where spans begin and end in the untouched text string
Browse files- essays_SuG.py +63 -24
essays_SuG.py
CHANGED
@@ -126,6 +126,8 @@ class Fancy(datasets.GeneratorBasedBuilder):
|
|
126 |
)
|
127 |
),
|
128 |
"text": datasets.Value("string"),
|
|
|
|
|
129 |
}
|
130 |
)
|
131 |
elif (
|
@@ -145,6 +147,8 @@ class Fancy(datasets.GeneratorBasedBuilder):
|
|
145 |
)
|
146 |
),
|
147 |
"text": datasets.Value("string"),
|
|
|
|
|
148 |
}
|
149 |
)
|
150 |
elif (
|
@@ -166,6 +170,8 @@ class Fancy(datasets.GeneratorBasedBuilder):
|
|
166 |
)
|
167 |
),
|
168 |
"text": datasets.Value("string"),
|
|
|
|
|
169 |
}
|
170 |
)
|
171 |
elif self.config.name == "sep_tok":
|
@@ -185,6 +191,8 @@ class Fancy(datasets.GeneratorBasedBuilder):
|
|
185 |
)
|
186 |
),
|
187 |
"text": datasets.Value("string"),
|
|
|
|
|
188 |
}
|
189 |
)
|
190 |
elif self.config.name == "sep_tok_full_labels":
|
@@ -206,6 +214,8 @@ class Fancy(datasets.GeneratorBasedBuilder):
|
|
206 |
)
|
207 |
),
|
208 |
"text": datasets.Value("string"),
|
|
|
|
|
209 |
}
|
210 |
)
|
211 |
|
@@ -252,11 +262,15 @@ class Fancy(datasets.GeneratorBasedBuilder):
|
|
252 |
# check for whether the data folder is in cwd.
|
253 |
# if it isnt, change cwd to its parent directory
|
254 |
# do this three times only (dont want infinite recursion)
|
255 |
-
for _ in range(
|
256 |
-
if Path.is_dir(cwd / "
|
257 |
-
|
|
|
|
|
|
|
|
|
258 |
# input(f"returning {cwd / 'data'}")
|
259 |
-
return cwd / "
|
260 |
cwd = cwd.parent
|
261 |
raise FileNotFoundError("data directory has not been found")
|
262 |
|
@@ -340,7 +354,14 @@ class Fancy(datasets.GeneratorBasedBuilder):
|
|
340 |
tokens = []
|
341 |
for sentence in essay["sentences"]:
|
342 |
for token in sentence["tokens"]:
|
343 |
-
tokens.append(
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
344 |
return tokens
|
345 |
|
346 |
def _get_label_dict(self, essay):
|
@@ -361,33 +382,44 @@ class Fancy(datasets.GeneratorBasedBuilder):
|
|
361 |
def _match_tokens(self, tokens, label_dict):
|
362 |
text = []
|
363 |
labels = []
|
364 |
-
|
|
|
|
|
|
|
365 |
# for each token, unpack it into its surface and gid
|
366 |
# then match the gid to the label and pack them back together
|
367 |
|
368 |
-
|
369 |
-
|
370 |
-
|
371 |
-
|
372 |
-
|
373 |
-
|
374 |
-
|
375 |
-
#
|
376 |
text.append("<s>")
|
377 |
labels.append("O")
|
378 |
-
|
379 |
-
|
380 |
-
|
381 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
382 |
):
|
383 |
-
# if
|
384 |
-
# of a span
|
385 |
-
# intert end of sequence token (EOS) and "O" label
|
386 |
text.append("</s>")
|
387 |
labels.append("O")
|
|
|
388 |
|
389 |
# always append the surface form
|
390 |
text.append(surface)
|
|
|
391 |
|
392 |
# append the correct type of label, depending on the config
|
393 |
if self.config.name == "full_labels":
|
@@ -407,7 +439,7 @@ class Fancy(datasets.GeneratorBasedBuilder):
|
|
407 |
|
408 |
else:
|
409 |
raise KeyError()
|
410 |
-
return text, labels
|
411 |
|
412 |
def _get_text(self, essay):
|
413 |
return essay["text"]
|
@@ -419,11 +451,18 @@ class Fancy(datasets.GeneratorBasedBuilder):
|
|
419 |
# input(tokens)
|
420 |
label_dict = self._get_label_dict(essay)
|
421 |
# input(label_dict)
|
422 |
-
tokens, labels = self._match_tokens(tokens, label_dict)
|
423 |
# input(tokens)
|
424 |
# input(labels)
|
425 |
text = self._get_text(essay)
|
426 |
-
return {
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
427 |
|
428 |
# method parameters are unpacked from `gen_kwargs` as given in `_split_generators`
|
429 |
def _generate_examples(self, data, id_range):
|
|
|
126 |
)
|
127 |
),
|
128 |
"text": datasets.Value("string"),
|
129 |
+
"span_begins": datasets.Sequence(datasets.Value("int16")),
|
130 |
+
"span_ends": datasets.Sequence(datasets.Value("int16")),
|
131 |
}
|
132 |
)
|
133 |
elif (
|
|
|
147 |
)
|
148 |
),
|
149 |
"text": datasets.Value("string"),
|
150 |
+
"span_begins": datasets.Sequence(datasets.Value("int16")),
|
151 |
+
"span_ends": datasets.Sequence(datasets.Value("int16")),
|
152 |
}
|
153 |
)
|
154 |
elif (
|
|
|
170 |
)
|
171 |
),
|
172 |
"text": datasets.Value("string"),
|
173 |
+
"span_begins": datasets.Sequence(datasets.Value("int16")),
|
174 |
+
"span_ends": datasets.Sequence(datasets.Value("int16")),
|
175 |
}
|
176 |
)
|
177 |
elif self.config.name == "sep_tok":
|
|
|
191 |
)
|
192 |
),
|
193 |
"text": datasets.Value("string"),
|
194 |
+
"span_begins": datasets.Sequence(datasets.Value("int16")),
|
195 |
+
"span_ends": datasets.Sequence(datasets.Value("int16")),
|
196 |
}
|
197 |
)
|
198 |
elif self.config.name == "sep_tok_full_labels":
|
|
|
214 |
)
|
215 |
),
|
216 |
"text": datasets.Value("string"),
|
217 |
+
"span_begins": datasets.Sequence(datasets.Value("int16")),
|
218 |
+
"span_ends": datasets.Sequence(datasets.Value("int16")),
|
219 |
}
|
220 |
)
|
221 |
|
|
|
262 |
# check for whether the data folder is in cwd.
|
263 |
# if it isnt, change cwd to its parent directory
|
264 |
# do this three times only (dont want infinite recursion)
|
265 |
+
for _ in range(5):
|
266 |
+
if Path.is_dir(cwd / "essays_SuG"):
|
267 |
+
print(f"found 'essays_SuG' folder at {cwd}")
|
268 |
+
# input(f"returning {cwd / 'essays_SuG'}")
|
269 |
+
return cwd / "essays_SuG"
|
270 |
+
if Path.is_dir(cwd / "data"):
|
271 |
+
print(f"found 'data' folder at {cwd}")
|
272 |
# input(f"returning {cwd / 'data'}")
|
273 |
+
return cwd / "data"
|
274 |
cwd = cwd.parent
|
275 |
raise FileNotFoundError("data directory has not been found")
|
276 |
|
|
|
354 |
tokens = []
|
355 |
for sentence in essay["sentences"]:
|
356 |
for token in sentence["tokens"]:
|
357 |
+
tokens.append(
|
358 |
+
(
|
359 |
+
token["surface"],
|
360 |
+
token["gid"],
|
361 |
+
token["characterOffsetBegin"],
|
362 |
+
token["characterOffsetEnd"],
|
363 |
+
)
|
364 |
+
)
|
365 |
return tokens
|
366 |
|
367 |
def _get_label_dict(self, essay):
|
|
|
382 |
def _match_tokens(self, tokens, label_dict):
|
383 |
text = []
|
384 |
labels = []
|
385 |
+
begins = []
|
386 |
+
ends = []
|
387 |
+
last_end = 0
|
388 |
+
for surface, gid, begin, end in tokens:
|
389 |
# for each token, unpack it into its surface and gid
|
390 |
# then match the gid to the label and pack them back together
|
391 |
|
392 |
+
if label_dict.get(gid, "O")[0] == "B":
|
393 |
+
# if we are at the beginning of a span
|
394 |
+
# insert begin of sequence token (BOS) and "O" label
|
395 |
+
if (
|
396 |
+
self.config.name == "sep_tok"
|
397 |
+
or self.config.name == "sep_tok_full_labels"
|
398 |
+
):
|
399 |
+
# if the config requires separator tokens
|
400 |
text.append("<s>")
|
401 |
labels.append("O")
|
402 |
+
begins.append(begin)
|
403 |
+
elif (
|
404 |
+
label_dict.get(gid, "O") == "O"
|
405 |
+
and len(labels) != 0
|
406 |
+
and labels[-1][0] != "O"
|
407 |
+
):
|
408 |
+
# if we are not in a span, and the previous label was
|
409 |
+
# of a span
|
410 |
+
# intert end of sequence token (EOS) and "O" label
|
411 |
+
if (
|
412 |
+
self.config.name == "sep_tok"
|
413 |
+
or self.config.name == "sep_tok_full_labels"
|
414 |
):
|
415 |
+
# if the config requires separator tokens
|
|
|
|
|
416 |
text.append("</s>")
|
417 |
labels.append("O")
|
418 |
+
ends.append(last_end)
|
419 |
|
420 |
# always append the surface form
|
421 |
text.append(surface)
|
422 |
+
last_end = end
|
423 |
|
424 |
# append the correct type of label, depending on the config
|
425 |
if self.config.name == "full_labels":
|
|
|
439 |
|
440 |
else:
|
441 |
raise KeyError()
|
442 |
+
return text, labels, begins, ends
|
443 |
|
444 |
def _get_text(self, essay):
|
445 |
return essay["text"]
|
|
|
451 |
# input(tokens)
|
452 |
label_dict = self._get_label_dict(essay)
|
453 |
# input(label_dict)
|
454 |
+
tokens, labels, begins, ends = self._match_tokens(tokens, label_dict)
|
455 |
# input(tokens)
|
456 |
# input(labels)
|
457 |
text = self._get_text(essay)
|
458 |
+
return {
|
459 |
+
"id": id,
|
460 |
+
"tokens": tokens,
|
461 |
+
"ner_tags": labels,
|
462 |
+
"text": text,
|
463 |
+
"span_begins": begins,
|
464 |
+
"span_ends": ends,
|
465 |
+
}
|
466 |
|
467 |
# method parameters are unpacked from `gen_kwargs` as given in `_split_generators`
|
468 |
def _generate_examples(self, data, id_range):
|