def sentence_to_spans(tokenized_sentence: list[str]) -> dict[str, int]: | |
span_to_token_id_ = dict() | |
start = 0 | |
for index, token_ in enumerate(tokenized_sentence): | |
span_to_token_id_[f"{start + index}-{start + index + len(token_)}"] = index | |
start += len(token_) | |
return span_to_token_id_ | |