def sentence_to_spans(tokenized_sentence: list[str]) -> dict[str, int]: span_to_token_id_ = dict() start = 0 for index, token_ in enumerate(tokenized_sentence): span_to_token_id_[f"{start + index}-{start + index + len(token_)}"] = index start += len(token_) return span_to_token_id_