Dataset Viewer
Full Screen
The dataset viewer is not available for this split.
Cannot extract the features (columns) for the split 'train' of the config 'default' of the dataset.
Error code:   FeaturesError
Exception:    ArrowInvalid
Message:      JSON parse error: Column() changed from object to array in row 0
Traceback:    Traceback (most recent call last):
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/packaged_modules/json/json.py", line 160, in _generate_tables
                  df = pandas_read_json(f)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/packaged_modules/json/json.py", line 38, in pandas_read_json
                  return pd.read_json(path_or_buf, **kwargs)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/pandas/io/json/_json.py", line 815, in read_json
                  return json_reader.read()
                File "/src/services/worker/.venv/lib/python3.9/site-packages/pandas/io/json/_json.py", line 1025, in read
                  obj = self._get_object_parser(self.data)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/pandas/io/json/_json.py", line 1051, in _get_object_parser
                  obj = FrameParser(json, **kwargs).parse()
                File "/src/services/worker/.venv/lib/python3.9/site-packages/pandas/io/json/_json.py", line 1187, in parse
                  self._parse()
                File "/src/services/worker/.venv/lib/python3.9/site-packages/pandas/io/json/_json.py", line 1403, in _parse
                  ujson_loads(json, precise_float=self.precise_float), dtype=None
              ValueError: Trailing data
              
              During handling of the above exception, another exception occurred:
              
              Traceback (most recent call last):
                File "/src/services/worker/src/worker/job_runners/split/first_rows.py", line 233, in compute_first_rows_from_streaming_response
                  iterable_dataset = iterable_dataset._resolve_features()
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/iterable_dataset.py", line 2831, in _resolve_features
                  features = _infer_features_from_batch(self.with_format(None)._head())
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/iterable_dataset.py", line 1845, in _head
                  return _examples_to_batch(list(self.take(n)))
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/iterable_dataset.py", line 2012, in __iter__
                  for key, example in ex_iterable:
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/iterable_dataset.py", line 1507, in __iter__
                  for key_example in islice(self.ex_iterable, self.n - ex_iterable_num_taken):
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/iterable_dataset.py", line 268, in __iter__
                  for key, pa_table in self.generate_tables_fn(**gen_kwags):
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/packaged_modules/json/json.py", line 163, in _generate_tables
                  raise e
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/packaged_modules/json/json.py", line 137, in _generate_tables
                  pa_table = paj.read_json(
                File "pyarrow/_json.pyx", line 308, in pyarrow._json.read_json
                File "pyarrow/error.pxi", line 154, in pyarrow.lib.pyarrow_internal_check_status
                File "pyarrow/error.pxi", line 91, in pyarrow.lib.check_status
              pyarrow.lib.ArrowInvalid: JSON parse error: Column() changed from object to array in row 0

Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.

Dataset Card for Book_Stitch

This dataset contains books from Project Gutenberg, tokenized into 1020-token chunks with markers that indicate the section and book unique identifier (UID). These markers serve as both prefix and suffix for the sections, ensuring that the sequential nature of each book is preserved and facilitating later text reconstruction. The book_stitch dataset is designed for training AI models to handle long texts in sections, retaining context for tasks like summarization, text stitching, and document analysis.

This dataset card aims to be a base template for new datasets. It has been generated using this raw template.

Dataset Details

Dataset Description

The book_stitch dataset is part of a series designed to teach AI models how to handle large documents, including stitching and unstitching sections of text. Each book is tokenized into fixed-size chunks of 1020 tokens, with markers appended to indicate the beginning and end of each section. This dataset works in conjunction with the context_stitch and train_stitch datasets to allow models to maintain long-range context across different sections of a document, enabling comprehensive text analysis and reassembly.

Curated by: [Robert McNarland, McNarland Software Consultation Inc.] Funded by None Shared by None Language(s) (NLP): [ English (books from Project Gutenberg)] License: [MIT] Dataset Sources [optional] Repository: [R3troR0b/book_stitch] Paper [optional]: [More Information Needed] Demo [optional]: [More Information Needed]

Uses

Direct Use

  • Document Classification and Reconstruction: The book_stitch dataset is key for teaching models to classify books and reconstruct them from tokenized chunks. The stitched-together sections allow for complex tasks like document-level classification and content retrieval.

  • Text Stitching and Unstitching: Models can learn how to stitch and unstitch text segments, with markers indicating where sections begin and end, supporting tasks like reassembling fragmented documents or summarizing long-form text.

  • Long-Document Modeling: The dataset helps train models to process long texts efficiently, maintaining contextual understanding across multiple sections by using section markers and UIDs.

  • Contextual Inference: By identifying relationships between text sections, models can better infer meaning and connections in lengthy documents, supporting tasks such as question-answering, summarization, and complex search.

Out-of-Scope Use

The dataset is not intended for use cases unrelated to document classification, text reassembly, or handling long-range context in texts. It may not be applicable to non-English texts.

Dataset Structure

Each entry in the dataset consists of:

Label: This includes both the prefix and suffix stitch markers indicating the section number and the book's UID (e.g., [/SEC:1;B-5]). Text: A 1020-token chunk of the book's content. Example:

{
  "label": "[/SEC:1;B-5] [/SEC:2;B-5]",
  "text": "CHAPTER I: START OF THE PROJECT GUTENBERG EBOOK..."
}

Dataset Creation

Curation Rationale

The book_stitch dataset was created to help AI models understand the structure of long-form text. By breaking books into consistent tokenized chunks with markers, models can be trained to stitch and unstitch sections, allowing for sophisticated text handling.

Source Data

Data Collection and Processing

The dataset was generated using a custom-built tokenizer to split Project Gutenberg books into fixed 1020-token chunks, adding section and book UIDs as markers. The end portion of each book may be smaller than previous sections. The table of contents and appendices are treated as part of the text.

Who are the source data producers?

Titles in the book_stitch dataset are from Project Gutenberg's English collection.

Annotations [optional]

Annotation process

This dataset does not include any additional annotations beyond the section markers and UIDs.

Who are the annotators?

There are no annotators for this dataset, as it relies solely on automated tokenization and marking.

Personal and Sensitive Information

This dataset does not contain personal, sensitive, or private information. All books are sourced from Project Gutenberg, a collection of public domain works.

Bias, Risks, and Limitations

Recommendations

Users should be aware that:

The dataset is limited to English books from Project Gutenberg and may not generalize well to non-English or non-literary domains. Since these books are public domain, the texts may reflect biases inherent in historical works.

Citation [optional]

BibTeX:

[More Information Needed]

APA:

[More Information Needed]

Glossary [optional]

  • Book UID (BUID): A unique identifier assigned to each book.
  • Stitch Markers: Markers appended to each section of the text to indicate the section number and the book's UID (e.g., [/SEC:1;B-5]).
  • Contextual Stitching: The process of stitching together sections of text while maintaining continuity.

More Information [optional]

[More Information Needed]

Dataset Card Authors [optional]

[Robert McNarland, McNarland Software Consultation Inc.]

Dataset Card Contact

[email protected]

Downloads last month
32