The dataset viewer is not available for this subset.
Exception: SplitsNotFoundError Message: The split names could not be parsed from the dataset config. Traceback: Traceback (most recent call last): File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/inspect.py", line 289, in get_dataset_config_info for split_generator in builder._split_generators( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/packaged_modules/folder_based_builder/folder_based_builder.py", line 191, in _split_generators for pa_metadata_table in self._read_metadata(downloaded_metadata_file, metadata_ext=metadata_ext): File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/packaged_modules/folder_based_builder/folder_based_builder.py", line 296, in _read_metadata for df in csv_file_reader: File "/src/services/worker/.venv/lib/python3.9/site-packages/pandas/io/parsers/readers.py", line 1843, in __next__ return self.get_chunk() File "/src/services/worker/.venv/lib/python3.9/site-packages/pandas/io/parsers/readers.py", line 1985, in get_chunk return self.read(nrows=size) File "/src/services/worker/.venv/lib/python3.9/site-packages/pandas/io/parsers/readers.py", line 1923, in read ) = self._engine.read( # type: ignore[attr-defined] File "/src/services/worker/.venv/lib/python3.9/site-packages/pandas/io/parsers/c_parser_wrapper.py", line 234, in read chunks = self._reader.read_low_memory(nrows) File "parsers.pyx", line 850, in pandas._libs.parsers.TextReader.read_low_memory File "parsers.pyx", line 905, in pandas._libs.parsers.TextReader._read_rows File "parsers.pyx", line 874, in pandas._libs.parsers.TextReader._tokenize_rows File "parsers.pyx", line 891, in pandas._libs.parsers.TextReader._check_tokenize_status File "parsers.pyx", line 2061, in pandas._libs.parsers.raise_parser_error pandas.errors.ParserError: Error tokenizing data. C error: Expected 1 fields in line 6, saw 2 The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/src/services/worker/src/worker/job_runners/config/split_names.py", line 65, in compute_split_names_from_streaming_response for split in get_dataset_split_names( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/inspect.py", line 343, in get_dataset_split_names info = get_dataset_config_info( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/inspect.py", line 294, in get_dataset_config_info raise SplitsNotFoundError("The split names could not be parsed from the dataset config.") from err datasets.inspect.SplitsNotFoundError: The split names could not be parsed from the dataset config.
Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.
Save the README content in English as a .txt file
TTS Dataset - Indonesian + English
Overview: This dataset is curated for fine-tuning and training expressive TTS (Text-to-Speech) models. It combines expressive Indonesian speech and some English lines. All data was collected from various online sources for research and non-commercial purposes only.
Dataset Statistics:
- Total Samples : 84,641
- Total Duration : 66 hours, 51 minutes, and 50 seconds
- Average Duration : ~1.58 seconds per sample
- Minimum Duration : 1.08 seconds
- Maximum Duration : 9.29 seconds
- English Content : ~6 hours (tagged with [en] and [/en])
- Indonesian Content : ~60+ hours
Metadata Format: Sentences in English are clearly marked using [en] and [/en] tags:
Example: [en]This is an English sentence.[/en] Ini adalah kalimat Bahasa Indonesia.
License & Usage:
- This dataset is provided strictly for research, educational, and personal experimental use.
- You are not allowed to sell, redistribute, or use it for commercial purposes.
- Voice content was collected from public online content and is not intended for voice cloning or impersonation.
Credit: Thanks to all anonymous content sources. The intention is to improve TTS for underrepresented languages and styles.
Contact: For takedown requests, issues, or contributions, please contact the dataset maintainer on HuggingFace.
π₯ How to Download This Dataset
You can use the provided download.py
script to automatically download and extract all dataset parts.
π§ Step-by-step
# Download the script
wget https://huggingface.co/datasets/PapaRazi/id-tts-v2/resolve/main/download.py
# Run it (downloads to ./id-tts-v2 by default)
python download.py
# Or specify a custom output path
python download.py /data/custom_path
- Downloads last month
- 322