dataset_info:
features:
- name: blob_id
dtype: string
- name: repo_name
dtype: string
- name: path
dtype: string
- name: length_bytes
dtype: int64
- name: score
dtype: float64
- name: int_score
dtype: int64
- name: text
dtype: string
- name: download_success
dtype: bool
splits:
- name: train
num_bytes: 13499266964
num_examples: 7678448
download_size: 6086016638
dataset_size: 13499266964
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
About This Dataset
The HuggingFaceTB team has released an impressive series of models called smollm (V1/V2) (paper: https://arxiv.org/abs/2502.02737).
According to their documentation, they used Stack-Edu as the code field corpus for pretraining and published https://huggingface.co/datasets/HuggingFaceTB/smollm-corpus.
However for some reason, only a Python-Edu subset is accessible and there's no content/text field in it.
The full dataset is stored on AWS S3; downloading it requires an AWS EC2 instance, or you will be blocked by AWS's rate limits under which you will never download it.
Fortunately, the py-edu subset is relatively small (~7 million files) for me to afford; downloading the entire set takes approximately one hour.
I am publishing the complete py-edu dataset here for anyone who needs it.
If this release inadvertently causes any issues for the HuggingFaceTB team, please reach out to me and I will remove it immediately.