--- dataset_info: features: - name: blob_id dtype: string - name: repo_name dtype: string - name: path dtype: string - name: length_bytes dtype: int64 - name: score dtype: float64 - name: int_score dtype: int64 - name: text dtype: string - name: download_success dtype: bool splits: - name: train num_bytes: 13499266964 num_examples: 7678448 download_size: 6086016638 dataset_size: 13499266964 configs: - config_name: default data_files: - split: train path: data/train-* --- # About This Dataset The HuggingFaceTB team has released an impressive series of models called smollm (V1/V2) (paper: https://arxiv.org/abs/2502.02737). According to their documentation, they used Stack-Edu as the code field corpus for pretraining and published https://huggingface.co/datasets/HuggingFaceTB/smollm-corpus. However for some reason, only a Python-Edu subset is accessible and there's no content/text field in it. The full dataset is stored on AWS S3; downloading it requires an AWS EC2 instance, or you will be blocked by AWS's rate limits under which you will never download it. Fortunately, the py-edu subset is relatively small (~7 million files) for me to afford; downloading the entire set takes approximately one hour. I am publishing the complete py-edu dataset here for anyone who needs it. If this release inadvertently causes any issues for the HuggingFaceTB team, please reach out to me and I will remove it immediately.