--- # Example metadata to be added to a dataset card. # Full dataset card template at https://github.com/huggingface/huggingface_hub/blob/main/src/huggingface_hub/templates/datasetcard_template.md language: - en license: mit # Example: apache-2.0 or any license from https://hf.co/docs/hub/repositories-licenses tags: - robotics - manipulation - rearrangement - computer-vision - reinforcement-learning - imitation-learning - rgbd - rgb - depth - low-level-control - whole-body-control - home-assistant - simulation - maniskill annotations_creators: - machine-generated # Generated from RL policies with filtering language_creators: - machine-generated language_details: en-US pretty_name: ManiSkill-HAB TidyHouse Dataset size_categories: - 100K Whole-body, low-level control/manipulation demonstration dataset for ManiSkill-HAB TidyHouse. ## Dataset Details ### Dataset Description Demonstration dataset for ManiSkill-HAB TidyHouse. Each subtask/object combination (e.g pick 002_master_chef_can) has 1000 successful episodes (200 samples/demonstration) gathered using [RL policies](https://huggingface.co/arth-shukla/mshab_checkpoints) fitered for safe robot behavior with a rule-based event labeling system. TidyHouse contains the Pick and Place subtasks. Relative to the other MS-HAB long-horizon tasks (PrepareGroceries, SetTable), TidyHouse Pick is approximately medium difficulty, while TidyHouse Place is medium-to-hard difficulty (on a scale of easy-medium-hard). ### Related Datasets Full information about the MS-HAB datasets (size, difficulty, links, etc), including the other long horizon tasks, are available [on the ManiSkill-HAB website](https://arth-shukla.github.io/mshab/#dataset-section). - [ManiSkill-HAB PrepareGroceries Dataset](https://huggingface.co/datasets/arth-shukla/MS-HAB-PrepareGroceries) - [ManiSkill-HAB SetTable Dataset](https://huggingface.co/datasets/arth-shukla/MS-HAB-SetTable) ## Uses ### Direct Use This dataset can be used to train vision-based learning from demonstrations and imitation learning methods, which can be evaluated with the [MS-HAB environments](https://github.com/arth-shukla/mshab). This dataset may be useful as synthetic data for computer vision tasks as well. ### Out-of-Scope Use While blind state-based policies can be trained on this dataset, it is recommended to train vision-based policies to handle collisions and obstructions. ## Dataset Structure Each subtask/object combination has files `[SUBTASK]/[OBJECT].json` and `[SUBTASK]/[OBJECT].h5`. The JSON file contains episode metadata, event labels, etc, while the HDF5 file contains the demonstration data. ## Dataset Creation The data is gathered using [RL policies](https://huggingface.co/arth-shukla/mshab_checkpoints) fitered for safe robot behavior with a rule-based event labeling system. ## Bias, Risks, and Limitations The dataset is purely synthetic. While MS-HAB supports high-quality ray-traced rendering, this dataset uses ManiSkill's default rendering for data generation due to efficiency. However, users can generate their own data with the [data generation code](https://github.com/arth-shukla/mshab/blob/main/mshab/utils/gen/gen_data.py).