--- tags: - robotics - grasping - simulation - nvidia task_categories: - other license: "cc-by-4.0" --- # GraspGen: Scaling Sim2Real Grasping GraspGen is a large-scale simulated grasp dataset for multiple robot embodiments and grippers. We release over 57 million grasps, computed for a subset of 8515 objects from the [Objaverse XL](https://objaverse.allenai.org/) (LVIS) dataset. These grasps are specific to three grippers: Franka Panda, the Robotiq-2f-140 industrial gripper, and a single-contact suction gripper (30mm radius). ## Dataset Format The dataset is released in the [WebDataset](https://github.com/webdataset/webdataset) format. The folder structure of the dataset is as follows: ``` grasp_data/ franka/shard_{0-7}.tar robotiq2f140/shard_{0-7}.tar suction/shard_{0-7}.tar splits/ franka/{train/valid}_scenes.json robotiq2f140/{train/valid}_scenes.json suction/{train/valid}_scenes.json ``` We release test-train splits along with the grasp dataset. The splits are made randomly based on object instances. Each json file in the shard has the following data in a python dictionary. Note that `num_grasps=2000` per object. ``` ‘object’/ ‘scale’ # This is the scale of the asset, float ‘grasps’/ ‘object_in_gripper’ # boolean mask indicating grasp success, [num_grasps X 1] ‘transforms’ # Pose of the gripper in homogenous matrices, [num_grasps X 4 X 4] ``` The coordinate frame convention for the three grippers are provided below: ## Visualizing the dataset We have provided some minimal, standalone scripts for visualizing this dataset. See the header of the [visualize_dataset.py](scripts/visualize_dataset.py) for installation instructions. Before running any of the visualization scripts, remember to start meshcat-server in a separate terminal: ``` shell meshcat-server ``` To visualize a single object from the dataset, alongside its grasps: ```shell cd scripts/ && python visualize_dataset.py --dataset_path /path/to/dataset --object_uuid {object_uuid} --object_file /path/to/mesh --gripper_name {choose from: franka, suction, robotiq2f140} ``` To sequentially visualize a list of objects with its grasps: ```shell cd scripts/ && python visualize_dataset.py --dataset_path /path/to/dataset --uuid_list {path to a splits.json file} --uuid_object_paths_file {path to a json file mapping uuid to absolute path of meshes} --gripper_name {choose from: franka, suction, robotiq2f140} ``` ## Objaverse dataset Please download the Objaverse XL (LVIS) objects separately. See the helper script [download_objaverse.py](scripts/download_objaverse.py) for instructions and usage. Note that running this script autogenerates a file that maps from `UUID` to the asset mesh path, which you can pass in as input `uuid_object_paths_file` to the `visualize_dataset.py` script. ## License License Copyright © 2025, NVIDIA Corporation & affiliates. All rights reserved. Both the dataset and visualization code is released under a CC-BY 4.0 [License](LICENSE_DATASET). For business inquiries, please submit the form [NVIDIA Research Licensing](https://www.nvidia.com/en-us/research/inquiries/). ## Contact Please reach out to [Adithya Murali](http://adithyamurali.com) (admurali@nvidia.com) and [Clemens Eppner](https://clemense.github.io/) (ceppner@nvidia.com) for further enquiries.