MatterGen
MatterGen is a generative model for inorganic materials design.
Model Details
Model Description
MatterGen is a generative model for inorganic materials design. It is a diffusion model which jointly predicts a material’s atomic fractional coordinates, elements, as well as unit cell lattice vectors. Besides unconditional generation of material candidates, MatterGen can also be trained or fine-tuned for conditional generation based on target property values, e.g., bulk modulus, chemical system, or magnetic density.
- Developed by: Materials Design Team, Microsoft Research AI for Science
- Model type: Diffusion model
- License: MIT
Model Sources
- Repository: https://github.com/microsoft/mattergen
- Paper: https://www.nature.com/articles/s41586-025-08628-5
Uses
Direct Use
- Generate inorganic materials candidates without property condition.
- Fine-tune the base model on user-provided data with property-labeled materials.
- Generate inorganic materials candidates with target property, e.g., bulk modulus, chemical system, magnetic density, or user-defined target properties after fine-tuning.
Out-of-Scope Use
- Generate materials with more than 20 atoms inside the unit cell.
- Generate organic crystals or non-crystalline materials.
- Generate crystals containing noble gas elements, radioactive elements, or elements with atomic number greater than 84 – these elements were removed from the training data.
Bias, Risks, and Limitations
MatterGen was only trained on and evaluated on up to 20 atoms inside the unit cell; more atoms are currently not supported. MatterGen’s training data is materials below 0.1 eV/atoms below the reference convex hull. Therefore, it is expected that the fraction of generated materials on or below the convex hull is significantly lower than the fraction of materials within 0.1 eV/atom above the convex hull.
Recommendations
The performance on property-guided generation heavily depends on the quality and quantity of the property labels used to train MatterGen. For extreme property values where there are few training structures with similar values, the performance may degrade.
For fine-tuning the model on a new property, use a sufficient amount of labeled property data for training, i.e., at least several thousands of labeled structures. Also ensure good coverage of property values in the range of values which are intended for property-guided generation.
How to Get Started with the Model
Clone the repository and follow the README instructions.
Training Details
Training Data
MatterGen was trained on crystalline materials from the following data sources: 1. MP (https://next-gen.materialsproject.org/; v2022.10.28, Creative Commons Attribution 4.0 International License), an open-access resource containing DFT-relaxed crystal structures obtained from a variety of sources, but largely based upon experimentally-known crystals. 2. The Alexandria dataset (https://alexandria.icams.rub.de/; Creative Commons Attribution 4.0 International License), an open-access resource containing DFT-relaxed crystal structures from a variety of sources, including a large quantity of hypothetical crystal structures generated by ML methods or other algorithmic means. To train MatterGen, we select only structures with up to 20 atoms and whose energy above hull is below 0.1 eV/atom. Further, we remove structures that contain noble gas elements, elements with atomic number higher than 84 (which includes most radioactive elements), or the radioactive elements “Tc” and “Pm” from the training data. For more information, see paper, Supplementary C.1.
Training Procedure
Preprocessing
We relax structures from the above data sources with DFT and select only those structures whose energy above the combined convex hull is below 0.1 eV/atom. MatterGen is trained solely on primitive structures. We further select only structures with up to 20 atoms inside the unit cell. We use the Niggli reduction to preprocess the unit cell lattices, followed by the polar decomposition to ensure the lattice matrices are symmetric matrices. See the paper for more detailed information.
Training Hyperparameters
- Starting learning rate 1e-4, reduces successively by a factor of 0.6 when training loss does not reduce within 100 epochs, up to 1e-6.
- Batch size 512
- float32 precision
Speeds, Sizes, Times
- MatterGen contains 46.8M parameters
- One training epoch of around 600K training samples takes around 6 minutes on 8 NVIDIA A100 GPUs
- Sampling 1,000 structures takes around two hours using a single NVIDIA V100 GPU
Evaluation
Testing Data, Factors & Metrics
Metrics
MatterGen was evaluated on unconditional generation across the following metrics:
- The percentage of stable, novel, and unique (S.U.N.) structures among 1,024 generated samples.
- Stable means a structure’s energy is less than 0.1 eV/atom above the reference convex hull
- Novel means a structure does not match any structure in our reference dataset with the disordered structure matcher presented in the paper.
- Unique means that there is no other structure among the generated ones which matches a given structure.
- The average root mean square distance (RMSD) of generated structures and their DFT-relaxed local energy minima, measured in Angstrom.
Results
MatterGen achieves 38.57 % S.U.N. rate among generated structures, and the average RMSD of its samples is 0.021 Angstrom. For more details see Section 2.2 of the MatterGen paper. We also evaluate MatterGen on property-conditioned generation. • For generation conditioned on chemical system, MatterGen produces 83 % S.U.N. structures on well-explored chemical systems, 65 % on partially explored systems, and 49 % on unexplored chemical systems. For more details, see Section 2.3 of the MatterGen paper. • Conditioning on a bulk modulus value of 400 GPa, MatterGen produces 106 S.U.N. structures with > 400 GPa bulk modulus given a budget of 180 DFT property calculations. For more details, see Section 2.4 of the MatterGen paper. • Conditioning on magnetic density of > 0.2 Angstrom-3, MatterGen produces 18 S.U.N. structures complying with the condition given a budget of 180 DFT property calculations. For more details, see Section 2.4 of the MatterGen paper.
Summary
MatterGen is able to produce novel, unique, and stable material candidates both with and without property conditions. For property-guided generation, MatterGen is able to produce S.U.N. structures with extreme property values such as 400 GPa bulk modulus, where there are only two such structures in the labeled reference set. MatterGen outperforms both classical as well as recent deep generative model baselines. For more details on the performance of MatterGen, see the paper.
Technical Specifications
Model Architecture and Objective
The model architecture is based on GemNet (Gasteiger et al. 2021).
Citation
BibTeX:
@article{MatterGen2025,
author = {Zeni, Claudio and Pinsler, Robert and Z{\"u}gner, Daniel and Fowler, Andrew and Horton, Matthew and Fu, Xiang and Wang, Zilong and Shysheya, Aliaksandra and Crabb{\'e}, Jonathan and Ueda, Shoko and Sordillo, Roberto and Sun, Lixin and Smith, Jake and Nguyen, Bichlien and Schulz, Hannes and Lewis, Sarah and Huang, Chin-Wei and Lu, Ziheng and Zhou, Yichi and Yang, Han and Hao, Hongxia and Li, Jielan and Yang, Chunlei and Li, Wenjie and Tomioka, Ryota and Xie, Tian},
journal = {Nature},
title = {A generative model for inorganic materials design},
year = {2025},
doi = {10.1038/s41586-025-08628-5},
}
Model Card Authors
Daniel Zügner ([email protected])
Model Card Contact
Daniel Zügner ([email protected]) Tian Xie ([email protected])