A young Steve Jobs once called computers ‘bicycles for the mind’—he was referring to the dramatic decrease in the energetic cost of transportation that could be obtained with the bicycle, which breaks all scaling laws for how efficiently an animal can achieve motion. The creativity of new approaches to old materials science problems facilitated by the dramatic uptake of machine learning (ML) is a testament to this notion. The uptake of ML in our subject has been facilitated by exceptional efforts to provide open datasets, e.g. Novel Materials Discovery (NOMAD)1, Materials Project2, Joint Automated Repository for Various Integrated Simulations (JARVIS)3, Automatic FLOW for Materials Discovery (AFLOW)4, Open Quantum Materials Database (OQMD)5 and many others, as well as the extremely high quality of openly available software packages such as scikit-learn6, PyTorch7, (Just After eXecution) JAX8, Quantum Espresso9, and Large-scale Atomic/Molecular Massively Parallel Simulator (LAMMPS)10. These resources, along with continued extraordinary developments in hardware, are helping a wide-range of researchers integrate ML into their work.
The dramatic rise in ML application started in the early 2010s. The ImageNet moment in 2012, when deep learning methods (AlexNet11), accelerated by GPUs and trained on a large, labelled dataset first outperformed traditional image classification methods is often identified as a key event. Since then, there has been a proliferation of ML methods applied to all aspects of society and research. In the sciences, the initial report of the performance of AlphaFold12 in the CASP14 contest captured the wider scientific imagination and made many researchers experts in protein folding overnight. While computational materials science has not had a moment as dramatic as AlphaFold, the field was a relatively early adopter of ML, and the prevalence of data-driven studies has steadily grown. In an ad hoc experiment (that would make a proper statistician shudder), we ranked the titles of articles from this journal from some recent years, to see how many papers appeared to be largely ML-focused. As early as 2017 we counted 7 out of 42 papers titles that had an ML flavour (~16%), in 2019 this was 28 out of 113 (~25%), this year so far that number is 49/112 (~42%). Performing this analysis it also becomes apparent just how diverse and creative the applications of ML in materials science have been. Figure 1 shows a slightly more systematic look at the rise in popularity of machine learning in materials science (and vice versa).
The bar chart shows the percentage of materials science papers including machine learning (green) and the percentage of machine learning papers including materials science (orange). The line plot shows the total number of publications for each of ‘materials science’ and ‘machine learning’ as obtained from the Web of Science. The underlying data and the notebook to generate the plot are available online (https://github.com/ML-Materials-Standards/ml-materials-checklist/tree/master/editorial-data).
In another famous aphorism credited by Francis Crick to Linus Pauling ‘The best way to have a good idea is to have lots of ideas’. However, Pauling’s quote has an important second clause ‘[…] and throw away the bad ones’. While the ease of adoption of ML techniques is broadly a welcome development, we need to be aware that these are highly-sophisticated, nuanced methods, and without careful practice one can often obtain results that seem exciting, but are actually not as useful as they appear. To take a materials science example, it is becoming common to find predictive models for materials properties that have limited domains due to their training data, and where their applicability to broader data and associated problems is highly uncertain. In a recent article published in npj Computational Materials Omee et al. demonstrate how some of the most performant modern graph neural networks suffer significant prediction drops when moved outside of this training distribution13, highlighting the need for rigorous validation in any model application.
At npj Computational Materials Science we are proud to have been at the forefront of ML adoption in materials science and we are committed to continuing to provide a widely read and respected platform for the latest and best developments in materials informatics. A key aspect of this commitment is ensuring the highest standards of scientific rigour in the material published in the journal. As new methods and technologies are rapidly adopted by large numbers of practitioners there is a well-known tendency for inflated expectations and unsubstantiated hype leading eventually to a ‘trough of despair’ period where disillusionment with the broken promises abounds. Luckily in scientific publishing, we have a strong safeguard against unchecked expectation inflation and hype; the peer review process. While many argue that peer review is not perfect (and it surely isn’t), it provides an opportunity for ideas to be strenuously tested by experts with an interest in preserving the integrity of a particular research field.
However, no matter how expert peer reviewers are, it is impossible to keep abreast of all the latest developments in a field. When we are working on an inherently multidisciplinary topic like ML in materials science, this challenge becomes even greater. It is increasingly difficult to identify individuals who are qualified to comment on all aspects of the latest research papers. In recognition of this challenge and considering our commitment to continuing to publish the best ML for materials science research, our editorial team has developed a checklist for reviewers (and authors) that can be used to guide assessments of papers with significant ML content. The guidelines that we developed are designed to be as general as possible and therefore we expect them to apply to a wide-range of datasets and ML methods. The list is available here: https://github.com/ML-Materials-Standards/ml-materials-checklist.
Our checklist is heavily indebted to existing resources14,15,16, and is aligned with general efforts in the community towards ensuring best practice17,18, we have adapted the guidelines to reflect what we see as the needs and uses of the computational materials science community. Some of the items on the checklist are certainly required, such as clear descriptions of models, data and training procedures. We expressly do not want this checklist to be seen as a barrier to publication and we do not see it as a prescription for what a paper must adhere to for publication. For example, we strongly recommend that datasets used in studies should be made open and available, however, if authors have valid reasons such as anonymity of subjects or safety concerns around sharing data then this will be considered. By making honest best efforts to follow these guidelines we hope that this will improve the quality and impact of a study rather than present a nuisance. We also do not think that these standards should only apply to ML-based studies, but striving for openness and reproducibility should be central to all computational materials science.
Finally, we recognise that this field moves fast, new technologies arrive with great frequency. As an example, the development of automated experimentation in materials synthesis and characterisation presents new challenges for reproducibility and benchmarking. These technologies have the potential to drastically change the landscape, so we don’t want to see this checklist become a dead letter. Therefore, we are treating the checklist as a living document, to be regularly assessed and updated. For this to be a truly useful and relevant resource we rely heavily on the input from our readership. In that regard, we always welcome your input and opinions. To facilitate this we are hosting the checklist as a GitHub repository, where you can comment and make suggestions on the document and we will regularly review and update the contents (https://github.com/ML-Materials-Standards/ml-materials-checklist). The leap forward facilitated by ML has been achieved because of the highest standards of openness and reproducibility and we intend to replicate that spirit in the research that we publish.
Data availability
All data associated with the plots in this paper are available at https://github.com/ML-Materials-Standards/ml-materials-checklist/tree/master/editorial-data.
Code availability
All code associated with the plots in this paper is available at https://github.com/ML-Materials-Standards/ml-materials-checklist/tree/master/editorial-data.
References
Scheidgen, M. et al. NOMAD: A distributed web-based platform for managing materials science research data. J. Open Source Softw. 8, 5388 (2023).
Jain, A. et al. Commentary: The materials project: a materials genome approach to accelerating materials innovation. APL Mater. 1, 011002 (2013).
Choudhary, K. et al. The joint automated repository for various integrated simulations (JARVIS) for data-driven materials design. npj Comput. Mater. 6, 1–13 (2020).
Curtarolo, S. et al. AFLOW: An automatic framework for high-throughput materials discovery. Comput. Mater. Sci. 58, 218–226 (2012).
Kirklin, S. et al. The Open Quantum Materials Database (OQMD): assessing the accuracy of DFT formation energies. npj Comput. Mater. 1, 1–15 (2015).
Pedregosa, F. Scikit-learn: Machine Learning in Python. J. Mach. Learn. Res. 12, 2825–2830 (2011).
Paszke, A. et al. Pytorch: An imperative style, high-performance deep learning library. Adv. neural inf. process. syst. 32, (2019).
google/jax. (Aug. 28, 2024). Python. Google. Accessed: Aug. 28, 2024. [Online]. Available: https://github.com/google/jax.
Giannozzi, P. et al. Quantum ESPRESSO: a modular and open-source software project for quantum simulations of materials. J. Phys. Condens. Matter 21, 395502 (2009).
Thompson, A. P. et al. LAMMPS—a flexible simulation tool for particle-based materials modeling at the atomic, meso, and continuum scales. Comput. Phys. Commun. 271, 108171 (2022).
Krizhevsky, A., Sutskever, I. & Hinton, G. E. ImageNet classification with deep convolutional neural networks. Commun. ACM 60, 84–90 (2017).
Jumper, J. et al. Highly accurate protein structure prediction with AlphaFold. Nature 596, 583–589 (2021).
Omee, S. S., Fu, N., Dong, R., Hu, M. & Hu, J. Structure-based out-of-distribution (OOD) materials property prediction: a benchmark study. npj Comput. Mater. 10, 1–14 (2024).
Artrith, N. et al. Best practices in machine learning for chemistry, Nat. Chem. 13 https://doi.org/10.1038/s41557-021-00716-z (2021).
Kapoor, S. et al. REFORMS: Consensus-based Recommendations for Machine-learning-based Science. Accessed: Aug. 07, [Online]. Available: https://reforms.cs.princeton.edu (2024).
Pineau, J. The Machine Learning Reproducibility Checklist. Accessed Aug 07, [Online] Available: https://www.cs.mcgill.ca/~jpineau/ReproducibilityChecklist.pdf (2024).
Wiley-VCH. Machine Learning Research Data Reporting Checklist. Accessed: Sept. 02. [Online]. Available: https://onlinelibrary.wiley.com/pb-assets/hub-assets/chemistry-europe/checklists/Machine_Learning_Data_Reporting_Checklist-1695894214660.pdf 2024).
Choudhary, K. et al. JARVIS-Leaderboard: a large scale benchmark of materials design methods. npj Comput Mater. 10, 93 https://doi.org/10.1038/s41524-024-01259-w (2024).
Acknowledgements
Certain commercial equipment, instruments, software, or materials are identified in this paper in order to specify the experimental procedure adequately. Such identifications are not intended to imply recommendation or endorsement by NIST, nor are they intended to imply that the materials or equipment identified are necessarily the best available for the purpose. The opinions stated are those of the author and do not represent an official position of NIST or the U S Government.
Author information
Authors and Affiliations
Contributions
K.T.B. conceived and wrote the first draft of the editorial. K.C., G.C., A.M.G., S.V.K. and D.M. edited, expanded and refined the manuscript subsequent to the first draft.
Corresponding author
Ethics declarations
Competing interests
The authors declare the following competing financial interests: GC has an equity interest in Symmetric Group LLP that licences force fields commercially and in Ångström AI, and the following Competing Non-Financial Interests: KC is involved in the development of benchmarking efforts for materials informatics on the JARVIS project.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Butler, K.T., Choudhary, K., Csanyi, G. et al. Setting standards for data driven materials science. npj Comput Mater 10, 231 (2024). https://doi.org/10.1038/s41524-024-01411-6
Received:
Accepted:
Published:
Version of record:
DOI: https://doi.org/10.1038/s41524-024-01411-6