Neural scaling of deep chemical models
Author(s)
Frey, Nathan C; Soklaski, Ryan; Axelrod, Simon; Samsi, Siddharth; Gómez-Bombarelli, Rafael; Coley, Connor W; Gadepally, Vijay; ... Show more Show less
DownloadPublished version (1.673Mb)
Publisher with Creative Commons License
Publisher with Creative Commons License
Creative Commons Attribution
Terms of use
Metadata
Show full item recordAbstract
Massive scale, in terms of both data availability and computation, enables important breakthroughs in key application areas of deep learning such as natural language processing and computer vision. There is emerging evidence that scale may be a key ingredient in scientific deep learning, but the importance of physical priors in scientific domains makes the strategies and benefits of scaling uncertain. Here we investigate neural-scaling behaviour in large chemical models by varying model and dataset sizes over many orders of magnitude, studying models with over one billion parameters, pre-trained on datasets of up to ten million datapoints. We consider large language models for generative chemistry and graph neural networks for machine-learned interatomic potentials. We investigate the interplay between physical priors and scale and discover empirical neural-scaling relations for language models in chemistry with a scaling exponent of 0.17 for the largest dataset size considered, and a scaling exponent of 0.26 for equivariant graph neural network interatomic potentials.
Date issued
2023Department
Lincoln Laboratory; Massachusetts Institute of Technology. Department of Materials Science and Engineering; Massachusetts Institute of Technology. Department of Chemical Engineering; Massachusetts Institute of Technology. Department of Electrical Engineering and Computer ScienceJournal
Nature Machine Intelligence
Publisher
Springer Science and Business Media LLC
Citation
Frey, N.C., Soklaski, R., Axelrod, S. et al. Neural scaling of deep chemical models. Nat Mach Intell 5, 1297–1305 (2023).
Version: Final published version