Transfer learning with graph neural networks for improved molecular property prediction in the multi-fidelity setting DOI Creative Commons
David Buterez, Jon Paul Janet, Steven J. Kiddle

et al.

Nature Communications, Journal Year: 2024, Volume and Issue: 15(1)

Published: Feb. 26, 2024

Abstract We investigate the potential of graph neural networks for transfer learning and improving molecular property prediction on sparse expensive to acquire high-fidelity data by leveraging low-fidelity measurements as an inexpensive proxy a targeted interest. This problem arises in discovery processes that rely screening funnels trading off overall costs against throughput accuracy. Typically, individual stages these are loosely connected each one generates at different scale fidelity. consider this setup holistically demonstrate empirically existing techniques generally unable harness information from multi-fidelity cascades. Here, we propose several effective strategies study them transductive inductive settings. Our analysis involves collection more than 28 million unique experimental protein-ligand interactions across 37 targets drug high-throughput 12 quantum properties dataset QMugs. The results indicate can improve performance tasks up eight times while using order magnitude less training data. Moreover, proposed methods consistently outperform graph-structured mechanics datasets.

Language: Английский

Physics-Inspired Structural Representations for Molecules and Materials DOI Creative Commons
Félix Musil, Andrea Grisafi, Albert P. Bartók

et al.

Chemical Reviews, Journal Year: 2021, Volume and Issue: 121(16), P. 9759 - 9815

Published: July 26, 2021

The first step in the construction of a regression model or data-driven analysis, aiming to predict elucidate relationship between atomic-scale structure matter and its properties, involves transforming Cartesian coordinates atoms into suitable representation. development representations has played, continues play, central role success machine-learning methods for chemistry materials science. This review summarizes current understanding nature characteristics most commonly used structural chemical descriptions atomistic structures, highlighting deep underlying connections different frameworks ideas that lead computationally efficient universally applicable models. It emphasizes link their physical chemistry, mathematical description, provides examples recent applications diverse set science problems, outlines open questions promising research directions field.

Language: Английский

Citations

439

Atomistic Line Graph Neural Network for improved materials property predictions DOI Creative Commons
Kamal Choudhary, Brian DeCost

npj Computational Materials, Journal Year: 2021, Volume and Issue: 7(1)

Published: Nov. 15, 2021

Abstract Graph neural networks (GNN) have been shown to provide substantial performance improvements for atomistic material representation and modeling compared with descriptor-based machine learning models. While most existing GNN models predictions are based on atomic distance information, they do not explicitly incorporate bond angles, which critical distinguishing many structures. Furthermore, properties known be sensitive slight changes in angles. We present an Atomistic Line Neural Network (ALIGNN), a architecture that performs message passing both the interatomic graph its line corresponding demonstrate angle information can efficiently included, leading improved multiple prediction tasks. ALIGNN predicting 52 solid-state molecular available JARVIS-DFT, Materials project, QM9 databases. outperform some previously reported tasks better or comparable model training speed.

Language: Английский

Citations

409

Graph neural networks for materials science and chemistry DOI Creative Commons
Patrick Reiser,

Marlen Neubert,

André Eberhard

et al.

Communications Materials, Journal Year: 2022, Volume and Issue: 3(1)

Published: Nov. 26, 2022

Abstract Machine learning plays an increasingly important role in many areas of chemistry and materials science, being used to predict properties, accelerate simulations, design new structures, synthesis routes materials. Graph neural networks (GNNs) are one the fastest growing classes machine models. They particular relevance for as they directly work on a graph or structural representation molecules therefore have full access all relevant information required characterize In this Review, we provide overview basic principles GNNs, widely datasets, state-of-the-art architectures, followed by discussion wide range recent applications GNNs concluding with road-map further development application GNNs.

Language: Английский

Citations

350

Artificial intelligence in drug discovery: recent advances and future perspectives DOI Creative Commons
José Jiménez-Luna, Francesca Grisoni, Nils Weskamp

et al.

Expert Opinion on Drug Discovery, Journal Year: 2021, Volume and Issue: 16(9), P. 949 - 959

Published: March 29, 2021

Introduction: Artificial intelligence (AI) has inspired computer-aided drug discovery. The widespread adoption of machine learning, in particular deep multiple scientific disciplines, and the advances computing hardware software, among other factors, continue to fuel this development. Much initial skepticism regarding applications AI pharmaceutical discovery started vanish, consequently benefitting medicinal chemistry.Areas covered: current status chemoinformatics is reviewed. topics discussed herein include quantitative structure-activity/property relationship structure-based modeling, de novo molecular design, chemical synthesis prediction. Advantages limitations learning are highlighted, together with a perspective on next-generation for discovery.Expert opinion: Deep learning-based approaches have only begun address some fundamental problems Certain methodological advances, such as message-passing models, spatial-symmetry-preserving networks, hybrid innovative paradigms, will likely become commonplace help most challenging questions. Open data sharing model development play central role advancement AI.

Language: Английский

Citations

307

SpookyNet: Learning force fields with electronic degrees of freedom and nonlocal effects DOI Creative Commons
Oliver T. Unke, Stefan Chmiela, Michael Gastegger

et al.

Nature Communications, Journal Year: 2021, Volume and Issue: 12(1)

Published: Dec. 14, 2021

Machine-learned force fields (ML-FFs) combine the accuracy of ab initio methods with efficiency conventional fields. However, current ML-FFs typically ignore electronic degrees freedom, such as total charge or spin state, and assume chemical locality, which is problematic when molecules have inconsistent states, nonlocal effects play a significant role. This work introduces SpookyNet, deep neural network for constructing explicit treatment freedom quantum nonlocality. Chemically meaningful inductive biases analytical corrections built into architecture allow it to properly model physical limits. SpookyNet improves upon state-of-the-art (or achieves similar performance) on popular chemistry data sets. Notably, able generalize across conformational space can leverage learned insights, e.g. by predicting unknown thus helping close further important remaining gap today's machine learning models in chemistry.

Language: Английский

Citations

207

Benchmarking graph neural networks for materials chemistry DOI Creative Commons
Victor Fung, Jiaxin Zhang, Eric Juarez

et al.

npj Computational Materials, Journal Year: 2021, Volume and Issue: 7(1)

Published: June 3, 2021

Abstract Graph neural networks (GNNs) have received intense interest as a rapidly expanding class of machine learning models remarkably well-suited for materials applications. To date, number successful GNNs been proposed and demonstrated systems ranging from crystal stability to electronic property prediction surface chemistry heterogeneous catalysis. However, consistent benchmark these remains lacking, hindering the development evaluation new in field. Here, we present workflow testing platform, MatDeepLearn, quickly reproducibly assessing comparing other models. We use this platform optimize evaluate selection top performing on several representative datasets computational chemistry. From our investigations note importance hyperparameter find roughly similar performances once optimized. identify strengths over conventional cases with compositionally diverse its overall flexibility respect inputs, due learned rather than defined representations. Meanwhile weaknesses are also observed including high data requirements, suggestions further improvement applications discussed.

Language: Английский

Citations

201

Neural Network Potentials: A Concise Overview of Methods DOI Open Access
Emir Kocer, Tsz Wai Ko, Jörg Behler

et al.

Annual Review of Physical Chemistry, Journal Year: 2022, Volume and Issue: 73(1), P. 163 - 186

Published: Jan. 4, 2022

In the past two decades, machine learning potentials (MLPs) have reached a level of maturity that now enables applications to large-scale atomistic simulations wide range systems in chemistry, physics, and materials science. Different algorithms been used with great success construction these MLPs. this review, we discuss an important group MLPs relying on artificial neural networks establish mapping from atomic structure potential energy. spite common feature, there are conceptual differences among MLPs, which concern dimensionality systems, inclusion long-range electrostatic interactions, global phenomena like nonlocal charge transfer, type descriptor represent structure, can be either predefined or learnable. A concise overview is given along discussion open challenges field.

Language: Английский

Citations

196

Machine Learning Methods for Small Data Challenges in Molecular Science DOI

Bozheng Dou,

Zailiang Zhu,

Ekaterina Merkurjev

et al.

Chemical Reviews, Journal Year: 2023, Volume and Issue: 123(13), P. 8736 - 8780

Published: June 29, 2023

Small data are often used in scientific and engineering research due to the presence of various constraints, such as time, cost, ethics, privacy, security, technical limitations acquisition. However, big have been focus for past decade, small their challenges received little attention, even though they technically more severe machine learning (ML) deep (DL) studies. Overall, challenge is compounded by issues, diversity, imputation, noise, imbalance, high-dimensionality. Fortunately, current era characterized technological breakthroughs ML, DL, artificial intelligence (AI), which enable data-driven discovery, many advanced ML DL technologies developed inadvertently provided solutions problems. As a result, significant progress has made decade. In this review, we summarize analyze several emerging potential molecular science, including chemical biological sciences. We review both basic algorithms, linear regression, logistic regression (LR),

Language: Английский

Citations

184

Provably efficient machine learning for quantum many-body problems DOI
Hsin-Yuan Huang, Richard Kueng, Giacomo Torlai

et al.

Science, Journal Year: 2022, Volume and Issue: 377(6613)

Published: Sept. 22, 2022

Classical machine learning (ML) provides a potentially powerful approach to solving challenging quantum many-body problems in physics and chemistry. However, the advantages of ML over traditional methods have not been firmly established. In this work, we prove that classical algorithms can efficiently predict ground-state properties gapped Hamiltonians after from other same phase matter. By contrast, under widely accepted conjecture, do learn data cannot achieve guarantee. We also classify wide range phases. Extensive numerical experiments corroborate our theoretical results variety scenarios, including Rydberg atom systems, two-dimensional random Heisenberg models, symmetry-protected topological phases, topologically ordered

Language: Английский

Citations

174

Perspective on integrating machine learning into computational chemistry and materials science DOI Open Access
Julia Westermayr, Michael Gastegger, Kristof T. Schütt

et al.

The Journal of Chemical Physics, Journal Year: 2021, Volume and Issue: 154(23)

Published: June 21, 2021

Machine learning (ML) methods are being used in almost every conceivable area of electronic structure theory and molecular simulation. In particular, ML has become firmly established the construction high-dimensional interatomic potentials. Not a day goes by without another proof principle published on how can represent predict quantum mechanical properties-be they observable, such as polarizabilities, or not, atomic charges. As is becoming pervasive simulation, we provide an overview atomistic computational modeling transformed incorporation approaches. From perspective practitioner field, assess common workflows to structure, dynamics, spectroscopy affected ML. Finally, discuss tighter lasting integration with chemistry materials science be achieved what it will mean for research practice, software development, postgraduate training.

Language: Английский

Citations

169