Machine-learning coupled cluster properties through a density tensor representation

Author(s):  
Ben Peyton
2020 ◽  
Vol 124 (23) ◽  
pp. 4861-4871 ◽  
Author(s):  
Benjamin G. Peyton ◽  
Connor Briggs ◽  
Ruhee D’Cunha ◽  
Johannes T. Margraf ◽  
T. Daniel Crawford

2020 ◽  
Author(s):  
Benjamin Peyton ◽  
Connor Briggs ◽  
Ruhee D'Cunha ◽  
Johannes T. Margraf ◽  
Thomas Crawford

<div> <div> <div> <p>The introduction of machine-learning (ML) algorithms to quantum mechanics enables rapid evaluation of otherwise intractable expressions at the cost of prior training on appropriate benchmarks. Many computational bottlenecks in the evaluation of accurate electronic structure theory could potentially benefit from the application of such models, from reducing the complexity of the underlying wave function parameter space to circumventing the complications of solving the electronic Schrödinger equation entirely. Applications of ML to electronic structure have thus far been focused on learning molecular properties (mainly the energy) from geometric representations. While this line of study has been quite successful, highly accurate models typically require a “big data” approach with thousands of train- ing data points. Herein, we propose a general, systematically improvable scheme for wave function-based ML of arbitrary molecular properties, inspired by the underlying equations that govern the canonical approach to computing the properties. To this end, we combine the established ML machinery of the t-amplitude tensor representation with a new reduced density matrix representation. The resulting model provides quantitative accuracy in both the electronic energy and dipoles of small molecules using only a few dozen training points per system. </p> </div> </div> </div>


2020 ◽  
Author(s):  
Benjamin Peyton ◽  
Connor Briggs ◽  
Ruhee D'Cunha ◽  
Johannes T. Margraf ◽  
Thomas Crawford

<div> <div> <div> <p>The introduction of machine-learning (ML) algorithms to quantum mechanics enables rapid evaluation of otherwise intractable expressions at the cost of prior training on appropriate benchmarks. Many computational bottlenecks in the evaluation of accurate electronic structure theory could potentially benefit from the application of such models, from reducing the complexity of the underlying wave function parameter space to circumventing the complications of solving the electronic Schrödinger equation entirely. Applications of ML to electronic structure have thus far been focused on learning molecular properties (mainly the energy) from geometric representations. While this line of study has been quite successful, highly accurate models typically require a “big data” approach with thousands of train- ing data points. Herein, we propose a general, systematically improvable scheme for wave function-based ML of arbitrary molecular properties, inspired by the underlying equations that govern the canonical approach to computing the properties. To this end, we combine the established ML machinery of the t-amplitude tensor representation with a new reduced density matrix representation. The resulting model provides quantitative accuracy in both the electronic energy and dipoles of small molecules using only a few dozen training points per system. </p> </div> </div> </div>


2021 ◽  
Vol 8 (1) ◽  
Author(s):  
Alexander G. Donchev ◽  
Andrew G. Taube ◽  
Elizabeth Decolvenaere ◽  
Cory Hargus ◽  
Robert T. McGibbon ◽  
...  

AbstractAdvances in computational chemistry create an ongoing need for larger and higher-quality datasets that characterize noncovalent molecular interactions. We present three benchmark collections of quantum mechanical data, covering approximately 3,700 distinct types of interacting molecule pairs. The first collection, which we refer to as DES370K, contains interaction energies for more than 370,000 dimer geometries. These were computed using the coupled-cluster method with single, double, and perturbative triple excitations [CCSD(T)], which is widely regarded as the gold-standard method in electronic structure theory. Our second benchmark collection, a core representative subset of DES370K called DES15K, is intended for more computationally demanding applications of the data. Finally, DES5M, our third collection, comprises interaction energies for nearly 5,000,000 dimer geometries; these were calculated using SNS-MP2, a machine learning approach that provides results with accuracy comparable to that of our coupled-cluster training data. These datasets may prove useful in the development of density functionals, empirically corrected wavefunction-based approaches, semi-empirical methods, force fields, and models trained using machine learning methods.


2020 ◽  
Author(s):  
Dakota Folmsbee ◽  
Geoffrey Hutchison

We have performed a large-scale evaluation of current computational methods, including conventional small-molecule force fields, semiempirical, density functional, ab initio electronic structure methods, and current machine learning (ML) techniques to evaluate relative single-point energies. Using up to 10 local minima geometries across ~700 molecules, each optimized by B3LYP-D3BJ with single-point DLPNO-CCSD(T) triple-zeta energies, we consider over 6,500 single points to compare the correlation between different methods for both relative energies and ordered rankings of minima. We find promise from current ML methods and recommend methods at each tier of the accuracy-time tradeoff, particularly the recent GFN2 semiempirical method, the B97-3c density functional approximation, and RI-MP2 for accurate conformer energies. The ANI family of ML methods shows promise, particularly the ANI-1ccx variant trained in part on coupled-cluster energies. Multiple methods suggest continued improvements should be expected in both performance and accuracy.


2020 ◽  
Author(s):  
Dakota Folmsbee ◽  
Geoffrey Hutchison

We have performed a large-scale evaluation of current computational methods, including conventional small-molecule force fields, semiempirical, density functional, ab initio electronic structure methods, and current machine learning (ML) techniques to evaluate relative single-point energies. Using up to 10 local minima geometries across ~700 molecules, each optimized by B3LYP-D3BJ with single-point DLPNO-CCSD(T) triple-zeta energies, we consider over 6,500 single points to compare the correlation between different methods for both relative energies and ordered rankings of minima. We find promise from current ML methods and recommend methods at each tier of the accuracy-time tradeoff, particularly the recent GFN2 semiempirical method, the B97-3c density functional approximation, and RI-MP2 for accurate conformer energies. The ANI family of ML methods shows promise, particularly the ANI-1ccx variant trained in part on coupled-cluster energies. Multiple methods suggest continued improvements should be expected in both performance and accuracy.


2020 ◽  
Vol 34 (04) ◽  
pp. 4527-4534
Author(s):  
Sören Laue ◽  
Matthias Mitterreiter ◽  
Joachim Giesen

Computing derivatives of tensor expressions, also known as tensor calculus, is a fundamental task in machine learning. A key concern is the efficiency of evaluating the expressions and their derivatives that hinges on the representation of these expressions. Recently, an algorithm for computing higher order derivatives of tensor expressions like Jacobians or Hessians has been introduced that is a few orders of magnitude faster than previous state-of-the-art approaches. Unfortunately, the approach is based on Ricci notation and hence cannot be incorporated into automatic differentiation frameworks like TensorFlow, PyTorch, autograd, or JAX that use the simpler Einstein notation. This leaves two options, to either change the underlying tensor representation in these frameworks or to develop a new, provably correct algorithm based on Einstein notation. Obviously, the first option is impractical. Hence, we pursue the second option. Here, we show that using Ricci notation is not necessary for an efficient tensor calculus and develop an equally efficient method for the simpler Einstein notation. It turns out that turning to Einstein notation enables further improvements that lead to even better efficiency.


Sign in / Sign up

Export Citation Format

Share Document