Research articles using hippynn

hippynn implements a variety of methods from the research literature. Some of the earlier research was created with an older, internal implementation of HIP-NN using theano. However, the capabilities are available in hippynn.

One of the main components of hippynn is the implementaiton of HIP-NN, or the Hierarchical Interacting Particle Neural Network, was introduced in Lubbers et al. [LSB18] for the modeling of molecular energies and forces from atomistic configuration data. HIP-NN was also used to help validate results for potential energy surfaces in Suwa et al. [SSL+19] and Smith et al. [SNM+21], and was later extended to a more flexible functional form, HIP-NN with Tensor Sensitivities, or HIP-NN-TS, in Chigaev et al. [CSA+23]. Fedik et al. [FLL+24] critically examined the performance of this improved functional form for transitions states and transition path sampling. Matin et al. [MAS+24] demonstrated a method for improving the performance of potentials with respect to experiment by incorporating experimental structural data. Burrill et al. [BLT+24] showed how a linear combination of semi-empirical and machine learning models can be more powerful than either model alone. Shinkle et al. [SPB+24] demonstrated that HIP-NN can model free energies for coarse-grained models using force-matching, and that these many-body models provide improved transferability between thermodynamic states.

HIP-NN is also useful for modeling properties aside from energy/forces. It was adapted to learn charges in Nebgen et al. [NLS+18] and to learn charge predictions from dipole information in Sifain et al. [SLN+18]. Bond order regression to predict two-body quantities was explored in Magedov et al. [MKM+21]. The atom (charge) and two-body (bond) regressions were combined to build Huckel-type quantum Hamiltonians in Zubatiuk et al. [ZNL+21]. This was extended to semi-empirical Hamiltonians in Zhou et al. [ZLB+22] by combining the facilities of hippynn with another pytorch code, PYSEQM, developed by Zhou et al. [ZNL+20], which provides quantum calculations that are differentiable by pytorch.

Another avenue of work has been to model excited state dynamics with HIP-NN. In Sifain et al. [SLM+21], a localization layer was used to predict both the energy and location of singlet-triplet excitations in organic materials. In Habib et al. [HLTN23], HIP-NN was used in a dynamical setting to learn the dynamics of excitons in nanoparticles. In this mode, the predictions of a model produce inputs for the next time step, and training takes place by backpropagating through multiple steps of prediction. :cite:t`li2024machine` used the framework to predict several excited state properties; energy, transition dipole, and non-adiabatic coupling vectors were predicted for several excited states in a molecular system.

[BLT+24]

Daniel Burrill, Chang Liu, Michael Taylor, Marc Cawkwell, Nicholas Lubbers, Danny Perez, Enrique Batista, and Ping Yang. Mltb: enhancing transferability and extensibility of density functional tight binding theory with multi-body interaction corrections. chemarXiv preprint, 2024.

[CSA+23]

Michael Chigaev, Justin S Smith, Steven Anaya, Benjamin Nebgen, Matthew Bettencourt, Kipton Barros, and Nicholas Lubbers. Lightweight and effective tensor sensitivity for atomistic neural networks. The Journal of Chemical Physics, 2023.

[FLL+24]

Nikita Fedik, Wei Li, Nicholas Lubbers, Benjamin Nebgen, Sergei Tretiak, and Ying Wai Li. Challenges and opportunities for machine learning potentials in transition path sampling: alanine dipeptide and azobenzene studies. chemarXiv preprint, 2024.

[HLTN23]

Adela Habib, Nicholas Lubbers, Sergei Tretiak, and Benjamin Nebgen. Machine learning models capture plasmon dynamics in ag nanoparticles. The Journal of Physical Chemistry A, 127(17):3768–3778, 2023.

[LSB18]

Nicholas Lubbers, Justin S Smith, and Kipton Barros. Hierarchical modeling of molecular energies using a deep neural network. The Journal of chemical physics, 2018.

[MKM+21]

Sergey Magedov, Christopher Koh, Walter Malone, Nicholas Lubbers, and Benjamin Nebgen. Bond order predictions using deep neural networks. Journal of Applied Physics, 2021.

[MAS+24]

Sakib Matin, Alice EA Allen, Justin Smith, Nicholas Lubbers, Ryan B Jadrich, Richard Messerly, Benjamin Nebgen, Ying Wai Li, Sergei Tretiak, and Kipton Barros. Machine learning potentials with the iterative boltzmann inversion: training to experiment. Journal of Chemical Theory and Computation, 20(3):1274–1281, 2024.

[NLS+18]

Benjamin Nebgen, Nicholas Lubbers, Justin S Smith, Andrew E Sifain, Andrey Lokhov, Olexandr Isayev, Adrian E Roitberg, Kipton Barros, and Sergei Tretiak. Transferable dynamic molecular charge assignment using deep neural networks. Journal of chemical theory and computation, 14(9):4687–4698, 2018.

[SPB+24]

Emily Shinkle, Aleksandra Pachalieva, Riti Bahl, Sakib Matin, Brendan Gifford, Galen T Craven, and Nicholas Lubbers. Thermodynamic transferability in coarse-grained force fields using graph neural networks. arXiv preprint arXiv:2406.12112, 2024.

[SLN+18]

Andrew E Sifain, Nicholas Lubbers, Benjamin T Nebgen, Justin S Smith, Andrey Y Lokhov, Olexandr Isayev, Adrian E Roitberg, Kipton Barros, and Sergei Tretiak. Discovering a transferable charge assignment model using machine learning. The journal of physical chemistry letters, 9(16):4495–4501, 2018.

[SLM+21]

Andrew E Sifain, Levi Lystrom, Richard A Messerly, Justin S Smith, Benjamin Nebgen, Kipton Barros, Sergei Tretiak, Nicholas Lubbers, and Brendan J Gifford. Predicting phosphorescence energies and inferring wavefunction localization with machine learning. Chemical Science, 12(30):10207–10217, 2021.

[SNM+21]

Justin S Smith, Benjamin Nebgen, Nithin Mathew, Jie Chen, Nicholas Lubbers, Leonid Burakovsky, Sergei Tretiak, Hai Ah Nam, Timothy Germann, Saryu Fensin, and others. Automated discovery of a robust interatomic potential for aluminum. Nature communications, 12(1):1257, 2021.

[SSL+19]

Hidemaro Suwa, Justin S Smith, Nicholas Lubbers, Cristian D Batista, Gia-Wei Chern, and Kipton Barros. Machine learning for molecular dynamics with strongly correlated electrons. Physical Review B, 99(16):161107, 2019.

[ZLB+22]

Guoqing Zhou, Nicholas Lubbers, Kipton Barros, Sergei Tretiak, and Benjamin Nebgen. Deep learning of dynamically responsive chemical hamiltonians with semiempirical quantum mechanics. Proceedings of the National Academy of Sciences, 119(27):e2120333119, 2022.

[ZNL+20]

Guoqing Zhou, Ben Nebgen, Nicholas Lubbers, Walter Malone, Anders MN Niklasson, and Sergei Tretiak. Graphics processing unit-accelerated semiempirical born oppenheimer molecular dynamics using pytorch. Journal of Chemical Theory and Computation, 16(8):4951–4962, 2020.

[ZNL+21]

Tetiana Zubatiuk, Benjamin Nebgen, Nicholas Lubbers, Justin S Smith, Roman Zubatyuk, Guoqing Zhou, Christopher Koh, Kipton Barros, Olexandr Isayev, and Sergei Tretiak. Machine learned hückel theory: interfacing physics and deep neural networks. The Journal of Chemical Physics, 2021.