Research articles using hippynn
hippynn implements a variety of methods from the research literature. Some of the earlier research was created with an older, internal implementation of HIP-NN using theano. However, the capabilities are available in hippynn.
One of the main components of hippynn is the implementaiton of HIP-NN, or the Hierarchical Interacting Particle Neural Network, was introduced in Lubbers et al. [LSB18] for the modeling of molecular energies and forces from atomistic configuration data. HIP-NN was also used to help validate results for potential energy surfaces in Suwa et al. [SSL+19] and Smith et al. [SNM+21], and was later extended to a more flexible functional form, HIP-NN with Tensor Sensitivities, or HIP-NN-TS, in Chigaev et al. [CSA+23]. Fedik et al. [FLL+24] critically examined the performance of this improved functional form for transitions states and transition path sampling. Matin et al. [MAS+24] demonstrated a method for improving the performance of potentials with respect to experiment by incorporating experimental structural data. Burrill et al. [BLT+24] showed how a linear combination of semi-empirical and machine learning models can be more powerful than either model alone. Shinkle et al. [SPB+24] demonstrated that HIP-NN can model free energies for coarse-grained models using force-matching, and that these many-body models provide improved transferability between thermodynamic states.
HIP-NN is also useful for modeling properties aside from energy/forces. It was adapted to learn charges in Nebgen et al. [NLS+18] and to learn charge predictions from dipole information in Sifain et al. [SLN+18]. Bond order regression to predict two-body quantities was explored in Magedov et al. [MKM+21]. The atom (charge) and two-body (bond) regressions were combined to build Huckel-type quantum Hamiltonians in Zubatiuk et al. [ZNL+21]. This was extended to semi-empirical Hamiltonians in Zhou et al. [ZLB+22] by combining the facilities of hippynn with another pytorch code, PYSEQM, developed by Zhou et al. [ZNL+20], which provides quantum calculations that are differentiable by pytorch.
Another avenue of work has been to model excited state dynamics with HIP-NN. In Sifain et al. [SLM+21], a localization layer was used to predict both the energy and location of singlet-triplet excitations in organic materials. In Habib et al. [HLTN23], HIP-NN was used in a dynamical setting to learn the dynamics of excitons in nanoparticles. In this mode, the predictions of a model produce inputs for the next time step, and training takes place by backpropagating through multiple steps of prediction. Li et al. [LLT+24] used the framework to predict several excited state properties; energy, transition dipole, and non-adiabatic coupling vectors were predicted for several excited states in a molecular system.
Daniel Burrill, Chang Liu, Michael Taylor, Marc Cawkwell, Nicholas Lubbers, Danny Perez, Enrique Batista, and Ping Yang. Mltb: enhancing transferability and extensibility of density functional tight binding theory with multi-body interaction corrections. chemarXiv preprint, 2024.
Michael Chigaev, Justin S Smith, Steven Anaya, Benjamin Nebgen, Matthew Bettencourt, Kipton Barros, and Nicholas Lubbers. Lightweight and effective tensor sensitivity for atomistic neural networks. The Journal of Chemical Physics, 2023.
Nikita Fedik, Wei Li, Nicholas Lubbers, Benjamin Nebgen, Sergei Tretiak, and Ying Wai Li. Challenges and opportunities for machine learning potentials in transition path sampling: alanine dipeptide and azobenzene studies. chemarXiv preprint, 2024.
Adela Habib, Nicholas Lubbers, Sergei Tretiak, and Benjamin Nebgen. Machine learning models capture plasmon dynamics in ag nanoparticles. The Journal of Physical Chemistry A, 127(17):3768–3778, 2023.
Xinyang Li, Nicholas Lubbers, Sergei Tretiak, Kipton Barros, and Yu Zhang. Machine learning framework for modeling exciton polaritons in molecular materials. Journal of Chemical Theory and Computation, 20(2):891–901, 2024.
Nicholas Lubbers, Justin S Smith, and Kipton Barros. Hierarchical modeling of molecular energies using a deep neural network. The Journal of chemical physics, 2018.
Sergey Magedov, Christopher Koh, Walter Malone, Nicholas Lubbers, and Benjamin Nebgen. Bond order predictions using deep neural networks. Journal of Applied Physics, 2021.
Sakib Matin, Alice EA Allen, Justin Smith, Nicholas Lubbers, Ryan B Jadrich, Richard Messerly, Benjamin Nebgen, Ying Wai Li, Sergei Tretiak, and Kipton Barros. Machine learning potentials with the iterative boltzmann inversion: training to experiment. Journal of Chemical Theory and Computation, 20(3):1274–1281, 2024.
Benjamin Nebgen, Nicholas Lubbers, Justin S Smith, Andrew E Sifain, Andrey Lokhov, Olexandr Isayev, Adrian E Roitberg, Kipton Barros, and Sergei Tretiak. Transferable dynamic molecular charge assignment using deep neural networks. Journal of chemical theory and computation, 14(9):4687–4698, 2018.
Emily Shinkle, Aleksandra Pachalieva, Riti Bahl, Sakib Matin, Brendan Gifford, Galen T Craven, and Nicholas Lubbers. Thermodynamic transferability in coarse-grained force fields using graph neural networks. arXiv preprint arXiv:2406.12112, 2024.
Andrew E Sifain, Nicholas Lubbers, Benjamin T Nebgen, Justin S Smith, Andrey Y Lokhov, Olexandr Isayev, Adrian E Roitberg, Kipton Barros, and Sergei Tretiak. Discovering a transferable charge assignment model using machine learning. The journal of physical chemistry letters, 9(16):4495–4501, 2018.
Andrew E Sifain, Levi Lystrom, Richard A Messerly, Justin S Smith, Benjamin Nebgen, Kipton Barros, Sergei Tretiak, Nicholas Lubbers, and Brendan J Gifford. Predicting phosphorescence energies and inferring wavefunction localization with machine learning. Chemical Science, 12(30):10207–10217, 2021.
Justin S Smith, Benjamin Nebgen, Nithin Mathew, Jie Chen, Nicholas Lubbers, Leonid Burakovsky, Sergei Tretiak, Hai Ah Nam, Timothy Germann, Saryu Fensin, and others. Automated discovery of a robust interatomic potential for aluminum. Nature communications, 12(1):1257, 2021.
Hidemaro Suwa, Justin S Smith, Nicholas Lubbers, Cristian D Batista, Gia-Wei Chern, and Kipton Barros. Machine learning for molecular dynamics with strongly correlated electrons. Physical Review B, 99(16):161107, 2019.
Guoqing Zhou, Nicholas Lubbers, Kipton Barros, Sergei Tretiak, and Benjamin Nebgen. Deep learning of dynamically responsive chemical hamiltonians with semiempirical quantum mechanics. Proceedings of the National Academy of Sciences, 119(27):e2120333119, 2022.
Guoqing Zhou, Ben Nebgen, Nicholas Lubbers, Walter Malone, Anders MN Niklasson, and Sergei Tretiak. Graphics processing unit-accelerated semiempirical born oppenheimer molecular dynamics using pytorch. Journal of Chemical Theory and Computation, 16(8):4951–4962, 2020.
Tetiana Zubatiuk, Benjamin Nebgen, Nicholas Lubbers, Justin S Smith, Roman Zubatyuk, Guoqing Zhou, Christopher Koh, Kipton Barros, Olexandr Isayev, and Sergei Tretiak. Machine learned hückel theory: interfacing physics and deep neural networks. The Journal of Chemical Physics, 2021.