Javascript must be enabled to continue!
AABBA: Atom–Atom Bond–Bond Bond–Atom Graph Kernel for Machine Learning on Molecules and Materials
View through CrossRef
Graphs are one of the most natural and powerful representations available for molecules; natural because they have an intuitive correspondence to skeletal formulas, the language used by chemists worldwide, and powerful, because they are highly expressive both globally (molecular topology) and locally (atomic properties). Graph kernels are used to transform molecular graphs into fixed-length vectors, which can be used as fingerprints in machine learning (ML) models. To date, kernels have mostly focused on the atomic nodes of the graph. In this work, we developed an extended graph kernel computing atom–atom, bond–bond, and bond–atom (AABBA) autocorrelations. The resulting AABBA representations were evaluated with a transition metal complex benchmark, motivated by the higher complexity of these compounds relative to organic molecules. In particular, we tested different flavors of the AABBA kernel in the prediction of the energy barriers and bond distances of the Vaska’s complex dataset (Friederich et al., Chem. Sci., 2020, 11, 4584). For a variety of ML models, including neural networks, gradient boosting machines, and Gaussian processes, we showed that AABBA outperforms the baseline including only atom–atom autocorrelations. Dimensionality reduction studies also showed that the bond–bond and bond–atom autocorrelations yield many of the most relevant features. We believe that the AABBA graph kernel can accelerate the discovery of chemical compounds and inspire novel molecular representations in which both atomic and bond properties play an important role.
American Chemical Society (ACS)
Title: AABBA: Atom–Atom Bond–Bond Bond–Atom Graph Kernel for Machine Learning on Molecules and Materials
Description:
Graphs are one of the most natural and powerful representations available for molecules; natural because they have an intuitive correspondence to skeletal formulas, the language used by chemists worldwide, and powerful, because they are highly expressive both globally (molecular topology) and locally (atomic properties).
Graph kernels are used to transform molecular graphs into fixed-length vectors, which can be used as fingerprints in machine learning (ML) models.
To date, kernels have mostly focused on the atomic nodes of the graph.
In this work, we developed an extended graph kernel computing atom–atom, bond–bond, and bond–atom (AABBA) autocorrelations.
The resulting AABBA representations were evaluated with a transition metal complex benchmark, motivated by the higher complexity of these compounds relative to organic molecules.
In particular, we tested different flavors of the AABBA kernel in the prediction of the energy barriers and bond distances of the Vaska’s complex dataset (Friederich et al.
, Chem.
Sci.
, 2020, 11, 4584).
For a variety of ML models, including neural networks, gradient boosting machines, and Gaussian processes, we showed that AABBA outperforms the baseline including only atom–atom autocorrelations.
Dimensionality reduction studies also showed that the bond–bond and bond–atom autocorrelations yield many of the most relevant features.
We believe that the AABBA graph kernel can accelerate the discovery of chemical compounds and inspire novel molecular representations in which both atomic and bond properties play an important role.
Related Results
The AABBA Graph Kernel: Atom–Atom, Bond–Bond, and Bond–Atom Autocorrelations for Machine Learning
The AABBA Graph Kernel: Atom–Atom, Bond–Bond, and Bond–Atom Autocorrelations for Machine Learning
Graphs are one of the most natural and powerful representations available for molecules; natural because they have an intuitive correspondence to skeletal formulas, the language us...
Physicochemical Properties of Wheat Fractionated by Wheat Kernel Thickness and Separated by Kernel Specific Density
Physicochemical Properties of Wheat Fractionated by Wheat Kernel Thickness and Separated by Kernel Specific Density
ABSTRACTTwo wheat cultivars, soft white winter wheat Yang‐mai 11 and hard white winter wheat Zheng‐mai 9023, were fractionated by kernel thickness into five sections; the fractiona...
Genetic Variation in Potential Kernel Size Affects Kernel Growth and Yield of Sorghum
Genetic Variation in Potential Kernel Size Affects Kernel Growth and Yield of Sorghum
Large‐seededness can increase grain yield in sorghum [Sorghum bicolor (L.) Moench] if larger kernel size more than compensates for the associated reduction in kernel number. The ai...
Sorghum Kernel Weight
Sorghum Kernel Weight
The influence of genotype and panicle position on sorghum [Sorghum bicolor (L.) Moench] kernel growth is poorly understood. In the present study, sorghum kernel weight (KW) differe...
Isolation Graph Kernel
Isolation Graph Kernel
A recent Wasserstein Weisfeiler-Lehman (WWL) Graph Kernel has a distinctive feature: Representing the distribution of Weisfeiler-Lehman (WL)-embedded node vectors of a graph in a h...
Machine Learning the Hydrogen Adsorption Capacity of Metal Organic Frameworks
Machine Learning the Hydrogen Adsorption Capacity of Metal Organic Frameworks
High-throughput virtual screening and machine learning (ML) are powerful tools for accelerating the discovery of nanoporous adsorbents for gas storage applications, including metal...
Abstract 902: Explainable AI: Graph machine learning for response prediction and biomarker discovery
Abstract 902: Explainable AI: Graph machine learning for response prediction and biomarker discovery
Abstract
Accurately predicting drug sensitivity and understanding what is driving it are major challenges in drug discovery. Graphs are a natural framework for captu...
Domination of Polynomial with Application
Domination of Polynomial with Application
In this paper, .We .initiate the study of domination. polynomial , consider G=(V,E) be a simple, finite, and directed graph without. isolated. vertex .We present a study of the Ira...

