Predicting CO$_2$ Absorption in Ionic Liquids with Molecular Descriptors
and Explainable Graph Neural Networks
- URL: http://arxiv.org/abs/2210.01120v1
- Date: Thu, 29 Sep 2022 18:31:12 GMT
- Title: Predicting CO$_2$ Absorption in Ionic Liquids with Molecular Descriptors
and Explainable Graph Neural Networks
- Authors: Yue Jian, Yuyang Wang, Amir Barati Farimani
- Abstract summary: Liquids (ILs) provide a promising solution for CO$$ capture and storage to mitigate global warming.
In this work, we develop both fingerprint-based Machine Learning models and Graph Neural Networks (GNNs) to predict the CO$$ in ILs.
Our method outperforms previous ML models by reaching a high accuracy (MAE of 0.0137, $R2$ of 0.9884)
- Score: 9.04563945965023
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Ionic Liquids (ILs) provide a promising solution for CO$_2$ capture and
storage to mitigate global warming. However, identifying and designing the
high-capacity IL from the giant chemical space requires expensive, and
exhaustive simulations and experiments. Machine learning (ML) can accelerate
the process of searching for desirable ionic molecules through accurate and
efficient property predictions in a data-driven manner. But existing
descriptors and ML models for the ionic molecule suffer from the inefficient
adaptation of molecular graph structure. Besides, few works have investigated
the explainability of ML models to help understand the learned features that
can guide the design of efficient ionic molecules. In this work, we develop
both fingerprint-based ML models and Graph Neural Networks (GNNs) to predict
the CO$_2$ absorption in ILs. Fingerprint works on graph structure at the
feature extraction stage, while GNNs directly handle molecule structure in both
the feature extraction and model prediction stage. We show that our method
outperforms previous ML models by reaching a high accuracy (MAE of 0.0137,
$R^2$ of 0.9884). Furthermore, we take the advantage of GNNs feature
representation and develop a substructure-based explanation method that
provides insight into how each chemical fragments within IL molecules
contribute to the CO$_2$ absorption prediction of ML models. We also show that
our explanation result agrees with some ground truth from the theoretical
reaction mechanism of CO$_2$ absorption in ILs, which can advise on the design
of novel and efficient functional ILs in the future.
Related papers
- Atomic and Subgraph-aware Bilateral Aggregation for Molecular
Representation Learning [57.670845619155195]
We introduce a new model for molecular representation learning called the Atomic and Subgraph-aware Bilateral Aggregation (ASBA)
ASBA addresses the limitations of previous atom-wise and subgraph-wise models by incorporating both types of information.
Our method offers a more comprehensive way to learn representations for molecular property prediction and has broad potential in drug and material discovery applications.
arXiv Detail & Related papers (2023-05-22T00:56:00Z) - Gibbs-Helmholtz Graph Neural Network: capturing the temperature
dependency of activity coefficients at infinite dilution [1.290382979353427]
We develop the Gibbs-Helmholtz Graph Neural Network (GH-GNN) model for predicting $ln gamma_ijinfty$ of molecular systems at different temperatures.
We analyze the performance of GH-GNN for continuous and discrete inter/extrapolation and give indications for the model's applicability domain and expected accuracy.
arXiv Detail & Related papers (2022-12-02T14:25:58Z) - Efficient Chemical Space Exploration Using Active Learning Based on
Marginalized Graph Kernel: an Application for Predicting the Thermodynamic
Properties of Alkanes with Molecular Simulation [10.339394156446982]
We use molecular dynamics simulation to generate data and graph neural network (GNN) to predict.
In specific, targeting 251,728 alkane molecules consisting of 4 to 19 carbon atoms and their liquid physical properties.
validation shows that only 313 molecules were sufficient to train an accurate GNN model with $rm R2 > 0.99$ for computational test sets and $rm R2 > 0.94$ for experimental test sets.
arXiv Detail & Related papers (2022-09-01T14:59:13Z) - HiGNN: Hierarchical Informative Graph Neural Networks for Molecular
Property Prediction Equipped with Feature-Wise Attention [5.735627221409312]
We propose a well-designed hierarchical informative graph neural networks framework (termed HiGNN) for predicting molecular property.
Experiments demonstrate that HiGNN achieves state-of-the-art predictive performance on many challenging drug discovery-associated benchmark datasets.
arXiv Detail & Related papers (2022-08-30T05:16:15Z) - Graph neural networks for the prediction of molecular structure-property
relationships [59.11160990637615]
Graph neural networks (GNNs) are a novel machine learning method that directly work on the molecular graph.
GNNs allow to learn properties in an end-to-end fashion, thereby avoiding the need for informative descriptors.
We describe the fundamentals of GNNs and demonstrate the application of GNNs via two examples for molecular property prediction.
arXiv Detail & Related papers (2022-07-25T11:30:44Z) - Chemical-Reaction-Aware Molecule Representation Learning [88.79052749877334]
We propose using chemical reactions to assist learning molecule representation.
Our approach is proven effective to 1) keep the embedding space well-organized and 2) improve the generalization ability of molecule embeddings.
Experimental results demonstrate that our method achieves state-of-the-art performance in a variety of downstream tasks.
arXiv Detail & Related papers (2021-09-21T00:08:43Z) - Do Large Scale Molecular Language Representations Capture Important
Structural Information? [31.76876206167457]
We present molecular embeddings obtained by training an efficient transformer encoder model, referred to as MoLFormer.
Experiments show that the learned molecular representation performs competitively, when compared to graph-based and fingerprint-based supervised learning baselines.
arXiv Detail & Related papers (2021-06-17T14:33:55Z) - BIGDML: Towards Exact Machine Learning Force Fields for Materials [55.944221055171276]
Machine-learning force fields (MLFF) should be accurate, computationally and data efficient, and applicable to molecules, materials, and interfaces thereof.
Here, we introduce the Bravais-Inspired Gradient-Domain Machine Learning approach and demonstrate its ability to construct reliable force fields using a training set with just 10-200 atoms.
arXiv Detail & Related papers (2021-06-08T10:14:57Z) - Self-Supervised Graph Transformer on Large-Scale Molecular Data [73.3448373618865]
We propose a novel framework, GROVER, for molecular representation learning.
GROVER can learn rich structural and semantic information of molecules from enormous unlabelled molecular data.
We pre-train GROVER with 100 million parameters on 10 million unlabelled molecules -- the biggest GNN and the largest training dataset in molecular representation learning.
arXiv Detail & Related papers (2020-06-18T08:37:04Z) - Multi-View Graph Neural Networks for Molecular Property Prediction [67.54644592806876]
We present Multi-View Graph Neural Network (MV-GNN), a multi-view message passing architecture.
In MV-GNN, we introduce a shared self-attentive readout component and disagreement loss to stabilize the training process.
We further boost the expressive power of MV-GNN by proposing a cross-dependent message passing scheme.
arXiv Detail & Related papers (2020-05-17T04:46:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.