Predicting the Temperature Dependence of Surfactant CMCs Using Graph
Neural Networks
- URL: http://arxiv.org/abs/2403.03767v1
- Date: Wed, 6 Mar 2024 15:03:04 GMT
- Title: Predicting the Temperature Dependence of Surfactant CMCs Using Graph
Neural Networks
- Authors: Christoforos Brozos, Jan G. Rittig, Sandip Bhattacharya, Elie Akanny,
Christina Kohlmann, Alexander Mitsos
- Abstract summary: classical QSPR and Graph Neural Networks (GNNs) have been successfully applied to predict the CMC of surfactants at room temperature.
We herein develop a GNN model for temperature-dependent CMC prediction of surfactants.
- Score: 38.39977540117143
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The critical micelle concentration (CMC) of surfactant molecules is an
essential property for surfactant applications in industry. Recently, classical
QSPR and Graph Neural Networks (GNNs), a deep learning technique, have been
successfully applied to predict the CMC of surfactants at room temperature.
However, these models have not yet considered the temperature dependency of the
CMC, which is highly relevant for practical applications. We herein develop a
GNN model for temperature-dependent CMC prediction of surfactants. We collect
about 1400 data points from public sources for all surfactant classes, i.e.,
ionic, nonionic, and zwitterionic, at multiple temperatures. We test the
predictive quality of the model for following scenarios: i) when CMC data for
surfactants are present in the training of the model in at least one different
temperature, and ii) CMC data for surfactants are not present in the training,
i.e., generalizing to unseen surfactants. In both test scenarios, our model
exhibits a high predictive performance of R$^2 \geq $ 0.94 on test data. We
also find that the model performance varies by surfactant class. Finally, we
evaluate the model for sugar-based surfactants with complex molecular
structures, as these represent a more sustainable alternative to synthetic
surfactants and are therefore of great interest for future applications in the
personal and home care industries.
Related papers
- Predicting the Temperature-Dependent CMC of Surfactant Mixtures with Graph Neural Networks [36.814181034608666]
Surfactants are key ingredients in foaming and cleansing products across various industries.
In practice, surfactant mixtures are typically used due to to performance, environmental, and cost reasons.
We develop a graph neural network framework for surfactant mixtures to predict the temperature-dependent CMC.
arXiv Detail & Related papers (2024-11-04T16:17:57Z) - Pre-trained Molecular Language Models with Random Functional Group Masking [54.900360309677794]
We propose a SMILES-based underlineem Molecular underlineem Language underlineem Model, which randomly masking SMILES subsequences corresponding to specific molecular atoms.
This technique aims to compel the model to better infer molecular structures and properties, thus enhancing its predictive capabilities.
arXiv Detail & Related papers (2024-11-03T01:56:15Z) - Graph Neural Networks for Surfactant Multi-Property Prediction [38.39977540117143]
Graph Neural Networks (GNNs) have exhibited a great predictive performance for property prediction of ionic liquids, polymers and drugs in general.
We create the largest available CMC database with 429 molecules and the first large data collection for surface excess concentration.
GNN yields highly accurate predictions for CMC, showing great potential for future industrial applications.
arXiv Detail & Related papers (2024-01-03T18:32:25Z) - Activity Cliff Prediction: Dataset and Benchmark [20.41770222873952]
We first introduce ACNet, a large-scale dataset for AC prediction.
ACNet curates over 400K Matched Molecular Pairs (MMPs) against 190 targets.
We propose a baseline framework to benchmark the predictive performance of molecular representations encoded by deep neural networks for AC prediction.
arXiv Detail & Related papers (2023-02-15T09:19:07Z) - Calibration and generalizability of probabilistic models on low-data
chemical datasets with DIONYSUS [0.0]
We perform an extensive study of the calibration and generalizability of probabilistic machine learning models on small chemical datasets.
We analyse the quality of their predictions and uncertainties in a variety of tasks (binary, regression) and datasets.
We offer practical insights into model and feature choice for modelling small chemical datasets, a common scenario in new chemical experiments.
arXiv Detail & Related papers (2022-12-03T08:19:06Z) - SPT-NRTL: A physics-guided machine learning model to predict
thermodynamically consistent activity coefficients [0.12352483741564477]
We introduce SPT-NRTL, a machine learning model to predict thermodynamically consistent activity coefficients.
SPT-NRTL achieves higher accuracy than UNIFAC in the prediction of activity coefficients across all functional groups.
arXiv Detail & Related papers (2022-09-09T06:21:05Z) - Graph Neural Networks for Temperature-Dependent Activity Coefficient
Prediction of Solutes in Ionic Liquids [58.720142291102135]
We present a GNN to predict temperature-dependent infinite dilution ACs of solutes in ILs.
We train the GNN on a database including more than 40,000 AC values and compare it to a state-of-the-art MCM.
The GNN and MCM achieve similar high prediction performance, with the GNN additionally enabling high-quality predictions for ACs of solutions that contain ILs and solutes not considered during training.
arXiv Detail & Related papers (2022-06-23T15:27:29Z) - Closed-form Continuous-Depth Models [99.40335716948101]
Continuous-depth neural models rely on advanced numerical differential equation solvers.
We present a new family of models, termed Closed-form Continuous-depth (CfC) networks, that are simple to describe and at least one order of magnitude faster.
arXiv Detail & Related papers (2021-06-25T22:08:51Z) - Assessing Graph-based Deep Learning Models for Predicting Flash Point [52.931492216239995]
Graph-based deep learning (GBDL) models were implemented in predicting flash point for the first time.
Average R2 and Mean Absolute Error (MAE) scores of MPNN are, respectively, 2.3% lower and 2.0 K higher than previous comparable studies.
arXiv Detail & Related papers (2020-02-26T06:10:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.