Attention-Based Learning on Molecular Ensembles
- URL: http://arxiv.org/abs/2011.12820v1
- Date: Wed, 25 Nov 2020 15:23:52 GMT
- Title: Attention-Based Learning on Molecular Ensembles
- Authors: Kangway V. Chuang, Michael J. Keiser
- Abstract summary: We describe an end-to-end deep learning approach that operates directly on small-moleculeal ensembles.
We show how attention-based pooling can elucidate key conformational poses in tasks based on molecular geometry.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The three-dimensional shape and conformation of small-molecule ligands are
critical for biomolecular recognition, yet encoding 3D geometry has not
improved ligand-based virtual screening approaches. We describe an end-to-end
deep learning approach that operates directly on small-molecule conformational
ensembles and identifies key conformational poses of small-molecules. Our
networks leverage two levels of representation learning: 1) individual
conformers are first encoded as spatial graphs using a graph neural network,
and 2) sampled conformational ensembles are represented as sets using an
attention mechanism to aggregate over individual instances. We demonstrate the
feasibility of this approach on a simple task based on bidentate coordination
of biaryl ligands, and show how attention-based pooling can elucidate key
conformational poses in tasks based on molecular geometry. This work
illustrates how set-based learning approaches may be further developed for
small molecule-based virtual screening.
Related papers
- UniIF: Unified Molecule Inverse Folding [67.60267592514381]
We propose a unified model UniIF for inverse folding of all molecules.
Our proposed method surpasses state-of-the-art methods on all tasks.
arXiv Detail & Related papers (2024-05-29T10:26:16Z) - Learning Over Molecular Conformer Ensembles: Datasets and Benchmarks [44.934084652800976]
We introduce the first MoleculAR Conformer Ensemble Learning benchmark to thoroughly evaluate the potential of learning on conformer ensembles.
Our findings reveal that direct learning from an conformer space can improve performance on a variety of tasks and models.
arXiv Detail & Related papers (2023-09-29T20:06:46Z) - Geometry-aware Line Graph Transformer Pre-training for Molecular
Property Prediction [4.598522704308923]
Geometry-aware line graph transformer (Galformer) pre-training is a novel self-supervised learning framework.
Galformer consistently outperforms all baselines on both classification and regression tasks.
arXiv Detail & Related papers (2023-09-01T14:20:48Z) - Unified Molecular Modeling via Modality Blending [35.16755562674055]
We introduce a novel "blend-then-predict" self-supervised learning method (MoleBLEND)
MoleBLEND blends atom relations from different modalities into one unified relation for matrix encoding, then recovers modality-specific information for both 2D and 3D structures.
Experiments show that MoleBLEND achieves state-of-the-art performance across major 2D/3D benchmarks.
arXiv Detail & Related papers (2023-07-12T15:27:06Z) - Implicit Geometry and Interaction Embeddings Improve Few-Shot Molecular
Property Prediction [53.06671763877109]
We develop molecular embeddings that encode complex molecular characteristics to improve the performance of few-shot molecular property prediction.
Our approach leverages large amounts of synthetic data, namely the results of molecular docking calculations.
On multiple molecular property prediction benchmarks, training from the embedding space substantially improves Multi-Task, MAML, and Prototypical Network few-shot learning performance.
arXiv Detail & Related papers (2023-02-04T01:32:40Z) - A 3D-Shape Similarity-based Contrastive Approach to Molecular
Representation Learning [0.7340017786387767]
We propose a new contrastive-learning procedure for graph neural networks, Molecular Contrastive Learning from Shape Similarity (MolCLaSS)
Rather than directly encoding or targeting three-dimensional poses, MolCLaSS matches a similarity objective based on Gaussian overlays to learn a meaningful representation of molecular shape.
arXiv Detail & Related papers (2022-11-03T20:10:46Z) - Self-Similarity Priors: Neural Collages as Differentiable Fractal
Representations [73.14227103400964]
We investigate the role of learning in the automated discovery of self-similarity and in its utilization for downstream tasks.
We design a novel class of implicit operators, Neural Collages, which represent data as the parameters of a self-referential, structured transformation.
We investigate how to leverage the representations produced by Neural Collages in various tasks, including data compression and generation.
arXiv Detail & Related papers (2022-04-15T22:54:23Z) - ATOM3D: Tasks On Molecules in Three Dimensions [91.72138447636769]
Deep neural networks have recently gained significant attention.
In this work we present ATOM3D, a collection of both novel and existing datasets spanning several key classes of biomolecules.
We develop three-dimensional molecular learning networks for each of these tasks, finding that they consistently improve performance.
arXiv Detail & Related papers (2020-12-07T20:18:23Z) - Primal-Dual Mesh Convolutional Neural Networks [62.165239866312334]
We propose a primal-dual framework drawn from the graph-neural-network literature to triangle meshes.
Our method takes features for both edges and faces of a 3D mesh as input and dynamically aggregates them.
We provide theoretical insights of our approach using tools from the mesh-simplification literature.
arXiv Detail & Related papers (2020-10-23T14:49:02Z) - A Trainable Optimal Transport Embedding for Feature Aggregation and its
Relationship to Attention [96.77554122595578]
We introduce a parametrized representation of fixed size, which embeds and then aggregates elements from a given input set according to the optimal transport plan between the set and a trainable reference.
Our approach scales to large datasets and allows end-to-end training of the reference, while also providing a simple unsupervised learning mechanism with small computational cost.
arXiv Detail & Related papers (2020-06-22T08:35:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.