One Transformer Can Understand Both 2D & 3D Molecular Data
- URL: http://arxiv.org/abs/2210.01765v4
- Date: Tue, 28 Mar 2023 03:01:29 GMT
- Title: One Transformer Can Understand Both 2D & 3D Molecular Data
- Authors: Shengjie Luo, Tianlang Chen, Yixian Xu, Shuxin Zheng, Tie-Yan Liu,
Liwei Wang, Di He
- Abstract summary: We develop a novel Transformer-based Molecular model called Transformer-M.
It can take molecular data of 2D or 3D formats as input and generate meaningful semantic representations.
All empirical results show that Transformer-M can simultaneously achieve strong performance on 2D and 3D tasks.
- Score: 94.93514673086631
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Unlike vision and language data which usually has a unique format, molecules
can naturally be characterized using different chemical formulations. One can
view a molecule as a 2D graph or define it as a collection of atoms located in
a 3D space. For molecular representation learning, most previous works designed
neural networks only for a particular data format, making the learned models
likely to fail for other data formats. We believe a general-purpose neural
network model for chemistry should be able to handle molecular tasks across
data modalities. To achieve this goal, in this work, we develop a novel
Transformer-based Molecular model called Transformer-M, which can take
molecular data of 2D or 3D formats as input and generate meaningful semantic
representations. Using the standard Transformer as the backbone architecture,
Transformer-M develops two separated channels to encode 2D and 3D structural
information and incorporate them with the atom features in the network modules.
When the input data is in a particular format, the corresponding channel will
be activated, and the other will be disabled. By training on 2D and 3D
molecular data with properly designed supervised signals, Transformer-M
automatically learns to leverage knowledge from different data modalities and
correctly capture the representations. We conducted extensive experiments for
Transformer-M. All empirical results show that Transformer-M can simultaneously
achieve strong performance on 2D and 3D tasks, suggesting its broad
applicability. The code and models will be made publicly available at
https://github.com/lsj2408/Transformer-M.
Related papers
- MolMix: A Simple Yet Effective Baseline for Multimodal Molecular Representation Learning [17.93173928602627]
We propose a simple transformer-based baseline for multimodal molecular representation learning.
We integrate three distinct modalities: SMILES strings, 2D graph representations, and 3D conformers of molecules.
Despite its simplicity, our approach achieves state-of-the-art results across multiple datasets.
arXiv Detail & Related papers (2024-10-10T14:36:58Z) - 3D-VirtFusion: Synthetic 3D Data Augmentation through Generative Diffusion Models and Controllable Editing [52.68314936128752]
We propose a new paradigm to automatically generate 3D labeled training data by harnessing the power of pretrained large foundation models.
For each target semantic class, we first generate 2D images of a single object in various structure and appearance via diffusion models and chatGPT generated text prompts.
We transform these augmented images into 3D objects and construct virtual scenes by random composition.
arXiv Detail & Related papers (2024-08-25T09:31:22Z) - Multi-Type Point Cloud Autoencoder: A Complete Equivariant Embedding for Molecule Conformation and Pose [0.8886153850492464]
We develop, train, and evaluate a new type of autoencoder, molecular O(3) encoding net (Mo3ENet) for multi-type point clouds.
Mo3ENet is end-to-end equivariant, meaning the learned representation can be manipulated on O(3), a practical bonus for downstream learning tasks.
arXiv Detail & Related papers (2024-05-22T16:14:37Z) - Meta-Transformer: A Unified Framework for Multimodal Learning [105.77219833997962]
Multimodal learning aims to build models that process and relate information from multiple modalities.
Despite years of development in this field, it still remains challenging to design a unified network for processing various modalities.
We propose a framework, named Meta-Transformer, that leverages a $textbffrozen$ encoder to perform multimodal perception.
arXiv Detail & Related papers (2023-07-20T12:10:29Z) - MUDiff: Unified Diffusion for Complete Molecule Generation [104.7021929437504]
We present a new model for generating a comprehensive representation of molecules, including atom features, 2D discrete molecule structures, and 3D continuous molecule coordinates.
We propose a novel graph transformer architecture to denoise the diffusion process.
Our model is a promising approach for designing stable and diverse molecules and can be applied to a wide range of tasks in molecular modeling.
arXiv Detail & Related papers (2023-04-28T04:25:57Z) - Unified 2D and 3D Pre-Training of Molecular Representations [237.36667670201473]
We propose a new representation learning method based on a unified 2D and 3D pre-training.
Atom coordinates and interatomic distances are encoded and then fused with atomic representations through graph neural networks.
Our method achieves state-of-the-art results on 10 tasks, and the average improvement on 2D-only tasks is 8.3%.
arXiv Detail & Related papers (2022-07-14T11:36:56Z) - 3D-Transformer: Molecular Representation with Transformer in 3D Space [11.947499562836953]
3D-Transformer is a variant of the Transformer for molecular representations that incorporates 3D spatial information.
Our experiments show significant improvements over state-of-the-art models on the crystal property prediction task and the protein-ligand binding affinity prediction task.
arXiv Detail & Related papers (2021-10-04T05:11:23Z) - Dual-view Molecule Pre-training [186.07333992384287]
Dual-view molecule pre-training can effectively combine the strengths of both types of molecule representations.
DMP is tested on nine molecular property prediction tasks and achieves state-of-the-art performances on seven of them.
arXiv Detail & Related papers (2021-06-17T03:58:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.