MeLM, a generative pretrained language modeling framework that solves
forward and inverse mechanics problems
- URL: http://arxiv.org/abs/2306.17525v1
- Date: Fri, 30 Jun 2023 10:28:20 GMT
- Title: MeLM, a generative pretrained language modeling framework that solves
forward and inverse mechanics problems
- Authors: Markus J. Buehler
- Abstract summary: We report a flexible multi-modal mechanics language model, MeLM, applied to solve various nonlinear forward and inverse problems.
The framework is applied to various examples including bio-inspired hierarchical honeycomb design and carbon nanotube mechanics.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: We report a flexible multi-modal mechanics language model, MeLM, applied to
solve various nonlinear forward and inverse problems, that can deal with a set
of instructions, numbers and microstructure data. The framework is applied to
various examples including bio-inspired hierarchical honeycomb design, carbon
nanotube mechanics, and protein unfolding. In spite of the flexible nature of
the model-which allows us to easily incorporate diverse materials, scales, and
mechanical features-it performs well across disparate forward and inverse
tasks. Based on an autoregressive attention-model, MeLM effectively represents
a large multi-particle system consisting of hundreds of millions of neurons,
where the interaction potentials are discovered through graph-forming
self-attention mechanisms that are then used to identify relationships from
emergent structures, while taking advantage of synergies discovered in the
training data. We show that the model can solve complex degenerate mechanics
design problems and determine novel material architectures across a range of
hierarchical levels, providing an avenue for materials discovery and analysis.
Looking beyond the demonstrations reported in this paper, we discuss other
opportunities in applied mechanics and general considerations about the use of
large language models in modeling, design, and analysis that can span a broad
spectrum of material properties from mechanical, thermal, optical, to
electronic.
Related papers
- Retrieval-Enhanced Machine Learning: Synthesis and Opportunities [60.34182805429511]
Retrieval-enhancement can be extended to a broader spectrum of machine learning (ML)
This work introduces a formal framework of this paradigm, Retrieval-Enhanced Machine Learning (REML), by synthesizing the literature in various domains in ML with consistent notations which is missing from the current literature.
The goal of this work is to equip researchers across various disciplines with a comprehensive, formally structured framework of retrieval-enhanced models, thereby fostering interdisciplinary future research.
arXiv Detail & Related papers (2024-07-17T20:01:21Z) - AtomAgents: Alloy design and discovery through physics-aware multi-modal multi-agent artificial intelligence [0.0]
The proposed physics-aware generative AI platform, AtomAgents, synergizes the intelligence of large language models (LLM)
Our results enable accurate prediction of key characteristics across alloys and highlight the crucial role of solid solution alloying to steer the development of advanced metallic alloys.
arXiv Detail & Related papers (2024-07-13T22:46:02Z) - Cephalo: Multi-Modal Vision-Language Models for Bio-Inspired Materials Analysis and Design [0.0]
Cephalo is a series of vision large language models (V-LLMs) designed for materials science applications.
It is trained on integrated image and text data from thousands of scientific papers.
Generative applications include bio-inspired designs, including pollen-inspired architected materials.
arXiv Detail & Related papers (2024-05-29T13:34:32Z) - LVLM-Interpret: An Interpretability Tool for Large Vision-Language Models [50.259006481656094]
We present a novel interactive application aimed towards understanding the internal mechanisms of large vision-language models.
Our interface is designed to enhance the interpretability of the image patches, which are instrumental in generating an answer.
We present a case study of how our application can aid in understanding failure mechanisms in a popular large multi-modal model: LLaVA.
arXiv Detail & Related papers (2024-04-03T23:57:34Z) - X-LoRA: Mixture of Low-Rank Adapter Experts, a Flexible Framework for Large Language Models with Applications in Protein Mechanics and Molecular Design [0.0]
We report a mixture of expert strategy to create fine-tuned large language models using a deep layer-wise token-level approach based on low-rank adaptation (LoRA)
The design is inspired by the biological principles of universality and diversity, where neural network building blocks are reused in different hierarchical manifestations.
We develop a tailored X-LoRA model that offers scientific capabilities including forward/inverse analysis tasks and enhanced reasoning capability, focused on biomaterial analysis, protein mechanics and design.
arXiv Detail & Related papers (2024-02-11T10:23:34Z) - ProtAgents: Protein discovery via large language model multi-agent
collaborations combining physics and machine learning [0.0]
ProtAgents is a platform for de novo protein design based on Large Language Models (LLMs)
Multiple AI agents with distinct capabilities collaboratively address complex tasks within a dynamic environment.
The flexibility in designing the agents, on one hand, and their capacity in autonomous collaboration through the dynamic LLM-based multi-agent environment unleashes great potentials.
arXiv Detail & Related papers (2024-01-27T20:19:49Z) - Generative Pretrained Autoregressive Transformer Graph Neural Network
applied to the Analysis and Discovery of Novel Proteins [0.0]
We report a flexible language-model based deep learning strategy, applied here to solve complex forward and inverse problems in protein modeling.
The model is applied to predict secondary structure content (per-residue level and overall content), protein solubility, and sequencing tasks.
We find that adding additional tasks yields emergent synergies that the model exploits in improving overall performance.
arXiv Detail & Related papers (2023-05-07T12:30:24Z) - Dynamic Latent Separation for Deep Learning [67.62190501599176]
A core problem in machine learning is to learn expressive latent variables for model prediction on complex data.
Here, we develop an approach that improves expressiveness, provides partial interpretation, and is not restricted to specific applications.
arXiv Detail & Related papers (2022-10-07T17:56:53Z) - Symmetry Group Equivariant Architectures for Physics [52.784926970374556]
In the domain of machine learning, an awareness of symmetries has driven impressive performance breakthroughs.
We argue that both the physics community and the broader machine learning community have much to understand.
arXiv Detail & Related papers (2022-03-11T18:27:04Z) - How to See Hidden Patterns in Metamaterials with Interpretable Machine
Learning [82.67551367327634]
We develop a new interpretable, multi-resolution machine learning framework for finding patterns in the unit-cells of materials.
Specifically, we propose two new interpretable representations of metamaterials, called shape-frequency features and unit-cell templates.
arXiv Detail & Related papers (2021-11-10T21:19:02Z) - Model-agnostic multi-objective approach for the evolutionary discovery
of mathematical models [55.41644538483948]
In modern data science, it is more interesting to understand the properties of the model, which parts could be replaced to obtain better results.
We use multi-objective evolutionary optimization for composite data-driven model learning to obtain the algorithm's desired properties.
arXiv Detail & Related papers (2021-07-07T11:17:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.