Quantum Many-Body Physics Calculations with Large Language Models
- URL: http://arxiv.org/abs/2403.03154v1
- Date: Tue, 5 Mar 2024 17:47:22 GMT
- Title: Quantum Many-Body Physics Calculations with Large Language Models
- Authors: Haining Pan, Nayantara Mudur, Will Taranto, Maria Tikhanovskaya,
Subhashini Venugopalan, Yasaman Bahri, Michael P. Brenner, Eun-Ah Kim
- Abstract summary: Large language models (LLMs) have demonstrated an unprecedented ability to perform complex tasks in multiple domains.
We focus on a broadly used approximation method in quantum physics: the Hartree-Fock method.
We design multi-step prompt templates that break down the analytic calculation into standardized steps.
We evaluate GPT-4's performance in executing the calculation for 15 research papers from the past decade.
- Score: 8.010378126123783
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: Large language models (LLMs) have demonstrated an unprecedented ability to
perform complex tasks in multiple domains, including mathematical and
scientific reasoning. We demonstrate that with carefully designed prompts, LLMs
can accurately carry out key calculations in research papers in theoretical
physics. We focus on a broadly used approximation method in quantum physics:
the Hartree-Fock method, requiring an analytic multi-step calculation deriving
approximate Hamiltonian and corresponding self-consistency equations. To carry
out the calculations using LLMs, we design multi-step prompt templates that
break down the analytic calculation into standardized steps with placeholders
for problem-specific information. We evaluate GPT-4's performance in executing
the calculation for 15 research papers from the past decade, demonstrating
that, with correction of intermediate steps, it can correctly derive the final
Hartree-Fock Hamiltonian in 13 cases and makes minor errors in 2 cases.
Aggregating across all research papers, we find an average score of 87.5 (out
of 100) on the execution of individual calculation steps. Overall, the
requisite skill for doing these calculations is at the graduate level in
quantum condensed matter theory. We further use LLMs to mitigate the two
primary bottlenecks in this evaluation process: (i) extracting information from
papers to fill in templates and (ii) automatic scoring of the calculation
steps, demonstrating good results in both cases. The strong performance is the
first step for developing algorithms that automatically explore theoretical
hypotheses at an unprecedented scale.
Related papers
- A Training Data Recipe to Accelerate A* Search with Language Models [3.037409201025504]
LM-based navigations are quite weak, incurring a high computational cost without a significant performance improvement.
Existing methods to learn theses do not consider the requirements of the planner, and typically need a lot of compute.
We reduce the number of iterations required to find the solutions by upto 13x, with a wall-clock speed-up of upto 5x.
arXiv Detail & Related papers (2024-07-13T19:21:44Z) - Predicting Probabilities of Error to Combine Quantization and Early Exiting: QuEE [68.6018458996143]
We propose a more general dynamic network that can combine both quantization and early exit dynamic network: QuEE.
Our algorithm can be seen as a form of soft early exiting or input-dependent compression.
The crucial factor of our approach is accurate prediction of the potential accuracy improvement achievable through further computation.
arXiv Detail & Related papers (2024-06-20T15:25:13Z) - The Fine-Grained Complexity of Gradient Computation for Training Large
Language Models [12.853829771559916]
Large language models (LLMs) have made fundamental contributions over the last a few years.
We show nearly identical results for the harder-seeming problem of computing the gradient of loss function of one layer attention network.
arXiv Detail & Related papers (2024-02-07T00:45:31Z) - The Search-and-Mix Paradigm in Approximate Nash Equilibrium Algorithms [9.037333651025351]
This work provides an automatic method for approximation analysis on a well-studied problem in theoretical computer science.
We observe that such algorithms can be reformulated into a search-and-mix paradigm.
By doing so, we are able to fully automate the procedure of designing and analyzing the mixing phase.
arXiv Detail & Related papers (2023-10-12T06:30:12Z) - The Languini Kitchen: Enabling Language Modelling Research at Different
Scales of Compute [66.84421705029624]
We introduce an experimental protocol that enables model comparisons based on equivalent compute, measured in accelerator hours.
We pre-process an existing large, diverse, and high-quality dataset of books that surpasses existing academic benchmarks in quality, diversity, and document length.
This work also provides two baseline models: a feed-forward model derived from the GPT-2 architecture and a recurrent model in the form of a novel LSTM with ten-fold throughput.
arXiv Detail & Related papers (2023-09-20T10:31:17Z) - QuIP: 2-Bit Quantization of Large Language Models With Guarantees [44.212441764241]
This work studies post-training parameter quantization in large language models (LLMs)
We introduce quantization with incoherence processing (QuIP), a new method based on the insight that quantization benefits from $textitincoherent$ weight and Hessian matrices.
arXiv Detail & Related papers (2023-07-25T07:44:06Z) - Evaluating and Improving Tool-Augmented Computation-Intensive Math
Reasoning [75.74103236299477]
Chain-of-thought prompting(CoT) and tool augmentation have been validated as effective practices for improving large language models.
We propose a new approach that can deliberate the reasoning steps with tool interfaces, namely textbfDELI.
Experimental results on CARP and six other datasets show that the proposed DELI mostly outperforms competitive baselines.
arXiv Detail & Related papers (2023-06-04T17:02:59Z) - Representation Learning with Multi-Step Inverse Kinematics: An Efficient
and Optimal Approach to Rich-Observation RL [106.82295532402335]
Existing reinforcement learning algorithms suffer from computational intractability, strong statistical assumptions, and suboptimal sample complexity.
We provide the first computationally efficient algorithm that attains rate-optimal sample complexity with respect to the desired accuracy level.
Our algorithm, MusIK, combines systematic exploration with representation learning based on multi-step inverse kinematics.
arXiv Detail & Related papers (2023-04-12T14:51:47Z) - End-to-end resource analysis for quantum interior point methods and portfolio optimization [63.4863637315163]
We provide a complete quantum circuit-level description of the algorithm from problem input to problem output.
We report the number of logical qubits and the quantity/depth of non-Clifford T-gates needed to run the algorithm.
arXiv Detail & Related papers (2022-11-22T18:54:48Z) - Quantum Sparse Coding [5.130440339897477]
We develop a quantum-inspired algorithm for sparse coding.
The emergence of quantum computers and Ising machines can potentially lead to more accurate estimations.
We conduct numerical experiments with simulated data on Lightr's quantum-inspired digital platform.
arXiv Detail & Related papers (2022-09-08T13:00:30Z) - Quantum Algorithms for Data Representation and Analysis [68.754953879193]
We provide quantum procedures that speed-up the solution of eigenproblems for data representation in machine learning.
The power and practical use of these subroutines is shown through new quantum algorithms, sublinear in the input matrix's size, for principal component analysis, correspondence analysis, and latent semantic analysis.
Results show that the run-time parameters that do not depend on the input's size are reasonable and that the error on the computed model is small, allowing for competitive classification performances.
arXiv Detail & Related papers (2021-04-19T00:41:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.