Harnessing the Power of Adversarial Prompting and Large Language Models
for Robust Hypothesis Generation in Astronomy
- URL: http://arxiv.org/abs/2306.11648v1
- Date: Tue, 20 Jun 2023 16:16:56 GMT
- Title: Harnessing the Power of Adversarial Prompting and Large Language Models
for Robust Hypothesis Generation in Astronomy
- Authors: Ioana Ciuc\u{a}, Yuan-Sen Ting, Sandor Kruk, Kartheik Iyer
- Abstract summary: We employ in-context prompting, supplying the model with up to 1000 papers from the NASA Astrophysics Data System.
Our findings point towards a substantial boost in hypothesis generation when using in-context prompting.
We illustrate how adversarial prompting empowers GPT-4 to extract essential details from a vast knowledge base to produce meaningful hypotheses.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This study investigates the application of Large Language Models (LLMs),
specifically GPT-4, within Astronomy. We employ in-context prompting, supplying
the model with up to 1000 papers from the NASA Astrophysics Data System, to
explore the extent to which performance can be improved by immersing the model
in domain-specific literature. Our findings point towards a substantial boost
in hypothesis generation when using in-context prompting, a benefit that is
further accentuated by adversarial prompting. We illustrate how adversarial
prompting empowers GPT-4 to extract essential details from a vast knowledge
base to produce meaningful hypotheses, signaling an innovative step towards
employing LLMs for scientific research in Astronomy.
Related papers
- At First Sight: Zero-Shot Classification of Astronomical Images with Large Multimodal Models [0.0]
Vision-Language multimodal Models (VLMs) offer the possibility for zero-shot classification in astronomy.
We investigate two models, GPT-4o and LLaVA-NeXT, for zero-shot classification of low-surface brightness galaxies and artifacts.
We show that with natural language prompts these models achieved significant accuracy (above 80 percent typically) without additional training/fine tuning.
arXiv Detail & Related papers (2024-06-24T18:17:54Z) - SpaRC and SpaRP: Spatial Reasoning Characterization and Path Generation for Understanding Spatial Reasoning Capability of Large Language Models [70.01883340129204]
spatial reasoning is a crucial component of both biological and artificial intelligence.
We present a comprehensive study of the capability of current state-of-the-art large language models (LLMs) on spatial reasoning.
arXiv Detail & Related papers (2024-06-07T01:06:34Z) - LLM and Simulation as Bilevel Optimizers: A New Paradigm to Advance Physical Scientific Discovery [141.39722070734737]
We propose to enhance the knowledge-driven, abstract reasoning abilities of Large Language Models with the computational strength of simulations.
We introduce Scientific Generative Agent (SGA), a bilevel optimization framework.
We conduct experiments to demonstrate our framework's efficacy in law discovery and molecular design.
arXiv Detail & Related papers (2024-05-16T03:04:10Z) - Simple Techniques for Enhancing Sentence Embeddings in Generative Language Models [3.0566617373924325]
Sentence embedding is a fundamental task within the realm of Natural Language Processing, finding extensive application in search engines, expert systems, and question-and-answer platforms.
With the continuous evolution of large language models such as LLaMA and Mistral, research on sentence embedding has recently achieved notable breakthroughs.
We propose two innovative prompt engineering techniques capable of further enhancing the expressive power of PLMs' raw embeddings.
arXiv Detail & Related papers (2024-04-05T07:07:15Z) - Scientific Language Modeling: A Quantitative Review of Large Language
Models in Molecular Science [27.874571056109758]
Large language models (LLMs) offer a fresh approach to tackle scientific problems from a natural language processing (NLP) perspective.
We propose a multi-modal benchmark, named ChEBI-20-MM, and perform 1263 experiments to assess the model's compatibility with data modalities and knowledge acquisition.
Our pioneering analysis offers an exploration of the learning mechanism and paves the way for advancing SLM in molecular science.
arXiv Detail & Related papers (2024-02-06T16:12:36Z) - SciGLM: Training Scientific Language Models with Self-Reflective
Instruction Annotation and Tuning [60.14510984576027]
SciGLM is a suite of scientific language models able to conduct college-level scientific reasoning.
We apply a self-reflective instruction annotation framework to generate step-by-step reasoning for unlabelled scientific questions.
We fine-tuned the ChatGLM family of language models with SciInstruct, enhancing their scientific and mathematical reasoning capabilities.
arXiv Detail & Related papers (2024-01-15T20:22:21Z) - The Impact of Large Language Models on Scientific Discovery: a
Preliminary Study using GPT-4 [0.0]
This report focuses on GPT-4, the state-of-the-art language model.
We evaluate GPT-4's knowledge base, scientific understanding, scientific numerical calculation abilities, and various scientific prediction capabilities.
arXiv Detail & Related papers (2023-11-13T14:26:12Z) - Large Language Models for Scientific Synthesis, Inference and
Explanation [56.41963802804953]
We show how large language models can perform scientific synthesis, inference, and explanation.
We show that the large language model can augment this "knowledge" by synthesizing from the scientific literature.
This approach has the further advantage that the large language model can explain the machine learning system's predictions.
arXiv Detail & Related papers (2023-10-12T02:17:59Z) - Large Language Models for Automated Open-domain Scientific Hypotheses Discovery [50.40483334131271]
This work proposes the first dataset for social science academic hypotheses discovery.
Unlike previous settings, the new dataset requires (1) using open-domain data (raw web corpus) as observations; and (2) proposing hypotheses even new to humanity.
A multi- module framework is developed for the task, including three different feedback mechanisms to boost performance.
arXiv Detail & Related papers (2023-09-06T05:19:41Z) - A Survey on Large Language Models for Recommendation [77.91673633328148]
Large Language Models (LLMs) have emerged as powerful tools in the field of Natural Language Processing (NLP)
This survey presents a taxonomy that categorizes these models into two major paradigms, respectively Discriminative LLM for Recommendation (DLLM4Rec) and Generative LLM for Recommendation (GLLM4Rec)
arXiv Detail & Related papers (2023-05-31T13:51:26Z) - Galactic ChitChat: Using Large Language Models to Converse with
Astronomy Literature [0.0]
We demonstrate the potential of the state-of-the-art OpenAI GPT-4 large language model to engage in meaningful interactions with Astronomy papers.
We employ a distillation technique that effectively reduces the size of the original input paper by 50%.
We then explore the model's responses using a multi-document context.
arXiv Detail & Related papers (2023-04-12T03:02:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.