On the Relationship between Sentence Analogy Identification and Sentence
Structure Encoding in Large Language Models
- URL: http://arxiv.org/abs/2310.07818v3
- Date: Tue, 6 Feb 2024 02:24:53 GMT
- Title: On the Relationship between Sentence Analogy Identification and Sentence
Structure Encoding in Large Language Models
- Authors: Thilini Wijesiriwardene, Ruwan Wickramarachchi, Aishwarya Naresh
Reganti, Vinija Jain, Aman Chadha, Amit Sheth, Amitava Das
- Abstract summary: We look at how Large Language Models' abilities to capture sentence analogies vary with their abilities to encode syntactic and semantic structures.
We find that the LLMs which capture syntactic structures better, also have higher abilities in identifying sentence analogies.
- Score: 7.716762867270514
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: The ability of Large Language Models (LLMs) to encode syntactic and semantic
structures of language is well examined in NLP. Additionally, analogy
identification, in the form of word analogies are extensively studied in the
last decade of language modeling literature. In this work we specifically look
at how LLMs' abilities to capture sentence analogies (sentences that convey
analogous meaning to each other) vary with LLMs' abilities to encode syntactic
and semantic structures of sentences. Through our analysis, we find that LLMs'
ability to identify sentence analogies is positively correlated with their
ability to encode syntactic and semantic structures of sentences. Specifically,
we find that the LLMs which capture syntactic structures better, also have
higher abilities in identifying sentence analogies.
Related papers
- Semantic Structure-Mapping in LLM and Human Analogical Reasoning [14.412456982731467]
Analogical reasoning is considered core to human learning and cognition.
Recent studies have compared the analogical reasoning abilities of human subjects and Large Language Models (LLMs) on abstract symbol manipulation tasks.
This ability to draw analogies that link language to non-linguistic domains is thought to play a crucial role in language acquisition and broader cognitive development.
arXiv Detail & Related papers (2024-06-19T20:07:37Z) - How Proficient Are Large Language Models in Formal Languages? An In-Depth Insight for Knowledge Base Question Answering [52.86931192259096]
Knowledge Base Question Answering (KBQA) aims to answer natural language questions based on facts in knowledge bases.
Recent works leverage the capabilities of large language models (LLMs) for logical form generation to improve performance.
arXiv Detail & Related papers (2024-01-11T09:27:50Z) - CoVLM: Composing Visual Entities and Relationships in Large Language
Models Via Communicative Decoding [66.52659447360104]
CoVLM can guide the LLM to explicitly compose visual entities and relationships among the text.
We propose CoVLM, which can guide the LLM to explicitly compose visual entities and relationships among the text.
arXiv Detail & Related papers (2023-11-06T18:59:44Z) - StoryAnalogy: Deriving Story-level Analogies from Large Language Models
to Unlock Analogical Understanding [72.38872974837462]
We evaluate the ability to identify and generate analogies by constructing a first-of-its-kind large-scale story-level analogy corpus.
textscStory Analogy contains 24K story pairs from diverse domains with human annotations on two similarities from the extended Structure-Mapping Theory.
We observe that the data in textscStory Analogy can improve the quality of analogy generation in large language models.
arXiv Detail & Related papers (2023-10-19T16:29:23Z) - Why Do We Need Neuro-symbolic AI to Model Pragmatic Analogies? [6.8107181513711055]
A hallmark of intelligence is the ability to use a familiar domain to make inferences about a less familiar domain, known as analogical reasoning.
We discuss analogies at four distinct levels of complexity: lexical analogies, syntactic analogies, semantic analogies, and pragmatic analogies.
We employ Neuro-symbolic AI techniques that combine statistical and symbolic AI, informing the representation of unstructured text to highlight and augment relevant content, provide abstraction and guide the mapping process.
arXiv Detail & Related papers (2023-08-02T21:13:38Z) - ParaAMR: A Large-Scale Syntactically Diverse Paraphrase Dataset by AMR
Back-Translation [59.91139600152296]
ParaAMR is a large-scale syntactically diverse paraphrase dataset created by abstract meaning representation back-translation.
We show that ParaAMR can be used to improve on three NLP tasks: learning sentence embeddings, syntactically controlled paraphrase generation, and data augmentation for few-shot learning.
arXiv Detail & Related papers (2023-05-26T02:27:33Z) - ANALOGYKB: Unlocking Analogical Reasoning of Language Models with A Million-scale Knowledge Base [51.777618249271725]
ANALOGYKB is a million-scale analogy knowledge base derived from existing knowledge graphs (KGs)
It identifies two types of analogies from the KGs: 1) analogies of the same relations, which can be directly extracted from the KGs, and 2) analogies of analogous relations, which are identified with a selection and filtering pipeline enabled by large language models (LLMs)
arXiv Detail & Related papers (2023-05-10T09:03:01Z) - ANALOGICAL -- A Novel Benchmark for Long Text Analogy Evaluation in
Large Language Models [1.4546044532817048]
ANALOGICAL is a new benchmark to intrinsically evaluate large language models.
Our evaluation finds that it is increasingly challenging for LLMs to identify analogies when going up the analogy taxonomy.
arXiv Detail & Related papers (2023-05-08T21:12:20Z) - Syntactic Persistence in Language Models: Priming as a Window into
Abstract Language Representations [0.38498574327875945]
We investigate the extent to which modern, neural language models are susceptible to syntactic priming.
We introduce a novel metric and release Prime-LM, a large corpus where we control for various linguistic factors which interact with priming strength.
We report surprisingly strong priming effects when priming with multiple sentences, each with different words and meaning but with identical syntactic structure.
arXiv Detail & Related papers (2021-09-30T10:38:38Z) - A Comparative Study of Lexical Substitution Approaches based on Neural
Language Models [117.96628873753123]
We present a large-scale comparative study of popular neural language and masked language models.
We show that already competitive results achieved by SOTA LMs/MLMs can be further improved if information about the target word is injected properly.
arXiv Detail & Related papers (2020-05-29T18:43:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.