On General Language Understanding
- URL: http://arxiv.org/abs/2310.18038v1
- Date: Fri, 27 Oct 2023 10:36:54 GMT
- Title: On General Language Understanding
- Authors: David Schlangen
- Abstract summary: This paper sketches the outlines of a model of understanding, which can ground questions of the adequacy of current methods of measurement of model quality.
The paper makes three claims: A) That different language use situation types have different characteristics, B) That language understanding is a multifaceted phenomenon, and C) That the choice of Understanding Indicator marks the limits of benchmarking.
- Score: 18.2932386988379
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Natural Language Processing prides itself to be an empirically-minded, if not
outright empiricist field, and yet lately it seems to get itself into
essentialist debates on issues of meaning and measurement ("Do Large Language
Models Understand Language, And If So, How Much?"). This is not by accident:
Here, as everywhere, the evidence underspecifies the understanding. As a
remedy, this paper sketches the outlines of a model of understanding, which can
ground questions of the adequacy of current methods of measurement of model
quality. The paper makes three claims: A) That different language use situation
types have different characteristics, B) That language understanding is a
multifaceted phenomenon, bringing together individualistic and social
processes, and C) That the choice of Understanding Indicator marks the limits
of benchmarking, and the beginnings of considerations of the ethics of NLP use.
Related papers
- LLMs as Function Approximators: Terminology, Taxonomy, and Questions for Evaluation [18.2932386988379]
This paper argues that the loss of clarity on what these models model leads to metaphors like "artificial general intelligences"
The proposal is to see their generality, and their potential value, in their ability to approximate specialist function, based on a natural language specification.
arXiv Detail & Related papers (2024-07-18T17:49:56Z) - From Form(s) to Meaning: Probing the Semantic Depths of Language Models Using Multisense Consistency [13.154753046052527]
We focus on consistency across languages as well as paraphrases.
We find that the model's multisense consistency is lacking and run several follow-up analyses to verify.
We conclude that, in this aspect, the understanding of LLMs is still quite far from being consistent and human-like.
arXiv Detail & Related papers (2024-04-18T12:48:17Z) - Epistemology of Language Models: Do Language Models Have Holistic Knowledge? [30.02796959216552]
This paper investigates the inherent knowledge in language models from the perspective of holism.
The purpose of this paper is to explore whether characteristics of language models exhibit consistent with holism.
arXiv Detail & Related papers (2024-03-19T16:06:10Z) - Situated Natural Language Explanations [54.083715161895036]
Natural language explanations (NLEs) are among the most accessible tools for explaining decisions to humans.
Existing NLE research perspectives do not take the audience into account.
Situated NLE provides a perspective and facilitates further research on the generation and evaluation of explanations.
arXiv Detail & Related papers (2023-08-27T14:14:28Z) - Natural Language Decompositions of Implicit Content Enable Better Text
Representations [56.85319224208865]
We introduce a method for the analysis of text that takes implicitly communicated content explicitly into account.
We use a large language model to produce sets of propositions that are inferentially related to the text that has been observed.
Our results suggest that modeling the meanings behind observed language, rather than the literal text alone, is a valuable direction for NLP.
arXiv Detail & Related papers (2023-05-23T23:45:20Z) - Language Models as Inductive Reasoners [125.99461874008703]
We propose a new paradigm (task) for inductive reasoning, which is to induce natural language rules from natural language facts.
We create a dataset termed DEER containing 1.2k rule-fact pairs for the task, where rules and facts are written in natural language.
We provide the first and comprehensive analysis of how well pretrained language models can induce natural language rules from natural language facts.
arXiv Detail & Related papers (2022-12-21T11:12:14Z) - A Linguistic Investigation of Machine Learning based Contradiction
Detection Models: An Empirical Analysis and Future Perspectives [0.34998703934432673]
We analyze two Natural Language Inference data sets with respect to their linguistic features.
The goal is to identify those syntactic and semantic properties that are particularly hard to comprehend for a machine learning model.
arXiv Detail & Related papers (2022-10-19T10:06:03Z) - Analyzing the Limits of Self-Supervision in Handling Bias in Language [52.26068057260399]
We evaluate how well language models capture the semantics of four tasks for bias: diagnosis, identification, extraction and rephrasing.
Our analyses indicate that language models are capable of performing these tasks to widely varying degrees across different bias dimensions, such as gender and political affiliation.
arXiv Detail & Related papers (2021-12-16T05:36:08Z) - The Singleton Fallacy: Why Current Critiques of Language Models Miss the
Point [3.096615629099618]
We discuss the current critique against neural network-based Natural Language Understanding (NLU) solutions known as language models.
We will argue that there are many different types of language use, meaning and understanding, and that (current) language models are build with the explicit purpose of acquiring and representing one type of structural understanding of language.
arXiv Detail & Related papers (2021-02-08T16:12:36Z) - Probing Pretrained Language Models for Lexical Semantics [76.73599166020307]
We present a systematic empirical analysis across six typologically diverse languages and five different lexical tasks.
Our results indicate patterns and best practices that hold universally, but also point to prominent variations across languages and tasks.
arXiv Detail & Related papers (2020-10-12T14:24:01Z) - Information-Theoretic Probing for Linguistic Structure [74.04862204427944]
We propose an information-theoretic operationalization of probing as estimating mutual information.
We evaluate on a set of ten typologically diverse languages often underrepresented in NLP research.
arXiv Detail & Related papers (2020-04-07T01:06:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.