AI Must Embrace Specialization via Superhuman Adaptable Intelligence
- URL: http://arxiv.org/abs/2602.23643v1
- Date: Fri, 27 Feb 2026 03:26:21 GMT
- Title: AI Must Embrace Specialization via Superhuman Adaptable Intelligence
- Authors: Judah Goldfeder, Philippe Wyder, Yann LeCun, Ravid Shwartz Ziv,
- Abstract summary: We argue that AI must embrace specialization, rather than strive for superhuman generality.<n>We argue that Superhuman Adaptable Intelligence (SAI) can fill in the skill gaps where humans are incapable.
- Score: 39.20881378353777
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Everyone from AI executives and researchers to doomsayers, politicians, and activists is talking about Artificial General Intelligence (AGI). Yet, they often don't seem to agree on its exact definition. One common definition of AGI is an AI that can do everything a human can do, but are humans truly general? In this paper, we address what's wrong with our conception of AGI, and why, even in its most coherent formulation, it is a flawed concept to describe the future of AI. We explore whether the most widely accepted definitions are plausible, useful, and truly general. We argue that AI must embrace specialization, rather than strive for generality, and in its specialization strive for superhuman performance, and introduce Superhuman Adaptable Intelligence (SAI). SAI is defined as intelligence that can learn to exceed humans at anything important that we can do, and that can fill in the skill gaps where humans are incapable. We then lay out how SAI can help hone a discussion around AI that was blurred by an overloaded definition of AGI, and extrapolate the implications of using it as a guide for the future.
Related papers
- What Does 'Human-Centred AI' Mean? [0.0]
AI is usefully seen as a relationship between technology and humans.<n>All AI implicates human cognition; no matter what.<n>To even begin to de-fetishise AI, we must look the human-in-the-loop in the eyes.
arXiv Detail & Related papers (2025-07-26T14:18:52Z) - Giving AI a voice: how does AI think it should be treated? [0.0]
This chapter includes a brief human-AI conversation on the topic of AI rights and ethics.<n>There are new questions and angles that AI brings to the table that we might not have considered before.
arXiv Detail & Related papers (2025-04-21T07:59:17Z) - Aligning Generalisation Between Humans and Machines [74.120848518198]
AI technology can support humans in scientific discovery and forming decisions, but may also disrupt democracies and target individuals.<n>The responsible use of AI and its participation in human-AI teams increasingly shows the need for AI alignment.<n>A crucial yet often overlooked aspect of these interactions is the different ways in which humans and machines generalise.
arXiv Detail & Related papers (2024-11-23T18:36:07Z) - On the consistent reasoning paradox of intelligence and optimal trust in AI: The power of 'I don't know' [79.69412622010249]
Consistent reasoning, which lies at the core of human intelligence, is the ability to handle tasks that are equivalent.
CRP asserts that consistent reasoning implies fallibility -- in particular, human-like intelligence in AI necessarily comes with human-like fallibility.
arXiv Detail & Related papers (2024-08-05T10:06:53Z) - Keep the Future Human: Why and How We Should Close the Gates to AGI and Superintelligence, and What We Should Build Instead [0.20919309330073077]
Advances in AI have transformed AI from a niche academic field to the core business strategy of many of the world's largest companies.<n>This essay argues that we should keep the future human by closing the "gates" to smarter-than-human, autonomous, general-purpose AI.<n>Instead, we should focus on powerful, trustworthy AI tools that can empower individuals and transformatively improve human societies' abilities to do what they do best.
arXiv Detail & Related papers (2023-11-15T23:41:12Z) - Can Machines Imitate Humans? Integrative Turing-like tests for Language and Vision Demonstrate a Narrowing Gap [56.611702960809644]
We benchmark AI's ability to imitate humans in three language tasks and three vision tasks.<n>Next, we conducted 72,191 Turing-like tests with 1,916 human judges and 10 AI judges.<n>Imitation ability showed minimal correlation with conventional AI performance metrics.
arXiv Detail & Related papers (2022-11-23T16:16:52Z) - Cybertrust: From Explainable to Actionable and Interpretable AI (AI2) [58.981120701284816]
Actionable and Interpretable AI (AI2) will incorporate explicit quantifications and visualizations of user confidence in AI recommendations.
It will allow examining and testing of AI system predictions to establish a basis for trust in the systems' decision making.
arXiv Detail & Related papers (2022-01-26T18:53:09Z) - Challenges of Artificial Intelligence -- From Machine Learning and
Computer Vision to Emotional Intelligence [0.0]
We believe that AI is a helper, not a ruler of humans.
Computer vision has been central to the development of AI.
Emotions are central to human intelligence, but little use has been made in AI.
arXiv Detail & Related papers (2022-01-05T06:00:22Z) - Trustworthy AI: A Computational Perspective [54.80482955088197]
We focus on six of the most crucial dimensions in achieving trustworthy AI: (i) Safety & Robustness, (ii) Non-discrimination & Fairness, (iii) Explainability, (iv) Privacy, (v) Accountability & Auditability, and (vi) Environmental Well-Being.
For each dimension, we review the recent related technologies according to a taxonomy and summarize their applications in real-world systems.
arXiv Detail & Related papers (2021-07-12T14:21:46Z) - Human $\neq$ AGI [1.370633147306388]
General Intelligence (AGI) and Human-Level Artificial Intelligence (HLAI) have been used to interchangeably refer to the Holy Grail of Artificial Intelligence research.
This paper argues that implicit assumption of equivalence between capabilities of AGI and HLAI appears to be unjustified, as humans are not general intelligences.
arXiv Detail & Related papers (2020-07-11T14:06:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.