Deliberative Technology for Alignment
- URL: http://arxiv.org/abs/2312.03893v1
- Date: Wed, 6 Dec 2023 20:34:32 GMT
- Title: Deliberative Technology for Alignment
- Authors: Andrew Konya, Deger Turan, Aviv Ovadya, Lina Qui, Daanish Masood,
Flynn Devine, Lisa Schirch, Isabella Roberts, Deliberative Alignment Forum
- Abstract summary: Deliberative technology is already being used across these institutions to help align governance with human will.
The race to superhuman AGI is already underway, and the AI systems it gives rise to may become the most powerful systems of the future.
Right now, there is a window of opportunity to use deliberative technology to align the impact of powerful AI with the will of humanity.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: For humanity to maintain and expand its agency into the future, the most
powerful systems we create must be those which act to align the future with the
will of humanity. The most powerful systems today are massive institutions like
governments, firms, and NGOs. Deliberative technology is already being used
across these institutions to help align governance and diplomacy with human
will, and modern AI is poised to make this technology significantly better. At
the same time, the race to superhuman AGI is already underway, and the AI
systems it gives rise to may become the most powerful systems of the future.
Failure to align the impact of such powerful AI with the will of humanity may
lead to catastrophic consequences, while success may unleash abundance. Right
now, there is a window of opportunity to use deliberative technology to align
the impact of powerful AI with the will of humanity. Moreover, it may be
possible to engineer a symbiotic coupling between powerful AI and deliberative
alignment systems such that the quality of alignment improves as AI
capabilities increase.
Related papers
- Gradual Disempowerment: Systemic Existential Risks from Incremental AI Development [15.701299669203618]
We analyze how even incremental improvements in AI capabilities can undermine human influence over large-scale systems that society depends on.
We argue that this dynamic could lead to an effectively irreversible loss of human influence over crucial societal systems, precipitating an existential catastrophe through the permanent disempowerment of humanity.
arXiv Detail & Related papers (2025-01-28T13:45:41Z) - Societal Adaptation to Advanced AI [1.2607853680700076]
Existing strategies for managing risks from advanced AI systems often focus on affecting what AI systems are developed and how they diffuse.
We urge a complementary approach: increasing societal adaptation to advanced AI.
We introduce a conceptual framework which helps identify adaptive interventions that avoid, defend against and remedy potentially harmful uses of AI systems.
arXiv Detail & Related papers (2024-05-16T17:52:12Z) - Now, Later, and Lasting: Ten Priorities for AI Research, Policy, and Practice [63.20307830884542]
Next several decades may well be a turning point for humanity, comparable to the industrial revolution.
Launched a decade ago, the project is committed to a perpetual series of studies by multidisciplinary experts.
We offer ten recommendations for action that collectively address both the short- and long-term potential impacts of AI technologies.
arXiv Detail & Related papers (2024-04-06T22:18:31Z) - Managing extreme AI risks amid rapid progress [171.05448842016125]
We describe risks that include large-scale social harms, malicious uses, and irreversible loss of human control over autonomous AI systems.
There is a lack of consensus about how exactly such risks arise, and how to manage them.
Present governance initiatives lack the mechanisms and institutions to prevent misuse and recklessness, and barely address autonomous systems.
arXiv Detail & Related papers (2023-10-26T17:59:06Z) - Fairness in AI and Its Long-Term Implications on Society [68.8204255655161]
We take a closer look at AI fairness and analyze how lack of AI fairness can lead to deepening of biases over time.
We discuss how biased models can lead to more negative real-world outcomes for certain groups.
If the issues persist, they could be reinforced by interactions with other risks and have severe implications on society in the form of social unrest.
arXiv Detail & Related papers (2023-04-16T11:22:59Z) - Examining the Differential Risk from High-level Artificial Intelligence
and the Question of Control [0.0]
The extent and scope of future AI capabilities remain a key uncertainty.
There are concerns over the extent of integration and oversight of AI opaque decision processes.
This study presents a hierarchical complex systems framework to model AI risk and provide a template for alternative futures analysis.
arXiv Detail & Related papers (2022-11-06T15:46:02Z) - AI Governance and Ethics Framework for Sustainable AI and Sustainability [0.0]
There are many emerging AI risks for humanity, such as autonomous weapons, automation-spurred job loss, socio-economic inequality, bias caused by data and algorithms, privacy violations and deepfakes.
Social diversity, equity and inclusion are considered key success factors of AI to mitigate risks, create values and drive social justice.
In our journey towards an AI-enabled sustainable future, we need to address AI ethics and governance as a priority.
arXiv Detail & Related papers (2022-09-28T22:23:10Z) - Cybertrust: From Explainable to Actionable and Interpretable AI (AI2) [58.981120701284816]
Actionable and Interpretable AI (AI2) will incorporate explicit quantifications and visualizations of user confidence in AI recommendations.
It will allow examining and testing of AI system predictions to establish a basis for trust in the systems' decision making.
arXiv Detail & Related papers (2022-01-26T18:53:09Z) - Trustworthy AI: A Computational Perspective [54.80482955088197]
We focus on six of the most crucial dimensions in achieving trustworthy AI: (i) Safety & Robustness, (ii) Non-discrimination & Fairness, (iii) Explainability, (iv) Privacy, (v) Accountability & Auditability, and (vi) Environmental Well-Being.
For each dimension, we review the recent related technologies according to a taxonomy and summarize their applications in real-world systems.
arXiv Detail & Related papers (2021-07-12T14:21:46Z) - Building Bridges: Generative Artworks to Explore AI Ethics [56.058588908294446]
In recent years, there has been an increased emphasis on understanding and mitigating adverse impacts of artificial intelligence (AI) technologies on society.
A significant challenge in the design of ethical AI systems is that there are multiple stakeholders in the AI pipeline, each with their own set of constraints and interests.
This position paper outlines some potential ways in which generative artworks can play this role by serving as accessible and powerful educational tools.
arXiv Detail & Related papers (2021-06-25T22:31:55Z) - On Controllability of AI [1.370633147306388]
We present arguments as well as supporting evidence indicating that advanced AI can't be fully controlled.
Consequences of uncontrollability of AI are discussed with respect to future of humanity and research on AI, and AI safety and security.
arXiv Detail & Related papers (2020-07-19T02:49:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.