Multinational AGI Consortium (MAGIC): A Proposal for International
Coordination on AI
- URL: http://arxiv.org/abs/2310.09217v1
- Date: Fri, 13 Oct 2023 16:12:26 GMT
- Title: Multinational AGI Consortium (MAGIC): A Proposal for International
Coordination on AI
- Authors: Jason Hausenloy, Andrea Miotti, Claire Dennis
- Abstract summary: MAGIC would be the only institution in the world permitted to develop advanced AI.
We propose one positive vision of the future, where MAGIC, as a global governance regime, can lay the groundwork for long-term, safe regulation of advanced AI.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper proposes a Multinational Artificial General Intelligence
Consortium (MAGIC) to mitigate existential risks from advanced artificial
intelligence (AI). MAGIC would be the only institution in the world permitted
to develop advanced AI, enforced through a global moratorium by its signatory
members on all other advanced AI development. MAGIC would be exclusive,
safety-focused, highly secure, and collectively supported by member states,
with benefits distributed equitably among signatories. MAGIC would allow narrow
AI models to flourish while significantly reducing the possibility of
misaligned, rogue, breakout, or runaway outcomes of general-purpose systems. We
do not address the political feasibility of implementing a moratorium or
address the specific legislative strategies and rules needed to enforce a ban
on high-capacity AGI training runs. Instead, we propose one positive vision of
the future, where MAGIC, as a global governance regime, can lay the groundwork
for long-term, safe regulation of advanced AI.
Related papers
- Using AI Alignment Theory to understand the potential pitfalls of regulatory frameworks [55.2480439325792]
This paper critically examines the European Union's Artificial Intelligence Act (EU AI Act)
Uses insights from Alignment Theory (AT) research, which focuses on the potential pitfalls of technical alignment in Artificial Intelligence.
As we apply these concepts to the EU AI Act, we uncover potential vulnerabilities and areas for improvement in the regulation.
arXiv Detail & Related papers (2024-10-10T17:38:38Z) - Can apparent bystanders distinctively shape an outcome? Global south
countries and global catastrophic risk-focused governance of artificial
intelligence [0.0]
We argue that global south countries like India and Singapore could be fairly consequential in the global catastrophic risk-focused governance of AI.
We also suggest some ways through which global south countries can play a positive role in designing, strengthening and operationalizing global catastrophic risk-focused AI governance.
arXiv Detail & Related papers (2023-12-07T18:54:16Z) - The risks of risk-based AI regulation: taking liability seriously [46.90451304069951]
The development and regulation of AI seems to have reached a critical stage.
Some experts are calling for a moratorium on the training of AI systems more powerful than GPT-4.
This paper analyses the most advanced legal proposal, the European Union's AI Act.
arXiv Detail & Related papers (2023-11-03T12:51:37Z) - Taking control: Policies to address extinction risks from AI [0.0]
We argue that voluntary commitments from AI companies would be an inappropriate and insufficient response.
We describe three policy proposals that would meaningfully address the threats from advanced AI.
arXiv Detail & Related papers (2023-10-31T15:53:14Z) - Managing extreme AI risks amid rapid progress [171.05448842016125]
We describe risks that include large-scale social harms, malicious uses, and irreversible loss of human control over autonomous AI systems.
There is a lack of consensus about how exactly such risks arise, and how to manage them.
Present governance initiatives lack the mechanisms and institutions to prevent misuse and recklessness, and barely address autonomous systems.
arXiv Detail & Related papers (2023-10-26T17:59:06Z) - AI Regulation in Europe: From the AI Act to Future Regulatory Challenges [3.0821115746307663]
It argues for a hybrid regulatory strategy that combines elements from both philosophies.
The paper examines the AI Act as a pioneering legislative effort to address the multifaceted challenges posed by AI.
It advocates for immediate action to create protocols for regulated access to high-performance, potentially open-source AI systems.
arXiv Detail & Related papers (2023-10-06T07:52:56Z) - International Institutions for Advanced AI [47.449762587672986]
International institutions may have an important role to play in ensuring advanced AI systems benefit humanity.
This paper identifies a set of governance functions that could be performed at an international level to address these challenges.
It groups these functions into four institutional models that exhibit internal synergies and have precedents in existing organizations.
arXiv Detail & Related papers (2023-07-10T16:55:55Z) - Fairness in AI and Its Long-Term Implications on Society [68.8204255655161]
We take a closer look at AI fairness and analyze how lack of AI fairness can lead to deepening of biases over time.
We discuss how biased models can lead to more negative real-world outcomes for certain groups.
If the issues persist, they could be reinforced by interactions with other risks and have severe implications on society in the form of social unrest.
arXiv Detail & Related papers (2023-04-16T11:22:59Z) - Fairness in Agreement With European Values: An Interdisciplinary
Perspective on AI Regulation [61.77881142275982]
This interdisciplinary position paper considers various concerns surrounding fairness and discrimination in AI, and discusses how AI regulations address them.
We first look at AI and fairness through the lenses of law, (AI) industry, sociotechnology, and (moral) philosophy, and present various perspectives.
We identify and propose the roles AI Regulation should take to make the endeavor of the AI Act a success in terms of AI fairness concerns.
arXiv Detail & Related papers (2022-06-08T12:32:08Z) - The Limits of Global Inclusion in AI Development [7.421135890148154]
Extant global inequality has motivated Western institutions to involve more diverse groups in the development and application of AI systems.
We argue that more focus should be placed on the redistribution of power, rather than just on including underrepresented groups.
arXiv Detail & Related papers (2021-02-02T02:53:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.