AI Governance InternationaL Evaluation Index (AGILE Index) 2025
- URL: http://arxiv.org/abs/2507.11546v2
- Date: Wed, 30 Jul 2025 07:35:02 GMT
- Title: AI Governance InternationaL Evaluation Index (AGILE Index) 2025
- Authors: Yi Zeng, Enmeng Lu, Xiaoyang Guo, Cunqing Huangfu, Jiawei Xie, Yu Chen, Zhengqi Wang, Dongqi Liang, Gongce Cao, Jin Wang, Zizhe Ruan, Xin Guan, Ammar Younas,
- Abstract summary: AI Governance InternationaL Evaluation Index (AGILE Index) project launched in 2023.<n>AGILE Index 2025 incorporates systematic refinements to better balance scientific rigor with practical adaptability.<n>AGILE Index 2025 evaluates 40 countries across income levels, regions, and technological development stages.
- Score: 13.374492753616067
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: The year 2024 witnessed accelerated global AI governance advancements, marked by strengthened multilateral frameworks and proliferating national regulatory initiatives. This acceleration underscores an unprecedented need to systematically track governance progress--an imperative that drove the launch of the AI Governance InternationaL Evaluation Index (AGILE Index) project since 2023. The inaugural AGILE Index, released in February 2024 after assessing 14 countries, established an operational and comparable baseline framework. Building on pilot insights, AGILE Index 2025 incorporates systematic refinements to better balance scientific rigor with practical adaptability. The updated methodology expands data diversity while enhancing metric validity and cross-national comparability. Reflecting both research advancements and practical policy evolution, AGILE Index 2025 evaluates 40 countries across income levels, regions, and technological development stages, with 4 Pillars, 17 Dimensions and 43 Indicators. In compiling the data, the team integrates multi-source evidence including policy documents, governance practices, research outputs, and risk exposure to construct a unified comparison framework. This approach maps global disparities while enabling countries to identify governance strengths, gaps, and systemic constraints. Through ongoing refinement and iterations, we hope the AGILE Index will fundamentally advance transparency and measurability in global AI governance, delivering data-driven assessments that depict national AI governance capacity, assist governments in recognizing their maturation stages and critical governance issues, and ultimately provide actionable insights for enhancing AI governance systems nationally and globally.
Related papers
- Toward Generalizable Evaluation in the LLM Era: A Survey Beyond Benchmarks [229.73714829399802]
This survey probes the core challenges that the rise of Large Language Models poses for evaluation.<n>We identify and analyze two pivotal transitions: (i) from task-specific to capability-based evaluation, which reorganizes benchmarks around core competencies such as knowledge, reasoning, instruction following, multi-modal understanding, and safety.<n>We will dissect this issue, along with the core challenges of the above two transitions, from the perspectives of methods, datasets, evaluators, and metrics.
arXiv Detail & Related papers (2025-04-26T07:48:52Z) - Policy Regularization on Globally Accessible States in Cross-Dynamics Reinforcement Learning [53.9544543607396]
We propose a novel framework that integrates reward rendering with Imitation from Observation (IfO)<n>By instantiating F-distance in different ways, we derive two theoretical analysis and develop a practical algorithm called Accessible State Oriented Policy Regularization (ASOR)<n>ASOR serves as a general add-on module that can be incorporated into various approaches RL, including offline RL and off-policy RL.
arXiv Detail & Related papers (2025-03-10T03:50:20Z) - AI Governance InternationaL Evaluation Index (AGILE Index) [15.589972522113754]
The rapid advancement of Artificial Intelligence (AI) technology is profoundly transforming human society.<n>Since 2022, the extensive deployment of generative AI, particularly large language models, marked a new phase in AI governance.<n>As consensus on international governance continues to be established and put into action, the practical importance of conducting a global assessment of the state of AI governance is progressively coming to light.<n>The inaugural evaluation of the AGILE Index commences with an exploration of four foundational pillars: the development level of AI, the AI governance environment, the AI governance instruments, and the AI governance effectiveness.
arXiv Detail & Related papers (2025-02-21T10:16:56Z) - AILuminate: Introducing v1.0 of the AI Risk and Reliability Benchmark from MLCommons [62.374792825813394]
This paper introduces AILuminate v1.0, the first comprehensive industry-standard benchmark for assessing AI-product risk and reliability.<n>The benchmark evaluates an AI system's resistance to prompts designed to elicit dangerous, illegal, or undesirable behavior in 12 hazard categories.
arXiv Detail & Related papers (2025-02-19T05:58:52Z) - OmniEval: An Omnidirectional and Automatic RAG Evaluation Benchmark in Financial Domain [62.89809156574998]
We introduce an omnidirectional and automatic RAG benchmark, OmniEval, in the financial domain.<n>Our benchmark is characterized by its multi-dimensional evaluation framework.<n>Our experiments demonstrate the comprehensiveness of OmniEval, which includes extensive test datasets.
arXiv Detail & Related papers (2024-12-17T15:38:42Z) - The Global AI Vibrancy Tool [4.31370360897884]
The Global AI Vibrancy Tool (GVT) is an interactive suite of visualizations designed to facilitate the comparison of AI vibrancy across 36 countries.<n>Using weights for indicators and pillars developed by AI Index's panel of experts, the Global AI Vibrancy Ranking for 2023 places the United States first by a significant margin.<n>The ranking also highlights the rise of smaller nations such as Singapore when evaluated on both absolute and per capita bases.
arXiv Detail & Related papers (2024-11-21T01:41:17Z) - A Large-Scale Study of Relevance Assessments with Large Language Models: An Initial Look [52.114284476700874]
This paper reports on the results of a large-scale evaluation (the TREC 2024 RAG Track) where four different relevance assessment approaches were deployed.
We find that automatically generated UMBRELA judgments can replace fully manual judgments to accurately capture run-level effectiveness.
Surprisingly, we find that LLM assistance does not appear to increase correlation with fully manual assessments, suggesting that costs associated with human-in-the-loop processes do not bring obvious tangible benefits.
arXiv Detail & Related papers (2024-11-13T01:12:35Z) - Future of Algorithmic Organization: Large-Scale Analysis of Decentralized Autonomous Organizations (DAOs) [45.02792904507959]
Decentralized Autonomous Organizations (DAOs) resemble early online communities, particularly those centered around open-source projects.
In just a few years, the deployment of governance tokens surged with a total of $24.5 billion and 11.1M governance token holders collectively managing decisions across over 13,000s as of 2024.
We examine factors such as voting power, participation, and characteristics dictating the level of decentralization, thus, the efficiency of management structures.
arXiv Detail & Related papers (2024-10-16T23:45:10Z) - Sustainable Visions: Unsupervised Machine Learning Insights on Global Development Goals [0.3764231189632788]
The 2030 Agenda for Sustainable Development outlines 17 goals for countries of the world to address global challenges in their development.<n>We have used a novel data-driven methodology to analyze time-series data for over 20 years (2000-2022) from 107 countries using unsupervised machine learning (ML) techniques.<n>Our findings show that progress toward the SDGs is heavily influenced by geographical, cultural and socioeconomic factors.
arXiv Detail & Related papers (2024-09-19T03:10:49Z) - Navigating Governance Paradigms: A Cross-Regional Comparative Study of Generative AI Governance Processes & Principles [19.25514463100802]
This paper aims toDepict the nuances of nascent and diverse governance approaches based on risks, rules, outcomes, principles, or a mix across different regions around the globe.
Our research introduces a Harmonized GenAI Framework, "H-GenAIGF," based on the current governance approaches of six regions: European Union (EU), United States (US), China (CN), Canada (CA), United Kingdom (UK), and Singapore (SG)
arXiv Detail & Related papers (2024-08-14T08:16:44Z) - Worldwide AI Ethics: a review of 200 guidelines and recommendations for
AI governance [0.0]
This paper conducts a meta-analysis of 200 governance policies and ethical guidelines for AI usage published by public bodies, academic institutions, private companies, and civil society organizations worldwide.
We identify at least 17 resonating principles prevalent in the policies and guidelines of our dataset, released as an open-source database and tool.
We present the limitations of performing a global scale analysis study paired with a critical analysis of our findings, presenting areas of consensus that should be incorporated into future regulatory efforts.
arXiv Detail & Related papers (2022-06-23T18:03:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.