A multi-component framework for the analysis and design of explainable
artificial intelligence
- URL: http://arxiv.org/abs/2005.01908v1
- Date: Tue, 5 May 2020 01:48:40 GMT
- Title: A multi-component framework for the analysis and design of explainable
artificial intelligence
- Authors: S. Atakishiyev, H. Babiker, N. Farruque, R. Goebel1, M-Y. Kima, M.H.
Motallebi, J. Rabelo, T. Syed, O. R. Za\"iane
- Abstract summary: The rapid growth of research in explainable artificial intelligence (XAI) follows on two substantial developments.
The emergence of concern for creating trusted AI systems, including the creation of regulatory principles to ensure transparency and trust of AI systems.
Here we intend to provide a strategic inventory of XAI requirements, demonstrate their connection to a history of XAI ideas, and synthesize those ideas into a simple framework to calibrate five successive levels of XAI.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The rapid growth of research in explainable artificial intelligence (XAI)
follows on two substantial developments. First, the enormous application
success of modern machine learning methods, especially deep and reinforcement
learning, which have created high expectations for industrial, commercial and
social value. Second, the emergence of concern for creating trusted AI systems,
including the creation of regulatory principles to ensure transparency and
trust of AI systems.These two threads have created a kind of "perfect storm" of
research activity, all eager to create and deliver it any set of tools and
techniques to address the XAI demand. As some surveys of current XAI suggest,
there is yet to appear a principled framework that respects the literature of
explainability in the history of science, and which provides a basis for the
development of a framework for transparent XAI. Here we intend to provide a
strategic inventory of XAI requirements, demonstrate their connection to a
history of XAI ideas, and synthesize those ideas into a simple framework to
calibrate five successive levels of XAI.
Related papers
- Computational Safety for Generative AI: A Signal Processing Perspective [65.268245109828]
computational safety is a mathematical framework that enables the quantitative assessment, formulation, and study of safety challenges in GenAI.
We show how sensitivity analysis and loss landscape analysis can be used to detect malicious prompts with jailbreak attempts.
We discuss key open research challenges, opportunities, and the essential role of signal processing in computational AI safety.
arXiv Detail & Related papers (2025-02-18T02:26:50Z) - PXGen: A Post-hoc Explainable Method for Generative Models [0.5266869303483376]
generative AI (XAI) plays a crucial role in ensuring the responsible development and deployment of generative AI technologies.
Recent research emphasizes that a proficient XAI method should adhere to a set of criteria, primarily focusing on two key areas.
We propose PXGen, a post-hoc explainable method for generative models.
arXiv Detail & Related papers (2025-01-21T02:10:50Z) - Generative Artificial Intelligence Meets Synthetic Aperture Radar: A Survey [49.29751866761522]
This paper aims to investigate the intersection of GenAI and SAR.
First, we illustrate the common data generation-based applications in SAR field.
Then, an overview of the latest GenAI models is systematically reviewed.
Finally, the corresponding applications in SAR domain are also included.
arXiv Detail & Related papers (2024-11-05T03:06:00Z) - Explainable Artificial Intelligence: A Survey of Needs, Techniques, Applications, and Future Direction [5.417632175667161]
Explainable Artificial Intelligence (XAI) addresses challenges by providing explanations for how these models make decisions and predictions.
Existing studies have examined the fundamental concepts of XAI, its general principles, and the scope of XAI techniques.
This paper provides a comprehensive literature review encompassing common terminologies and definitions, the need for XAI, beneficiaries of XAI, a taxonomy of XAI methods, and the application of XAI methods in different application areas.
arXiv Detail & Related papers (2024-08-30T21:42:17Z) - Applications of Explainable artificial intelligence in Earth system science [12.454478986296152]
This review aims to provide a foundational understanding of explainable AI (XAI)
XAI offers a set of powerful tools that make the models more transparent.
We identify four significant challenges that XAI faces within the Earth system science (ESS)
A visionary outlook for ESS envisions a harmonious blend where process-based models govern the known, AI models explore the unknown, and XAI bridges the gap by providing explanations.
arXiv Detail & Related papers (2024-06-12T15:05:29Z) - Evolutionary Computation and Explainable AI: A Roadmap to Understandable Intelligent Systems [37.02462866600066]
Evolutionary computation (EC) offers significant potential to contribute to explainable AI (XAI)
This paper provides an introduction to XAI and reviews current techniques for explaining machine learning models.
We then explore how EC can be leveraged in XAI and examine existing XAI approaches that incorporate EC techniques.
arXiv Detail & Related papers (2024-06-12T02:06:24Z) - Usable XAI: 10 Strategies Towards Exploiting Explainability in the LLM Era [77.174117675196]
XAI is being extended towards Large Language Models (LLMs)
This paper analyzes how XAI can benefit LLMs and AI systems.
We introduce 10 strategies, introducing the key techniques for each and discussing their associated challenges.
arXiv Detail & Related papers (2024-03-13T20:25:27Z) - Explainable AI (XAI): A Systematic Meta-Survey of Current Challenges and
Future Opportunities [0.0]
Explainable AI (XAI) has been proposed to make AI more transparent and thus advance the adoption of AI in critical domains.
This study presents a systematic meta-survey for challenges and future research directions in XAI.
arXiv Detail & Related papers (2021-11-11T19:06:13Z) - Counterfactual Explanations as Interventions in Latent Space [62.997667081978825]
Counterfactual explanations aim to provide to end users a set of features that need to be changed in order to achieve a desired outcome.
Current approaches rarely take into account the feasibility of actions needed to achieve the proposed explanations.
We present Counterfactual Explanations as Interventions in Latent Space (CEILS), a methodology to generate counterfactual explanations.
arXiv Detail & Related papers (2021-06-14T20:48:48Z) - An interdisciplinary conceptual study of Artificial Intelligence (AI)
for helping benefit-risk assessment practices: Towards a comprehensive
qualification matrix of AI programs and devices (pre-print 2020) [55.41644538483948]
This paper proposes a comprehensive analysis of existing concepts coming from different disciplines tackling the notion of intelligence.
The aim is to identify shared notions or discrepancies to consider for qualifying AI systems.
arXiv Detail & Related papers (2021-05-07T12:01:31Z) - Distributed and Democratized Learning: Philosophy and Research
Challenges [80.39805582015133]
We propose a novel design philosophy called democratized learning (Dem-AI)
Inspired by the societal groups of humans, the specialized groups of learning agents in the proposed Dem-AI system are self-organized in a hierarchical structure to collectively perform learning tasks more efficiently.
We present a reference design as a guideline to realize future Dem-AI systems, inspired by various interdisciplinary fields.
arXiv Detail & Related papers (2020-03-18T08:45:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.