Self-Certification of High-Risk AI Systems: The Example of AI-based Facial Emotion Recognition
- URL: http://arxiv.org/abs/2601.08295v1
- Date: Tue, 13 Jan 2026 07:34:28 GMT
- Title: Self-Certification of High-Risk AI Systems: The Example of AI-based Facial Emotion Recognition
- Authors: Gregor Autischer, Kerstin Waxnegger, Dominik Kowald,
- Abstract summary: The European Union's Artificial Intelligence Act establishes comprehensive requirements for high-risk AI systems.<n>We investigate the practical application of the Fraunhofer AI assessment catalogue as a certification framework.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The European Union's Artificial Intelligence Act establishes comprehensive requirements for high-risk AI systems, yet the harmonized standards necessary for demonstrating compliance remain not fully developed. In this paper, we investigate the practical application of the Fraunhofer AI assessment catalogue as a certification framework through a complete self-certification cycle of an AI-based facial emotion recognition system. Beginning with a baseline model that has deficiencies, including inadequate demographic representation and prediction uncertainty, we document an enhancement process guided by AI certification requirements. The enhanced system achieves higher accuracy with improved reliability metrics and comprehensive fairness across demographic groups. We focused our assessment on two of the six Fraunhofer catalogue dimensions, reliability and fairness, the enhanced system successfully satisfies the certification criteria for these examined dimensions. We find that the certification framework provides value as a proactive development tool, driving concrete technical improvements and generating documentation naturally through integration into the development process. However, fundamental gaps separate structured self-certification from legal compliance: harmonized European standards are not fully available, and AI assessment frameworks and catalogues cannot substitute for them on their own. These findings establish the Fraunhofer AI assessment catalogue as a valuable preparatory tool that complements rather than replaces formal compliance requirements at this time.
Related papers
- Nishpaksh: TEC Standard-Compliant Framework for Fairness Auditing and Certification of AI Models [4.881152405850494]
We propose Nishpaksh, an indigenous fairness evaluation tool that operationalizes the Telecommunication Engineering Centre (TEC) Standard for the Evaluation and Rating of Artificial Intelligence Systems.<n>Nishpaksh integrates survey-based risk, contextual threshold determination, and quantitative fairness evaluation into a unified, web-based dashboard.
arXiv Detail & Related papers (2026-01-23T17:35:05Z) - AI-NativeBench: An Open-Source White-Box Agentic Benchmark Suite for AI-Native Systems [52.65695508605237]
We introduce AI-NativeBench, the first application-centric and white-box AI-Native benchmark suite grounded in Model Context Protocol (MCP) and Agent-to-Agent (A2A) standards.<n>By treating agentic spans as first-class citizens within distributed traces, our methodology enables granular analysis of engineering characteristics beyond simple capabilities.<n>This work provides the first systematic evidence to guide the transition from measuring model capability to engineering reliable AI-Native systems.
arXiv Detail & Related papers (2026-01-14T11:32:07Z) - Safe and Certifiable AI Systems: Concepts, Challenges, and Lessons Learned [45.44933002008943]
This white paper presents the T"UV AUSTRIA Trusted AI framework.<n>It is an end-to-end audit catalog and methodology for assessing and certifying machine learning systems.<n>Building on three pillars - Secure Software Development, Functional Requirements, and Ethics & Data Privacy - it translates the high-level obligations of the EU AI Act into specific, testable criteria.
arXiv Detail & Related papers (2025-09-08T17:52:08Z) - Criteria for Credible AI-assisted Carbon Footprinting Systems: The Cases of Mapping and Lifecycle Modeling [0.0]
We present a set of criteria to validate AI-assisted systems that calculate greenhouse gas (GHG) emissions for products and materials.<n>This approach may be used as a foundation for practitioners, auditors, and standards bodies to evaluate AI-assisted environmental assessment tools.
arXiv Detail & Related papers (2025-08-29T21:05:19Z) - The Architecture of Trust: A Framework for AI-Augmented Real Estate Valuation in the Era of Structured Data [0.0]
The Uniform Appraisal dataset (UAD) 3.6's mandatory 2026 implementation transforms residential property valuation from narrative reporting to machine-readable formats.<n>This paper provides the first comprehensive analysis of this regulatory shift alongside concurrent AI advances in computer vision, natural language processing, and autonomous systems.<n>We develop a three-layer framework for AI-augmented valuation addressing technical implementation and institutional trust requirements.
arXiv Detail & Related papers (2025-08-04T05:24:25Z) - The AI Imperative: Scaling High-Quality Peer Review in Machine Learning [49.87236114682497]
We argue that AI-assisted peer review must become an urgent research and infrastructure priority.<n>We propose specific roles for AI in enhancing factual verification, guiding reviewer performance, assisting authors in quality improvement, and supporting ACs in decision-making.
arXiv Detail & Related papers (2025-06-09T18:37:14Z) - Image Quality Assessment for Embodied AI [103.66095742463195]
Embodied AI has developed rapidly in recent years, but it is still mainly deployed in laboratories.<n>There is no IQA method to assess the usability of an image in embodied tasks, namely, the perceptual quality for robots.
arXiv Detail & Related papers (2025-05-22T15:51:07Z) - AILuminate: Introducing v1.0 of the AI Risk and Reliability Benchmark from MLCommons [62.374792825813394]
This paper introduces AILuminate v1.0, the first comprehensive industry-standard benchmark for assessing AI-product risk and reliability.<n>The benchmark evaluates an AI system's resistance to prompts designed to elicit dangerous, illegal, or undesirable behavior in 12 hazard categories.
arXiv Detail & Related papers (2025-02-19T05:58:52Z) - Practical Application and Limitations of AI Certification Catalogues in the Light of the AI Act [0.1433758865948252]
This work focuses on the practical application and limitations of existing certification catalogues in the light of the AI Act.<n>We use the AI Assessment Catalogue as a comprehensive tool to systematically assess an AI model's compliance with certification standards.<n>We observe the limitations of an AI system that has no active development team anymore and highlight the importance of complete system documentation.
arXiv Detail & Related papers (2025-01-20T15:54:57Z) - No Trust without regulation! [0.0]
The explosion in performance of Machine Learning (ML) and the potential of its applications are encouraging us to consider its use in industrial systems.
It is still leaving too much to one side the issue of safety and its corollary, regulation and standards.
The European Commission has laid the foundations for moving forward and building solid approaches to the integration of AI-based applications that are safe, trustworthy and respect European ethical values.
arXiv Detail & Related papers (2023-09-27T09:08:41Z) - Toward Trustworthy AI Development: Mechanisms for Supporting Verifiable
Claims [59.64274607533249]
AI developers need to make verifiable claims to which they can be held accountable.
This report suggests various steps that different stakeholders can take to improve the verifiability of claims made about AI systems.
We analyze ten mechanisms for this purpose--spanning institutions, software, and hardware--and make recommendations aimed at implementing, exploring, or improving those mechanisms.
arXiv Detail & Related papers (2020-04-15T17:15:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.