AI Governance Library

METHODOLOGY FOR THE RISK AND IMPACT ASSESSMENT OF ARTIFICIAL INTELLIGENCE SYSTEMS FROM THE POINT OF VIEW OF HUMAN RIGHTS, DEMOCRACY AND THE RULE OF LAW (HUDERIA METHODOLOGY)

HUDERIA is a methodology adopted by the Council of Europe in November 2024 for assessing AI systems from the perspective of human rights, democracy, and the rule of law.
METHODOLOGY FOR THE RISK AND IMPACT ASSESSMENT OF ARTIFICIAL INTELLIGENCE SYSTEMS FROM THE POINT OF VIEW OF HUMAN RIGHTS, DEMOCRACY AND THE RULE OF LAW (HUDERIA METHODOLOGY)

What’s Covered?

The HUDERIA (Human Rights, Democracy, Rule of Law Impact Assessment) is not a checklist. It’s a detailed, structured framework designed to help public and private actors think through how an AI system could impact human rights and democratic values—both in theory and in the messiness of real-world deployment.

It’s built around four core elements:

  1. Context-Based Risk Analysis (COBRA)A structured process for scoping out the AI system, identifying contextual risk factors, mapping potential impacts, and triaging systems based on severity and probability of harm.It examines three risk domains:
    • Application context (e.g., purpose, legal environment)
    • Design/development context (e.g., data quality, model explainability)
    • Deployment context (e.g., safeguards, training, misuse risks)
  2. Stakeholder Engagement Process (SEP)Recommends participatory processes tailored to the system’s risk profile. This includes stakeholder mapping, positionality reflection (how the developers’ identity and power shapes outcomes), and practical design of inclusive engagement methods.
  3. Risk and Impact Assessment (RIA)A two-step deep dive into the impacts previously identified, focusing on scale, scope, reversibility, and probability—especially for marginalized or vulnerable populations. Encourages reflection on cumulative and long-term effects.
  4. Mitigation Plan (MP)Introduces a mitigation hierarchy (avoid, reduce, restore, compensate), prioritizing prevention over remediation. Also covers access to remedies and outlines how to document and assign responsibility for mitigation.

The final section introduces Iterative Review, recognizing that AI systems evolve, and their real-world effects often shift after deployment. This component calls for continuous reassessment based on environmental, social, legal, and technical changes.

💡 Why it matters?

HUDERIA represents one of the first attempts to build a risk and impact methodology that’s grounded in human rights law and democratic theory, not just technical or economic concerns. It brings political legitimacy and legal nuance to AI governance, especially for public sector applications and states bound by European conventions.

What’s Missing?

  • Operational simplicity: HUDERIA is conceptually rich but resource-heavy. Without a clear lightweight version, it may be tough for smaller actors or less-resourced administrations to apply in practice.
  • Private sector usability: Though public and private actors can both use HUDERIA, it’s written primarily for state actors. A more explicit mapping to corporate governance models would boost uptake.
  • Examples and case studies: The guidance is dense with process but light on practical illustrations. Concrete examples would make it more approachable.
  • No integration with risk scoring: There’s no clear method for converting the rich qualitative assessments into a risk level that can trigger thresholds or governance obligations.

Best For:

Public authorities, regulators, legal experts, and AI governance professionals in Council of Europe member states looking for a structured way to address human rights risks. Especially useful for those developing national or sectoral AI oversight mechanisms.

Source Details:

Title: HUDERIA: Methodology for the Risk and Impact Assessment of Artificial Intelligence Systems from the Point of View of Human Rights, Democracy and the Rule of Law

Authors: Committee on Artificial Intelligence (CAI), Council of Europe

Adopted: 28 November 2024 (CAI(2024)16rev2)

Model origin: Developed with input from the Alan Turing Institute following earlier work of CAHAI

Type: Non-binding guidance, not a legal instrument

Relation to AI Convention: Complementary and optional, but aligns with Chapter V of the Council of Europe Framework Convention on AI

About the author
Jakub Szarmach

AI Governance Library

Curated Library of AI Governance Resources

AI Governance Library

Great! You’ve successfully signed up.

Welcome back! You've successfully signed in.

You've successfully subscribed to AI Governance Library.

Success! Check your email for magic link to sign-in.

Success! Your billing info has been updated.

Your billing was not updated.