The conversation about making governance operational – not just presentable – is exactly what's happening at Human x AI Europe on May 19 in Vienna, where implementation practitioners are gathering to share what survives contact with production systems.
The Problem With Most Governance Templates
Here's what happens in most organizations: someone downloads a governance framework, fills in the blanks, gets executive sign-off, and files it somewhere. Six months later, a model drifts, an incident occurs, and nobody can find the document – let alone use it to respond.
The gap between governance-as-document and governance-as-practice is where AI projects die. Templates aren't the problem. The problem is templates designed for auditors instead of operators.
A recent analysis on DEV Community puts it directly: The most important thing to understand about AI governance is that a single policy applied to everyone will be wrong for most of them. That's the starting point for any framework worth implementing.
Role-Based Governance: The First Decision That Matters
Before selecting a framework, answer this question: Who in the organization needs guardrails, and who needs accountability?
A front-line employee using an AI-powered tool to draft customer responses operates in a fundamentally different context than a developer building that tool. The first person wasn't hired to evaluate data handling risks. The second person was.
Blanket restrictions applied uniformly accomplish two things: they frustrate the people whose judgment the organization depends on, and they create a false sense of security about everyone else.
The DEV Community framework distinguishes explicitly between these populations:
- Front-line staff: Appropriate guardrails, firewall rules, restricted tool access. Not because they can't be trusted, but because they haven't been given the context to make good judgments in this domain.
- Technical professionals: Training, clear principles, and accountability. The question shifts from is this tool permitted to does this person have the context to make good decisions, and are they accountable for outcomes?
This distinction should appear in the first section of any governance document. If it doesn't, the template is already failing.
What a 12-Month Implementation Actually Looks Like
The EvalCommunity AI Governance Implementation Roadmap provides a structured 12-month approach that's worth examining – not because every organization should follow it exactly, but because it makes the phases explicit.
Quarter 1: Foundation
- Appoint a governance lead with actual authority
- Conduct an inventory of existing AI systems (most organizations skip this and regret it)
- Select frameworks and customize them to organizational context
Quarter 2: Design
- Build the ethics review process with clear criteria
- Create risk assessment methodology
- Develop the tools people will actually use: templates, checklists, dashboards
Quarter 3: Pilot
- Test governance with 3-5 real projects
- Train staff (target: 80% completion)
- Refine based on what breaks
Quarter 4: Scale
- Expand to all AI projects
- Integrate with existing project management lifecycle
- Establish reporting and continuous improvement
The key success metrics from this roadmap: 100% of AI projects under governance by month 12, 80% of staff trained, zero major incidents, 90% stakeholder satisfaction. These are measurable. That matters.
Framework Selection: What's Available
Several frameworks compete for attention. The right choice depends on organizational context, regulatory environment, and risk profile.
Current analysis identifies the major options:
- NIST AI Risk Management Framework (AI RMF): Four core functions – Govern, Map, Measure, Manage. Strong foundation for US organizations.
- ISO 42001: International standard for AI management systems. Useful for organizations needing certification.
- EU AI Act: Legally binding for high-risk systems. Risk-based classification with mandatory conformity assessments.
- OWASP Top 10 for LLMs: Security-focused, practical for development teams.
The Singapore Model AI Governance Framework offers a different approach: algorithm-agnostic, technology-agnostic, and sector-agnostic. It focuses on four areas: internal governance, decision-making models, operations management, and customer relationship management.
Most organizations will need to combine elements from multiple frameworks. The mistake is treating framework selection as a one-time decision rather than an ongoing adaptation.
The Documentation That Actually Matters
When an incident occurs, three questions need immediate answers:
- What does good enough look like for this system?
- Who gets paged when it breaks?
- How does rollback work?
If the governance documentation can't answer these questions quickly, it's not operational documentation – it's compliance theater.
The FINOS AI Governance Framework demonstrates what working governance documentation looks like: clear roles, explicit decision-making processes, defined approval thresholds, and documented escalation paths.
Beyond incident response, governance documentation should include:
- Model cards: What the system does, what it doesn't do, known limitations
- Risk classifications: Which systems are high-risk, what controls apply
- Data lineage: Where training data came from, how it was processed
- Change logs: What changed, when, and why
The FairNow AI Governance Policy Template provides a 16-page structure covering policy statements, roles and responsibilities, oversight mechanisms, and escalation procedures. It's designed for operationalization, not just compliance.
What Goes Wrong
The Databricks analysis of governance challenges identifies common failure modes:
- Unclear ownership: Responsibility fragmented across data, engineering, legal, and business teams. Models ship, but no one owns outcomes.
- Fragmented systems: Data, training, deployment, and monitoring live in separate systems. Oversight becomes impossible.
- Limited auditability: Proving how a system was trained, evaluated, and deployed requires documentation that doesn't exist.
The Federal AI Community of Practice Governance Toolkit addresses this by mapping stakeholders to specific governance functions. The Chief Privacy Officer (CPO) owns different aspects than the Chief Information Security Officer (CISO). Making these relationships explicit prevents the I thought someone else was handling that failure mode.
The Minimum Viable Governance Document
For organizations starting from zero, here's what the first version needs:
- Scope: Which AI systems are covered, which aren't
- Roles: Who approves, who monitors, who responds to incidents
- Risk classification: How systems are categorized, what controls apply to each tier
- Review process: How new AI projects get evaluated before deployment
- Monitoring requirements: What gets tracked, how often, by whom
- Incident response: What happens when something breaks
Everything else can be added later. These six elements are the foundation.
Frequently Asked Questions
Q: What is the difference between AI governance and AI ethics?
A: AI governance is the operational framework – policies, roles, processes, and controls that manage AI systems throughout their lifecycle. AI ethics provides the principles (fairness, transparency, accountability) that governance frameworks implement. Governance is how; ethics is why.
Q: How long does AI governance implementation typically take?
A: A structured implementation following the EvalCommunity roadmap takes approximately 12 months to reach full organizational coverage, with pilot projects beginning around month 7.
Q: Which AI governance framework should a European organization use?
A: European organizations must comply with the EU AI Act for high-risk systems. Most combine this with NIST AI RMF for operational guidance and ISO 42001 if certification is required. Framework selection depends on sector, risk profile, and existing compliance infrastructure.
Q: What documentation is required for AI governance compliance?
A: At minimum: system inventory with risk classifications, model cards documenting capabilities and limitations, data lineage records, change logs, incident response procedures, and role assignments with clear accountability.
Q: How should AI governance differ for developers versus end users?
A: Developers need training, clear principles, and accountability for outcomes. End users need appropriate guardrails and restricted access to high-risk tools. Applying identical restrictions to both populations undermines the judgment of technical professionals while providing false security for everyone else.
Q: What happens if an organization deploys AI without a governance framework?
A: Under the EU AI Act, violations involving high-risk systems can result in fines up to €35 million or 7% of global annual turnover. Beyond regulatory risk, ungoverned AI systems create operational risks: model drift goes undetected, incidents lack response procedures, and accountability gaps emerge when outcomes go wrong.