AI RISK ASSESSMENT
MARIA: Marginal Risk Assessment Without Ground Truth for AI Systems
Before deploying any AI system, organizations must ensure it delivers improvement without introducing undue risk. Traditional evaluations often fail due to the impracticality of obtaining reliable ground truth data. This paper introduces MARIA, a novel framework that shifts focus from absolute risk to marginal risk – the difference in risk between a new AI system and an existing baseline. By emphasizing relative performance across predictability, capability, and interaction dominance, MARIA equips teams with actionable insights to safely and responsibly adopt AI, even in complex domains without a definitive 'correct' answer.
Executive Impact
Understand the critical shifts and opportunities MARIA presents for enterprise AI adoption.
Deep Analysis & Enterprise Applications
Select a topic to dive deeper, then explore the specific findings from the research, rebuilt as interactive, enterprise-focused modules.
The Problem with Ground Truth
Traditional AI evaluation heavily relies on comparing system outputs to a known "ground truth." However, in real-world enterprise scenarios, obtaining this definitive truth is often impractical or impossible. MARIA identifies four key challenges:
- Unknowable: Real-world outcomes or long-term impacts are often inherently unpredictable or untraceable (e.g., social harm, counterfactuals in bail decisions).
- Delayed: Ground truth may only become apparent after long periods, rendering immediate evaluation impossible (e.g., vendor quality, policy effects).
- Expensive: Establishing ground truth can be prohibitively costly, requiring extensive expert labor, time, and resources at scale (e.g., multiple expert reviews for every submission).
- Withheld: Relevant data may exist but is not collected or disclosed due to privacy concerns, institutional omission, or unwillingness to track performance.
These challenges necessitate a new evaluation paradigm that focuses on relative performance rather than absolute metrics against an elusive truth.
MARIA: A New Paradigm for AI Risk Assessment
MARIA proposes a marginal risk assessment (MR) approach, defined as MR = AR = R_new - R_baseline. This framework evaluates the additional risk introduced or mitigated when a new AI system replaces an existing one, without assuming access to ground truth.
It focuses on three complementary dimensions of relative evaluation:
- Predictability: Measures how stable and internally coherent a system is under repetition, perturbation, and control changes. Metrics include Self-consistency, Cross-consensus, Input stability, Control stability, and Uncertainty governance.
- Capability: Assesses the functional competence of the underlying AI and the system built upon it. Metrics include Model-level benchmarks and System effectiveness & operational efficiency.
- Interaction Dominance: Evaluates how systems behave relative to each other in structured, symmetric games. Metrics include Persuasion duels, Prediction-surprise games, and Compression-reconstruction games.
By shifting to relative evaluation, MARIA provides actionable guidance for identifying where AI enhances outcomes and where it introduces new risks, facilitating responsible AI adoption.
Real-World Application: Document Evaluation
The paper showcases MARIA's utility through a case study on document evaluation, a common organizational workflow. Traditionally, this involves two human evaluators, with a third triggered for significant discrepancies. The goal was to assess the marginal risk of replacing one human with an AI system (LLM).
Key Findings:
- Predictability: Both human and AI reviewers showed high self-consistency, but AI exhibited slightly higher sensitivity to input variations. This highlights internal reliability but also areas for calibration.
- Performance: Human-AI agreement was moderate and lower than human-human, indicating discrepancies in specific criteria that need targeted calibration.
- Final Decision Consistency: Introducing AI shifted the frequency of third-review triggers, providing a practical signal for process-level risk monitoring.
- Fairness: Preliminary analysis revealed mild distributional shifts, underscoring the need for ongoing audits.
- Security & Efficiency: AI introduced new exposure points (e.g., prompt injection) but also significant efficiency gains (reduced manual review time).
This case study demonstrates how MARIA offers practical insights into where AI systems enhance outcomes and where they introduce new risks, enabling responsible integration without relying on an absolute, often unattainable, ground truth.
MARIA Workflow: A Phased Approach
| Feature | Traditional Evaluation | MARIA Framework |
|---|---|---|
| Core Assumption | Relies on absolute ground truth | Ground truth often unavailable or impractical |
| Evaluation Focus | Absolute performance & risk | Relative difference (R_new - R_baseline) |
| Key Challenges | Unknowable, delayed, expensive, withheld data | Addresses these challenges head-on |
| Metrics Used | Accuracy, F1-score (against ground truth) | Predictability, Capability, Interaction Dominance |
| Outcome | Often stalled or misleading | Actionable guidance for responsible AI adoption |
Deep Dive: Document Evaluation Case Study
The paper showcases MARIA's utility through a case study on document evaluation, a common organizational workflow. Traditionally, this involves two human evaluators, with a third triggered for significant discrepancies. The goal was to assess the marginal risk of replacing one human with an AI system (LLM).
Key Findings:
- Predictability: Both human and AI reviewers showed high self-consistency, but AI exhibited slightly higher sensitivity to input variations. This highlights internal reliability but also areas for calibration.
- Performance: Human-AI agreement was moderate and lower than human-human, indicating discrepancies in specific criteria that need targeted calibration.
- Final Decision Consistency: Introducing AI shifted the frequency of third-review triggers, providing a practical signal for process-level risk monitoring.
- Fairness: Preliminary analysis revealed mild distributional shifts, underscoring the need for ongoing audits.
- Security & Efficiency: AI introduced new exposure points (e.g., prompt injection) but also significant efficiency gains (reduced manual review time).
This case study demonstrates how MARIA offers practical insights into where AI systems enhance outcomes and where they introduce new risks, enabling responsible integration without relying on an absolute, often unattainable, ground truth.
Calculate Your Potential AI ROI
Estimate the efficiency gains and cost savings your enterprise could realize by implementing marginal risk-assessed AI solutions.
Your Journey to Responsible AI with MARIA
A structured approach to integrating marginal risk assessment into your enterprise AI strategy, ensuring safety and efficiency.
Phase 1: Define Scope & Baselines
Define the scope, identify baselines, and select appropriate metrics for marginal risk assessment, explicitly acknowledging ground truth limitations.
Phase 2: Comparative Analysis & Alignment
Generate system outputs under identical conditions, apply metric-specific evaluations (predictability, capability, interaction dominance), and analyze observed alignment and divergence.
Phase 3: Calibration & Risk Exploration
Validate metric reliability, adjust for systematic biases, and deeply explore comparative results, including game-based evaluations for interaction dominance, documenting all assumptions.
Phase 4: Aggregation & Interpretation
Normalize metric scales, aggregate results, form a relative ordering of system dominance, and prepare a transparent report detailing findings, assumptions, and actionable insights for responsible AI adoption.
Ready to Transform Your AI Evaluation?
Don't let the absence of perfect ground truth hinder your AI initiatives. Partner with us to implement the MARIA framework and navigate your AI adoption with confidence.