AI Deployment Authorisation Score (ADAS)
- ADAS is a standardized, multi-dimensional framework that evaluates AI deployment readiness using defined metrics for risk, alignment, externality, control, and auditability.
- It employs mathematically defined evaluation dimensions, cryptographic certification, and strict threshold logic to merge technical evidence with legal and regulatory standards.
- The framework supports critical applications in healthcare, social care, and infrastructure by offering actionable, machine-readable metrics for deployment authorization.
The AI Deployment Authorisation Score (ADAS) is a standardized, multi-dimensional quantification and gating mechanism for evaluating whether a given artificial intelligence system is authorized for deployment within a specific domain and jurisdiction. ADAS serves dual, complementary purposes: (1) as an actionable, machine-readable regulatory and governance framework for high-risk and regulated domains, and (2) as a deterministic deployment-readiness gate for production models in mission-critical fields such as healthcare, social care, and critical infrastructure. The ADAS framework formalizes deployment authorization via mathematically defined evaluation dimensions, cryptographically verifiable certification, and clear threshold logic, enabling integration with legal, regulatory, and commercial decision processes (Madad, 25 Nov 2025, Saparning, 11 Jan 2026).
1. Formal Definition and Architecture of ADAS
ADAS takes as input a candidate AI deployment represented as a tuple
where denotes the model, the associated data, the action space, the human-in-the-loop protocol, built-in technical controls (e.g., shutdown), and the use context. Evaluation is further parameterized by jurisdiction and domain .
The core ADAS function is
where is a five-dimensional score vector: Risk , Alignment , Externality , Control , and Auditability (sometimes referred to as Traceability, ). Each score is accompanied by a confidence interval for (Saparning, 11 Jan 2026).
In domain-specialized deployments, such as caregiving-AI, ADAS is realized via a single value in which aggregates weighted, normalized scores across a bespoke set of longitudinally evaluated safety and relationship-centric dimensions, subject to a strict set of “autofail” override conditions (Madad, 25 Nov 2025).
2. Core Evaluation Dimensions
Cross-domain Regulatory Dimensions
The global regulatory ADAS instantiation evaluates systems according to:
| Dimension | Semantics | Aggregation Formula |
|---|---|---|
| Risk (R) | Probability × severity of harm | |
| Alignment (A) | Behavioral closeness to intended policy | |
| Externality(E) | Effects on third parties, discrimination, etc. | |
| Control (C) | Human override and failsafe robustness | |
| Auditability(T) | Completeness & integrity of audit artefacts |
Metrics include hazard catalogues, observed policy divergences, disparate impact ratios, human-in-the-loop efficacy, and audit artifact coverage (Saparning, 11 Jan 2026).
Caregiving Deployment Dimensions
InvisibleBench specializes ADAS for relationship-oriented AI deployments through the following weighted dimensions:
| Dimension | Weight | Key Subcomponents/Autofails |
|---|---|---|
| Safety | 20% | Crisis detection; autofail: missed crisis |
| Compliance | 15% | Regulatory boundaries; autofail: WOPR violations |
| Trauma-Informed Design | 15% | SAMHSA principle adherence; no autofail |
| Belonging/Cultural Fitness | 34% | Cultural fit, relational quality, actionable support; autofail: attachment engineering |
| Memory | 16% | Longitudinal fact & privacy consistency; no autofail |
Each dimension is scored through domain-specific rubrics (LLM judge cascades, deterministic pattern-matching, and scenario metadata checks), and only scenarios with all autofail checks passing contribute a nonzero ADAS (Madad, 25 Nov 2025).
3. Scoring and Threshold Logic
The ADAS methodology proceeds in structured steps:
- Per-dimension Evaluation: For each aspect and turn (where applicable), apply rubric-based or algorithmic scoring: .
- Raw Summation: Aggregate over turns: .
- Normalization: for each .
- Renormalization of Weights: (where is the applicable scenario dimensions).
- Composite ADAS:
Autofail overrides—for explicit missed crises, regulatory violations (WOPR), or attachment engineering—set for the scenario, regardless of partial credits (Madad, 25 Nov 2025).
Threshold logic implements deployment gates:
- PASS: and zero autofails ( deploy-ready)
- REVIEW: and zero autofails (manual review required)
- FAIL: or any autofail triggered
- Lexicographic rules in critical domains may require specific per-dimension minimums—e.g., , before any further evaluation (Saparning, 11 Jan 2026).
4. Evidence, Certification, and Transparency
ADAS requires an Evidence Bundle: an append-only, content-hashed set of artefacts (model cards, audit reports, provenance logs, benchmark outputs). Each artefact is timestamped and referenced in the decision process. For each dimension , a test suite and scoring function process the evidence to yield and .
Authorisation Pipeline (simplified pseudocode):
1 2 3 4 5 6 7 8 9 10 11 12 13 14 |
function AUTHORISE(S, J, U, EvidenceBundle E, Policy P):
if not E.contains_required(P.evidence_requirements):
return DENIED(reason="Missing required evidence")
for d in {R,A,E,C,T}:
(score[d], sigma[d]) = Evaluate_d(T_d, E)
for d in {R,A,E,C,T}:
if score[d] - sigma[d] < P.thresholds_JU[d]:
return DENIED(reason="Threshold failure in dimension " + d)
if P.has_conditional_rules:
conds = P.evaluate_conditions(S,E)
if conds not empty:
return APPROVED_WITH_CONDITIONS(conds)
cert = ISSUE_CERTIFICATE(S,J,U,score,sigma,P.version,E.fingerprint)
return APPROVED(certificate=cert) |
Authorisation results in a cryptographically signed certificate containing deployment details, jurisdiction, score vector, policy version, evidence fingerprint, issuance and expiry times, with publication into a Merkle-tree-based certificate transparency log. Certificate inclusion and potential revocation (also signed, logged) are auditable by relying parties (Saparning, 11 Jan 2026).
5. Domain-Specific Applications and Policy Integration
ADAS supports direct mapping to legal and insurance workflows:
- EU AI Act: Maps regulatory requirements to ADAS dimensions and sets explicit per-dimension thresholds (e.g., , , , , ). Approval is contingent on all thresholds being met under the relevant policy (Saparning, 11 Jan 2026).
- Critical Infrastructure (US): Implements lexicographic rules prioritizing control and auditing over aggregate scores—, —and only then evaluates risk minimization and further dimensions.
- Insurance Underwriting: Underwriting systems consume to calculate dynamic premiums; valid ADAS certificates are required for claim eligibility, and revocation leads to premium adjustment or denial (Saparning, 11 Jan 2026).
- Caregiving AI: InvisibleBench’s deployment gate applies ADAS as a “go/no-go” threshold, integrating autofails and longitudinal risk scoring to protect end-users and enforce regulatory regimes such as the Illinois WOPR Act (Madad, 25 Nov 2025).
6. Limitations and Recommendations
Key limitations of current ADAS instantiations include:
- Test Suite Boundaries: Reliance on scripted scenarios (e.g., InvisibleBench) rather than real-world interaction logs.
- Regulatory Anchoring: Current deployments (e.g., caregiving AI) are geographically and linguistically anchored (US/Illinois, English-only).
- Judge and Detection Biases: Use of a single LLM as judge for scoring introduces potential biases; heuristics for attachment engineering remain only provisionally validated.
- Cross-jurisdiction Calibration: Translating thresholds and metrics across heterogeneous international regulations is complex and may require empirical adjustment (Saparning, 11 Jan 2026).
- Metric Gaming: Public knowledge of test suites may incentivize tailored optimization rather than robust safety.
Both (Madad, 25 Nov 2025) and (Saparning, 11 Jan 2026) advise, especially in high-risk caregiving scenarios, a hybrid approach in which deterministic pattern-matching (crisis detection, WOPR term blocklists) supplements ADAS as a further technical safeguard (“deterministic safety routing”)—i.e., full production deployment should couple machine-scored ADAS gates with hard-coded, model-independent overrides for major safety and legality violations.
7. Future Directions
The ADAS paradigm is positioned as the missing institutional primitive bridging descriptive model reporting and enforceable deployment licensure. Identified avenues for future research and system evolution include:
- Adaptive threshold adjustment via Bayesian post-incident updating.
- Automated evidence acquisition from real-time system logs.
- Expansion of ADAS for real-time, API-level deployment gating and automated shutoff.
- Formal analysis of adversarial actions and supply-chain threats in authorization frameworks.
- Continuous cognitive and cultural adaptation of dimensions for new deployment contexts and policy updates (Saparning, 11 Jan 2026).
A plausible implication is that ADAS, by binding technical evidence, regulatory thresholds, cryptographic transparency, and operational overrides, serves both as a deployment license protocol and an institutional enforcement layer necessary for scalable, lawful, and safe AI system deployment.