🤖 AI Summary
This work addresses critical gaps—namely, the lack of quantifiable security risk assessment, incomplete coverage, and misalignment with compliance requirements—in three major AI governance standards: NIST AI RMF 1.0, EU ALTAI, and UK ICO guidelines. We propose the first computable security gap measurement framework tailored for AI standards, featuring a four-dimensional metric system (RSI, AVPI, CSGP, RCVS) grounded in formal risk modeling, cross-standard auditing, and root-cause analysis to systematically identify 136 security concerns. Results reveal that NIST omits 69.23% of critical risks; ALTAI exhibits the most severe attack surface exposure (AVPI = 0.51); and ICO suffers an 80.00% compliance–security gap. Crucially, we uncover inter-standard disparities in risk response capability and root-cause–level vulnerability distributions—e.g., ALTAI’s insufficient process definitions and NIST’s weak implementation guidance. The study delivers actionable control enhancement recommendations, establishing a quantitative evaluation paradigm for AI governance standardization.
📝 Abstract
As AI systems integrate into critical infrastructure, security gaps in AI compliance frameworks demand urgent attention. This paper audits and quantifies security risks in three major AI governance standards: NIST AI RMF 1.0, UK's AI and Data Protection Risk Toolkit, and the EU's ALTAI. Using a novel risk assessment methodology, we develop four key metrics: Risk Severity Index (RSI), Attack Potential Index (AVPI), Compliance-Security Gap Percentage (CSGP), and Root Cause Vulnerability Score (RCVS). Our analysis identifies 136 concerns across the frameworks, exposing significant gaps. NIST fails to address 69.23 percent of identified risks, ALTAI has the highest attack vector vulnerability (AVPI = 0.51) and the ICO Toolkit has the largest compliance-security gap, with 80.00 percent of high-risk concerns remaining unresolved. Root cause analysis highlights under-defined processes (ALTAI RCVS = 033) and weak implementation guidance (NIST and ICO RCVS = 0.25) as critical weaknesses. These findings emphasize the need for stronger, enforceable security controls in AI compliance. We offer targeted recommendations to enhance security posture and bridge the gap between compliance and real-world AI risks.