aicomply.
STD-AI-008

AI Accuracy, Robustness & Security Standard

Ensure AI systems achieve appropriate levels of accuracy, robustness, and cybersecurity.

2

Controls

0

Compliant

2

In Progress

0

Not Started

Overall Progress
40%
Implementation Guidance
Detailed guidance for implementing this standard

AI Accuracy, Robustness and Security Standard

Document Type: Standard
Standard ID: STD-AI-008
Standard Title: AI Accuracy, Robustness and Security Standard
Version: 1.0
Effective Date: 2025-08-01
Next Review Date: 2026-08-01
Review Frequency: Annually or upon regulatory change
Parent Policy: POL-AI-001 - Artificial Intelligence Policy
Owner: Chief Technology Officer
Approved By: AI Governance Committee Chair
Status: Draft
Classification: Internal Use Only


TABLE OF CONTENTS

  1. Document History
  2. Objective
  3. Scope and Applicability
  4. Control Standard
  5. Supporting Procedures
  6. Compliance
  7. Roles and Responsibilities
  8. Exceptions
  9. Enforcement
  10. Key Performance Indicators (KPIs)
  11. Training Requirements
  12. Definitions
  13. Link with AI Act and ISO42001

DOCUMENT HISTORY

VersionDateAuthorChangesApproval DateApproved By
0.12025-07-05David Lee, Chief Technology OfficerInitial draft--
0.22025-07-20David Lee, Chief Technology OfficerAdded security details--
0.32025-08-01David Lee, Chief Technology OfficerIncorporated security feedback--
1.02025-08-01David Lee, Chief Technology OfficerFinal version approved - GRC restructured2025-07-25Jane Doe, AI Governance Committee Chair

OBJECTIVE

This standard defines requirements for ensuring AI systems achieve appropriate levels of accuracy, robustness, and cybersecurity in compliance with EU AI Act Article 15.

Primary Goals:

  • Ensure AI systems achieve appropriate accuracy levels per Article 15(1)
  • Ensure AI systems are robust per Article 15(4)
  • Ensure AI systems are resilient against cybersecurity threats per Article 15(5)

SCOPE AND APPLICABILITY

2.1 Mandatory Applicability

This standard is mandatory for:

  • All high-risk AI systems (EU AI Act Article 15)
  • All AI systems with significant impact on health, safety, or fundamental rights

2.2 Recommended Applicability

This standard is recommended for:

  • All AI systems for best practices
  • Limited-risk and minimal-risk AI systems (voluntary)

2.3 Requirements Covered

  • Accuracy requirements, testing, and monitoring
  • Robustness requirements, testing, and drift detection
  • Cybersecurity requirements, threat assessment, and security testing

2.4 Out of Scope

  • General software quality (covered by software development standards)
  • Non-AI system security (covered by IT security standards)
  • Requirements outside EU AI Act scope

CONTROL STANDARD

Control ARS-001: Accuracy Requirements Definition

Control ID: ARS-001
Control Name: Accuracy Requirements and Metrics Definition
Control Type: Preventive
Control Frequency: Per high-risk AI system, during design phase
Risk Level: High

Control Objective

Define accuracy requirements based on intended purpose per Article 15(1) to ensure AI systems achieve appropriate accuracy levels for their use case, enabling safe and effective deployment.

Control Requirements

CR-001.1: Accuracy Requirements Documentation

Define and document accuracy requirements with appropriate metrics and thresholds.

Accuracy Metrics:

MetricDefinitionUse CaseTarget (High-Risk)
Overall Accuracy% of correct predictionsGeneral classification≥95%
Precision% of positive predictions that are correctWhen false positives costly≥95%
Recall% of actual positives correctly identifiedWhen false negatives costly≥95%
F1 ScoreHarmonic mean of precision and recallBalanced metric≥95%
ROC-AUCArea under ROC curveBinary classification≥0.95
Domain-SpecificCustom metrics for specific domainDomain-specific use casesDefined per use case

Mandatory Actions:

  • Define accuracy requirements based on intended purpose
  • Select appropriate metrics
  • Set accuracy thresholds
  • Document requirements
  • Obtain AI System Owner approval
  • Review and update annually

Accuracy Requirements by Use Case:

Use CaseAccuracy MetricThresholdJustification
Medical DiagnosisPrecision, Recall≥98%High safety impact
Credit ScoringROC-AUC, Precision≥0.95, ≥95%Financial impact
Fraud DetectionPrecision, Recall≥95%, ≥90%Balance false positives/negatives
Image ClassificationOverall Accuracy≥95%General use case

Evidence Required:

  • Accuracy Requirements Document (DOC-AI-ARS-001)
  • Metrics definitions
  • Threshold specifications
  • Approval records
  • Annual review records

Audit Verification:

  • Verify accuracy requirements defined for all high-risk AI
  • Confirm metrics appropriate for use case
  • Check thresholds set and justified
  • Validate approval obtained
  • Verify annual review completed

Control ARS-002: Accuracy Testing

Control ID: ARS-002
Control Name: Accuracy Testing and Validation
Control Type: Preventive
Control Frequency: Before deployment, after model updates
Risk Level: High

Control Objective

Test AI system accuracy before deployment to verify it meets defined accuracy requirements, ensuring safe and effective operation.

Control Requirements

CR-002.1: Comprehensive Accuracy Testing

Conduct comprehensive accuracy testing using representative test data.

Testing Requirements:

RequirementSpecificationImplementation
Test DataRepresentative of production dataTest dataset validation
Use Case CoverageTest across all use casesUse case test matrix
User Group CoverageTest for all user groupsUser group test matrix
Metric CalculationCalculate all defined metricsAutomated metric calculation
Threshold VerificationVerify all thresholds metThreshold compliance check
DocumentationDocument all test resultsTest report

Mandatory Actions:

  • Plan accuracy testing
  • Prepare representative test data
  • Execute tests across all use cases
  • Test for all user groups
  • Calculate all defined metrics
  • Verify threshold compliance
  • Document test results
  • Block deployment if thresholds not met

Accuracy Test Plan:

Test ComponentDescriptionSuccess Criteria
Overall Accuracy TestTest overall accuracyMeets threshold
Precision TestTest precisionMeets threshold
Recall TestTest recallMeets threshold
Use Case TestsTest each use caseAll use cases meet thresholds
User Group TestsTest each user groupAll groups meet thresholds
Edge Case TestsTest edge casesAcceptable performance

Evidence Required:

  • Accuracy Test Plan (PLAN-AI-ARS-001)
  • Test results (TEST-AI-ARS-001)
  • Analysis reports
  • Threshold compliance verification
  • Approval records

Audit Verification:

  • Verify accuracy testing conducted before deployment
  • Confirm representative test data used
  • Check all use cases tested
  • Validate all user groups tested
  • Verify thresholds met
  • Check deployment blocked if thresholds not met

Control ARS-003: Accuracy Monitoring

Control ID: ARS-003
Control Name: Production Accuracy Monitoring
Control Type: Detective
Control Frequency: Continuous
Risk Level: Medium

Control Objective

Monitor accuracy in production to detect accuracy degradation, identify issues early, and enable timely corrective actions to maintain AI system performance.

Control Requirements

CR-003.1: Continuous Accuracy Monitoring

Implement continuous accuracy monitoring with alerting and analysis.

Monitoring Requirements:

RequirementSpecificationImplementation
Real-Time MonitoringMonitor accuracy metrics in real-timeReal-time monitoring system
Threshold AlertingAlert when accuracy drops below thresholdAutomated alerting
Trend AnalysisAnalyze accuracy trendsTrend analysis tools
Issue InvestigationInvestigate accuracy issues promptlyInvestigation procedures
Corrective ActionsImplement corrective actionsCorrective action procedures

Mandatory Actions:

  • Implement accuracy monitoring
  • Set up alerting for threshold breaches
  • Monitor continuously
  • Track accuracy trends
  • Investigate accuracy issues
  • Implement corrective actions
  • Document monitoring results

Accuracy Monitoring Metrics:

MetricThresholdAlert LevelFrequency
Overall Accuracy< 95% (high-risk)WarningDaily
Precision< 95%WarningDaily
Recall< 95%WarningDaily
Accuracy Degradation> 2% dropCriticalReal-time
Accuracy Stability> 2% variationWarningWeekly

Evidence Required:

  • Accuracy monitoring dashboard
  • Alert logs
  • Trend analysis reports
  • Investigation records
  • Corrective action records
  • Monthly accuracy reports

Audit Verification:

  • Verify accuracy monitoring implemented
  • Confirm alerting configured
  • Check monitoring reviewed regularly
  • Validate issues investigated
  • Verify corrective actions implemented

Control ARS-004: Robustness Requirements

Control ID: ARS-004
Control Name: Robustness Requirements Definition
Control Type: Preventive
Control Frequency: Per high-risk AI system, during design phase
Risk Level: High

Control Objective

Define robustness requirements per Article 15(4) to ensure AI systems are resilient to errors, faults, inconsistencies, and adversarial conditions, maintaining performance across diverse scenarios.

Control Requirements

CR-004.1: Comprehensive Robustness Requirements

Define robustness requirements across all relevant dimensions.

Robustness Dimensions:

DimensionDescriptionRequirementsMetrics
Technical RobustnessResilience to errors, faults, inconsistenciesError handling, fault toleranceError rate, fault recovery time
Input RobustnessHandling of edge cases, outliers, adversarial inputsInput validation, outlier handlingEdge case success rate, adversarial robustness
Environmental RobustnessPerformance across different conditionsEnvironmental adaptationPerformance across conditions
Temporal RobustnessStability over time (no drift)Drift detection, model stabilityDrift score, performance stability

Mandatory Actions:

  • Define robustness requirements
  • Establish robustness metrics
  • Set robustness thresholds
  • Document requirements
  • Obtain approval
  • Review and update annually

Robustness Requirements by Dimension:

DimensionMetricThresholdJustification
Technical RobustnessError rate< 1%Acceptable error handling
Input RobustnessEdge case success rate≥90%Handle edge cases
Adversarial RobustnessAdversarial success rate≥85%Resist adversarial attacks
Environmental RobustnessPerformance variation< 5%Consistent across conditions
Temporal RobustnessDrift score< 0.1Stable over time

Evidence Required:

  • Robustness Requirements Document (DOC-AI-ARS-002)
  • Metrics and thresholds
  • Approval records
  • Annual review records

Audit Verification:

  • Verify robustness requirements defined
  • Confirm all dimensions covered
  • Check thresholds set
  • Validate approval obtained

Control ARS-005: Robustness Testing

Control ID: ARS-005
Control Name: Robustness Testing and Validation
Control Type: Preventive
Control Frequency: Before deployment, after model updates
Risk Level: High

Control Objective

Test AI system robustness to verify it meets defined robustness requirements, ensuring resilience to errors, faults, and adversarial conditions.

Control Requirements

CR-005.1: Comprehensive Robustness Testing

Conduct comprehensive robustness testing across all dimensions.

Test Types:

Test TypePurposeTest MethodSuccess Criteria
Edge Case TestingTest handling of edge casesEdge case test suite≥90% success rate
Stress TestingTest under extreme conditionsStress test scenariosSystem remains functional
Adversarial TestingTest resistance to adversarial attacksAdversarial attack simulations≥85% success rate
Fault Injection TestingTest fault toleranceFault injection scenariosSystem recovers gracefully
Environmental Variation TestingTest across different conditionsEnvironmental test matrix< 5% performance variation
Drift TestingTest temporal stabilityDrift simulationDrift score < 0.1

Mandatory Actions:

  • Plan robustness testing
  • Execute all test types
  • Analyze results
  • Verify compliance with requirements
  • Document results
  • Block deployment if requirements not met

Robustness Test Plan:

Test ComponentDescriptionSuccess Criteria
Edge Case TestsTest edge case handling≥90% success rate
Stress TestsTest under stressSystem functional
Adversarial TestsTest adversarial resistance≥85% success rate
Fault Injection TestsTest fault toleranceGraceful recovery
Environmental TestsTest across conditions< 5% variation
Drift TestsTest temporal stabilityDrift score < 0.1

Evidence Required:

  • Robustness Test Plan (PLAN-AI-ARS-002)
  • Test results (TEST-AI-ARS-002)
  • Analysis reports
  • Compliance verification
  • Approval records

Audit Verification:

  • Verify robustness testing conducted
  • Confirm all test types executed
  • Check results meet requirements
  • Validate deployment blocked if requirements not met

Control ARS-006: Model Drift Detection

Control ID: ARS-006
Control Name: Model Drift Detection and Response
Control Type: Detective
Control Frequency: Continuous
Risk Level: Medium

Control Objective

Monitor for data drift and concept drift to detect performance degradation early and enable timely model updates or retraining to maintain AI system performance.

Control Requirements

CR-006.1: Drift Detection Implementation

Implement comprehensive drift detection with alerting and response procedures.

Drift Types:

Drift TypeDescriptionDetection MethodThreshold
Data DriftChange in input data distributionStatistical distribution comparison> 0.1
Concept DriftChange in relationship between inputs and outputsPerformance monitoring> 5% performance drop
Label DriftChange in output distributionOutput distribution comparison> 0.1

Mandatory Actions:

  • Implement drift detection
  • Monitor continuously
  • Alert on drift detection
  • Investigate drift causes
  • Retrain or update model if needed
  • Document drift events and responses

Drift Detection Configuration:

Detection MethodMetricThresholdAlert Level
Statistical DistanceKL divergence, Wasserstein distance> 0.1Warning
Performance MonitoringAccuracy, precision, recall> 5% dropCritical
Distribution ComparisonDistribution similarity< 0.9Warning

Evidence Required:

  • Drift detection configuration
  • Drift monitoring dashboard
  • Alert logs
  • Drift investigation records
  • Retraining records
  • Monthly drift reports

Audit Verification:

  • Verify drift detection implemented
  • Confirm monitoring continuous
  • Check alerts configured
  • Validate drift investigated
  • Verify model retrained when needed

Control ARS-007: Cybersecurity Requirements

Control ID: ARS-007
Control Name: Cybersecurity Requirements and Threat Assessment
Control Type: Preventive
Control Frequency: Per high-risk AI system, during design phase
Risk Level: High

Control Objective

Define cybersecurity requirements and assess AI-specific security threats per Article 15(5) to ensure AI systems are resilient against cybersecurity threats and protect against AI-specific attack vectors.

Control Requirements

CR-007.1: Security Requirements and Threat Assessment

Define security requirements and assess AI-specific threats.

Security Dimensions:

DimensionRequirementImplementationVerification
ConfidentialityProtect sensitive data and modelsEncryption, access controlsSecurity testing
IntegrityPrevent unauthorized modificationsIntegrity checks, access controlsSecurity testing
AvailabilityEnsure system availabilityRedundancy, DDoS protectionAvailability testing
AuthenticationVerify user identitiesAuthentication mechanismsSecurity testing
AuthorizationControl access appropriatelyRole-based access controlSecurity testing

AI-Specific Security Threats:

Threat TypeDescriptionDefense StrategyTesting Method
Adversarial AttacksMalicious inputs to fool modelAdversarial training, input validationAdversarial testing
Model ExtractionStealing model through queriesRate limiting, query monitoringModel extraction testing
Data PoisoningCorrupting training dataData validation, provenance trackingData quality checks
Model InversionExtracting training data from modelDifferential privacy, access controlsModel inversion testing
Backdoor AttacksHidden malicious behaviorModel verification, testingBackdoor detection testing

Mandatory Actions:

  • Define security requirements
  • Conduct threat modeling
  • Assess AI-specific threats
  • Implement security controls
  • Test security controls
  • Document security architecture

Evidence Required:

  • Security Requirements Document (DOC-AI-ARS-003)
  • Threat model
  • AI threat assessment
  • Security control documentation
  • Test results

Audit Verification:

  • Verify security requirements defined
  • Confirm threat modeling conducted
  • Check AI-specific threats assessed
  • Validate security controls implemented
  • Verify security testing completed

Control ARS-008: Security Testing

Control ID: ARS-008
Control Name: Cybersecurity Testing and Validation
Control Type: Preventive
Control Frequency: Before deployment, annually, after security updates
Risk Level: High

Control Objective

Test AI system security to verify it meets security requirements and is resilient against cybersecurity threats, including AI-specific attack vectors.

Control Requirements

CR-008.1: Comprehensive Security Testing

Conduct comprehensive security testing including AI-specific tests.

Test Types:

Test TypePurposeTest MethodSuccess Criteria
Penetration TestingTest system securityPenetration testNo critical vulnerabilities
Vulnerability ScanningIdentify vulnerabilitiesAutomated scanningNo critical vulnerabilities
Adversarial Robustness TestingTest resistance to adversarial attacksAdversarial attack simulations≥85% success rate
Security Code ReviewReview code for security issuesManual and automated reviewNo critical issues
Dependency ScanningIdentify vulnerable dependenciesDependency scanning toolsNo critical vulnerabilities
Model Extraction TestingTest resistance to model extractionModel extraction attemptsModel protected
Data Poisoning TestingTest resistance to data poisoningData poisoning simulationsSystem detects poisoning

Mandatory Actions:

  • Plan security testing
  • Execute all test types
  • Remediate findings
  • Verify remediation
  • Document results
  • Block deployment if critical vulnerabilities found

Security Test Plan:

Test ComponentDescriptionSuccess Criteria
Penetration TestsTest system securityNo critical vulnerabilities
Vulnerability ScansScan for vulnerabilitiesNo critical vulnerabilities
Adversarial TestsTest adversarial resistance≥85% success rate
Code ReviewReview code securityNo critical issues
Dependency ScansScan dependenciesNo critical vulnerabilities
AI-Specific TestsTest AI-specific threatsThreats mitigated

Evidence Required:

  • Security Test Plan (PLAN-AI-ARS-003)
  • Test results (TEST-AI-ARS-003)
  • Remediation records
  • Verification results
  • Approval records

Audit Verification:

  • Verify security testing conducted
  • Confirm all test types executed
  • Check critical vulnerabilities remediated
  • Validate deployment blocked if critical vulnerabilities found

SUPPORTING PROCEDURES

This standard is implemented through the following detailed procedures:

Procedure PROC-AI-ARS-001: Accuracy Testing Procedure

Purpose: Define step-by-step process for accuracy testing
Owner: Chief Technology Officer
Implements: Controls ARS-001, ARS-002

Procedure Steps:

  1. Define accuracy requirements - Control ARS-001
  2. Prepare test data
  3. Plan accuracy testing
  4. Execute accuracy tests - Control ARS-002
  5. Analyze results
  6. Verify threshold compliance
  7. Document results
  8. Obtain approval

Outputs:

  • Accuracy Requirements Document
  • Accuracy Test Plan
  • Test results
  • Approval records

Procedure PROC-AI-ARS-002: Robustness Testing Procedure

Purpose: Define process for robustness testing
Owner: Chief Technology Officer
Implements: Controls ARS-004, ARS-005, ARS-006

Procedure Steps:

  1. Define robustness requirements - Control ARS-004
  2. Plan robustness testing
  3. Execute robustness tests - Control ARS-005
  4. Implement drift detection - Control ARS-006
  5. Analyze results
  6. Verify compliance
  7. Document results

Outputs:

  • Robustness Requirements Document
  • Robustness Test Plan
  • Test results
  • Drift detection configuration

Procedure PROC-AI-ARS-003: Cybersecurity Assessment Procedure

Purpose: Define process for cybersecurity assessment
Owner: Chief Technology Officer
Implements: Controls ARS-007, ARS-008

Procedure Steps:

  1. Define security requirements - Control ARS-007
  2. Conduct threat modeling
  3. Assess AI-specific threats
  4. Plan security testing
  5. Execute security tests - Control ARS-008
  6. Remediate findings
  7. Verify remediation
  8. Document results

Outputs:

  • Security Requirements Document
  • Threat model
  • Security Test Plan
  • Test results
  • Remediation records

Procedure PROC-AI-ARS-004: Performance Monitoring Procedure

Purpose: Define process for continuous performance monitoring
Owner: Chief Technology Officer
Implements: Controls ARS-003, ARS-006

Procedure Steps:

  1. Configure accuracy monitoring - Control ARS-003
  2. Configure drift detection - Control ARS-006
  3. Set up alerting
  4. Monitor continuously
  5. Investigate issues
  6. Implement corrective actions
  7. Report metrics

Outputs:

  • Monitoring configuration
  • Alert logs
  • Investigation records
  • Corrective action records
  • Performance reports

COMPLIANCE

5.1 Compliance Monitoring

Monitoring Approach: Continuous automated monitoring supplemented by monthly manual reviews and quarterly comprehensive audits.

Compliance Metrics:

MetricTargetMeasurement MethodFrequencyOwner
Accuracy Requirements Coverage100%% of high-risk AI with accuracy requirementsMonthlyChief Technology Officer
Accuracy Threshold Compliance100%% of AI systems meeting accuracy thresholdsDailyChief Technology Officer
Robustness Requirements Coverage100%% of high-risk AI with robustness requirementsMonthlyChief Technology Officer
Robustness Threshold Compliance100%% of AI systems meeting robustness thresholdsMonthlyChief Technology Officer
Security Requirements Coverage100%% of high-risk AI with security requirementsMonthlyIT Security Manager
Security Vulnerability Count0 criticalNumber of critical vulnerabilitiesMonthlyIT Security Manager
Drift Detection Coverage100%% of AI systems with drift detectionMonthlyChief Technology Officer
Security Testing Completion100%% of required security tests completedQuarterlyIT Security Manager

Monitoring Tools:

  • Performance Monitoring Dashboard
  • Security Monitoring Dashboard
  • Compliance Reports
  • Monthly compliance reports
  • Quarterly AI Governance Committee reviews

5.2 Internal Audit Requirements

Audit Frequency: Annually (minimum)

Audit Scope:

  • Accuracy requirements completeness
  • Accuracy testing quality
  • Robustness requirements completeness
  • Robustness testing quality
  • Security requirements completeness
  • Security testing quality
  • Drift detection effectiveness
  • Controls effectiveness (ARS-001 through ARS-008)

Audit Activities:

  • Review 100% of high-risk AI for requirements
  • Sample 20% of tests for quality review
  • Test accuracy monitoring
  • Test drift detection
  • Test security controls
  • Review security test results
  • Interview key personnel

Audit Outputs:

  • Annual Accuracy, Robustness, and Security Audit Report
  • Findings and recommendations
  • Corrective action plans for deficiencies

5.3 External Audit / Regulatory Inspection

Preparation:

  • Maintain audit-ready documentation at all times
  • Designate Chief Technology Officer and IT Security Manager as regulatory liaisons
  • Prepare standard response procedures for authority requests

Provide to Auditors/Regulators:

  • Accuracy requirements and test results
  • Robustness requirements and test results
  • Security requirements and test results
  • Threat models
  • Performance monitoring reports
  • Security test reports
  • Internal audit reports
  • Evidence of controls execution

Authority Request Response:

  • Acknowledge request within 1 business day
  • Provide requested documentation within 5 business days
  • Coordinate through Legal, Chief Technology Officer, and IT Security Manager
  • Document all interactions with authorities

ROLES AND RESPONSIBILITIES

6.1 RACI Matrix

ActivityChief Technology OfficerAI System OwnerData ScienceIT Security ManagerQA/TestingOperations
Accuracy RequirementsR/AARICI
Accuracy TestingRARIRI
Accuracy MonitoringRACIIR
Robustness RequirementsRARCCI
Robustness TestingRARCRI
Drift DetectionRARIIC
Security RequirementsRACR/AII
Security TestingRACRRI

RACI Legend:

  • R = Responsible (does the work)
  • A = Accountable (ultimately answerable)
  • C = Consulted (provides input)
  • I = Informed (kept up-to-date)

6.2 Role Descriptions

Chief Technology Officer

  • Primary Responsibility: Owns accuracy, robustness, and security framework
  • Key Activities:
    • Establishes requirements framework
    • Approves requirements
    • Monitors compliance
    • Reports metrics
  • Required Competencies: EU AI Act Article 15, AI system architecture, performance management

AI System Owner

  • Primary Responsibility: Accountable for accuracy, robustness, and security of their AI system
  • Key Activities:
    • Ensures requirements defined
    • Ensures testing completed
    • Monitors performance
    • Participates in reviews
  • Required Competencies: AI system knowledge, performance requirements

Data Science

  • Primary Responsibility: Implements accuracy and robustness requirements
  • Key Activities:
    • Defines accuracy/robustness requirements
    • Implements models meeting requirements
    • Conducts testing
    • Monitors performance
  • Required Competencies: Machine learning, model development, performance optimization

IT Security Manager

  • Primary Responsibility: Owns security requirements and testing
  • Key Activities:
    • Defines security requirements
    • Conducts threat modeling
    • Manages security testing
    • Monitors security
  • Required Competencies: Cybersecurity, AI security, threat modeling

QA/Testing

  • Primary Responsibility: Conducts testing
  • Key Activities:
    • Plans testing
    • Executes tests
    • Reports results
  • Required Competencies: Testing methodologies, quality assurance

Operations

  • Primary Responsibility: Monitors performance in production
  • Key Activities:
    • Monitors accuracy
    • Monitors drift
    • Alerts on issues
  • Required Competencies: Operations monitoring, alerting

EXCEPTIONS

7.1 Exception Philosophy

Accuracy, robustness, and security are critical regulatory compliance activities for high-risk AI systems. Exceptions are granted restrictively and only where compensating controls adequately mitigate risks.


7.2 Allowed Exceptions

The following exceptions may be granted with proper justification and approval:

Exception TypeJustification RequiredMaximum DurationApproval AuthorityCompensating Controls
Reduced Accuracy Threshold (Minimal-Risk AI)AI system clearly minimal-risk; lower accuracy acceptablePermanentChief Technology OfficerDocument rationale; Annual re-confirmation
Extended Security Testing TimelineResource constraints prevent timely testing30 daysChief Technology Officer + IT Security ManagerInterim security measures; Accelerated plan

7.3 Prohibited Exceptions

The following exceptions cannot be granted under any circumstances:

Skipping accuracy requirements for high-risk AI - Mandatory per Article 15(1), no exceptions
Skipping robustness requirements for high-risk AI - Mandatory per Article 15(4), no exceptions
Skipping security requirements for high-risk AI - Mandatory per Article 15(5), no exceptions
Deploying with accuracy below threshold - Creates safety and compliance risks
Deploying with critical security vulnerabilities - Creates security risks


7.4 Exception Request Process

Step 1: Submit Exception Request

  • Complete Exception Request Form (FORM-AI-EXCEPTION-001)
  • Include business justification
  • Propose compensating controls
  • Specify duration requested
  • Attach risk assessment

Step 2: Risk Assessment

  • Chief Technology Officer assesses risk of granting exception
  • Evaluates adequacy of compensating controls
  • Documents residual risk

Step 3: Approval

  • Route to appropriate approval authority based on exception type
  • Chief Technology Officer approval: Minor exceptions
  • Chief Technology Officer + AI Governance Committee: Significant exceptions
  • AI Governance Committee: Critical exceptions

Step 4: Documentation and Monitoring

  • Document exception in Exception Register
  • Assign exception owner
  • Set review date
  • Monitor compensating controls
  • Report exceptions quarterly to AI Governance Committee

Step 5: Exception Review and Closure

  • Review exception at specified review date
  • Assess if exception still needed
  • Close exception when normal requirements met
  • Document lessons learned

ENFORCEMENT

8.1 Non-Compliance Consequences

ViolationSeverityConsequenceRemediation Required
High-risk AI without accuracy requirementsCriticalImmediate suspension until requirements definedDefine requirements within 5 business days; Root cause analysis
Deploying with accuracy below thresholdCriticalImmediate halt deployment; Compliance gap assessmentImprove accuracy; Re-test; Re-approve
Missing robustness requirementsHighEscalation to AI Governance CommitteeDefine requirements within 10 business days
Missing security requirementsCriticalImmediate suspensionDefine requirements within 5 business days
Critical security vulnerabilitiesCriticalImmediate correction; Security investigationRemediate within 24 hours; Security review
Missing drift detectionMediumWritten warningImplement drift detection within 10 business days

8.2 Escalation Procedures

Level 1: Chief Technology Officer

  • Minor procedural violations
  • Documentation deficiencies
  • Timeline delays < 5 days
  • Action: Written warning, corrective action required

Level 2: Chief Technology Officer + AI Governance Committee

  • Repeated violations
  • Missing requirements
  • Performance below thresholds
  • Action: Formal review, corrective action plan, management notification

Level 3: AI Governance Committee

  • High-risk AI without requirements
  • Critical security vulnerabilities
  • Critical compliance failures
  • Action: Immediate AI system suspension, investigation, disciplinary action

Level 4: Executive Management + Legal

  • Potential regulatory enforcement action
  • Significant legal liability
  • Reputational risk
  • Action: Executive crisis management, legal strategy, regulatory engagement

8.3 Immediate Escalation Triggers

Escalate immediately to AI Governance Committee + Legal if:

  • ⚠️ High-risk AI system operating without accuracy/robustness/security requirements
  • ⚠️ Critical security vulnerability identified
  • ⚠️ Security breach or attack
  • ⚠️ Regulatory inquiry or inspection related to accuracy/robustness/security
  • ⚠️ Significant accuracy degradation affecting safety

8.4 Disciplinary Actions

Individuals responsible for violations may be subject to:

  • Verbal or written warning
  • Mandatory retraining
  • Performance improvement plan
  • Reassignment of responsibilities
  • Suspension (with pay during investigation)
  • Termination (for egregious violations, e.g., knowingly deploying with critical vulnerabilities)

Factors Considered:

  • Intent (knowing violation vs. honest mistake)
  • Severity of violation
  • Impact (actual or potential)
  • Cooperation with remediation
  • Prior violation history

KEY PERFORMANCE INDICATORS (KPIs)

9.1 Accuracy, Robustness, and Security KPIs

KPI IDKPI NameDefinitionTargetMeasurement MethodFrequencyOwnerReporting To
KPI-ARS-001Accuracy Requirements Coverage% of high-risk AI with accuracy requirements100%(# AI with requirements / # high-risk AI) × 100MonthlyChief Technology OfficerAI Governance Committee
KPI-ARS-002Accuracy Threshold Compliance% of AI systems meeting accuracy thresholds100%(# AI meeting thresholds / # total AI) × 100DailyChief Technology OfficerManagement
KPI-ARS-003Accuracy StabilityAccuracy variation over time< 2%Standard deviation of accuracyWeeklyChief Technology OfficerManagement
KPI-ARS-004Robustness Requirements Coverage% of high-risk AI with robustness requirements100%(# AI with requirements / # high-risk AI) × 100MonthlyChief Technology OfficerAI Governance Committee
KPI-ARS-005Robustness ScoreComposite robustness score≥90%Weighted average of robustness metricsMonthlyChief Technology OfficerManagement
KPI-ARS-006Drift Detection Coverage% of AI systems with drift detection100%(# AI with drift detection / # total AI) × 100MonthlyChief Technology OfficerManagement
KPI-ARS-007Drift Detection RateNumber of drift events detectedMonitor trendsCount of drift eventsWeeklyChief Technology OfficerManagement
KPI-ARS-008Security Requirements Coverage% of high-risk AI with security requirements100%(# AI with requirements / # high-risk AI) × 100MonthlyIT Security ManagerAI Governance Committee
KPI-ARS-009Security Vulnerability CountNumber of critical vulnerabilities0Count of critical vulnerabilitiesMonthlyIT Security ManagerAI Governance Committee
KPI-ARS-010Security Testing Completion% of required security tests completed100%(# tests completed / # required tests) × 100QuarterlyIT Security ManagerManagement

9.2 KPI Dashboards and Reporting

Real-Time Dashboard (Chief Technology Officer and IT Security Manager access)

  • Current accuracy metrics
  • Robustness scores
  • Drift detection status
  • Security vulnerability status
  • Performance trends

Monthly Management Report

  • KPI-ARS-001, 002, 003, 004, 005, 006, 007, 008, 009
  • Trend analysis (vs. previous month)
  • Issues and risks
  • Planned actions

Quarterly AI Governance Committee Report

  • All KPIs
  • Accuracy performance assessment
  • Robustness performance assessment
  • Security posture assessment
  • Internal audit findings (if conducted)
  • Exception register review

Annual Executive Report

  • Full-year KPI performance
  • Performance maturity assessment
  • Strategic recommendations
  • Regulatory outlook

9.3 KPI Thresholds and Alerts

KPIGreen (Good)Yellow (Warning)Red (Critical)Alert Action
Accuracy Threshold Compliance100%95-99%< 95%Red: Immediate escalation to AI Governance Committee Chair
Accuracy Stability< 2%2-5%> 5%Red: Escalate to AI Governance Committee
Robustness Score≥90%85-89%< 85%Yellow: Improvement plan; Red: Escalate to AI Governance Committee
Security Vulnerability Count01-2> 2Red: Immediate escalation to IT Security Manager + AI Governance Committee

TRAINING REQUIREMENTS

10.1 Training Program Overview

All personnel involved in accuracy, robustness, and security must complete role-specific training to ensure competency in EU AI Act Article 15 requirements, performance management, and security practices.


10.2 Role-Based Training Requirements

RoleTraining CourseDurationContentFrequencyAssessment Required
Chief Technology OfficerPerformance Management Expert Training16 hoursEU AI Act Article 15; Accuracy requirements; Robustness requirements; Security requirementsInitial + annuallyYes - Written exam (≥90%)
Data ScientistsAccuracy and Robustness Training12 hoursAccuracy metrics; Robustness testing; Drift detection; Model optimizationInitial + annuallyYes - Practical exercise
IT Security ManagerAI Security Expert Training16 hoursAI-specific security threats; Security testing; Threat modeling; Security controlsInitial + annuallyYes - Written exam (≥90%)
QA/TestingPerformance Testing Training8 hoursAccuracy testing; Robustness testing; Security testing; Test methodologiesInitial + annuallyYes - Practical exercise
AI System OwnersPerformance Overview4 hoursAccuracy requirements; Robustness requirements; Security requirements; ResponsibilitiesAt onboarding + annuallyYes - Knowledge check (≥80%)
All AI Development StaffPerformance Awareness2 hoursAccuracy basics; Robustness basics; Security basics; RequirementsAt onboarding + annuallyYes - Knowledge check (≥80%)

10.3 Training Content by Topic

EU AI Act Article 15 Requirements

  • Accuracy requirements (Article 15(1))
  • Robustness requirements (Article 15(4))
  • Security requirements (Article 15(5))
  • Compliance obligations

Accuracy Management

  • Accuracy metrics
  • Accuracy testing
  • Accuracy monitoring
  • Performance optimization

Robustness Management

  • Robustness dimensions
  • Robustness testing
  • Drift detection
  • Model stability

AI Security

  • AI-specific security threats
  • Security controls
  • Security testing
  • Threat modeling

10.4 Training Delivery Methods

Initial Training:

  • Instructor-led classroom or virtual training
  • Includes interactive exercises and case studies
  • Hands-on practice with testing tools
  • Group discussions of complex scenarios

Annual Refresher:

  • E-learning modules for core content review
  • Live update sessions for regulatory changes
  • Case study reviews of recent performance activities
  • Knowledge assessment

On-the-Job Training:

  • Mentoring for new performance staff
  • Job shadowing during testing
  • Supervised testing for first 5 AI systems

Just-in-Time Training:

  • Quick reference guides and job aids
  • Video tutorials on specific topics
  • Help desk support from experienced staff

10.5 Training Effectiveness Measurement

Assessment Methods:

  • Written exams for knowledge retention
  • Practical exercises for skill application
  • On-the-job observations for competency validation
  • Feedback surveys for training quality

Competency Validation:

  • Data Scientists: Must demonstrate ability to define requirements and conduct testing for 1 sample AI system with 100% compliance before independent work
  • All staff: Must pass knowledge assessments with minimum required scores

Training Metrics:

MetricTargetFrequency
Training completion rate100%Quarterly
Assessment pass rate (first attempt)≥ 90%Per training
Training effectiveness score (survey)≥ 4.0/5.0Per training
Time to competency (Data Scientists)< 30 daysPer person

10.6 Training Records

Records Maintained:

  • Training attendance records
  • Assessment scores
  • Competency validations
  • Refresher training completion
  • Individual training transcripts

Retention: 10 years (to align with EU AI Act documentation retention)

Access: HR, Chief Technology Officer, Internal Audit, Competent Authorities (upon request)


DEFINITIONS

TermDefinitionSource
AccuracyDegree to which AI system outputs are correctEU AI Act Article 15(1)
RobustnessAbility of AI system to maintain performance despite errors, faults, or adversarial conditionsEU AI Act Article 15(4)
CybersecurityProtection of AI systems against cybersecurity threatsEU AI Act Article 15(5)
Precision% of positive predictions that are correctThis Standard
Recall% of actual positives correctly identifiedThis Standard
F1 ScoreHarmonic mean of precision and recallThis Standard
ROC-AUCArea under Receiver Operating Characteristic curveThis Standard
Data DriftChange in input data distribution over timeThis Standard
Concept DriftChange in relationship between inputs and outputs over timeThis Standard
Adversarial AttackMalicious input designed to fool AI modelThis Standard
Model ExtractionAttack to steal model through queriesThis Standard
Data PoisoningAttack to corrupt training dataThis Standard
Model InversionAttack to extract training data from modelThis Standard
Backdoor AttackAttack to introduce hidden malicious behaviorThis Standard

LINK WITH AI ACT AND ISO42001

12.1 EU AI Act Regulatory Mapping

This standard implements the following EU AI Act requirements:

EU AI Act ProvisionArticleRequirement SummaryImplemented By (Controls)
Accuracy, Robustness and CybersecurityArticle 15Requirements for high-risk AIAll controls (ARS-001 through ARS-008)
AccuracyArticle 15(1)Appropriate accuracy levelsARS-001, ARS-002, ARS-003
RobustnessArticle 15(4)Resilience to errors and faultsARS-004, ARS-005, ARS-006
CybersecurityArticle 15(5)Resilience against cybersecurity threatsARS-007, ARS-008

12.2 ISO/IEC 42001:2023 Alignment

This standard aligns with ISO/IEC 42001:2023 as follows:

ISO 42001 ClauseRequirementImplementation in This Standard
Clause 6.1.2: AI system impact assessmentAssess accuracy, robustness, security impactsARS-001, ARS-004, ARS-007
Clause 8.2: AI system risk assessmentAccuracy, robustness, security in risk managementAll controls
Clause 9.1: Monitoring, measurement, analysis, and evaluationMonitor performanceARS-003, ARS-006

12.3 Relationship to Other Standards

This accuracy, robustness, and security standard integrates with other AI Act standards:

Related StandardIntegration PointRationale
STD-AI-001: ClassificationClassification determines if requirements applyHigh-risk AI requires Article 15 requirements
STD-AI-002: Risk ManagementAccuracy, robustness, security risks in risk assessmentRisk management identifies performance risks
STD-AI-004: Technical DocumentationPerformance metrics documented in Annex IVPerformance information in technical documentation
STD-AI-013: Incident ManagementPerformance issues may trigger incidentsPerformance degradation may be incidents

12.4 References and Related Documents

EU AI Act (Regulation (EU) 2024/1689):

  • Article 15: Accuracy, Robustness and Cybersecurity
  • Article 15(1): Accuracy
  • Article 15(3): Accuracy Declaration in Instructions for Use (cross-reference to STD-AI-006 Control TRANS-001)
  • Article 15(4): Robustness
  • Article 15(5): Cybersecurity

ISO/IEC Standards:

  • ISO/IEC 42001:2023: Information technology — Artificial intelligence — Management system
  • ISO/IEC 23894:2023: Information technology — Artificial intelligence — Guidance on risk management
  • ISO/IEC 27001:2022: Information security management systems

Internal Documents:

  • POL-AI-001: Artificial Intelligence Policy (parent policy)
  • STD-AI-001: AI System Classification Standard
  • STD-AI-002: AI Risk Management Standard
  • STD-AI-004: AI Technical Documentation Standard
  • PROC-AI-ARS-001, -002, -003, -004: Performance procedures

APPROVAL AND AUTHORIZATION

RoleNameTitleSignatureDate
Prepared ByDavid LeeChief Technology Officer_________________________
Reviewed ByIT Security ManagerIT Security Manager_________________________
Reviewed BySarah JohnsonAI Act Program Manager_________________________
Reviewed ByJane DoeChief Strategy & Risk Officer_________________________
Approved ByJane DoeAI Governance Committee Chair_________________________

Effective Date: 2025-08-01
Next Review Date: 2026-08-01
Review Frequency: Annually or upon regulatory change


END OF STANDARD STD-AI-008


This standard is a living document. Feedback and improvement suggestions should be directed to the Chief Technology Officer.

Standard Details

Standard ID

STD-AI-008

Version

1.0

Status

draft

Owner

CTO

Effective Date

2025-08-01

Applicability

High-risk AI systems

EU AI Act References
Article 15
ISO 42001 Mapping
Clause 8.2