aicomply.
Lesson12 minChapter 3 of 8

Real-World Testing

Article 60 provisions for testing AI in real conditions.

Real-World Testing (Article 60)

Learning Objectives

By the end of this chapter, you will be able to:

  • Understand when and how real-world testing is permitted under Article 60
  • Design testing plans that meet regulatory requirements
  • Implement informed consent procedures compliant with the AI Act
  • Establish monitoring and safeguard frameworks for live testing
  • Document real-world testing to support conformity assessment

Introduction: Testing in the Real World

Laboratory testing and simulations have limits. At some point, AI systems must be tested in real-world conditions to validate performance. Article 60 provides a framework for such testing—balancing the need for realistic validation against the protection of affected persons.

Expert Insight

Real-world testing is where theory meets practice. The AI Act doesn't prohibit it—it requires that you do it responsibly. A well-designed testing program actually strengthens your conformity assessment by providing real-world performance evidence.


Legal Framework (Article 60)

When Real-World Testing Applies

ConditionArticle 60 Requirement
System typeHigh-risk AI systems under Annex III
Lifecycle stageBefore placing on market or putting into service
PurposeTesting performance under real-world conditions
Approval pathwayEither within a sandbox OR with approved testing plan
SafeguardsSubject to specific protections for affected persons

Article 60 vs. Sandbox Testing

AspectRegulatory Sandbox (Articles 57-58)Real-World Testing (Article 60)
ScopeBroader development and validationFocused performance testing
DurationLonger (6-24 months typical)Shorter, specific testing period
Regulatory engagementContinuous supervisionPlan approval + monitoring
Best forNovel systems, compliance uncertaintyValidating known system performance
RelationshipCan include real-world testingCan be standalone or within sandbox

Prerequisites for Real-World Testing

Mandatory Requirements

RequirementArticle ReferenceWhat It Means
Testing planArticle 60(4)Detailed plan approved by market surveillance authority
Informed consentArticle 60(4)(i)Freely-given informed consent from test subjects
MonitoringArticle 60(4)(a)Effective oversight with ability to intervene
ReversibilityArticle 60(4)(d)AI decisions can be disregarded or reversed
Risk mitigationArticle 60(4)(b)Safeguards to prevent harm
No significant riskArticle 60(4)(b)Testing must not create serious health/safety risks
LiabilityArticle 60(9)Provider remains liable under applicable Union and national liability law

Testing Plan Contents

Plan ElementDescriptionAuthority Review Focus
System descriptionTechnical details of the AI being testedUnderstanding what's being tested
Testing objectivesWhat will be validated, success criteriaClarity and measurability
Testing methodologyHow testing will be conductedScientific validity
Subject selectionWho will participate, how recruitedRepresentativeness, vulnerability
Informed consent processHow consent will be obtained and documentedAdequacy and voluntariness
SafeguardsProtections for test subjectsSufficiency of protections
Monitoring proceduresHow testing will be supervisedAbility to detect and respond to issues
Intervention triggersWhen testing will be stoppedClear thresholds for action
Data handlingHow data will be collected, used, protectedGDPR compliance
Duration and scopeHow long, how many subjects, what contextsProportionality
Incident proceduresHow incidents will be handled and reportedResponse capability

Informed Consent Framework

Consent Requirements (Article 60(4)(i), see also Article 61)

RequirementImplementation
Freely-givenNo coercion, pressure, or undue inducement
InformedSubject understands what they're consenting to, with information per Article 61(1)
DocumentedWritten or electronic record of consent
WithdrawableSubject can withdraw at any time without consequence

Information to Provide

Before consent, test subjects must be informed about:

Information ElementExample Content
Nature of the AI system"This AI analyses your responses to assess creditworthiness"
Purpose of testing"We are testing the system's accuracy before market launch"
How they will be affected"The AI will evaluate your application alongside our standard process"
Risks and safeguards"There is a risk of incorrect assessment; human review verifies all decisions"
Data collection and use"Your data will be used for testing only and deleted after 12 months"
Duration of participation"Testing lasts 3 months; your involvement is for one application"
Right to withdraw"You can withdraw at any time; your application continues normally"
Contact for questions"Contact our Data Protection Officer at dpo@company.com"
Testing identification numberThe Union-wide unique single identification number of the testing per Article 61(1)(e)

Consent Exceptions

SituationException BasisRequirements
Law enforcement/migrationArticle 60(4)(i) exceptionTesting must not negatively affect subjects; personal data deleted after test
Emergency situationsNot explicitly addressedLikely requires post-hoc consent or exemption

Compliance Note

The consent exception for law enforcement and migration is narrow and requires enhanced safeguards. Don't assume it applies—seek legal advice before using this exception.


Monitoring and Safeguards

Monitoring Framework

Monitoring ElementImplementationPurpose
Real-time oversightDashboard, alerts, human supervisorDetect issues immediately
Performance trackingAccuracy, fairness, drift metricsValidate system performance
Incident detectionAutomated and manual detectionIdentify problems early
Subject feedbackChannels for concerns/complaintsCapture subjective impacts
DocumentationComprehensive loggingEvidence for conformity assessment

Intervention Triggers

Define clear thresholds for action:

TriggerResponse
Performance below thresholdPause testing, investigate, remediate
Bias detectedSuspend testing for affected groups, analyse
Harm to subjectStop testing immediately, support subject, report
Subject withdrawalRemove from testing, honour consent revocation
Authority requestImmediate compliance with authority instructions

Human Oversight During Testing

Oversight TypeWhen RequiredImplementation
Pre-decision reviewHigh-stakes decisions (e.g., credit, employment)Human reviews before action
Concurrent monitoringAll testingSupervisor monitors in real-time
Post-decision reviewAll AI decisionsHuman reviews outcomes
Override capabilityAlwaysAbility to disregard AI output

Special Categories of Testing

Vulnerable Populations

Testing involving vulnerable persons requires enhanced safeguards:

Vulnerable GroupAdditional Requirements
ChildrenParental/guardian consent, age-appropriate information
ElderlyAccessible information, capacity verification
EmployeesNo workplace coercion, union consultation if applicable
PatientsClinical ethics oversight, medical safeguards
Economically dependentEnsure no exploitation of financial vulnerability

Law Enforcement and Migration

Article 60(4)(i) provides limited exceptions:

RequirementImplementation
No negative effectTesting and its outcomes must not have any negative effect on subjects
Personal data deletionPersonal data shall be deleted after the test is performed
Testing plan approvalMarket surveillance authority must still approve
Enhanced safeguardsGreater protections than standard testing
DocumentationComprehensive records for accountability

Duration and Scope Limits

Proportionality Requirements

FactorConsideration
Maximum duration6 months, extendable to 12 months per Article 60(4)(f)
DurationNo longer than necessary to achieve testing objectives
Subject numbersMinimum needed for statistical validity
Scope of decisionsLimited to what's necessary to test
Geographic scopeAppropriate to testing objectives

Typical Testing Parameters

Testing TypeTypical DurationTypical Scale
Pilot testing1-3 months50-500 subjects
Extended validation3-6 months500-5,000 subjects
Pre-launch testing1-2 months1,000-10,000 subjects

Documentation Requirements

Real-Time Documentation

DocumentContentsUpdate Frequency
Testing logAll testing activities, decisions, eventsContinuous
Incident registerIssues, near-misses, complaintsAs they occur
Consent recordsAll consent documentationPer subject
Performance dataAccuracy, fairness, other metricsDaily/weekly
Monitoring reportsSupervisor observationsWeekly

Post-Testing Documentation

DocumentPurposeRetention
Testing summary reportOverall findings, conclusions10 years minimum
Performance validationEvidence for conformity assessment10 years minimum
Incident summaryAll issues and resolutions10 years minimum
Subject outcomesWhat happened to test subjectsAs required by GDPR

Authority Notification and Approval

Approval Process

Ongoing Reporting

Report TypeTimingContents
Progress reportsMonthly during testingActivities, metrics, issues
Incident reportsImmediately upon occurrenceDetails, response, remediation
Completion reportEnd of testingSummary, findings, recommendations

Liability and Insurance

Liability Framework (Article 60(9))

Liability AspectProvider Responsibility
Harm to test subjectsProvider remains fully liable under applicable Union and national liability law
Insurance (recommended)While not explicitly required by the AI Act, adequate insurance is a prudent practice
No liability transferCannot contract away liability to subjects

Insurance Considerations

Coverage TypeWhat It Covers
Product liabilityHarm caused by the AI system
Professional indemnityErrors in testing design or execution
Clinical trials (if applicable)Medical testing-specific coverage
Cyber liabilityData breaches during testing

Real-World Testing Checklist

Pre-Testing

  • Develop comprehensive testing plan
  • Identify and assess risks to test subjects
  • Design safeguards and monitoring procedures
  • Create informed consent materials
  • Establish intervention triggers and procedures
  • Obtain liability insurance/coverage
  • Submit plan to market surveillance authority
  • Obtain authority approval

During Testing

  • Obtain informed consent from all subjects
  • Activate monitoring systems
  • Document all activities and decisions
  • Submit regular progress reports
  • Report incidents immediately
  • Maintain human oversight
  • Respond to any authority requests

Post-Testing

  • Complete testing summary report
  • Compile performance validation evidence
  • Document all incidents and resolutions
  • Notify authority of testing completion
  • Retain all documentation (10+ years)
  • Use findings in conformity assessment

What You Learned

Key concepts from this chapter

**Real-world testing is permitted** but subject to specific safeguards under Article 60

**Informed consent is mandatory** (with limited exceptions for law enforcement/migration)

**Testing plans must be approved** by market surveillance authorities before testing begins

**Monitoring must be effective** with clear triggers for intervention

**AI decisions must be reversible** or disregardable during testing

Chapter Complete

Innovation Pathways

3/8

chapters