aicomply.
STD-AI-006

AI Transparency Standard

Provide clear, comprehensive information to deployers and users about AI system capabilities and limitations.

2

Controls

0

Compliant

2

In Progress

0

Not Started

Overall Progress
60%
Implementation Guidance
Detailed guidance for implementing this standard

AI Transparency Standard

Document Type: Standard
Standard ID: STD-AI-006
Standard Title: AI Transparency Standard
Version: 1.0
Effective Date: 2025-08-01
Next Review Date: 2026-08-01
Review Frequency: Annually or upon regulatory change
Parent Policy: POL-AI-001 - Artificial Intelligence Policy
Owner: Product Director
Approved By: AI Governance Committee Chair
Status: Draft
Classification: Internal Use Only


TABLE OF CONTENTS

  1. Document History
  2. Objective
  3. Scope and Applicability
  4. Control Standard
  5. Supporting Procedures
  6. Compliance
  7. Roles and Responsibilities
  8. Exceptions
  9. Enforcement
  10. Key Performance Indicators (KPIs)
  11. Training Requirements
  12. Definitions
  13. Link with AI Act and ISO42001

DOCUMENT HISTORY

VersionDateAuthorChangesApproval DateApproved By
0.12025-06-30Product DirectorInitial draft--
0.22025-07-15Product DirectorAdded Article 50 details--
0.32025-07-28Product DirectorIncorporated user feedback--
1.02025-08-01Product DirectorFinal version approved - GRC restructured2025-07-25Jane Doe, AI Governance Committee Chair

OBJECTIVE

This standard defines requirements for providing clear, comprehensive information to deployers and users about AI system capabilities and limitations in compliance with EU AI Act Articles 13 and 50.

Primary Goals:

  • Provide comprehensive instructions for use to deployers per Article 13
  • Ensure users are aware they are interacting with AI per Article 50
  • Enable users to understand AI decisions through explainability

SCOPE AND APPLICABILITY

2.1 Mandatory Applicability

This standard is mandatory for:

  • All high-risk AI systems (EU AI Act Article 13)
  • AI systems with transparency obligations (EU AI Act Article 50)
  • All deployers and end users

2.2 Recommended Applicability

This standard is recommended for:

  • All AI systems for best practices
  • Limited-risk and minimal-risk AI systems (voluntary transparency)

2.3 Transparency Requirements Covered

  • Instructions for Use (Article 13)
  • User transparency disclosures (Article 50)
  • Explainability and decision explanations
  • Transparency notice design and implementation

2.4 Out of Scope

  • General product documentation (covered by product documentation standards)
  • Non-AI system transparency (covered by other transparency standards)
  • Transparency outside EU AI Act scope

CONTROL STANDARD

Control TRANS-001: Instructions for Use Content

Control ID: TRANS-001
Control Name: Instructions for Use Creation and Distribution
Control Type: Preventive
Control Frequency: Per high-risk AI system, after substantial modifications
Risk Level: High

Control Objective

Provide comprehensive instructions for use to deployers per Article 13(3) to ensure deployers understand AI system characteristics, capabilities, limitations, and proper use, enabling safe and effective deployment.

Control Requirements

CR-001.1: Complete Instructions for Use

Create and distribute complete Instructions for Use containing all required elements per Article 13(3).

Required Elements (Article 13(3)):

ElementDescriptionMandatoryExample
Provider IdentityIdentity and contact details of providerYESCompany name, address, contact
CharacteristicsCharacteristics, capabilities, and limitationsYESDetailed description
Performance MetricsAccuracy, robustness metricsYESAccuracy: 95%, robustness: 92%
Changes and UpdatesInformation about changes and updatesYESUpdate history, change log
Human OversightHuman oversight measuresYESOversight procedures, triggers
Expected LifetimeExpected lifetime of AI systemYES5 years, maintenance schedule
Maintenance and CareMaintenance and care instructionsYESMaintenance procedures, schedules
InstallationInstallation instructionsRECOMMENDEDStep-by-step installation guide (best practice, not required by Art. 13(3))
Technical SpecificationsTechnical specificationsYESHardware, software requirements
Intended PurposeIntended purpose and use casesYESPurpose statement, use cases
Foreseeable MisuseReasonably foreseeable misuseYESMisuse scenarios, prevention
Known RisksKnown risks and mitigation measuresYESRisk list, mitigation strategies
Training RequirementsTraining requirements for deployersRECOMMENDEDTraining courses, certifications (best practice, not a discrete Art. 13(3) requirement)

Mandatory Actions:

  • Create Instructions for Use per Article 13(3)
  • Include all required elements
  • Use clear, understandable language
  • Obtain legal review
  • Obtain Product Director approval
  • Distribute to deployers before deployment
  • Maintain distribution records

Evidence Required:

  • Instructions for Use document (TECH-AI-IFU-XXX)
  • Completeness checklist (completed)
  • Legal review records
  • Approval records
  • Distribution records
  • Deployer acknowledgment records

Audit Verification:

  • Verify Instructions for Use created for all high-risk AI
  • Confirm all required elements included
  • Check legal review completed
  • Validate approval obtained
  • Verify distribution to deployers
  • Check deployer acknowledgments received

Control TRANS-002: Instructions for Use Updates

Control ID: TRANS-002
Control Name: Instructions for Use Maintenance and Updates
Control Type: Preventive
Control Frequency: As triggered by changes
Risk Level: Medium

Control Objective

Keep Instructions for Use current and accurate by updating them when AI system changes occur, ensuring deployers always have current information about AI system capabilities, limitations, and risks.

Control Requirements

CR-002.1: Change-Driven Updates

Update Instructions for Use promptly when changes occur.

Update Triggers:

  • Substantial modifications to AI system
  • Changes to intended purpose
  • New risks identified
  • Changes to capabilities or limitations
  • Changes to performance metrics
  • Changes to human oversight measures
  • Changes to maintenance requirements
  • Regulatory requirement changes

Mandatory Actions:

  • Monitor for update triggers
  • Assess impact on Instructions for Use
  • Update affected sections
  • Maintain version history
  • Obtain approval for updates
  • Communicate updates to deployers
  • Track deployer acknowledgments
  • Update within 30 days of change

Update Communication Process:

  1. Identify change requiring update
  2. Update Instructions for Use
  3. Obtain approval
  4. Notify all deployers
  5. Provide updated Instructions for Use
  6. Track acknowledgments
  7. Follow up with non-acknowledgers

Evidence Required:

  • Updated Instructions for Use (versioned)
  • Update notifications sent
  • Deployer acknowledgments
  • Version history
  • Approval records
  • Update completion records

Audit Verification:

  • Verify Instructions for Use updated within 30 days of changes
  • Confirm all affected sections updated
  • Check deployers notified
  • Validate acknowledgments tracked
  • Verify version history maintained

Control TRANS-003: AI Disclosure to Users

Control ID: TRANS-003
Control Name: User AI Disclosure Implementation
Control Type: Preventive
Control Frequency: Continuous, at each user interaction
Risk Level: High

Control Objective

Ensure users are aware they are interacting with AI per Article 50 to enable informed decision-making and protect user rights, preventing deception and ensuring transparency.

Control Requirements

CR-003.1: Mandatory AI Disclosure

Implement AI disclosure notices per Article 50 requirements.

Disclosure Requirements (Article 50):

AI System TypeDisclosure RequirementWhenFormat
Chatbots/Conversational AIInform users they are interacting with AI (unless obvious)Before or at start of interactionText notice, audio announcement
Synthetic ContentLabel AI-generated content (deepfakes, etc.)With contentVisible label, watermark
Emotion RecognitionInform users of emotion recognition useBefore useClear notice, consent
Biometric CategorizationInform users of biometric categorizationBefore useClear notice, consent

Mandatory Actions:

  • Implement AI disclosure notices
  • Display prominently (before or at interaction)
  • Use clear, understandable language
  • Obtain user acknowledgment (where required)
  • Document disclosure
  • Monitor disclosure effectiveness

Disclosure Design Requirements:

  • Clear and understandable language
  • Prominent display (visible, not hidden)
  • Timely (before or at interaction)
  • Accessible format (WCAG 2.1 AA)
  • Multiple languages if needed
  • Persistent (for ongoing interactions)

Evidence Required:

  • Transparency notices (NOTICE-AI-TRANS-XXX)
  • Disclosure screenshots/recordings
  • User acknowledgment logs
  • Disclosure effectiveness metrics
  • Accessibility verification records

Audit Verification:

  • Verify AI disclosure implemented for all applicable AI systems
  • Confirm disclosure displayed prominently
  • Check clear language used
  • Validate user acknowledgments obtained (where required)
  • Verify accessibility requirements met

Control TRANS-004: Transparency Notice Design

Control ID: TRANS-004
Control Name: Transparency Notice Design and Implementation
Control Type: Preventive
Control Frequency: Per AI system, continuous monitoring
Risk Level: Medium

Control Objective

Design and implement effective transparency notices that are clear, prominent, accessible, and understandable to ensure users are properly informed about AI use.

Control Requirements

CR-004.1: Notice Design and Testing

Design transparency notices following best practices and test with users.

Design Requirements:

RequirementSpecificationImplementation
ClarityClear, understandable language (grade 8 reading level)Plain language review
ProminenceVisible, not hidden (above fold, sufficient size)UI/UX design guidelines
TimelinessBefore or at interactionImplementation timing
AccessibilityWCAG 2.1 AA compliantAccessibility testing
LanguagesMultiple languages if neededTranslation services
PersistencePersistent for ongoing interactionsUI implementation

Mandatory Actions:

  • Design transparency notices
  • Test with users (usability testing)
  • Implement in UI/UX
  • Verify visibility and prominence
  • Test accessibility (WCAG 2.1 AA)
  • Monitor effectiveness
  • Update based on feedback

User Testing Requirements:

  • Test with representative users
  • Verify notice visibility
  • Verify notice understandability
  • Verify notice effectiveness
  • Document test results
  • Implement improvements

Evidence Required:

  • Transparency notice designs
  • User testing results
  • Implementation verification
  • Accessibility test results
  • Effectiveness metrics
  • User feedback records

Audit Verification:

  • Verify transparency notices designed
  • Confirm user testing conducted
  • Check accessibility requirements met
  • Validate notices implemented in UI
  • Verify effectiveness monitored

Control TRANS-005: Decision Explainability

Control ID: TRANS-005
Control Name: AI Decision Explanations and Explainability
Control Type: Preventive
Control Frequency: On-demand, when requested
Risk Level: Medium

Control Objective

Enable users to understand AI decisions by providing explanations when requested, supporting user rights to explanation and enabling informed decision-making.

Control Requirements

CR-005.1: Explanation Capability

Implement explanation capability for AI decisions.

Explanation Requirements:

RequirementSpecificationImplementation
On-DemandProvide explanations when requestedExplanation API/UI
Key FactorsExplain key factors influencing decisionFeature importance, decision factors
UnderstandableUse understandable languagePlain language, visualizations
Human ReviewEnable human review of decisionsHuman review process
TimelyProvide explanations promptlyResponse time < 5 seconds

Explanation Content:

  • Decision outcome
  • Key factors influencing decision
  • Confidence/probability scores
  • Alternative outcomes considered
  • Data used in decision
  • Model version used

Mandatory Actions:

  • Implement explanation capability
  • Design explanation format
  • Test understandability with users
  • Enable on-demand access
  • Monitor explanation usage
  • Improve explanations based on feedback

Explanation Formats:

  • Text explanations
  • Visual explanations (charts, graphs)
  • Interactive explanations
  • Step-by-step explanations
  • Summary and detailed views

Evidence Required:

  • Explanation capability documentation
  • Explanation examples
  • User testing results
  • Explanation usage metrics
  • User feedback records
  • Improvement records

Audit Verification:

  • Verify explanation capability implemented
  • Confirm explanations provided on-demand
  • Check explanations understandable
  • Validate human review enabled
  • Verify explanation usage monitored

SUPPORTING PROCEDURES

This standard is implemented through the following detailed procedures:

Procedure PROC-AI-TRANS-001: Instructions for Use Creation Procedure

Purpose: Define step-by-step process for creating Instructions for Use
Owner: Product Director
Implements: Controls TRANS-001, TRANS-002

Procedure Steps:

  1. Gather AI system information
  2. Create Instructions for Use template
  3. Complete all required elements - Control TRANS-001
  4. Conduct legal review
  5. Obtain approval
  6. Distribute to deployers
  7. Track acknowledgments
  8. Update as needed - Control TRANS-002

Outputs:

  • Complete Instructions for Use
  • Distribution records
  • Acknowledgment records

Procedure PROC-AI-TRANS-002: Transparency Notice Implementation Procedure

Purpose: Define process for implementing transparency notices
Owner: Product Director
Implements: Controls TRANS-003, TRANS-004

Procedure Steps:

  1. Identify AI system type requiring disclosure
  2. Design transparency notice - Control TRANS-004
  3. Test with users
  4. Implement in UI/UX - Control TRANS-003
  5. Verify visibility and accessibility
  6. Monitor effectiveness
  7. Update based on feedback

Outputs:

  • Transparency notices
  • User testing results
  • Implementation verification
  • Effectiveness metrics

Procedure PROC-AI-TRANS-003: Transparency Information Update Procedure

Purpose: Define process for updating transparency information
Owner: Product Director
Implements: Control TRANS-002

Procedure Steps:

  1. Monitor for update triggers
  2. Assess impact on transparency information
  3. Update Instructions for Use
  4. Update transparency notices (if needed)
  5. Obtain approval
  6. Communicate updates
  7. Track acknowledgments

Outputs:

  • Updated transparency information
  • Update notifications
  • Acknowledgment records

COMPLIANCE

5.1 Compliance Monitoring

Monitoring Approach: Continuous automated monitoring supplemented by monthly manual reviews and quarterly comprehensive audits.

Compliance Metrics:

MetricTargetMeasurement MethodFrequencyOwner
Instructions for Use Completeness100%% of high-risk AI with complete Instructions for UseMonthlyProduct Director
Instructions for Use Currency100%% of Instructions for Use current (< 1 year old)MonthlyProduct Director
Transparency Notice Display Rate100%% of user interactions with disclosureDailyProduct Director
User Awareness≥90%% of users aware of AI use (survey)QuarterlyProduct Director
Update Timeliness100%% of updates completed within 30 daysMonthlyProduct Director
Explanation Availability100%% of AI systems with explanation capabilityQuarterlyProduct Director
Accessibility Compliance100%% of notices meeting WCAG 2.1 AAQuarterlyProduct Director

Monitoring Tools:

  • Transparency Dashboard
  • User Survey Results
  • Compliance Reports
  • Monthly compliance reports
  • Quarterly AI Governance Committee reviews

5.2 Internal Audit Requirements

Audit Frequency: Annually (minimum)

Audit Scope:

  • Instructions for Use completeness and quality
  • Instructions for Use currency
  • Transparency notice implementation
  • Transparency notice effectiveness
  • Explanation capability implementation
  • Accessibility compliance
  • Controls effectiveness (TRANS-001 through TRANS-005)

Audit Activities:

  • Review 100% of high-risk AI for Instructions for Use
  • Sample 20% of Instructions for Use for quality review
  • Test transparency notices in production
  • Survey users for awareness
  • Test explanation capability
  • Test accessibility compliance
  • Interview key personnel

Audit Outputs:

  • Annual Transparency Audit Report
  • Findings and recommendations
  • Corrective action plans for deficiencies

5.3 External Audit / Regulatory Inspection

Preparation:

  • Maintain audit-ready transparency documentation at all times
  • Designate Product Director and Legal as regulatory liaisons
  • Prepare standard response procedures for authority requests

Provide to Auditors/Regulators:

  • Instructions for Use (sample or all)
  • Transparency notices (screenshots/recordings)
  • User awareness survey results
  • Explanation capability documentation
  • Accessibility test results
  • Transparency procedures
  • Internal audit reports
  • Evidence of controls execution

Authority Request Response:

  • Acknowledge request within 1 business day
  • Provide requested documentation within 5 business days
  • Coordinate through Legal and Product Director
  • Document all interactions with authorities

ROLES AND RESPONSIBILITIES

6.1 RACI Matrix

ActivityProduct DirectorAI System OwnerUX/UI DesignLegal & ComplianceMarketingCustomer Support
Instructions for Use CreationR/AACCCI
Instructions for Use UpdatesRACCCI
Transparency Notice DesignRARCCC
Transparency Notice ImplementationRARCII
Explanation CapabilityRACIII
User TestingRARICC
Accessibility ComplianceRARIII
Deployer CommunicationRAICRR
User CommunicationRACCRR

RACI Legend:

  • R = Responsible (does the work)
  • A = Accountable (ultimately answerable)
  • C = Consulted (provides input)
  • I = Informed (kept up-to-date)

6.2 Role Descriptions

Product Director

  • Primary Responsibility: Owns transparency framework, ensures compliance
  • Key Activities:
    • Establishes transparency framework
    • Approves Instructions for Use
    • Approves transparency notices
    • Reports transparency metrics
    • Coordinates with authorities
  • Required Competencies: EU AI Act Articles 13 and 50, product management, user experience

AI System Owner

  • Primary Responsibility: Accountable for transparency of their AI system
  • Key Activities:
    • Ensures Instructions for Use created
    • Ensures transparency notices implemented
    • Participates in user testing
    • Coordinates updates
  • Required Competencies: AI system knowledge, transparency requirements

UX/UI Design

  • Primary Responsibility: Designs and implements transparency notices
  • Key Activities:
    • Designs transparency notices
    • Implements notices in UI/UX
    • Tests accessibility
    • Conducts user testing
  • Required Competencies: UX/UI design, accessibility (WCAG), user testing

Legal & Compliance

  • Primary Responsibility: Reviews for regulatory compliance
  • Key Activities:
    • Reviews Instructions for Use
    • Reviews transparency notices
    • Advises on regulatory requirements
    • Manages authority requests
  • Required Competencies: EU AI Act legal expertise, regulatory compliance

Marketing

  • Primary Responsibility: Communicates transparency to deployers and users
  • Key Activities:
    • Communicates Instructions for Use to deployers
    • Communicates transparency to users
    • Manages deployer relationships
  • Required Competencies: Communication, stakeholder management

Customer Support

  • Primary Responsibility: Supports users with transparency questions
  • Key Activities:
    • Answers user questions about AI use
    • Provides explanations
    • Escalates complex questions
  • Required Competencies: Customer service, AI system knowledge

EXCEPTIONS

7.1 Exception Philosophy

Transparency is a critical regulatory compliance activity for high-risk AI systems and AI systems with transparency obligations. Exceptions are granted restrictively and only where compensating controls adequately mitigate risks.


7.2 Allowed Exceptions

The following exceptions may be granted with proper justification and approval:

Exception TypeJustification RequiredMaximum DurationApproval AuthorityCompensating Controls
Simplified Instructions (Minimal-Risk AI)AI system clearly minimal-risk; simplified instructions sufficientPermanentProduct DirectorDocument rationale; Annual re-confirmation
Extended Update TimelineResource constraints prevent timely update60 daysProduct Director + AI Governance CommitteeInterim communication; Accelerated plan

7.3 Prohibited Exceptions

The following exceptions cannot be granted under any circumstances:

Skipping Instructions for Use for high-risk AI - Mandatory per Article 13, no exceptions
Skipping AI disclosure for Article 50 systems - Mandatory per Article 50, no exceptions
Hiding transparency notices - Violates transparency requirements
Using unclear language - Violates understandability requirements
Skipping accessibility requirements - Violates accessibility laws


7.4 Exception Request Process

Step 1: Submit Exception Request

  • Complete Exception Request Form (FORM-AI-EXCEPTION-001)
  • Include business justification
  • Propose compensating controls
  • Specify duration requested
  • Attach risk assessment

Step 2: Risk Assessment

  • Product Director assesses risk of granting exception
  • Evaluates adequacy of compensating controls
  • Documents residual risk

Step 3: Approval

  • Route to appropriate approval authority based on exception type
  • Product Director approval: Minor exceptions
  • Product Director + AI Governance Committee: Significant exceptions
  • AI Governance Committee: Critical exceptions

Step 4: Documentation and Monitoring

  • Document exception in Exception Register
  • Assign exception owner
  • Set review date
  • Monitor compensating controls
  • Report exceptions quarterly to AI Governance Committee

Step 5: Exception Review and Closure

  • Review exception at specified review date
  • Assess if exception still needed
  • Close exception when normal transparency completed
  • Document lessons learned

ENFORCEMENT

8.1 Non-Compliance Consequences

ViolationSeverityConsequenceRemediation Required
High-risk AI without Instructions for UseCriticalImmediate suspension until Instructions createdCreate Instructions within 10 business days; Root cause analysis
Missing AI disclosure (Article 50)CriticalImmediate correction; Compliance gap assessmentImplement disclosure within 5 business days
Outdated Instructions for UseHighEscalation to AI Governance Committee; Freeze on new featuresUpdate Instructions within 30 business days
Inaccessible transparency noticesHighImmediate correction; Legal riskFix accessibility within 5 business days
Unclear transparency languageMediumWritten warningImprove language within 10 business days
Missing explanation capabilityMediumWritten warningImplement explanation capability within 30 business days

8.2 Escalation Procedures

Level 1: Product Director

  • Minor procedural violations
  • Documentation deficiencies
  • Timeline delays < 5 days
  • Action: Written warning, corrective action required

Level 2: Product Director + AI Governance Committee

  • Repeated violations
  • Missing Instructions for Use
  • Missing transparency notices
  • Action: Formal review, corrective action plan, management notification

Level 3: AI Governance Committee

  • High-risk AI without Instructions for Use
  • Missing Article 50 disclosures
  • Critical compliance failures
  • Action: Immediate AI system suspension, investigation, disciplinary action

Level 4: Executive Management + Legal

  • Potential regulatory enforcement action
  • Significant legal liability
  • Reputational risk
  • Action: Executive crisis management, legal strategy, regulatory engagement

8.3 Immediate Escalation Triggers

Escalate immediately to AI Governance Committee + Legal if:

  • ⚠️ High-risk AI system operating without Instructions for Use
  • ⚠️ Article 50 disclosure missing for applicable systems
  • ⚠️ Regulatory inquiry or inspection related to transparency
  • ⚠️ User complaint about lack of transparency
  • ⚠️ Media attention to transparency issues

8.4 Disciplinary Actions

Individuals responsible for transparency violations may be subject to:

  • Verbal or written warning
  • Mandatory retraining
  • Performance improvement plan
  • Reassignment of responsibilities
  • Suspension (with pay during investigation)
  • Termination (for egregious violations, e.g., knowingly hiding AI use)

Factors Considered:

  • Intent (knowing violation vs. honest mistake)
  • Severity of violation
  • Impact (actual or potential)
  • Cooperation with remediation
  • Prior violation history

KEY PERFORMANCE INDICATORS (KPIs)

9.1 Transparency KPIs

KPI IDKPI NameDefinitionTargetMeasurement MethodFrequencyOwnerReporting To
KPI-TRANS-001Instructions for Use Completeness% of high-risk AI with complete Instructions for Use100%(# high-risk AI with IFU / # high-risk AI) × 100MonthlyProduct DirectorAI Governance Committee
KPI-TRANS-002Instructions for Use Currency% of Instructions for Use current (< 1 year old)100%(# current IFU / # total IFU) × 100MonthlyProduct DirectorManagement
KPI-TRANS-003Transparency Notice Display Rate% of user interactions with disclosure100%(# interactions with disclosure / # total interactions) × 100DailyProduct DirectorManagement
KPI-TRANS-004User Awareness% of users aware of AI use (survey)≥90%Survey resultsQuarterlyProduct DirectorAI Governance Committee
KPI-TRANS-005Update Timeliness% of updates completed within 30 days100%(# updates within 30 days / # total updates) × 100MonthlyProduct DirectorManagement
KPI-TRANS-006Explanation Availability% of AI systems with explanation capability100%(# AI systems with explanations / # total AI systems) × 100QuarterlyProduct DirectorAI Governance Committee
KPI-TRANS-007Accessibility Compliance% of notices meeting WCAG 2.1 AA100%(# compliant notices / # total notices) × 100QuarterlyProduct DirectorAI Governance Committee
KPI-TRANS-008Deployer Acknowledgment Rate% of deployers acknowledging Instructions for Use100%(# acknowledgments / # deployers) × 100MonthlyProduct DirectorManagement
KPI-TRANS-009Explanation Usage Rate% of users requesting explanations≥10%(# explanation requests / # total decisions) × 100MonthlyProduct DirectorManagement
KPI-TRANS-010User Satisfaction (Transparency)User satisfaction score with transparency≥4.0/5.0User survey scoreQuarterlyProduct DirectorAI Governance Committee

9.2 KPI Dashboards and Reporting

Real-Time Dashboard (Product Director access)

  • Current Instructions for Use completeness
  • Transparency notice display rates
  • User awareness metrics
  • Explanation usage
  • Accessibility compliance

Monthly Management Report

  • KPI-TRANS-001, 002, 003, 005, 008, 009
  • Trend analysis (vs. previous month)
  • Issues and risks
  • Planned actions

Quarterly AI Governance Committee Report

  • All KPIs
  • User awareness survey results
  • Explanation capability status
  • Accessibility compliance status
  • Internal audit findings (if conducted)
  • Exception register review

Annual Executive Report

  • Full-year KPI performance
  • Transparency maturity assessment
  • Strategic recommendations
  • Regulatory outlook

9.3 KPI Thresholds and Alerts

KPIGreen (Good)Yellow (Warning)Red (Critical)Alert Action
Instructions for Use Completeness100%95-99%< 95%Red: Immediate escalation to AI Governance Committee Chair
Transparency Notice Display Rate100%95-99%< 95%Red: Immediate escalation to AI Governance Committee
User Awareness≥90%80-89%< 80%Yellow: Improve transparency; Red: Escalate to AI Governance Committee
Update Timeliness100%90-99%< 90%Red: Escalate to AI Governance Committee

TRAINING REQUIREMENTS

10.1 Training Program Overview

All personnel involved in transparency must complete role-specific training to ensure competency in EU AI Act Articles 13 and 50 requirements, transparency design, and transparency procedures.


10.2 Role-Based Training Requirements

RoleTraining CourseDurationContentFrequencyAssessment Required
Product DirectorTransparency Management Expert Training12 hoursEU AI Act Articles 13 and 50; Instructions for Use; Transparency notices; ExplainabilityInitial + annuallyYes - Written exam (≥90%)
UX/UI DesignersTransparency Notice Design8 hoursArticle 50 requirements; Notice design; Accessibility; User testingInitial + annuallyYes - Practical design exercise
AI System OwnersTransparency Overview4 hoursTransparency requirements; Responsibilities; Instructions for UseAt onboarding + annuallyYes - Knowledge check (≥80%)
MarketingTransparency Communication4 hoursCommunicating transparency; Deployer communication; User communicationInitial + annuallyYes - Knowledge check (≥80%)
Customer SupportTransparency Support3 hoursAnswering transparency questions; Providing explanationsInitial + annuallyYes - Knowledge check (≥80%)
All AI Development StaffTransparency Awareness2 hoursTransparency basics; When transparency needed; RequirementsAt onboarding + annuallyYes - Knowledge check (≥80%)

10.3 Training Content by Topic

EU AI Act Article 13 (Instructions for Use)

  • Required elements
  • Content requirements
  • Distribution requirements
  • Update requirements

EU AI Act Article 50 (User Transparency)

  • Disclosure requirements
  • Chatbot/conversational AI disclosure
  • Synthetic content labeling
  • Emotion recognition disclosure
  • Biometric categorization disclosure

Transparency Notice Design

  • Design principles
  • Accessibility requirements (WCAG 2.1 AA)
  • User testing
  • Effectiveness measurement

Explainability

  • Explanation requirements
  • Explanation formats
  • User understanding
  • Human review

10.4 Training Delivery Methods

Initial Training:

  • Instructor-led classroom or virtual training
  • Includes interactive exercises and case studies
  • Hands-on practice with transparency tools
  • Group discussions of complex scenarios

Annual Refresher:

  • E-learning modules for core content review
  • Live update sessions for regulatory changes
  • Case study reviews of recent transparency activities
  • Knowledge assessment

On-the-Job Training:

  • Mentoring for new transparency staff
  • Job shadowing during transparency implementation
  • Supervised transparency work for first 3 AI systems

Just-in-Time Training:

  • Quick reference guides and job aids
  • Video tutorials on specific topics
  • Help desk support from experienced transparency staff

10.5 Training Effectiveness Measurement

Assessment Methods:

  • Written exams for knowledge retention
  • Practical exercises for skill application
  • On-the-job observations for competency validation
  • Feedback surveys for training quality

Competency Validation:

  • Product Directors: Must create Instructions for Use for 1 sample AI system with 100% completeness before independent creation
  • UX/UI Designers: Must design transparency notice meeting all requirements
  • All staff: Must pass knowledge assessments with minimum required scores

Training Metrics:

MetricTargetFrequency
Training completion rate100%Quarterly
Assessment pass rate (first attempt)≥ 90%Per training
Training effectiveness score (survey)≥ 4.0/5.0Per training
Time to competency (Product Directors)< 30 daysPer person

10.6 Training Records

Records Maintained:

  • Training attendance records
  • Assessment scores
  • Competency validations
  • Refresher training completion
  • Individual training transcripts

Retention: 10 years (to align with EU AI Act documentation retention)

Access: HR, Product Director, Internal Audit, Competent Authorities (upon request)


DEFINITIONS

TermDefinitionSource
Instructions for UseComprehensive information provided to deployers about AI system characteristics, capabilities, limitations, and proper useEU AI Act Article 13(3)
TransparencyProviding clear information to users about AI system use and capabilitiesEU AI Act Article 50
AI DisclosureInforming users they are interacting with AIEU AI Act Article 50
Synthetic ContentAI-generated content such as deepfakes, synthetic mediaEU AI Act Article 50(2)
Emotion RecognitionAI system that infers emotions from biometric dataEU AI Act Article 50(3)
Biometric CategorizationAI system that categorizes persons based on biometric dataEU AI Act Article 50(3)
ExplainabilityAbility to explain AI decisions in understandable termsThis Standard
DeployerNatural or legal person using AI system under its authorityEU AI Act Article 3(4)
UserEnd user interacting with AI systemThis Standard
AccessibilityDesign that enables use by people with disabilities (WCAG 2.1 AA)WCAG 2.1

LINK WITH AI ACT AND ISO42001

12.1 EU AI Act Regulatory Mapping

This standard implements the following EU AI Act requirements:

EU AI Act ProvisionArticleRequirement SummaryImplemented By (Controls)
Transparency and Information ProvisionArticle 13Instructions for Use for high-risk AITRANS-001, TRANS-002
Transparency ObligationsArticle 50User disclosure for certain AI systemsTRANS-003, TRANS-004
AI DisclosureArticle 50(1)Inform users of AI interactionTRANS-003
Synthetic Content LabelingArticle 50(2)Label AI-generated content (obligation on providers)TRANS-003
Emotion Recognition DisclosureArticle 50(3)Inform persons of emotion recognition use (obligation on deployers)TRANS-003
Biometric Categorization DisclosureArticle 50(3)Inform persons of biometric categorization use (obligation on deployers)TRANS-003
Deep Fake DisclosureArticle 50(4)Disclose that content is artificially generated or manipulated (obligation on deployers)TRANS-003

12.2 ISO/IEC 42001:2023 Alignment

This standard aligns with ISO/IEC 42001:2023 as follows:

ISO 42001 ClauseRequirementImplementation in This Standard
Clause 5.2: AI policyPolicy includes transparency commitmentInstructions for Use and transparency notices
Clause 6.1.2: AI system impact assessmentAssess transparency impactsTransparency notice design and testing
Clause 7.4: CommunicationCommunicate with interested partiesInstructions for Use, transparency notices
Clause 8.2: AI system risk assessmentTransparency-related risk assessmentTransparency requirements in risk assessment

12.3 Relationship to Other Standards

This transparency standard integrates with other AI Act standards:

Related StandardIntegration PointRationale
STD-AI-001: ClassificationClassification determines transparency requirementsHigh-risk AI requires Article 13; Article 50 applies to specific AI types
STD-AI-002: Risk ManagementKnown risks documented in Instructions for UseRisk management outputs feed into Instructions for Use
STD-AI-004: Technical DocumentationInstructions for Use referenced in technical documentationInstructions for Use linked to Annex IV documentation
STD-AI-007: Human OversightHuman oversight measures in Instructions for UseHuman oversight information included in Instructions for Use

12.4 References and Related Documents

EU AI Act (Regulation (EU) 2024/1689):

  • Article 13: Transparency and Information Provision
  • Article 13(3): Instructions for Use Content
  • Article 50: Transparency Obligations
  • Article 50(1): AI Disclosure Requirements (chatbot/conversational AI disclosure)
  • Article 50(2): Synthetic Content Marking (obligation on providers)
  • Article 50(3): Emotion Recognition and Biometric Categorisation Disclosure (obligation on deployers)
  • Article 50(4): Deep Fake Disclosure (obligation on deployers)

ISO/IEC Standards:

  • ISO/IEC 42001:2023: Information technology — Artificial intelligence — Management system
  • ISO/IEC 23894:2023: Information technology — Artificial intelligence — Guidance on risk management

Accessibility Standards:

  • WCAG 2.1: Web Content Accessibility Guidelines

Internal Documents:

  • POL-AI-001: Artificial Intelligence Policy (parent policy)
  • STD-AI-001: AI System Classification Standard
  • STD-AI-002: AI Risk Management Standard
  • STD-AI-004: AI Technical Documentation Standard
  • PROC-AI-TRANS-001, -002, -003: Transparency procedures

APPROVAL AND AUTHORIZATION

RoleNameTitleSignatureDate
Prepared ByProduct DirectorProduct Director_________________________
Reviewed ByMichael BrownChief Legal Officer_________________________
Reviewed BySarah JohnsonAI Act Program Manager_________________________
Reviewed ByJane DoeChief Strategy & Risk Officer_________________________
Approved ByJane DoeAI Governance Committee Chair_________________________

Effective Date: 2025-08-01
Next Review Date: 2026-08-01
Review Frequency: Annually or upon regulatory change


END OF STANDARD STD-AI-006


This standard is a living document. Feedback and improvement suggestions should be directed to the Product Director.

Standard Details

Standard ID

STD-AI-006

Version

1.0

Status

draft

Owner

Product Directors

Effective Date

2025-08-01

Applicability

High-risk AI systems, systems with transparency obligations

EU AI Act References
Article 13Article 50
ISO 42001 Mapping
Clause 4.2Clause 7.4