1. Annex IV Technical Documentation: Overview
Article 11 of the EU AI Act requires providers of high-risk AI systems to draw up technical documentation before the system is placed on the market or put into service. The content and structure of this documentation is prescribed in Annex IV of the regulation — a detailed list of information elements that must be recorded, maintained, and made available to national competent authorities upon request.
The purpose of technical documentation is threefold. First, it enables conformity assessment — whether self-assessed under Annex VI or audited by a notified body under Annex VII, the assessor needs comprehensive documentation to evaluate compliance. Second, it enables market surveillance — authorities can request documentation at any point during the system's market life to verify ongoing compliance. Third, it enables transparency and accountability — deployers, affected persons, and oversight bodies can understand how the system works and what safeguards are in place.
Scope of Obligation
Technical documentation requirements apply to all providers of high-risk AI systems — regardless of organization size. While Article 62 provides simplified documentation provisions for SMEs (small and medium-sized enterprises) and startups, these simplifications relate to format and proportionality, not exemption from the core information requirements. Every element of Annex IV must still be addressed.
2. What Must Be Documented
Annex IV organizes documentation into eight major categories. Each category addresses a distinct aspect of the AI system's design, development, testing, and deployment:
A. General Description of the AI System
- The system's intended purpose, provider identity, and version number
- How the AI system interacts with hardware, software, or other AI systems
- The versions of relevant software or firmware and any requirements on third-party software
- A description of the forms in which the AI system is placed on the market (embedded in hardware, downloadable software, API)
- The underlying technology, including general logic and algorithms used
B. Detailed Description of System Elements
- The development process, including design specifications and architecture decisions
- The computational and hardware resources used for development, training, testing, and validation
- Where relevant, data requirements in terms of datasheets describing training methodologies and techniques
- Assessment of human oversight measures, including technical measures to facilitate interpretation of outputs
- Pre-determined changes to the system and its performance, if any
C. Risk Management Documentation
- The risk management system established per Article 9, including identified risks and mitigation measures
- Residual risks and their justification
- Results of risk assessment iterations throughout the development process
- Risk mitigation measures and their effectiveness testing
D. Data Governance and Management
- Training, validation, and testing data descriptions — origin, scope, characteristics, and any preprocessing or labelling
- Data quality metrics and selection criteria
- Relevant design choices — collection methods, identification of gaps or shortcomings, and how they were addressed
- Bias detection, prevention, and mitigation measures applied to datasets
E. Testing and Validation
- Testing and validation procedures used, including metrics and test logs
- Performance against accuracy, robustness, and cybersecurity benchmarks (Article 15)
- Potentially discriminatory impact testing results
- Testing under foreseeable conditions of misuse
F. Monitoring and Control
- Measures for human oversight (Article 14 compliance documentation)
- Logging capabilities — what events are recorded, in what format, for how long
- Description of the quality management system (Article 17)
| Category | Key Articles | Typical Volume |
|---|---|---|
| General description | Art. 11, Annex IV(1) | 10-20 pages |
| System elements & development | Art. 11, Annex IV(2) | 30-80 pages |
| Risk management | Art. 9, Annex IV(3) | 20-50 pages |
| Data governance | Art. 10, Annex IV(4) | 15-40 pages |
| Testing & validation | Art. 15, Annex IV(5) | 20-60 pages |
| Monitoring & control | Art. 12, 14, 17 | 10-30 pages |
3. Format and Structure
The EU AI Act does not prescribe a single mandatory format for technical documentation. However, several structural principles are either required or strongly implied by the regulation and the emerging guidance from the European Commission:
- Clear and comprehensive. Article 11(1) states documentation must be “drawn up in such a way as to demonstrate that the high-risk AI system complies with the requirements” and “provide national competent authorities and notified bodies with all the necessary information to assess the compliance”
- Kept up to date. Documentation is not a one-time snapshot — it must be updated whenever the system undergoes substantial modification
- Language requirements. Documentation must be in an official EU language determined by the member state(s) where the system is placed on the market. In practice, most providers maintain documentation in English as the primary language with translations as required
- Machine-readable where possible. While not a strict legal requirement, the trend toward automated compliance checking favors structured, machine-readable documentation formats (JSON, XML, standardized templates) over unstructured prose
Harmonized Standards
CEN and CENELEC are developing harmonized standards (under standardization request M/593) that will provide detailed templates and formats for AI system documentation. Once published and referenced in the Official Journal of the EU, compliance with these standards will confer a presumption of conformity with the corresponding documentation requirements. Until then, providers should follow the Annex IV structure as closely as possible and monitor CEN/CENELEC progress.
Practical Tip
Structure your documentation to mirror Annex IV section by section. Use a clear table of contents, version control, and cross-references to supporting evidence. Even before harmonized standards are finalized, this approach ensures your documentation is auditable, traceable, and easy for assessors to navigate.
4. When Documentation Must Be Created
The EU AI Act is unambiguous on timing: technical documentation must be drawn up before the high-risk AI system is placed on the market or put into service (Article 11(1)). This is a prerequisite, not a post-deployment requirement. The conformity assessment cannot begin without complete documentation, and CE marking cannot be applied without a completed conformity assessment.
Documentation as a Development Discipline
In practice, this means documentation must be an integral part of the development lifecycle — not something compiled retroactively after development is complete. The most effective approach is continuous documentation: recording design decisions, risk assessments, data provenance, and test results as they occur during development, rather than reconstructing them from memory after the fact.
Retroactive documentation is not only more expensive and error-prone — it also fails to capture the iterative reasoning that led to specific design choices. The risk management system required by Article 9, for instance, mandates a “continuous iterative process” — documenting only the final state of the risk assessment misses the iterations that demonstrate due diligence.
| Development Phase | Documentation Activities |
|---|---|
| Requirements & design | Intended purpose, risk classification rationale, initial risk identification, architecture decisions, data sourcing strategy |
| Data preparation | Dataset provenance, quality assessment, bias detection results, preprocessing steps, labelling methodology |
| Training & development | Algorithm selection rationale, hyperparameter choices, training logs, iterative risk assessment updates |
| Testing & validation | Test plans, performance metrics, accuracy benchmarks, robustness testing, cybersecurity assessments, discrimination testing |
| Pre-market | Final documentation compilation, conformity assessment, Declaration of Conformity, quality management system documentation |
5. Keeping Documentation Updated Throughout the Lifecycle
Technical documentation is a living artifact. The EU AI Act establishes several triggers that require documentation updates after initial market placement:
Substantial Modification
Article 6(1)(c) defines substantial modification as a change to the AI system “which is not foreseen or planned in the initial conformity assessment and as a result of which the compliance of the system with the requirements is affected.” When a substantial modification occurs, the provider must:
- Update the technical documentation to reflect the modification
- Conduct a new conformity assessment (or update the existing one)
- If applicable, issue a new Declaration of Conformity
- Update the EU database registration
Post-Market Monitoring Findings
Article 72 requires providers to establish a post-market monitoring system. Findings from this monitoring — including performance degradation, new risks identified in production, incidents, or near-misses — must be reflected in the documentation. The risk management system documentation, in particular, must evolve based on real-world evidence.
Continuous Learning Systems
AI systems that continue learning after deployment pose a particular documentation challenge. The EU AI Act acknowledges this by requiring that pre-determined changes — changes to the system that are anticipated and designed into the initial architecture — be documented upfront, including their expected impact on performance and compliance. Changes that fall outside the pre-determined envelope constitute substantial modifications and trigger full re-assessment.
Version Control Is Non-Negotiable
Every version of the technical documentation must be preserved and traceable to the corresponding version of the AI system. When an authority requests documentation for a system deployed in 2027, you must be able to produce the documentation that was current at that deployment date — not just the latest version. Treat documentation with the same versioning discipline as source code.
6. How Automated Tools Reduce the Burden
The documentation requirements of Annex IV are extensive. For a typical high-risk AI system, producing complete technical documentation can require 100-300 pages of structured content, covering everything from data provenance to cybersecurity testing results. Maintaining this documentation across system versions and through post-market monitoring multiplies the effort further.
Automated documentation tools address this burden at multiple levels:
Code-to-Documentation Extraction
Modern tools can parse source code, configuration files, and ML pipeline definitions to automatically extract system architecture, data flow descriptions, model parameters, and dependency graphs. This eliminates the most labor-intensive part of documentation — manually transcribing what is already expressed in code into human-readable documentation.
Regulatory Mapping
Tools that understand the structure of regulatory requirements can automatically map system components to the specific Annex IV information elements they satisfy. This ensures completeness — no requirement goes undocumented because it was overlooked during manual compilation.
The TAMR+ Approach
Traceable Agentic Multi-hop Reasoning (TAMR+) takes a fundamentally different approach to compliance documentation. Rather than treating documentation as a static text-generation task, TAMR+ represents the AI system, its regulatory requirements, and the evidence of compliance as a knowledge graph. This graph-based representation enables:
- Multi-hop regulatory reasoning — traversing from a system component through its risk assessment to the applicable Article requirements and back to the evidence trail, in a single query
- 74% accuracy on regulatory questions — when asked whether a specific compliance requirement is met, TAMR+ provides accurate, traceable answers three-quarters of the time, with full citation chains
- Continuous compliance monitoring — because the graph updates as the codebase changes, documentation stays current automatically rather than requiring periodic manual refresh
- Audit trail generation — every compliance claim is backed by a traceable path through the knowledge graph, from assertion to evidence
Why Graph Intelligence Outperforms Vector Search for Compliance
Traditional RAG (Retrieval-Augmented Generation) approaches use vector similarity to find relevant regulatory text. This works for simple lookups but fails on multi-hop compliance questions — “Is our data governance process compliant with Article 10, considering the exception in Article 10(5) for data generated in regulatory sandboxes?” TAMR+ traverses the regulatory graph to chain these connections, producing structured, verifiable answers rather than probabilistic text similarity matches.
7. Retention Requirements
The EU AI Act establishes explicit retention periods for different categories of documentation:
| Document Type | Retention Period | Legal Basis |
|---|---|---|
| Technical documentation (Annex IV) | 10 years | Article 18(1) |
| EU Declaration of Conformity | 10 years | Article 47(2) |
| Quality management system documentation | 10 years | Article 17(2) |
| Automatically generated logs | Min. 6 months (deployer) | Article 26(6) |
| Post-market monitoring data | Proportionate to lifecycle | Article 72 |
The 10-year retention period begins from the date the AI system is placed on the market or put into service — whichever is later. For systems with long operational lives (critical infrastructure, medical devices), this means documentation may need to be retained for 15-20+ years in practice.
Interaction with GDPR
Where technical documentation or logs contain personal data — which is common for AI systems processing human-related data — the retention obligations under the EU AI Act must be balanced against GDPR's data minimization and storage limitation principles (Article 5(1)(c)(e) GDPR). The AI Act explicitly addresses this tension: personal data in documentation should be limited to what is strictly necessary for compliance verification, and appropriate technical and organizational measures must protect it throughout the retention period.
Practical Retention Strategy
Implement a documentation archival system with versioned, immutable snapshots tied to system versions. Use cryptographic hashing to ensure document integrity over the retention period. Separate documentation containing personal data from purely technical documentation to apply different retention and access policies. Automate retention period tracking and disposal scheduling.
8. Frequently Asked Questions
What must be included in AI system technical documentation under the EU AI Act?▾
When must technical documentation be created for an AI system?▾
How long must AI system documentation be retained?▾
Can automated tools help with EU AI Act documentation requirements?▾
Related in the AI Risk Cluster
AI Risk Assessment Framework
Comprehensive risk assessment aligned with EU AI Act Article 9 and NIST AI RMF
Human Oversight for AI: EU Obligations
Article 14 requirements and practical implementation patterns for human oversight
The Complete EU AI Act Compliance Guide
Comprehensive overview of EU AI Act timelines, classifications, and requirements
Conformity Assessment Guide
Step-by-step guide to self-assessment and notified body audits
