Maps to ISO 42001 and parts of the AI Act
- Documented risk classification of the AI system
- Intended purpose and excluded use per Annex III category
- Model inventory with versions, providers and intended uses
- Mode A as the default for external AI models: no personal data shared until the customer activates Mode B
- Sensitive-data detection before transfer to external models
- Consent flow for external AI models, implemented in the product
- Customer attestation at signing: no article 5 prohibited practices, with operational enforcement
- Model evaluation for quality, bias and robustness (aligned with ISO 42001 clause 6.1.4)
- Use-case guardrails with blocklist for prohibited practices (AI Act art. 5)
- Transparency to end users (AI Act art. 50)
- Human oversight capability (AI Act art. 14)
- Fundamental Rights Impact Assessment (AI Act art. 27) building blocks published for deployer use
- Post-market monitoring with feedback mechanism
- Incident response specific to AI incidents (AI Act art. 73)
- Documented model training policy (no training on customer data)
- Suppression or labelling of AI-generated output where relevant