Docs

AI Transparency Requirements

Transparency rules require businesses to disclose when users interact with AI or encounter AI-generated content. This guide covers EU AI Act Article 50, chatbot identification, content labeling, watermarking, deepfake rules, and the right to know you are talking to AI.

EU AI Act Article 50

Effective August 2, 2026 — Providers and deployers of certain AI systems must meet transparency obligations.

AI that interacts with people — Must clearly indicate AI involvement at first interaction.

Generative AI outputs — Should be marked in machine-readable format and detectable as AI-generated where technically feasible.

Deepfakes — Deployers must disclose when image, audio, or video is AI-generated or manipulated.

Text on public interest — AI-generated text published on matters of public interest must be disclosed as artificially generated.

Emotion recognition and biometric categorization — Users must be informed when these systems are used.

Chatbot Identification

Principle — Users have a right to know they are talking to AI, not a human.

Implementation — Clear disclosure at the start of a conversation. Not buried in terms. Not implied. Explicit.

Examples — I am an AI assistant. This chat is powered by AI. Before you continue, note that you are speaking with an AI.

AI-Generated Content Labeling

When — Content that could be mistaken for human-created. Especially for public-facing or high-stakes use.

How — Labels, watermarks, or metadata. Machine-readable where possible for detection tools.

Where — Depends on context. Marketing, support, publishing. Err on the side of disclosure when in doubt.

Watermarking and Detection

Watermarking — Embedding signals in AI output (image, audio, video, text) to indicate AI origin. Some visible; some imperceptible.

Detection — Tools to verify AI-generated content. Providers may offer APIs for third-party verification.

Limitations — Not all content can be watermarked. Detection is not perfect. Use as one layer, not the only layer.

Deepfake Rules

Disclosure — AI-generated or manipulated image, audio, or video must be disclosed as artificial when deployed.

Exceptions — Law enforcement (authorized), artistic/creative/satirical (limited disclosure). Check jurisdiction.

Practical — If you use AI to generate or alter media for marketing or other deployment, disclose it.

What Businesses Need to Do

Customer-facing AI — Disclose at first interaction. Chatbots, virtual agents, support AI.

AI-generated content — Label or disclose when publishing. Marketing, articles, social. Especially for public interest topics.

Deepfakes — Disclose when deploying. Do not pass off AI media as real without disclosure.

Vendor review — Ensure AI vendors support transparency. Ask about disclosure features and compliance.

The Bottom Line

Transparency is a core principle: users should know when they interact with AI or see AI-generated content. Implement disclosure for chatbots and AI-generated content. Plan for Article 50 by August 2026. When in doubt, disclose.

Related Reading