EU AI ACT

EU AI Act Compliance

The world's first comprehensive AI regulation is here. Understand what's required, when it's required, and how to prove compliance.

€35M
Maximum fine for prohibited AI
Or 7% global revenue
Aug 2026
High-risk requirements effective
17 months to prepare
100%
PolicyGuard audit trail coverage
Prove every acknowledgment
OVERVIEW

What Is the EU AI Act?

The EU AI Act is the European Union's landmark regulation establishing a comprehensive legal framework for artificial intelligence. Adopted in 2024, it takes a risk-based approach that categorizes AI systems by their potential for harm and applies corresponding requirements.

The Act applies to providers who develop AI systems, deployers who use AI systems, and importers and distributors who bring AI systems into the EU market. Crucially, it applies regardless of where the organization is based if the AI system is used within the EU or affects EU residents.

For most organizations, the immediate concerns are the prohibited practices (effective February 2025), AI literacy requirements (effective February 2025), and high-risk system obligations (effective August 2026). Organizations using AI tools like ChatGPT, Claude, or Copilot must ensure employees understand acceptable use and that governance frameworks are in place.

CLASSIFICATION

Risk-Based Classification

Unacceptable Risk

Prohibited
Banned as of February 2025

Examples

Social scoring by governments, real-time biometric identification in public spaces (with limited exceptions), AI that manipulates behavior to cause harm, exploitation of vulnerabilities.

What It Means

These AI practices are prohibited outright. Organizations must not develop or deploy these systems.

High Risk

Strict Requirements
Requirements effective August 2026

Examples

AI in employment decisions, credit scoring, educational assessment, law enforcement, migration and border control, critical infrastructure.

What It Means

Subject to conformity assessments, risk management systems, data governance, documentation, human oversight, and accuracy requirements.

Limited Risk

Transparency Obligations
Effective now

Examples

Chatbots, emotion recognition systems, deepfake generators, AI-generated content.

What It Means

Must inform users they are interacting with AI or that content is AI-generated.

Minimal Risk

No Specific Requirements
No additional obligations

Examples

AI-enabled video games, spam filters, most business productivity AI.

What It Means

No specific EU AI Act requirements, but general laws and internal policies still apply.

TIMELINE

EU AI Act Timeline

August 2024Already Required
  • EU AI Act enters into force
February 2, 2025Already Required
  • Prohibited AI practices banned
  • AI literacy training required (Article 4)
  • Governance structures must be in place
August 2, 2025Upcoming
  • Governance rules for general-purpose AI
  • Notified body requirements
August 2, 2026Critical Deadline
  • High-risk AI system requirements
  • Conformity assessments required
  • Full documentation obligations
  • Human oversight requirements
August 2, 2027Ongoing
  • Requirements for high-risk AI embedded in regulated products
REQUIREMENTS

Key Requirements for Organizations

What It Requires

Organizations must ensure staff and others dealing with AI systems have sufficient AI literacy. This includes understanding AI capabilities, limitations, and risks appropriate to their role.

How PolicyGuard Helps

Built-in training modules with quizzes provide verifiable AI literacy education. Track completion by employee and department. Export training records for auditors.

What It Requires

Users must be informed when interacting with AI systems like chatbots. AI-generated content must be disclosed. Emotion recognition and biometric systems require explicit notice.

How PolicyGuard Helps

Policy templates include transparency disclosure requirements. Acknowledgment tracking proves employees understand disclosure obligations.

What It Requires

High-risk systems require technical documentation, risk management systems, data governance, human oversight measures, and accuracy monitoring.

How PolicyGuard Helps

PolicyGuard governs employee AI usage rather than AI system development. For employees using third-party AI tools in high-risk contexts, our policies establish acceptable use boundaries and audit trails.

What It Requires

Organizations must establish governance frameworks for AI, assign responsibilities, and maintain accountability for AI systems they deploy.

How PolicyGuard Helps

Department-based policy assignment, role-specific training, and comprehensive audit trails establish clear governance structure with documented accountability.

ENFORCEMENT

Penalties for Non-Compliance

€35M
or 7% of global annual revenue
Using prohibited AI practices
€15M
or 3% of global annual revenue
Non-compliance with high-risk AI requirements
€7.5M
or 1.5% of global annual revenue
Providing incorrect information, other violations

Small and medium enterprises face proportionally lower maximums. The regulation also allows for warnings and corrective measures before fines.

CHECKLIST

EU AI Act Compliance Checklist

Immediate Actions (Required Now)

Before August 2026

Ongoing

PLATFORM

Achieve EU AI Act Compliance with PolicyGuard

AI Literacy Training (Article 4)

Our built-in training modules satisfy the EU AI Act's AI literacy requirement. Employees complete video-based training with quizzes, and you get timestamped completion records proving compliance.

Policy Acknowledgment Trail

When auditors ask "how do you know employees follow your AI policy?", show them timestamped acknowledgments from every employee. Our browser extension captures acknowledgments at the point of AI tool access.

EU AI Act Policy Templates

Expert-curated policy templates specifically designed for EU AI Act compliance. Covering acceptable use, prohibited practices, transparency obligations, and incident reporting requirements.

Start Your EU AI Act Compliance Journey

August 2026 is closer than you think. Build your compliance foundation today.

See it in action. EU AI Act templates included. Setup in minutes.

FAQ

Frequently Asked Questions

Yes, if your AI systems are used within the EU or produce outputs affecting EU residents. The Act has extraterritorial reach similar to GDPR. US companies serving EU customers or with EU employees should prepare for compliance.

Generally no. Using general-purpose AI tools for productivity is typically minimal or limited risk. However, if you use AI to make or support consequential decisions (employment, credit, etc.), those applications may be high-risk. The risk classification depends on the use case, not the tool itself.

AI literacy means ensuring staff understand AI capabilities, limitations, and appropriate use for their role. Prove it with documented training programs, completion records, and assessments. PolicyGuard's training modules provide this with timestamped completion tracking.

Providers develop or commission AI systems and place them on the market. Deployers use AI systems under their authority. Most organizations using tools like ChatGPT are deployers. Both have obligations, but provider obligations are more extensive.

Some requirements are already in effect (February 2025): prohibited practices ban, AI literacy, and governance requirements. High-risk system requirements take effect August 2026. Start now to avoid scrambling as deadlines approach.

PolicyGuard provides EU AI Act-aligned policy templates, AI literacy training with verifiable completion records, and audit-ready evidence of policy acknowledgment. While we don't handle technical conformity assessment for AI systems you build, we ensure your workforce governance is compliant and documented.

Ready to govern every AI tool your team uses?

One platform to enforce policies, track compliance, and prove governance across 80+ AI tools.

Book a demo