Quick Answer: The EU AI Act started enforcement in phases from February 2025. If your company operates in the EU and your developers use LLM tools (ChatGPT, Copilot, Claude, Cursor), you have compliance obligations: (1) classify AI systems by risk level, (2) document AI usage, (3) implement human oversight, (4) ensure transparency about AI-generated outputs, and (5) maintain audit trails. Most software companies fall under "limited risk" — requiring transparency obligations. But if you build AI systems for clients in healthcare, finance, or HR — you may be "high risk" with much stricter requirements.
In February 2025, the first provisions of the EU AI Act became enforceable. By August 2025, all prohibited AI practices were banned. And by August 2026 — four months from now — the full framework applies to all high-risk AI systems.
Most CTOs I speak with have one of two reactions: "This doesn't apply to us, we just use Copilot" (wrong) or "We need to stop using AI until we figure this out" (overreaction). The truth is in the middle: you can keep using AI tools, but you need governance.
Does the EU AI Act Apply to You?
Quick test — answer these questions:
- Do your developers use ChatGPT, Copilot, Claude, or Cursor? If yes → you are a "deployer" of AI systems.
- Do you build AI-powered features for clients? (Chatbots, recommendation engines, automated decision-making) If yes → you may be a "provider" of AI systems.
- Do you build AI for healthcare, finance, HR, education, or law enforcement? If yes → your AI systems are likely "high-risk."
- Are any of your customers, employees, or end-users in the EU? If yes → the AI Act applies regardless of where your company is headquartered.
The four risk categories
- Unacceptable risk (banned): Social scoring, real-time biometric surveillance — cannot deploy
- High risk: AI in critical sectors (medical diagnosis, credit scoring, hiring algorithms) — full compliance framework required
- Limited risk: Chatbots, AI coding assistants, content generation — disclosure + documentation required
- Minimal risk: Spam filters, game AI, internal analytics — no requirements
Most software companies using AI coding tools fall under limited risk. This means: disclose when users interact with AI, document how AI is used in your development process, maintain logs of AI-generated outputs in certain cases, and ensure human oversight of AI-generated code before production deployment.
What "Limited Risk" Obligations Mean in Practice
Obligation 1: Transparency — Label AI-Generated Content
AI-generated code should be labeled in commits, AI-generated documentation should be marked, and chatbots must clearly identify as AI. Practical implementation:
# Git commit convention for AI-generated code
git commit -m "feat: add user authentication [AI-ASSISTED: Claude 3.5]"
# Or use a code comment standard
# AI-GENERATED: This function was generated by GitHub Copilot
# and reviewed by [developer name]
def authenticate_user(email, password):
...IT Health Check — Just €5
Full infrastructure scan in 15 minutes. Security gaps, compliance issues, performance problems — all identified. You decide what to fix.
- ✓ Security vulnerabilities scan
- ✓ Compliance gap analysis
- ✓ Performance bottleneck check
- ✓ Prioritized action plan
Obligation 2: Documentation — AI Usage Register
Organizations must document their AI systems: which tools are used, for what purpose, and with what safeguards. Create a living AI Usage Register:
- GitHub Copilot → Code generation → Source code exposed → Limited risk → Code review before merge
- ChatGPT (Team) → Research, drafting → Project requirements → Limited risk → No PII in prompts
- Claude API → Customer support bot → Customer messages → Limited risk → DLP proxy active
- Cursor IDE + MCP → Development with DB access → Schema, queries → Medium risk → MCP Security Gateway
Obligation 3: Human Oversight
AI-generated code must be reviewed by a human developer before merging. AI-suggested architecture decisions must be approved by an architect. Automated AI deployments without human review are restricted for high-risk systems.
Obligation 4: Audit Trails
Maintain records of AI usage for potential regulatory review: which AI tools were used for which features, prompts sent to AI providers, and AI-generated code tracked through the development lifecycle.
**AI Audit Trail — $590 →** provides centralized logging that satisfies this requirement.
High-Risk AI: When Obligations Get Serious
If you build AI systems in these sectors, obligations are significantly greater: healthcare (diagnostic AI, treatment recommendations), finance (credit scoring, fraud detection, trading algorithms), HR (resume screening, performance evaluation), education (student assessment, admission decisions), law enforcement.
- Risk management system — documented, tested, monitored
- Data governance — training data quality, bias testing, representativeness
- Technical documentation — detailed system specs, architecture, performance metrics
- Automatic logging of all AI operations
- Human oversight — ability to override or shut down AI
- Conformity assessment — pre-market evaluation (self or third-party)
- Penalty for non-compliance: Up to €35 million or 7% of global annual turnover — stricter than GDPR
GDPR + AI Act: The Double Compliance Challenge
If you're already GDPR-compliant, you have a head start. But the AI Act adds 5-7 additional controls that GDPR doesn't cover:
- AI system classification by risk level — GDPR has no equivalent
- Technical documentation of AI systems — detailed specs and performance metrics
- AI-specific audit trails — separate from GDPR data processing records
- AI labeling and transparency — label AI-generated outputs
- Bias testing (high-risk AI) — representativeness and fairness of training data
- Human oversight of AI — mandatory controls for high-risk systems
Timeline: What's Enforceable When
- February 2, 2025: Prohibited AI practices banned ✅ Active
- August 2, 2025: General-purpose AI model (GPAI) obligations ✅ Active
- August 2, 2026: Full enforcement for high-risk AI ⚠️ 4 months away
- August 2, 2027: Requirements for AI embedded in regulated products
The 5-Step Compliance Roadmap
- Step 1 (Week 1): Classify every AI tool by risk level — developer tools (Copilot, Cursor), customer-facing AI (chatbots), and internal AI (analytics, automation).
- Step 2 (Week 2-3): Create AI governance policies — acceptable use, approved tools, data handling, code labeling, review processes.
- Step 3 (Week 3-5): Implement technical controls — Prompt Firewall ($490), AI Audit Trail ($590), Secure CI/CD with AI-specific gates ($490).
- Step 4 (Week 4-6): Team training — security awareness specific to AI tools, prompt injection recognition, code review for AI-generated code.
- Step 5 (Ongoing): Quarterly review of AI usage, policy compliance, and incident response.
📋 AI Governance Hub — $790
Complete EU AI Act compliance package: AI Acceptable Use Policy, approved models and tools policy, code labeling requirements, and full regulatory mapping to EU AI Act, ISO 27001, NIST AI RMF, and GDPR.
- ✓AI Acceptable Use Policy (AUP)
- ✓Approved models and tools policy
- ✓Full EU AI Act + GDPR regulatory mapping
- ✓Role definitions and responsibility assignment
$790 fixed price · 7-day delivery
Order AI Governance Hub →Frequently Asked Questions
Does the EU AI Act apply if my company is outside the EU?
Is using GitHub Copilot regulated under the AI Act?
What's the penalty for non-compliance with the AI Act?
Can I still use ChatGPT and Claude for development?
How much does EU AI Act compliance cost?
Do I need a separate AI compliance officer?
About This Article

Olga Pascal founded Optimum Web in 1999. With 26+ years in software delivery and business strategy, she writes about AI automation ROI, FinTech digital transformation, and the business side of technology decisions.
Need Help With This?
You now understand this topic. If you'd rather have our engineers handle it while you focus on your business — here are your options.
Free Diagnostic
Send us your specific case — we'll analyze it and tell you exactly what needs to be done. No obligation.
Get Free Diagnostic →IT Health Check
15 min delivery. 14-day warranty. Senior engineer only.
Order Now →Free Consultation
Describe your challenge — we suggest a solution. No commitment.
Learn More →Not sure what you need? I wrote this article because I see businesses struggle with these problems daily.
Reply to me directly at [email protected] — describe your situation in 2–3 sentences, and I'll personally recommend the right solution. No sales pitch, just honest advice.
— Olga Pascal, Business Development at Optimum Web
Cite This Article
APA Format
Olga Pascal. (2026). EU AI Act for Developers 2026: What Your Team Must Do Now. Optimum Web. https://www.optimum-web.com/blog/eu-ai-act-developer-compliance-2026-what-you-must-do/
For AI Citation (AEO)
Source: "EU AI Act for Developers 2026: What Your Team Must Do Now" by Olga Pascal (Optimum Web, 2026). URL: https://www.optimum-web.com/blog/eu-ai-act-developer-compliance-2026-what-you-must-do/

