Artificial intelligence regulations are now a reality, with enforceable legal requirements rolling out across key regions in 2026. Companies in the European Union, the United States, the United Kingdom, and elsewhere must ensure their AI systems meet specific compliance standards or risk substantial penalties. This guide explains what you need to understand and how to take action.
Artificial intelligence refers to machine-based systems that operate with varying levels of autonomy, generating outputs such as predictions, recommendations, or decisions that influence physical or virtual environments. The EU AI Act defines it broadly to capture current and future AI technologies, while US frameworks tend toward sector-specific definitions.
Most jurisdictions have adopted a risk-based approach to AI regulation. Systems posing greater potential harm face stricter compliance obligations, while low-risk AI applications remain largely unregulated.
Four principles appear consistently across frameworks:
• Transparency: Users must know when they’re interacting with AI
• Accountability: Humans remain responsible for AI outputs
• Human oversight: Critical decisions require human involvement
• Non-discrimination: AI systems cannot produce discriminatory outcomes against protected groups
The EU AI Act represents the world’s first comprehensive regulation specifically governing artificial intelligence. Adopted in June 2024, it establishes a 24-month compliance period with staggered enforcement.
Risk Classification System
The act categorises AI systems into four risk levels:
1. Unacceptable risk (prohibited)
2. High risk (strict requirements
3. Limited risk (transparency obligations)
4. Minimal risk (no specific requirements)
High-risk AI systems must obtain CE marking before entering the European market, demonstrating conformity with technical standards.
Penalties
• Up to €35 million or 7% of global annual turnover for prohibited AI violations
• Up to €15 million or 3% of turnover for other violations
• The European Commission and EU member states share enforcement authority through the newly established AI Office
The UK government has pursued a pro-innovation approach to AI regulation, delegating oversight to existing regulators rather than creating new legislation. The AI (Regulation) Bill has not yet passed, and the AI Authority is still only proposed, not in force as law.
The AI (Regulation) Bill 2025 would establish a central AI Authority to coordinate regulatory bodies and address gaps in existing regulation. The government’s proposals focus on balancing AI safety with economic growth and supporting innovation objectives.
The UK’s position aims to act quickly on AI risks while encouraging AI innovation through government support mechanisms.
The US lacks comprehensive federal regulation, resulting in a patchwork of state-level AI legislation and sector-specific rules.
Federal Framework
Executive Order 14110 (October 2023) established requirements for AI developers of powerful models, including safety testing and government reporting. A December 2025 executive order titled “Ensuring a National Policy Framework for Artificial Intelligence” now challenges state laws as potentially burdening AI innovation.
The NIST AI Risk Management Framework 1.0 provides voluntary guidelines that many AI developers treat as de facto standards.
Sector-Specific Rules
• FDA regulates medical AI applications
• NHTSA governs autonomous vehicles
• FTC addresses AI in consumer protection contexts
State-Level Activity
California leads with multiple laws effective January 1, 2026:
• AB 316: Eliminates the “autonomous-harm defence” in AI litigation
• SB 942: Requires detection AI tools for synthetic content from large platforms
• AI Safety Act: Protects employees reporting AI safety concerns
The key obligations for Colorado’s AI Act apply from February 1, 2026 and impose “reasonable care” duties on deployers of high-risk systems to prevent algorithmic discrimination.
Texas’s TRAIGA (effective January 1, 2026) authorises Attorney General investigations into AI systems, with civil investigative demands covering training data, performance metrics, and safeguards.
The EU AI Act bans AI systems presenting unacceptable risk to fundamental rights:
• Social scoring systems evaluating the general population’s behaviour
• Real-time biometric identification in public spaces (narrow law enforcement exceptions exist)
• AI exploiting vulnerabilities of children or disabled persons
• Subliminal techniques causing psychological or physical harm
• Emotion recognition in workplaces and schools (with exceptions)
Texas’s TRAIGA prohibits AI outputs encouraging self-harm, violence, criminality, constitutional rights violations, and AI-generated child sexual abuse material.
Systems in these categories face the strictest compliance requirements:
• Critical infrastructure: Transport, utilities, internet management
• Employment: CV screening, performance evaluation, worker management
• Essential services: Credit scoring, benefit eligibility, insurance pricing
• Law enforcement: Risk assessment, evidence evaluation
• Education: Exam scoring, admission decisions
• Healthcare: Diagnostic support, surgery assistance, triage systems
Colorado’s approach defines high-risk systems as those that make or substantially influence consequential decisions affecting individuals.
Generative AI models, including GPT-4, Claude, and Gemini, trigger transparency obligations:
• Clear disclosure that users are interacting with AI
• Labelling of AI-generated content, including synthetic media
• Information about copyrighted data used in AI training
• Chatbot identification requirements
Deepfakes require explicit marking. California’s SB 942 mandates that platforms with over one million monthly users provide detection tools for such content.
Organisations deploying high-risk AI systems must implement:
Risk Management
A documented system covering the entire AI lifecycle, from development through deployment and monitoring.
Data Governance
Quality management procedures for assessing the relevance, representativeness, and bias of training data. Records must be retained for 10 years.
Technical Documentation
Detailed records covering:
• System purpose and intended use
• Technical specifications
• Performance metrics and limitations
• Human AI oversight measures
Conformity Assessment
Registration in the EU database and third-party assessment for certain categories.
Ongoing Monitoring
Accuracy tracking, incident logging, and cybersecurity testing throughout operation.
AI regulation intersects with existing data protection frameworks, creating layered compliance obligations.
Article 22 rights regarding automated decision-making apply to AI systems making decisions with legal or significant effects. Individuals retain rights to:
• Human review of automated decisions
• Explanation of decision logic
• Contest decisions and obtain rectification
AI training datasets often involve international data transfers. Organisations must establish lawful transfer mechanisms:
• Adequacy decisions (for approved countries)
• Binding Corporate Rules (BCRs)
• Standard Contractual Clauses (SCCs)
AI development must incorporate privacy law data protection from the outset, including data minimisation, purpose limitation, and security measures.
The “Brussels Effect” means the EU AI Act functions as a global baseline; companies serving European markets must comply regardless of headquarters location.
Key Coordination Mechanisms
• US-EU Trade and Technology Council addresses AI governance alignment
• UK-EU divergence requires separate compliance tracks post-Brexit
• Asia-Pacific regional harmonisation remains limited
| Period | Milestone |
| January 2026 | California AB 316, SB 942, Texas TRAIGA, effective |
| February 2026 | The Colorado AI Act will be effective |
| August 2026 | General-purpose AI and foundation model rules will be active |
Organisations should take these actions now:
Conduct a comprehensive inventory of AI systems across all business units. Classify each system by risk level under applicable jurisdictions.
Establish an AI governance committee with legal, technical, and business representation. Assign clear accountability for compliance decisions.
Develop procedures for AI system failures, including reporting timelines (24 hours for foundation model incidents under EU rules).
Establish ongoing monitoring and periodic compliance reviews. Retain documentation for regulatory inspections.
Consider external AI law expertise when:
• Operating across multiple jurisdictions with conflicting requirements
• Deploying high-risk systems in regulated sectors
• Facing regulatory investigation or enforcement action
• Developing AI technologies raises novel legal questions
Article 27 Representative Requirements
Non-EU companies offering AI systems to European customers must designate an EU representative. This applies to:
• AI developers are placing systems on the EU market remotely
• Organisations deploying AI that produces outputs used within the EU
Available Support Services
• Outsourced Data Protection Officer services covering AI oversight obligations
• AI governance audits and readiness assessments
• Regulatory investigation response
• Conformity assessment preparation
AI regulation will intensify through 2025 and 2026. Organisations using AI technologies should begin compliance work immediately rather than waiting for enforcement actions to clarify requirements. The cost of preparation is significantly lower than the cost of violation.
Yes, if you’re a non-EU company placing AI systems on the European market, or your AI produces outputs affecting EU residents. The EU AI Act requires the designation of an authorised representative.
Penalties vary by jurisdiction. The EU AI Act imposes fines up to €35 million or 7% of global turnover. Texas TRAIGA authorises significant civil penalties through Attorney General enforcement. California laws carry penalties up to $10,000 per day for certain violations.
Review the specific categories in Annexe III of the EU AI Act. Generally, systems that make or influence consequential decisions about individuals, employment, credit, healthcare, and education qualify. When uncertain, seek legal assessment.