Common Misunderstandings

Provider vs Deployer: Which AI Act Role Are You?

The AI Act assigns different obligations based on your role. Understanding whether you're a "provider" or "deployer" determines what compliance steps you must take.

AI Act Provider vs Deployer Roles

Provider vs Deployer: Which AI Act Role Are You?

One of the most fundamental questions for AI Act compliance is determining your role in the AI value chain. The obligations you face depend entirely on whether you're classified as a "provider," "deployer," or another role. Getting this wrong means missing critical compliance requirements.

Understanding the Roles

Provider Definition

A provider is an organization that:

  • Develops an AI system
  • Has an AI system developed and places it on the market or puts it into service under its own name or trademark
  • Makes substantial modifications to an AI system that changes its intended purpose

Key Point: Providers bear the primary responsibility for ensuring AI systems meet technical requirements, especially for high-risk systems.

Deployer Definition

A deployer is an organization that:

  • Uses an AI system under its own authority
  • Does not place the system on the market or put it into service for others
  • Uses the AI system for its intended purpose within its own operations

Key Point: Deployers must use AI responsibly according to instructions, monitor its performance, and ensure appropriate human oversight.

Why This Distinction Matters

The provider/deployer distinction determines:

  1. Compliance burden: Providers face more extensive technical obligations
  2. Liability exposure: Different responsibilities mean different risk profiles
  3. Documentation requirements: Providers create documentation; deployers maintain it
  4. Resource needs: Provider obligations typically require more technical expertise

Common Scenarios and Role Classification

Scenario 1: SaaS AI Tool for Internal Use

Situation: Your company subscribes to an AI-powered customer service platform.

Your Role: Deployer

  • You're using a third-party system as intended
  • The SaaS vendor is the provider
  • You must follow their instructions and monitor performance

Your Obligations:

  • Use the system according to provider instructions
  • Conduct impact assessments if processing personal data
  • Ensure human oversight where required
  • Monitor for adverse outcomes
  • Report serious incidents to the provider

Scenario 2: Custom AI Development for Your Business

Situation: You hire developers to build a custom AI system exclusively for your company's operations.

Your Role: Provider

  • You commissioned the development
  • The system operates under your responsibility
  • You control its design and deployment

Your Obligations:

  • Meet all provider requirements for your system's risk category
  • Conduct conformity assessments if high-risk
  • Maintain technical documentation
  • Ensure data quality and governance
  • Implement risk management systems

Scenario 3: AI-Powered Product Sold to Customers

Situation: Your company sells software with embedded AI functionality to end customers.

Your Role: Provider (for the product you sell)

Also: Deployer (if you use third-party AI tools internally)

Your Obligations:

  • Full provider responsibilities for your product's AI
  • Must ensure CE marking if high-risk
  • Provide instructions and documentation to customers
  • Separate deployer obligations for any AI you use internally

Scenario 4: White-Label AI Solution

Situation: You resell an AI system under your own brand name without modifying it.

Your Role: Provider

  • Placing it on the market under your name makes you the provider
  • Original developer may still have obligations, but you assume provider role
  • This applies even if you don't modify the technology

Your Obligations:

  • Full provider compliance responsibilities
  • Ensure the underlying system meets requirements
  • Maintain all required documentation
  • May need contractual protection from original developer

Scenario 5: Modifying Third-Party AI

Situation: You significantly modify a third-party AI system's intended purpose.

Your Role: Can become Provider

  • Substantial modifications that change intended purpose make you a provider
  • Minor configurations or customizations don't trigger provider status
  • The line between "modification" and "use" can be unclear

Your Obligations:

  • If substantial modification: full provider obligations
  • If minor customization: remain a deployer

Detailed Provider Obligations

For All AI Systems:

  1. Quality Management: Implement systems for compliance monitoring
  2. Documentation: Maintain technical documentation demonstrating compliance
  3. Record-Keeping: Keep logs of system operation (for high-risk systems)
  4. Transparency: Provide clear information to deployers and users
  5. Post-Market Monitoring: Track performance after deployment

Additional Requirements for High-Risk Systems:

  1. Risk Management: Continuous identification and mitigation of risks
  2. Data Governance: Ensure training data quality and relevance
  3. Technical Documentation: Comprehensive description of system design and operation
  4. Automatic Logging: Design systems to log events automatically
  5. Human Oversight: Enable meaningful human supervision
  6. Accuracy & Robustness: Ensure appropriate performance levels
  7. Cybersecurity: Implement resilience against attempts to alter use
  8. Conformity Assessment: Undergo assessment before market placement
  9. CE Marking: Affix marking to compliant high-risk systems
  10. Registration: Register high-risk systems in EU database

Detailed Deployer Obligations

For All AI Systems:

  1. Follow Instructions: Use system according to provider's directions
  2. Input Data: Ensure input data is relevant for intended purpose
  3. Monitor: Oversee AI system operation based on instructions
  4. Record-Keeping: Keep logs as specified by provider (high-risk systems)

Additional Requirements for High-Risk Systems:

  1. Human Oversight: Assign qualified individuals to supervise the system
  2. Impact Assessment: Conduct fundamental rights impact assessment before use
  3. Incident Reporting: Report serious incidents to provider and authorities
  4. Data Protection: Comply with GDPR, including DPIAs where required
  5. Suspend/Discontinue: Stop using system if non-conformity or risk identified
  6. Cooperation: Work with authorities on compliance matters

Deployer-Specific High-Risk Categories:

Deployers face additional obligations when using AI for:

  • Employment decisions (hiring, promotion, termination)
  • Access to education and vocational training
  • Access to essential private services and public assistance
  • Evaluating creditworthiness
  • Assessing risk in life/health insurance pricing

The Gray Areas: When Classification Is Unclear

Integration and Customization

Adding AI as a component in a larger system can create ambiguity:

  • Simply integrating third-party AI: Usually a deployer
  • Extensively customizing AI behavior: May become a provider
  • Creating new functionality: Likely a provider

Open-Source AI Models

Using open-source models raises questions:

  • Downloading and using as-is: Typically a deployer
  • Fine-tuning for specific purposes: Gray area, depends on extent
  • Distributing modified version: Likely a provider

AI-as-a-Service with Customization

Cloud AI services with significant customization options:

  • Using standard API: Deployer
  • Extensive model training/customization: May shift toward provider
  • Threshold depends on how much you control system behavior

How to Determine Your Role: Decision Framework

Step 1: Ask Core Questions

  1. Did we develop this AI system or have it developed for us?
  2. Do we place this system on the market for others?
  3. Do we use this under our own authority for our operations?
  4. Did we substantially modify an existing system?

Step 2: Apply the Tests

  • Development Test: If you controlled development → Provider
  • Market Test: If you sell/distribute to others → Provider
  • Authority Test: If you use for own operations only → Deployer
  • Modification Test: If you changed intended purpose → Provider

Step 3: Document Your Analysis

Record your reasoning because:

  • Authorities may question your classification
  • It guides your compliance approach
  • It clarifies vendor relationships
  • It helps allocate resources

Multiple Roles Simultaneously

Many organizations will be both providers and deployers:

Example: A company that:

  • Develops an AI-powered SaaS product (Provider for this)
  • Uses third-party AI for internal HR (Deployer for this)
  • Purchases AI analytics tools (Deployer for these)

Each use case requires separate role analysis and corresponding compliance measures.

Practical Implications for Compliance

If You're a Provider:

  • Budget significantly for compliance, especially if offering high-risk systems
  • Invest in technical expertise for conformity assessments
  • Expect longer timelines before market launch
  • Plan for ongoing obligations even after system deployment
  • Consider insurance for provider liability

If You're a Deployer:

  • Vet your AI vendors carefully – their compliance affects you
  • Maintain operational documentation of how you use systems
  • Implement monitoring processes for AI performance
  • Train personnel on proper AI use and oversight
  • Prepare impact assessments before high-risk deployments

Contractual Considerations

Provider-Deployer Contracts Should Address:

  1. Role Clarity: Explicit statement of who is provider vs. deployer
  2. Compliance Responsibilities: Clear allocation of obligations
  3. Information Sharing: Provider's duty to supply documentation
  4. Incident Procedures: How serious incidents are reported and handled
  5. Modification Rights: What changes deployer can make without becoming provider
  6. Liability Allocation: Who bears risk for compliance failures
  7. Audit Rights: Deployer's ability to verify provider compliance

Common Mistakes to Avoid

Mistake 1: Assuming "Just Using AI" Means No Obligations

Even deployers have significant responsibilities, especially for high-risk systems.

Mistake 2: Not Recognizing Provider Status

Organizations that customize AI extensively may inadvertently become providers with much greater obligations.

Mistake 3: Relying Solely on Vendor Claims

A vendor saying "we handle compliance" doesn't eliminate your obligations as a deployer.

Mistake 4: Ignoring Role Shifts

Your role can change over time – regular reassessment is necessary.

Key Takeaways

  1. Classification is fundamental: Everything else in your compliance program depends on correctly identifying your role.

  2. Most organizations are deployers: Using third-party AI tools makes you a deployer, which still carries obligations.

  3. Providers face heavier burdens: Development and market placement trigger extensive technical requirements.

  4. You can be both: Many organizations have different roles for different AI systems.

  5. Document your analysis: Record why you classified each system as you did.

  6. Review regularly: Roles can shift as systems evolve or uses change.

  7. Contracts matter: Clear agreements with vendors protect everyone and clarify responsibilities.

Understanding your role isn't just a compliance checkbox – it's the foundation of your entire AI governance approach. Get this right, and everything else falls into place more easily.

Ready to Take the Next Step?

Get the comprehensive guide or generate a customized AI policy for your organization.

Download Free Guide

118 pages + templates

Get the comprehensive EU AI Act compliance guide with actionable steps, risk frameworks, and ready-to-use templates.

Generate AI Policy

Customized for you

Create a professional, customized AI usage policy tailored to your organization's needs in minutes.

Both resources are designed specifically for mid-sized EU companies navigating AI governance.