How to Deploy Microsoft Purview DLP for Copilot and Generative AI

A Diagram Illustrating The Four Primary Ai Data Ingestion Paths—Type, Paste, Attach, And In-Document Use—And How They Map Across Two Protection Layers: Inside Microsoft 365 (Internal Ai) And Outside Microsoft 365 (External Web Ai).

Organisations adopting Microsoft 365 Copilot and generative AI are facing a common challenge. Data is no longer static. It is constantly moving, reused, and now being surfaced and generated by AI.

This changes the risk profile.

The questions I hear most often are:

  1. What are the real data security risks with AI?
  2. Which controls actually make a difference?
  3. How do we deploy them effectively in Microsoft Purview?

This guide builds on my ECS 2026 session and focuses on the controls that provide the greatest practical value. It cuts through the noise and shows how to use Microsoft Purview DLP for Copilot and generative AI, with a practical approach to reducing real-world data leakage risk across both internal and external AI tools.

What is Microsoft Purview DLP for Copilot?

Microsoft Purview Data Loss Prevention (DLP) helps organisations identify, monitor, and protect sensitive data across Microsoft 365.

In the context of Copilot and generative AI, DLP provides controls to:

  • Prevent sensitive data from being entered into prompts
  • Restrict AI access to sensitive content
  • Block data from being shared with external AI services

The Shift: AI Is Outpacing Security

Data security has always been a moving target, but generative AI has fundamentally altered the speed at which data can move, surface, and be leaked.

According to Microsoft’s 2026 Data Security Index report AI adoption is accelerating faster than traditional data security controls can adapt. Enterprise data is no longer static; it continuously flows across endpoints, hybrid environments, collaboration platforms, and AI prompt boxes.

In practice, this creates three distinct challenges for security teams:

  • No single view of where sensitive data exists.
  • Inconsistent controls across fragmented platforms.
  • Limited visibility into how AI tools interact with corporate data.

The Real Risk with AI Data Security

Generative AI primarily amplifies existing data governance risks rather than creating entirely new ones. These include:

  • Oversharing caused by legacy, excessive permissions
  • Sensitive data leakage and insider risk
  • Uncontrolled external sharing
  • AI-generated outputs containing confidential information

What changes with AI is the scale and speed. Within seconds, sensitive information can be typed into prompts, pasted into external web tools, or automatically surfaced by internal AI systems.


The Four AI Data Paths (and How to Protect Them)

A Diagram Illustrating The Four Primary Ai Data Ingestion Paths—Type, Paste, Attach, And In-Document Use—And How They Map Across Two Protection Layers: Inside Microsoft 365 (Internal Ai) And Outside Microsoft 365 (External Web Ai).

One of the simplest ways to understand AI data protection is to focus on how data enters AI systems. There are four primary paths that define most of your AI data exposure surface:

  1. Type: Entering text directly into prompt boxes
  2. Paste: Copying corporate data from documents and pasting it into AI tools.
  3. Attach: Uploading files, spreadsheets, or images directly into an AI session.
  4. In-document use: Utilising native AI features embedded directly within applications (e.g., Word, Excel, PowerPoint).

If you do not protect all four paths, critical security gaps will remain.

Mapping Your Defence: The Two Protection Layers

To effectively block data leakage across these four paths, your Microsoft Purview strategy must be deployed in two distinct, overlapping layers. Most organisations focus heavily on one layer while completely neglecting the other. Effective protection requires both.

Layer 1: Inside Microsoft 365 (Internal AI)

This layer governs what internal AI tools, specifically Microsoft 365 Copilot and Copilot chat, can interact with. Your controls must regulate what these services are allowed to:

  • Access within your tenant
  • Process during a user session
  • Generate in response to a prompt
  • Retrieve from organisational knowledge bases

Layer 2: Outside Microsoft 365 (External AI)

This layer governs what data your users can send to external, web-based generative AI platforms. Your controls must monitor and restrict interactions with external services such as:

  • ChatGPT
  • Gemini
  • Claude
  • DeepSeek
  • Other unmanaged web-based AI tools

By aligning the four ingestion paths with these two protection layers, you ensure comprehensive visibility and control over your entire corporate AI footprint.


The Hidden Risk: Oversharing

One of the biggest misconceptions around Microsoft 365 Copilot is that Copilot itself creates exposure.

In reality, Copilot respects existing Microsoft 365 permissions.

The real problem is that many environments already contain significant oversharing.

Common examples include:

  • Legacy SharePoint sites with excessive access
  • “Everyone except external users” permissions
  • Unused Microsoft Teams containing sensitive content
  • Poorly governed SharePoint permissions inheritance
  • Broad access groups accumulated over years

Before Copilot, much of this content remained effectively hidden because users had to manually search for it.

Copilot changes this dynamic completely.

AI dramatically reduces the effort required to discover, summarise, and reuse information across Microsoft 365.

This means overshared data becomes far easier to expose at scale.

In most environments I assess, this is one of the main causes of Microsoft 365 data oversharing and AI-related data leakage.


Why Classification Matters for AI Security

Before deploying DLP for AI, organisations need a reliable data classification foundation.

Without classification:

  • DLP policies become inconsistent
  • Sensitive data may be missed entirely
  • False positives increase
  • User frustration grows
  • Security teams lose confidence in enforcement

Sensitivity labels and Sensitive Information Types (SITs) are foundational to effective AI governance.

If your organisation cannot reliably identify sensitive data, your AI controls will not operate effectively.

Classification is not optional.


The Six Microsoft Purview DLP Controls That Matter Most for Copilot and Generative AI

Rather than deploying every available control, focus first on the capabilities that address the most common real-world leakage scenarios.

The six controls below provide a strong minimum viable protection model for AI.


Control 1: Block Sensitive Prompts in Copilot

This is the most important Copilot DLP control for preventing data leakage in prompts.

Licensing: Supported across Microsoft 365 Copilot DLP-supported licensing
Prerequisites: Sensitive Information Types configured

Purpose

Prevent users from entering sensitive information into Copilot prompts.

This is one of the most common AI data leakage paths.

Microsoft Purview Dlp Policy For Copilot Prompt ProtectionDlp Location: Microsoft 365 Copilot And Copilot Chat
Dlp Scope: All Users &Amp; Groups (You Can Start With A Pilot Group)
Microsoft Purview Dlp Policy For Copilot Prompt ProtectionDlp Condition: Contains One Or More Sits
Dlp Action: Restrict Copilot From Processing Content → Processing Prompts = Blocked

Configuration

Configure a DLP policy targeting:

  • Microsoft 365 Copilot
  • Copilot Chat

Use conditions such as:

  • Content contains one or more SITs

Configure the action:

  • Processing prompts = Block

Operational Considerations

Prompt processing and web grounding are separate controls and must be configured as separate rules.

This control is particularly effective for preventing exposure of:

  • Personally identifiable information (PII)
  • Financial data
  • Credentials
  • Regulated business data

Recommended Deployment Approach

  • Start in audit mode
  • Pilot with a limited user group
  • Tune SIT detection carefully
  • Review Activity Explorer data before broad enforcement
  • Communicate clearly with users to reduce confusion

Quick Start

Use the built-in Microsoft Purview recommendation:

“Safeguard sensitive data in Microsoft 365 Copilot interactions”

This provides one of the fastest ways to implement initial protection.

Start With The “Safeguard Sensitive Data In Microsoft 365 Copilot Interactions” One-Click Policy In Dlp → Recommendations To Deploy Prompt Protection In Minutes.

Control 2: Block Web Grounding for Sensitive Prompts

Licensing: Microsoft 365 E5 / Preview

Purpose

Prevent Copilot from using sensitive prompts in Bing-powered web grounding.

Web grounding allows Copilot to enhance responses using web search context.

Although Microsoft does not send the entire prompt externally, Copilot may derive summarised search queries from user input. Sensitive data may still be reflected in that process.

Microsoft Purview Dlp Policy To Control Copilot Web Grounding.Dlp Condition: Contains One Or More Sits
Dlp Action: Restrict Copilot From Processing Content → Performing Web Searches = Blocked

Configuration

Condition:

  • Content contains one or more SITs

Action:

  • Performing web searches = Block

Operational Considerations

  • Only the Block action is currently supported
  • This must be deployed as a separate rule
  • Start with high-risk SITs before broad expansion

Recommended Deployment Approach

This control works best when paired with prompt protection policies.

It allows organisations to:

  • Retain Copilot functionality internally
  • Reduce external exposure risks
  • Apply more targeted governance controls

Control 3: Block Copilot Access to Labelled Content

Licensing: Microsoft 365 E5 / Preview
Prerequisites: Sensitivity labels deployed

Purpose

Prevent Copilot from accessing highly sensitive labelled files and emails.

This creates a hard governance boundary around protected content.

Microsoft Purview Dlp Policy To Control Copilot Access To Content With Purview Sensitivity Labels.Dlp Condition: Contains One Or More Sensitivity Labels
Dlp Action: Restrict Copilot From Processing Content → Accessing Knowledge Sources

Configuration

Condition:

  • Content contains sensitivity labels

Action:

  • Accessing knowledge sources = Block

Operational Considerations

This capability is highly dependent on label maturity and coverage.

Many organisations underestimate how incomplete their labelling strategy is until they begin AI deployments.

A practical approach is to:

  • Start with highly confidential labels
  • Create dedicated “Exclude from Copilot” labels if required
  • Expand incrementally

Recommended Deployment Approach

This is one of the cleanest and most scalable ways to restrict AI access to sensitive business information.

If you can label the content reliably, you can govern Copilot access consistently.

Quick Start

Use the Purview DLP recommendation:

“Protect items from Microsoft 365 Copilot and agent processing”

Start With The “Protect Items From Microsoft 365 Copilot And Agent Processing” One-Click Policy In Dlp → Recommendations To Prevent Copilot From Accessing Labelled Files And Emails.

Control 4: Block Paste and Upload to GenAI Websites

This is a critical Endpoint DLP control for generative AI websites such as ChatGPT.

Licensing: Microsoft 365 E5
Prerequisites: Endpoint DLP onboarding

Purpose

Prevent users from pasting or uploading sensitive data into external generative AI services.

In many environments, this is where the first AI-related data leakage occurs.

Endpoint Dlp Configuration For Blocking Generative Ai WebsitesDlp Location: Devices
Dlp Scope: All Users, Groups, Devices, Device Groups (You Can Start With A Pilot Group)
Endpoint Dlp Configuration For Blocking Generative Ai WebsitesDlp Action: Audit Or Restrict On Devices
-Upload To A Restricted Cloud Services Domain Or Access From An Unallowed Browsers → Sensitive Service Domain Restrictions = Generative Ai Websites  → Audit/Block
- Paste To Supported Browsers → → Sensitive Service Domain Restrictions = Generative Ai Websites → Audit/Block
Endpoint Dlp Configuration For Blocking Generative Ai WebsitesSensitive Service Domain Restrictions = Generative Ai Websites

Configuration

Configure Endpoint DLP policies targeting:

  • Devices

Use conditions such as:

  • Content contains one or more SITs

Actions: Audit or restrict actions on devices

  • Upload to a restricted cloud domain or access from unallowed browsers
  • Paste to supported browsers

Apply to the Generative AI websites service domain category

Operational Considerations

Browser coverage and enforcement capabilities vary depending on:

  • Edge for Business
  • Supported browsers
  • Defender integration

Organisations should validate behaviour carefully during pilot phases.

Recommended Deployment Approach

  • Start with audit mode
  • Use pilot groups
  • Tune SIT logic aggressively
  • Monitor user behaviour patterns before enforcement

If you only secure Copilot but ignore external AI tools, users can bypass internal controls very easily.

👉 In practice, this is where most organisations experience their first AI-related data leak.


Control 5: Block Upload of Labelled Files

Licensing: Microsoft 365 E5
Prerequisites: Sensitivity labels and Endpoint DLP onboarding

Purpose

Prevent labelled files from being uploaded to external AI platforms.

Labels often capture business intent more accurately than content inspection alone.

Endpoint Dlp Configuration For Blocking Generative Ai Websites Accessing Content With Purview Sensitivity Labels.Dlp Condition: Contains One Or More Sensitivity Labels
Dlp Action: Audit Or R
- Upload To A Restricted Cloud Services Domain Or Access From An Unallowed Browsers → Sensitive Service Domain Restrictions = Generative Ai Websites  → Audit/Block
- Paste To Supported Browsers → → Sensitive Service Domain Restrictions = Generative Ai Websites → Audit/Block

Configuration

Configure Endpoint DLP policies using:

Condition:

  • Content has a sensitivity label

Apply restrictions for:

  • Upload to restricted cloud service domains
  • Access from unallowed browsers
  • Paste into supported browsers

Use the:

  • Generative AI websites service domain category

Actions:

  • Audit
  • Block

Operational Considerations

This control complements SIT-based detection extremely well.

It is particularly valuable for:

  • Board documents
  • Legal material
  • Financial reports
  • M&A data
  • Strategic content

Recommended Deployment Approach

  • Combine with SIT-based policies
  • Use clear business-aligned label naming
  • Align labels directly to risk classifications
  • Roll out incrementally

Labels protect scenarios where traditional pattern matching may fail.

Control 6: Inline Data Protection in Edge

Licensing: Microsoft 365 E5 + Microsoft Purview PAYG
Prerequisites: Edge for Business and policy enforcement

Purpose

Provide real-time browser-level DLP enforcement for AI interactions.

Most generative AI usage occurs in the browser, making browser enforcement increasingly important.

Endpoint Dlp Configuration For Blocking Generative Ai Websites Accessing Content In A Managed  Edge Browser
Endpoint Dlp Configuration For Blocking Generative Ai Websites Accessing Content In A Managed  Edge Browser

Configuration

Deploy:

  • Edge for Business
  • Microsoft Purview integration
  • Inline web traffic policies

Configure:

  • DLP Policy Type = Inline web traffic
  • Enforcement location = Edge for Business

Enable:

  • PAYG billing

Operational Considerations

This capability provides some of the closest enforcement to real user AI behaviour.

However, organisations should:

  • Monitor PAYG consumption carefully
  • Validate browser enforcement scope
  • Ensure Edge adoption strategies are realistic

Recommended Deployment Approach

Use this as a layered control alongside Endpoint DLP rather than as a standalone solution.

If AI usage primarily occurs in the browser, browser-level controls become strategically important.


Putting It All Together

The controls naturally divide into two protection layers.

Protecting Microsoft 365 Copilot

  • Control 1: Block sensitive prompts
  • Control 2: Block web grounding
  • Control 3: Restrict access to labelled content

Protecting External Generative AI Usage

  • Control 4: Block paste and upload
  • Control 5: Block labelled file uploads
  • Control 6: Inline browser protection

This layered approach ensures Microsoft Purview DLP controls cover both Copilot and generative AI data leakage scenarios.

👉 Most organisations implement Copilot controls first and miss the endpoint layer. That is where the real gap appears.


A Practical Deployment Strategy

Avoid deploying everything at once.

A phased deployment model works far better:

Phase 1: Assess

  • Identify oversharing
  • Review permissions
  • Assess label maturity
  • Identify high-risk data locations

Phase 2: Pilot

  • Deploy audit policies
  • Use limited user groups
  • Measure false positives
  • Validate operational impact

Phase 3: Tune

  • Refine SIT logic
  • Improve label coverage
  • Adjust policy scope
  • Reduce noise

Phase 4: Enforce

  • Adapt policies continuously
  • Gradually expand enforcement
  • Monitor user behaviour
  • Review incidents regularly

Monitoring and Optimisation

DLP is not a set-and-forget technology.

Successful AI governance requires continuous monitoring and refinement.

Key capabilities include:

  • DLP reporting
  • Activity Explorer
  • Alerts
  • Data Security Posture Management (DSPM)
  • Security Copilot investigations
  • Custom reporting and analytics

These capabilities help organisations:

  • Identify oversharing
  • Understand AI interaction patterns
  • Prioritise remediation
  • Tune controls over time

If oversharing is not addressed, Copilot will surface it.


Licensing and Preview Considerations

Before deployment, organisations should validate:

  • Which features are Generally Available (GA)
  • Which remain in Preview
  • Microsoft 365 E5 dependencies
  • Endpoint DLP prerequisites
  • PAYG billing requirements
  • Browser enforcement limitations

Preview features may change before general availability and should be tested carefully before large-scale production deployment.


Final Thoughts

AI security is ultimately a data governance challenge.

Microsoft 365 Copilot does not introduce new risks. It exposes what already exists. If data is overshared, poorly classified, or loosely governed, AI will make that visible at speed and at scale.

The organisations that succeed with Copilot and generative AI are not the ones deploying the most controls. They are the ones that:

  • Understand where their data is stored
  • Reduce oversharing across Microsoft 365
  • Classify information in a consistent and meaningful way
  • Apply layered protection across internal and external AI
  • Continuously monitor and refine their controls

Microsoft Purview DLP can provide effective protection for Copilot and generative AI. However, this only works when it is built on strong data governance foundations.

👉 Effective AI security starts with effective data governance.


Frequently Asked Questions

What is Microsoft Purview DLP for Copilot?

Microsoft Purview Data Loss Prevention (DLP) helps organisations prevent sensitive data from being exposed through Microsoft 365 Copilot.
It enables you to:
Block sensitive data in prompts
Restrict access to sensitive content
Control how data is used in AI responses

Can Microsoft 365 Copilot leak sensitive data?

Yes, but not because Copilot bypasses security.
Copilot respects existing Microsoft 365 permissions. If sensitive data is already accessible, Copilot can surface and summarise it.
👉 The real risk is existing oversharing, not Copilot itself.

What is the biggest risk when deploying Copilot?

The biggest risk is oversharing.
Most environments already contain excessive permissions, legacy SharePoint access, and broadly shared content. Copilot makes that data easier to find and use.
👉 In practice, oversharing is the single largest risk before Copilot is even enabled.

How do you prevent data leakage in Microsoft 365 Copilot?

You need layered controls:
1 Block sensitive prompts using DLP
2 Restrict access to labelled content
3 Review and remediate oversharing
👉 Start with prompt protection, then expand to classification and governance.

How do you stop users sending data to ChatGPT or other AI tools?

Use Endpoint DLP and browser controls to restrict data leaving managed devices.
Key controls include:
1 Blocking paste and upload to generative AI websites
2 Restricting uploads based on sensitivity labels
3 Applying browser-level protection
👉 Without these controls, users can bypass Copilot protections.

Do you need both sensitivity labels and SITs?

Yes.
Sensitive Information Types detect patterns such as PII and financial data
Sensitivity labels reflect business context
👉 You need both for effective and reliable coverage.

Should DLP policies be deployed in block mode immediately?

No.
Best practice is to:
1 Start in audit mode
2 Use pilot groups
3 Tune policies
4 Move to block gradually
👉 Deploying too quickly often leads to false positives and user frustration.

Where should I start with Microsoft Purview DLP?

👉 Block sensitive prompts in Copilot
Then:
Start with Control 1:
👉 Block sensitive prompts in Copilot
Then:
1 Validate behaviour
2 Improve classification
3 Expand to Endpoint DLP
4 Add browser-level controls
👉 You can reduce risk quickly without overengineering your environment.


Need Help?

If you are planning a Microsoft 365 Copilot rollout or broader AI governance programme, now is the time to establish the right foundations.

I help organisations to:

  • Identify oversharing risks
  • Design Microsoft Purview governance strategies
  • Implement practical AI data protection controls
  • Improve Microsoft 365 security and compliance maturity

If you want to turn AI governance strategy into an operational deployment, feel free to get in touch.


References

View the Slides


Nikki Chapple Biography

I’m a Principal Cloud Architect at CloudWay and a dual Microsoft MVP in Microsoft 365 and Security. I help enterprises secure data and govern AI across Microsoft 365, specialising in Microsoft Purview, Copilot readiness, and compliance strategy.

🎙️ All Things M365 Compliance – I co-host this video and podcast show with Ryan Murphy, covering Microsoft 365 security, governance, and compliance. 📺 Watch on YouTube · 🎧 Listen on Spotify

💼 Connect: https://www.linkedin.com/in/nikkichapple/ · nikkichapple.com

🔐 Preparing for Copilot? Worried about oversharing or data leaks? I help enterprises secure Microsoft 365 and get their data governance foundations right before enabling AI. Get in touch →


Nikki Chapple Biography

I’m a Principal Cloud Architect at CloudWay and a dual Microsoft MVP in Microsoft 365 and Security. I help enterprises secure data and govern AI across Microsoft 365, specialising in Microsoft Purview, Copilot readiness, and compliance strategy.

🎙️ All Things M365 Compliance – I co-host this video and podcast show with Ryan Murphy, covering Microsoft 365 security, governance, and compliance. 📺 Watch on YouTube · 🎧 Listen on Spotify

💼 Connect: https://www.linkedin.com/in/nikkichapple/ · nikkichapple.com

🔐 Preparing for Copilot? Worried about oversharing or data leaks? I help enterprises secure Microsoft 365 and get their data governance foundations right before enabling AI. Get in touch →

Keep Reading

PreviousNext