M365 Show with Mirko Peters - Microsoft 365 Digital Workplace Daily

Advanced Copilot Agent Governance with Microsoft Purview

Opening – Hook + Teaching Promise

You’re leaking data through Copilot Studio right now, and you don’t even know it.Every time one of your bright, shiny new Copilot Agents runs, it inherits your permissions—every SharePoint library, every Outlook mailbox, every Dataverse table. It rummages through corporate data like an overeager intern who found the master key card. And unlike that intern, it doesn’t get tired or forget where the confidential folders are.

That’s the part too many teams miss: Copilot Studio gives you power automation wrapped in charm, but under the hood, it behaves precisely like you. If your profile can see finance data, your chatbot can see finance data. If you can punch through a restricted connector, so can every conversation your coworkers start with “Hey Copilot.” The result? A quiet but consistent leak of context—those accidental overshares hidden inside otherwise innocent answers.

By the end of this podcast, you’ll know exactly how to stop that. You’ll understand how to apply real Data Loss Prevention (DLP) policies to Copilot Studio so your agents stop slurping up whatever they please.We’ll dissect why this happens, how Power Platform’s layered DLP enforcement actually works, and what Microsoft’s consent model means when your AI assistant suddenly decides it’s an archivist.

And yes, there’s one DLP rule that ninety percent of admins forget—the one that truly seals the gap. It isn’t hidden in a secret portal, it’s sitting in plain sight, quietly ignored. Let’s just say that after today, your agents will act less like unsupervised interns and more like disciplined employees who understand the word confidential.

Section 1: The Hidden Problem – Agents That Know Too Much

Here’s the uncomfortable truth: every Copilot Agent you publish behaves as an extension of the user who invokes it. Not a separate account. Not a managed identity unless you make it one. It borrows your token, impersonates your rights, and goes shopping in your data estate. It’s convenient—until someone asks about Q2 bonuses and the agent obligingly quotes from the finance plan.

Copilot Studio links connectors with evangelical enthusiasm. Outlook? Sure. SharePoint? Absolutely. Dataverse? Why not. Each connector seems harmless in isolation—just another doorway. Together, they form an entire complex of hallways with no security guard. The metaphor everyone loves is “digital intern”: energetic, fast, and utterly unsupervised. One minute it’s fetching customer details, the next it’s volunteering the full sales ledger to a chat window.

Here’s where competent organizations trip. They assume policy inheritance covers everything: if a user has DLP boundaries, surely their agents respect them. Unfortunately, that assumption dies at the boundary between the tenant and the Power Platform environment. Agents exist between those layers—too privileged for tenant restrictions, too autonomous for simple app policies. They occupy the gray space Microsoft engineers politely call “service context.” Translation: loophole.

Picture this disaster class scenario. A marketing coordinator connects the agent to Excel Online for campaign data, adds Dataverse for CRM insights, then saves without reviewing the connector classification. The DLP policy in that environment treats Excel as Business and Dataverse as Non‑Business. The moment someone chats, data crosses from one side to the other, and your compliance officer’s blood pressure spikes. Congratulations—your Copilot just built a makeshift export pipeline.

The paradox deepens because most admins configure DLP reactively. They notice trouble only after strange audit alerts appear or a curious manager asks, “Why is Copilot quoting private Teams posts?” By then the event logs show legitimate user tokens, meaning your so‑called leak looks exactly like proper usage. Nothing technically broke; it simply followed rules too loosely written.

This is why Microsoft keeps repeating that Copilot Studio doesn’t create new identities—it extends existing ones. So when you wonder who accessed that sensitive table, the answer may be depressing: you did, or at least your delegated shadow did. If your Copilot can see finance data, so can every curious chatbot session your employees open, because it doesn’t need to authenticate twice. It already sits inside your trusted session like a polite hitchhiker with full keychain access.

What most teams need to internalize is that “AI governance” isn’t just a fancy compliance bullet. It’s a survival layer. Permissions without containment lead to what auditors politely call “context inference.” That’s when a model doesn’t expose a file but paraphrases its contents from cache. Try explaining that to regulators.

Now, before you panic and start ripping out connectors, understand the goal isn’t to eliminate integration—it’s to shape it. DLP exists precisely to draw those bright lines: what counts as Business, what belongs in quarantine, what never touches network A if it speaks to network B. Done correctly, Copilot Studio becomes powerful and predictable. Done naively, it’s the world’s most enthusiastic leaker wrapped in a friendly chat interface.

So yes, the hidden problem isn’t malevolence; it’s inheritance. Your agents know too much because you granted them omniscience by design. The good news is that omniscience can be filtered. But to design the filter, you need to know how the data actually travels—through connectors, through logs, through analytic stores that never made it into your compliance diagram.

So, let’s dissect how data really moves inside your environment before we patch the leak—because until you understand the route, every DLP rule you write is just guesswork wrapped in false confidence.

Section 2: How Data Flows Through Copilot Studio

Let’s trace the route of one innocent‑looking question through Copilot Studio. A user types, “Show me our latest sales pipeline.” That request doesn’t travel in a straight line. It starts at the client interface—web, Teams, or embedded app—then passes through the Power Platform connector linked to a service like Dataverse. Dataverse checks the user’s token, retrieves the data, and delivers results back to the agent runtime. The runtime wraps those results into text and logs portions of the conversation for analytics. By the time the answer appears on‑screen, pieces of it have touched four different services and at least two separate audit systems.

That hopscotch path is the first vulnerability. Each junction—user token, connector, runtime, analytics—is a potential exfiltration point. When you grant a connector access, you’re not only allowing data retrieval. You’re creating a transit corridor where temporary cache, conversation snippets, and telemetry coexist. Those fragments may include sensitive values even when your output seems scrubbed. That’s why understanding the flow beats blindly trusting the UI’s cheerful checkboxes.

Now, connectors themselves come in varieties: Standard, Premium, and Custom. Standard connectors—SharePoint, Outlook, OneDrive—sit inside Microsoft’s managed envelope. Premium ones bridge into higher‑value systems like SQL Server or Salesforce. Custom connectors are the real wild cards; they can point anywhere an API and an access token exist. DLP treats each tier differently. A policy may forbid combining Custom with Business connectors, yet admins often test prototypes in mixed environments “just once.” Spoiler: “just once” quickly becomes “in production.”

Even connectors that feel safe—Excel Online, for instance—can betray you when paired with dynamic output. Suppose your agent queries an Excel sheet storing regional revenue, summarizes it, and pushes the result into a chat where context persists. The summarized numbers might later mingle with different data sources in analytics. The spreadsheet itself never left your tenant, but the meaning extracted from it did. That’s information leakage by inference, not by download.

Add another wrinkle: Microsoft’s defaults are scoped per environment, not across the tenant. Each Power Platform environment—Development, Test, Production—carries its own DLP configuration unless you deliberately replicate the policy. So when you say, “We already have a tenant‑wide DLP,” what you really have is a polite illusion. Unless you manually enforce the same classification each time a new environment spins up, your shiny Copilot in the sandbox might still pipe confidential records straight into a Non‑Business connector. Think of it as identical twins who share DNA but not discipline.

And environments multiply. Teams love spawning new ones for pilots, hackathons, or region‑specific bots. Every time they do, Microsoft helpfully clones permissions but not necessarily DLP boundaries. That’s why governance by memo—“Please remember to secure your environment”—fails. Data protection needs automation, not trust.

Let me illustrate with a story that’s become folklore in cautious IT circles. A global enterprise built a Copilot agent for customer support, proudly boasting an airtight app‑level policy. They assumed the DLP tied to that app extended to all sub‑components. When compliance later reviewed logs, they discovered the agent had been cross‑referencing CRM details stored in an unmanaged environment. The culprit? The DLP lived at the app layer; the agent executed at environment scope. The legal team used words not suitable for slides.

The truth is predictable yet ignored: DLP boundaries form at the connector‑e