You already know the meme: chatbots talk, agents act, multi-agent systems actually get stuff done. If you’ve ever begged a bot to fix Intune and got a poem instead, this one’s for you. In this episode, we go full Netflix hands-on: you watch, you snack, I poke the dangerous Service Principal things so nobody nukes prod. We build a mini self-healing, governed multi-agent system using Azure AI Foundry + Semantic Kernel, wired into real enterprise surfaces: IntuneEntra IDMicrosoft GraphAzure AutomationLog AnalyticsWe run one-agent vs multi-agent head-to-head on a real workflow: 12 minutes vs 3 minutes time-to-fix — with only my subscription credit on the line. You’ll see why one agent stalls while teams fly, and how to ship this pattern safely in your own tenant. 🔥 What You’ll Learn 1. Why a Single Agent Isn’t Enough in the Enterprise We start by tearing apart the “one giant agent” fantasy: Single agents are like gas-station Swiss Army knives: technically they have tools, practically they bend on the first real job.You stuff planning, reasoning, execution, approvals, and reporting into one prompt → context explodes, latency spikes, hallucinations creep in.One agent trying to:Plan a changeCall Graph and IntuneWrite remediation scriptsRequest approvalsVerify resultsDocument everything…is basically a help desk, change board, and postmortem crammed into one very tired intern. We break down what actually goes wrong: Context windows flooded with logs, policies, and MDM miscellanyImportant details get truncated or inventedToken usage and costs balloon“Fix” attempts that quietly break other things (like deleting the resource instead of rotating a secret 😬)Then we introduce the fix: Multi-agent = roles + boundaries + parallelism Planner focuses on intent & constraintsOperator focuses on tools & executionReviewer focuses on guardrails & approvalsEach agent gets a tight instruction set, minimal memory, and a focused toolset, passing around small structured messages, not a 50-page policy doc. 2. Multi-Agent Systems 101 (No Hype, Just The Pattern) We map out a clear, shippable mental model: think digital team, not one big brain. Roles: Planner — understands the goal, constraints, environment; outputs a stepwise plan with tool callsOperator — executes the plan via tools: Graph, Azure Automation, Functions, Logic Apps, etc.Reviewer — checks groundedness, scope, compliance, and safety before risky changesMessenger/Concierge — interacts with humans: approvals, status updates, and audit summariesCore concepts: Tools = handsREST APIs (Graph, internal services)Azure Automation runbooks (device scripts, remediation)Azure Functions & Logic Apps (glue & approvals)RAG via Azure AI Search (curated knowledge, not random web junk)Memory = budget, not magicMinimize per-agent contextUse external state (Search, state store, thread metadata)Only pass what’s needed for the next decisionPlanning vs ExecutionPlanner decomposes → Operator calls tools → Reviewer checks → Messenger tells humansThis is where Semantic Kernel shines: planners, skills, function catalogs, retries, cancellationSafety by designManaged Identities per agentRBAC split into read vs managePIM for destructive operationsTool calls logged to Log AnalyticsContent Safety + prompt shields to block jailbreaks & indirect injection3. How Azure AI Foundry Powers Multi-Agent Workflows We then show how Azure AI Foundry becomes the control room: You’ll see how to define agents with: Instructions — short, role-specific promptsDeployments — different models per role (GPT-4-class for planning, SLMs for extraction)Knowledge — Azure AI Search indexes, uploaded docs, optional web groundingActions — OpenAPI tools, Graph, Logic Apps, Functions, Azure Automation, Code InterpreterConnected agents — yes, one agent can call another like a toolWhy this matters: Foundry handles threads, safety, tracing, and evaluationsSemantic Kernel orchestrates the planner → operator → reviewer loop in codeYou keep prompts short and put power in tools with strict schemasModel strategy: Reasoning models for planning and complex decisionsSmall models (SLMs) for extraction, classification, parameter shapingMix serverless endpoints and managed compute depending on cost & residency needsSafety & observability: Content Safety on inputs and outputsPrompt shields against jailbreak and indirect injectionFull tracing of tool calls (who, what, where, how long)Application Insights + Log Analytics for performance & auditBuilt-in evaluation flows for groundedness, relevance, and fluency Become a supporter of this podcast: https://www.spreaker.com/podcast/m365-show-podcast--6704921/support. Follow us on: LInkedIn Substack