The New Stack Podcast

The New Stack

The New Stack Podcast is all about the developers, software engineers and operations people who build at-scale architectures that change the way we develop and deploy software. For more content from The New Stack, subscribe on YouTube at: https://www.youtube.com/c/TheNewStack

  1. 7 UUR GELEDEN

    Microsoft's answer to Kubernetes sprawl: Staged rollouts & cross-cluster workload mobility

    Managing Kubernetes at fleet scale introduces significant complexity, especially as organizations expand from a few clusters to hundreds or thousands across cloud, on-premises, and edge environments. While GitOps remains the dominant model for declarative management, its traditional one-to-one repository-to-cluster approach struggles to handle multi-cluster realities such as global traffic routing, shared secrets, and unified observability. AsStephane Erbrech, Principal Software Engineer at Microsoftexplains, the challenge shifts from deployment to governance—maintaining consistency, security, and compliance across a vast distributed system without manual intervention. This need is amplified by the rise of AI workloads at the edge, where inference is increasingly decentralized. To address these challenges,Microsoft Azure Kubernetes Fleet Managerenables coordinated, staged rollouts across clusters, allowing teams to validate updates in lower-risk environments before production. Supporting this,Cilium Cluster Meshprovides seamless cross-cluster connectivity, enabling workload mobility and efficient resource use, especially for scarce GPU capacity. Together, these tools help modern platform teams manage lifecycle, networking, and orchestration at scale.  Learn more from The New Stack around managing Kubernetes at fleet scale:  KubeFleet: The Future of Multicluster Kubernetes App Management Why Microsoft is betting on temporary identities to stop autonomous agents from going rogue Join our community of newsletter subscribers to stay on top of the news and at the top of your game.

    25 min.
  2. 1 DAG GELEDEN

    Why long-running AI agents break on HTTP and how Ably is fixing it

    In this episode ofThe New Stack Makers, Matthew O’Riordan, CEO of Ably, explains how infrastructure originally built for human collaboration is now well-suited for long-running AI agents. While Ably initially resisted positioning itself as an AI company, the rise of agents that reason, call tools, and operate over extended periods revealed a natural fit for its real-time communication platform. O’Riordan highlights the limitations of HTTP for these use cases. While effective for short, request-response interactions, HTTP struggles with persistent, stateful experiences—such as handling dropped connections, multi-device usage, or mid-task interruptions. To address this, a new “durable session” layer is emerging, enabling continuous synchronization between agents and users through shared state, presence, and recovery mechanisms. Ably’s solution, AI Transport, augments existing architectures by keeping HTTP for requests while shifting responses to durable sessions. Features like mutable message streams and “live objects” allow seamless reconnection and collaboration. The goal is to provide a drop-in layer that developers can adopt without rethinking their stack—moving beyond traditional pub/sub models. Learn more from The New Stack around Ably and AI Transport:  How MCP Uses Streamable HTTP for Real-Time AI Tool Interaction Ably Touts Real-Time Starter Kits for Vercel and Netlify AI Agents Need Help. Here’s 4 Ways To Ship Software Reliably Join our community of newsletter subscribers to stay on top of the news and at the top of your game.

    32 min.
  3. 1 DAG GELEDEN

    Why the Linux Foundation adopted MCP, with Jim Zemlin and Mazin Gilbert

    Agentic AI is advancing rapidly, with open-source projects racing to keep pace with real-world deployment. To accelerate progress, the Linux Foundation consolidated key technologies—Model Context Protocol (MCP), Goose, and AGENTS.md—under the newly formed Agentic AI Foundation (AAIF) in late 2025. At the MCP Dev Summit in New York City, Linux Foundation CEO Jim Zemlin and newly appointed AAIF executive director Mazin Gilbert discussed this transition. Zemlin explained that leading both organizations was unsustainable, prompting a careful search for a leader with both technical expertise and collaborative leadership skills. Gilbert now takes on the challenge of guiding AAIF as it shapes the emerging agentic AI ecosystem. While the foundation currently oversees three projects, its broader mission involves defining the future architecture of agent-driven systems—deciding what to build, when, and why. These decisions will influence the trajectory of open-source AI development. The conversation also highlights the importance of open collaboration, funding dynamics, and early adopters in shaping the agentic stack’s evolution.   Learn more from The New Stack around the latest in open-source projects and The Linux Foundation:  Anthropic Donates the MCP Protocol to the Agentic AI Foundation SAFE-MCP, a Community-Built Framework for AI Agent Security Google Donates the Agent2Agent Protocol to the Linux Foundation Join our community of newsletter subscribers to stay on top of the news and at the top of your game.

    33 min.
  4. 6 DGN GELEDEN

    Fresh data has us asking, does AI demand Kubernetes?

    Kubernetes is rapidly emerging as the de facto operating system for AI, with two-thirds of organizations using it for generative AI inference and 82% adopting it in production. Its ecosystem — including tools like Kubeflow — enables organizations to build, scale, and retain control of AI systems through open, community-driven infrastructure. Bob Killen of CNCF and Liam Bollmann-Dodd of SlashData shared insights from recent reports showing that AI success still hinges on strong engineering fundamentals—especially internal developer platforms and overall developer experience. While AI-generated code accelerates development, it shifts bottlenecks to DevOps, reliability, and security, increasing operational complexity. As a result, operator experience and well-defined guardrails have become critical to safely scaling AI. These controls help constrain both human and AI developers, reducing risk while enabling speed. At the same time, organizations are evolving team structures, expanding platform engineering groups to support internal users more effectively. Despite growing complexity, the core lesson remains consistent: open source innovation thrives on people, processes, and collaboration as much as on technology itself. Learn more from The New Stack around the latest in Kubernetes and its emergence as an operating system for AI:  Kubernetes and AI: Are They a Fit? How AI Is Pushing Kubernetes Storage Beyond Its Limits Kubernetes and AI Are Shaping the Next Generation of Platforms Join our community of newsletter subscribers to stay on top of the news and at the top of your game.

    23 min.
  5. 29 APR

    Cut AI token usage by 96%? Here’s how AWS Strands Agents does it.

    In this episode of The New Stack Makers, AWS developer advocate Morgan Willis demonstrates Strands Agents, an open source agentic framework with rapid adoption since its launch. Using a simple accounting API, she walks through three approaches to retrieving a customer’s latest invoice, highlighting how design choices dramatically impact efficiency. The initial method maps each API endpoint to a separate tool, requiring five chained calls and consuming about 52,000 tokens. By shifting to intent-based tools—focused on outcomes rather than individual data operations—the same task is completed in a single call using just 2,000 tokens, improving both efficiency and reasoning. In a third iteration, tools are hosted on a remote MCP server via AWS Agent Core Gateway, with semantic search limiting the agent’s toolset to only what’s relevant per query, further reducing token usage. Willis emphasizes that narrowly scoped agents outperform general-purpose ones, delivering better speed, accuracy, and context efficiency. Designing smaller, specialized agents with tailored tools is key as tool ecosystems expand. Learn more from The New Stack around the latest with Strands and MCP: AWS Launches Its Take on an Open Source AI Agents SDK What Is MCP? Game Changer or Just More Hype? MCP’s biggest growing pains for production use will soon be solved Join our community of newsletter subscribers to stay on top of the news and at the top of your game.

    28 min.
  6. 28 APR

    Why Broadcom is betting on a private cloud comeback

    Broadcom’s VMware Cloud Foundation (VCF) is evolving from a turnkey infrastructure stack into a modern application platform, balancing simplicity with the flexibility demanded by Kubernetes-driven environments. AtKubeCon + CloudNativeCon Europe 2026, Broadcom leaders highlighted how VCF is adapting to support platform engineering teams, cloud-native workloads, and large-scale operations. A key industry shift is the return to private cloud, driven by data sovereignty concerns and the growing impact of AI. Enterprises are bringing workloads back on-premises while still expecting a cloud-like operating model. Broadcom is responding by prioritizing on-prem stability and aligning closely with open source, reflecting its strong contributions toKubernetesand related projects. Kubernetes is no longer a bolt-on but the core control plane of VCF, enabling unified management of compute, storage, and networking through declarative APIs. At the same time, the distinction between virtual machines and containers is fading. The focus is shifting toward application-centric platforms, where developers interact through consistent abstractions, allowing infrastructure to be provisioned seamlessly behind the scenes. Learn more from The New Stack around the latest around Broadcom:  Broadcom ‘Doubles Down’ on Open Source, Donates Kubernetes Tool to CNCF Why Broadcom gave Velero to the CNCF Sandbox — and what it means for Kubernetes data protection Join our community of newsletter subscribers to stay on top of the news and at the top of your game.

    24 min.
  7. 25 APR

    Why Broadcom gave Velero to the CNCF Sandbox — and what it means for Kubernetes data protection

    Broadcom continues to expand its role as a major contributor to cloud-native open source, particularly within the Cloud Native Computing Foundation (CNCF) ecosystem. Its recent donation of Velero—originally developed by VMware—to the CNCF Sandbox reflects a strategic move to foster broader community trust and collaboration. By shifting governance away from vendor control, Broadcom aims to position Velero as a truly community-driven data protection standard for Kubernetes environments, encouraging wider adoption and contribution.  At the same time, the company is reinforcing its position as a full-stack Kubernetes provider across both cloud-native and private cloud environments. Despite Kubernetes’ dominance, many organizations still struggle with its complexity. Broadcom is addressing this by focusing on lifecycle management, long-term support, and deep integration with existing infrastructure like vSphere.  In a podcast recorded at KubeCon + CloudNativeCon Europe 2026, Dilpreet Bindra emphasized that open source success comes not just from code contributions, but also from relinquishing control to empower the broader ecosystem and drive sustainable innovation.  Learn more from The New Stack about the latest developments around Velero:  Broadcom donates Velero to CNCF — and it could reshape how Kubernetes users handle backup and disaster recovery  How AI Search Is Supporting Artistic Freedom  Join our community of newsletter subscribers to stay on top of the news and at the top of your game.

    23 min.

Info

The New Stack Podcast is all about the developers, software engineers and operations people who build at-scale architectures that change the way we develop and deploy software. For more content from The New Stack, subscribe on YouTube at: https://www.youtube.com/c/TheNewStack

Meer van The New Stack

Suggesties voor jou