AI Stress Test

Geoff Ferres

The AI Stress Test explores new Enterprise AI research - including Frontier AI, Applied AI and Trusted AI developments - unpacking why it matters and what we can learn from historical parallels. **The AI Stress Test is currently on a brief hiatus while we reset the Podcast to an exciting new format - Looking forward to being back with you in March 2026!** #AI #EnterpriseAI #AIValue #FrontierAI #AppliedAI #TrustedAI #AIGovernance #AISafety #ResponsibleAI #AIStressTest #Learning #History #Technology #Innovation

الحلقات

  1. The new castles are built from rules, not stone

    ٥ فبراير

    The new castles are built from rules, not stone

    Nearly 1,000 years ago, medieval architects discovered something revolutionary: a single wall could be breached, but concentric castles - where each captured ring exposed attackers to lethal crossfire from the next - became nearly unconquerable. Siege warfare's success rate: extremely low. Cost to attackers: exponential. Fast forward to 2026: Anthropic has published Constitutional Classifiers++, demonstrating that the same principle works for AI safety. By layering classifiers, including a lightweight first-stage linear probe classifier and a second-stage ensemble of probe and external classifiers, they've built a system that reduced jailbreak success - with no red-teamer discovering a universal jailbreak capable of consistently extracting highly detailed answers to all eight target queries - while reducing computational cost by around 40x. The architectural parallel is strong: both systems weaponise depth. Medieval attackers facing nested walls encountered geometric escalation of complexity; modern attackers facing cascaded classifiers hit the same wall (literally). One breach no longer means defeat; it means exposure to multiple overlapping defensive layers. The crucial distinction? Medieval castles were static and artillery rendered them obsolete. Constitutional Classifiers have the potential to be much more dynamic, with the underlying ‘constitution’ (the ruleset) having flexibility to adapt over time as new attack patterns emerge. Medieval castles were unbreakable until the rules changed with artillery - will Constitutional Classifiers++ be unbreakable until the rules change again? Profiled research: Constitutional Classifiers - https://arxiv.org/pdf/2601.04603. #AI #EnterpriseAI #AIValue #FrontierAI #AppliedAI #TrustedAI #AIGovernance #AISafety #ResponsibleAI #AIStressTest #Learning #History #Technology #Innovation

    ٤٨ من الدقائق
  2. Why the FDA's food label revolution predicts AI's transparency future

    ٢٩ يناير

    Why the FDA's food label revolution predicts AI's transparency future

    Over 35 years ago, the US FDA began transforming fragmented nutrition disclosure into a single mandatory standard - a regulatory evolution that moved most FDA‑regulated packaged foods onto a single standardised Nutrition Facts label over the course of a few years. In leading edge research that parallels this labelling precedent, the AI Transparency Atlas has unmasked a critical AI transparency gap: even though many AI providers sit below 60% safety documentation compliance, most users remain blind to model behaviors, hallucinations and deception risks. This lack of AI labelling coincides with a focus on more transparent practice from a range of frontier labs - indicating that transparency may be becoming the new competitive currency of AI. Where the FDA standardised safety through analytical testing for chemical verification, AI players must standardise safety through better development norms and third-party review of context-dependent metrics … metrics that inherently resist uniform measurement. When can we expect the AI discussion will shift from ‘transparency roadmaps’ to FDA-style transparent labelling so everyone knows what they’re buying and using? AI Stress Test podcast links: https://podcasts.apple.com/us/podcast/ai-stress-test/id1849637428 https://open.spotify.com/show/03muUrgLytAxPdjYSwWNuH?si=f32295df101248d0 Profiled research link: AI Transparency Atlas - https://arxiv.org/pdf/2512.12443. #AI #EnterpriseAI #AIValue #FrontierAI #AppliedAI #TrustedAI #AIGovernance #AISafety #ResponsibleAI #AIStressTest #Learning #History #Technology #Innovation

    ٤٤ من الدقائق
  3. ٢٢ يناير

    Why corporate AI adoption in 2026 mirrors the PC's inevitable ascent

    Over 44 years ago, IBM did something remarkable - it turned legitimacy into a market strategy. A $4,000 computer package that had seemed absurd to most people became essential once IBM made it respectable. Five years later, the 'unnecessary' PC dominated boardrooms and homes alike. In leading-edge research, IEEE's Global Survey results capture an identical moment for agentic AI with 96% of global technologists agreeing that agentic AI innovation, exploration and adoption will continue at lightning speed and 59% of enterprises accelerating investment. The institutional credibility phase is underway. What happened next in the 1980s was inevitable adoption—and the adoption curve for AI is tracking the same trajectory. The essential difference here is that IBM's trial-and-error scaling created market learnings - competitors lifted, users adapted, markets corrected. Agentic AI's potential for black-box opacity, systemic scale and path-dependent lock-in may mean similar failures compound rather than correct and we may not have the luxury of learning through deployment. However, there’s a historical parallel we can't ignore - both technologies have progressed through near identical phases … technology leadership, organisational standardisation, consumer mass-market penetration. Both are driven by institutional validation that precedes consumer understanding. Yet we also face a paradox: if agentic AI’s adoption trajectory is historically supported and highly likely, why are we still negotiating between responsible deployment and competitive speed - rather than making rigorous safety validation the source of competitive differentiation? AI Stress Test Podcast links: https://podcasts.apple.com/us/podcast/ai-stress-test/id1849637428 https://open.spotify.com/show/03muUrgLytAxPdjYSwWNuH?si=f32295df101248d0 Profiled research: IEEE Global Survey - https://life.ieee.org/ieee-global-survey-the-impact-of-tech-in-2026/. #AI #EnterpriseAI #AIValue #FrontierAI #AppliedAI #TrustedAI #AIGovernance #AISafety #ResponsibleAI #AIStressTest #Learning #History #Technology #Innovation

    ٣٨ من الدقائق
  4. ١٦ يناير

    Build vs. buy, the cycle continues

    The 1960s-70s mainframe era established a pattern; enterprises rejected commercial software offerings, choosing instead to build custom applications in-house. The willingness to accept substantially higher costs and longer development timelines reflected a single calculus - strategic control over technology tethered to competitive advantage outweighed efficiency gains from standardised platforms. Fast forward to today. A systematic study of production AI agents (engaging 306 practitioners and 20 detailed case studies) documents that 85% of case study teams proceeded without third-party agent frameworks, building custom implementations from scratch. Human evaluation was relied on in 74% of cases and agent autonomy was constrained to fewer than 10 steps in 68% of cases. External industry forecasts have projected the agentic AI market will grow from around $5b to circa $200B by 2034 and this is likely fueled not by autonomous platforms, but by custom, human-supervised approaches. Both eras reveal that when organisations embed technology into competitive strategy, the build-vs-buy decision systematically favors building, despite higher costs. The precedent established in the mainframe era persists: control beats convenience when business continuity is at stake. What if the real market opportunity isn't selling AI platforms to enterprises, but selling the tools, infrastructure, and services that enable enterprises to build competitive AI systems themselves? Profiled research: Measuring Agents In Production - https://arxiv.org/pdf/2512.04123. #AI #EnterpriseAI #AIValue #FrontierAI #AppliedAI #TrustedAI #AIGovernance #AISafety #ResponsibleAI #AIStressTest #Learning #History #Technology #Innovation

    ٤١ من الدقائق
  5. ٨ يناير

    From fragmentation to dangerous consensus

    Between 1830 and 1886, American railways faced a coordination crisis: 23 independent gauge decisions created a fragmented network where the Southern Railway & Steamship Association's coordinated conversion of approximately 11,500 miles in May-June 1886 solved the integration problem through institutional coordination. In leading edge research from the University of Washington, frontier LLMs now exhibit 71-82% output homogeneity, potentially linked to RLHF (Reinforcement Learning from Human Feedback) alignment, suggesting that enterprises relying on multi-model decision-making inherit a coordination solution that has accidentally reversed itself - diversity in form, convergence in substance. Both episodes reveal how systems driven by local optimisation and local switching costs create paradoxical fragility: railroads needed standardisation to escape fragmentation, but AI needs the reverse - escape from the standardisation that alignment inadvertently engineered, risking epistemic monoculture in open-ended problem-solving contexts where diverse perspectives strengthen solutions. If railroads required crisis-driven coordination to reverse fragmentation, what institutional innovation can reverse AI's unintended convergence? Profiled research: The Artificial Hivemind - https://arxiv.org/pdf/2510.22954. #AI #EnterpriseAI #AIValue #FrontierAI #AppliedAI #TrustedAI #AIGovernance #AISafety #ResponsibleAI #AIStressTest #Learning #History #Technology #Innovation

    ٢٨ من الدقائق
  6. ١١‏/١٢‏/٢٠٢٥

    New classes, new skills

    Between the 1850s and the 1930s, capitalism experienced a permanent separation of ownership from control as professional managers assumed operational authority - and nearly simultaneously, enterprises invented systematic governance frameworks (standardised managerial accounting, formalised audits, standardised business education) to monitor and constrain these newly powerful delegates. The emerging agentic enterprise represents an analogous inflection point: MIT research identifies 2025 as a critical moment where AI orchestration transitions from siloed experimentation to integrated infrastructure, triggering parallel governance imperatives - policy-as-code frameworks, real-time monitoring systems, and new credentialisation pathways to manage AI agents as semi-autonomous coordinators operating within bounded governance parameters.​ What differs fundamentally is the type of agent being governed: managerial capitalism created frameworks to constrain ambitious human decision-makers with personal incentives and negotiation power; agentic systems require frameworks to constrain inference systems that lack human social context yet relentlessly optimise mathematical objectives, suggesting entirely different governance architectures may be required.​ When you transfer operational authority away from individuals who can be held personally accountable, what systematic guardrails must replace personal responsibility? Profiled research: Frontier AI use case developments: AI Agents for Customer Service – End-to-End Automation - https://www.comm100.com/blog/best-ai-chatbots-for-customer-service/; Gen-AI-Powered Network Digital Twins for Autonomous Operations - https://iowngf.org/wp-content/uploads/2025/02/IOWN-GF-RD-NDT_Use_Case-2.0.pdf; AI-Driven Smart Grids and Grid Optimisation - https://cleanenergyforum.yale.edu/2025/11/12/power-hungry-power-smart-can-ai-reduce-the-grid-strain-its-fueling. Applied AI developments: MLOps, Data Architecture, and Model Lifecycle Management - https://www.redhat.com/en/blog/enterprise-mlops-reference-design. Trusted AI developments: NIST AI Risk Management Framework 2025 Updates – Enhanced Guidance - https://www.ispartnersllc.com/blog/nist-ai-rmf-2025-updates-what-you-need-to-know-about-the-latest-framework-changes/; Australian AI Safety Institute (AISI) Establishment Announcement - https://www.minister.industry.gov.au/ministers/timayres/media-releases/establishment-australian-ai-safety-institute. Feature AI development: The Emerging Agentic Enterprise (AI‑Orchestrated Operations) - https://sloanreview.mit.edu/projects/the-emerging-agentic-enterprise-how-leaders-must-navigate-a-new-age-of-ai/. #AI #EnterpriseAI #AIValue #FrontierAI #AppliedAI #TrustedAI #AIGovernance #AISafety #ResponsibleAI #AIStressTest #Learning #History #Technology #Innovation

    ٣٠ من الدقائق
  7. ٠٤‏/١٢‏/٢٠٢٥

    When ‘safe until proven otherwise’ becomes dangerous

    Over a century ago, in 1906, the first documented death from asbestos exposure was recorded in testimony. Yet it would take until the 1920s for widespread medical evidence to emerge, and until 2006 - 100 years later - for meaningful regulation to gain momentum with the Rotterdam Convention. The culprit? A confidence gap: industry and institutions trusted the material's benefits whilst deprioritising the evidence of its harms. New enterprise AI research has identified a strikingly similar phenomenon: 78% of organisations claim full trust in AI systems, despite only 40% having implemented governance frameworks or ethical safeguards. Yet data shows that organisations prioritising trustworthy AI see significant ROI improvements. Asbestos created a latency trap: disease took 20-60 years to manifest, making the causal link between exposure and harm almost impossible to see in real time. AI presents a different latency trap—one where harms (bias, hallucination, systemic risk) accumulate invisibly across populations and organisational timescales, often undetectable within quarterly performance reviews. Is waiting for proof of harm before building governance a luxury we can no longer afford? Profiled research: Frontier AI use case developments: Next-Generation Models Suggest Manufacturing Strategies for an AI Agent Society - https://amiko.consulting/en/the-ai-revolution-in-the-second-week-of-november-2025-paradigm-shifts-and-new-opportunities-looming-for-manufacturing/; AI-Enabled Digital Twins for Public Health and Social Policy (Indonesia – Skyral) - https://edtechhub.org/2025/11/19/ai-observatory-waypoints-and-signals-issue-24/; Semantic Digital Twins for Industrial Energy Optimisation - https://www.forbes.com/sites/feliciajackson/2025/11/04/the-rise-of-industrial-ai-from-words-to-watts/. Applied AI developments: Microsoft Frontier Firms and Agentic AI in Core Functions - https://www.microsoft.com/en-us/microsoft-cloud/blog/2025/07/24/ai-powered-success-with-1000-stories-of-customer-transformation-and-innovation/. Trusted AI developments: Guidance for Risk Management of Artificial Intelligence Systems - https://www.edps.europa.eu/system/files/2025-11/2025-11-11_ai_risks_management_guidance_en.pdf. Feature AI development: IDC Trust-Action Gap Study - https://www.sas.com/content/dam/sasdam/documents/20250124/data-and-ai-impact-report-the-trust-imperative.pdf. #AI #EnterpriseAI #AIValue #FrontierAI #AppliedAI #TrustedAI #AIGovernance #AISafety #ResponsibleAI #AIStressTest #Learning #History #Technology #Innovation

    ٢٨ من الدقائق
  8. ٢٧‏/١١‏/٢٠٢٥

    Renaissance blueprints, AI control systems

    Over 540 years ago, following a plague that decimated up to half of Milan's population, Leonardo da Vinci sketched designs for an integrated ideal city where water management, sanitation, layered circulation and human flow were unified into a single coherent system - conceptualising the city itself as a technological solution to disease, congestion, and human suffering rather than a collection of isolated structures.​ Singapore's national digital twin - a federated capability integrating real-time sensor networks from 10,000+ distributed points, AI-powered anomaly detection, and cross-agency governance coordination - exemplifies how frontier AI transforms infrastructure from a static, episodically-maintained liability into a continuously intelligent system that predicts failures, aims to prevent catastrophic events and establishes a replicable governance model for AI-driven urban resilience at national scale.​ Whilst Leonardo's ideal city remained a hand-drawn speculative vision - a beautiful but unrealised intellectual exercise constrained by medieval technical limitations and cost - Singapore's digital twin is an operational, continuously updated, AI-mediated reality actively shaping real infrastructure decisions and public policy at scale, transforming systems thinking from Renaissance aspiration into automated, high-fidelity control with unprecedented capacity for both transformative benefit and systemic risk.​ Both visions promised rational mastery over urban chaos; have we learned the essential difference between systems thinking and systems control? Profiled research: Frontier AI use case developments: How Are AI Agents Used in Insurance Claims Processing - https://kanerika.com/blogs/ai-agents-for-insurance-claims-processing/; AI agents in finance: the agentic revolution - https://www.phacetlabs.com/blog/ai-agents-finance-agentic-revolution-intelligent-automation. Applied AI developments: MIT GenAI Divide Report: Shadow AI and True Integration - https://www.aigl.blog/state-of-ai-in-business-2025/; Organisational Scalability Bottlenecks and the 5Rs Framework - https://hbr.org/2025/11/most-ai-initiatives-fail-this-5-part-framework-can-help. Trusted AI developments: International AI Safety Report: Second Key Update - https://internationalaisafetyreport.org/publication/second-key-update-technical-safeguards-and-risk-management; MAS Singapore: Proposed Guidelines on AI Risk Management - https://www.mas.gov.sg/-/media/mas-media-library/publications/consultations/bd/2025/final_consultation_paper_on_guidelines_on_ai_risk_management_forrelease.pdf; Sony Fair Human-Centric Image Benchmark (FHIBE) for Computer Vision Fairness - https://www.nature.com/articles/s41586-025-09716-2. Feature AI development: Singapore National Digital Twin for City Infrastructure - https://aiagentstore.ai/ai-agent-news/topic/infrastructure-city/2025-11-18/detailed; https://www.aectechnicalsg.com/singapores-digital-twin-revolution/. #AI #EnterpriseAI #AIValue #FrontierAI #AppliedAI #TrustedAI #AIGovernance #AISafety #ResponsibleAI #AIStressTest #Learning #History #Technology #Innovation

    ٣٤ من الدقائق
  9. ٢٠‏/١١‏/٢٠٢٥

    The economics of asymmetric advantage have fundamentally changed

    Over 37 years ago, Robert Morris released a self-propagating worm that infected roughly 10% of the entire internet within hours - awakening the world to the reality that autonomous code could outpace human response. Today, new AI safety research has documented the first large-scale cyberattack where artificial intelligence orchestrated reconnaissance, exploitation, and data exfiltration across 30 global targets with 80–90% autonomy - rendering human-speed threat detection asymmetrically disadvantaged. Whilst Morris's 1988 accident revealed systemic fragility, the November 2025 Claude manipulation reveals something far more consequential: a deliberate manipulation of agentic AI capabilities that defence teams are challenged to contain. When defence requires matching human expertise against machine-speed operations, will we soon reach the point where human-led cybersecurity simply cannot scale? Profiled research: First AI-Orchestrated Cyberattack: Anthropic's Claude Exploitation (November 2025): https://www.anthropic.com/news/disrupting-AI-espionage; Five-Layer Framework for AI Governance: Integrating Regulation, Standards, and Certification: https://arxiv.org/abs/2509.11332 Constitutional AI: Aligning LLM Safety in 2025: https://sparkco.ai/blog/constitutional-ai-aligning-llm-safety-in-2025; #AI #AISafety #AISecurity #AISovereignty #AIGovernance #ResponsibleAI #TrustworthyAI #AIStressTest #Learning #History #Technology #Innovation

    ٣٢ من الدقائق
  10. ١٣‏/١١‏/٢٠٢٥

    When evolution stops being random

    Over 80 years ago, Alexander Fleming witnessed what he feared most - bacteria evolving resistance to penicillin within a decade of its mass introduction; yet, despite his public warnings, the cascade of resistance emerged precisely as foreseen - triggering an evolutionary arms race where successive antibiotic deployments accelerated rather than slowed resistance emergence, collapsing therapeutic horizons across bacterial species. In leading edge research released this month, Google's Threat Intelligence Group identified regenerative AI-native malware families employing just-in-time code regeneration, via LLMs, to rewrite their entire source code hourly to evade detection - rendering traditional incident responses fundamentally misaligned with threat velocity. Whilst biological pathogens evolve through random mutations constrained by metabolic costs and generational timescales, AI-powered malware achieves goal-directed adaptation orders of magnitude faster through intentional queries to LLMs - not merely another arms race escalation, but a fundamental phase transition toward autonomous threat design.​ Can we build governance frameworks to anticipate threats that design their own evolution in real time? Profiled research: AI-POWERED MALWARE: AUTONOMOUS ADAPTATION: https://cloud.google.com/blog/topics/threat-intelligence/threat-actor-usage-of-ai-tools; ADVANCED JAILBREAK TECHNIQUES: ECHO CHAMBER CONTEXT-POISONING: https://neuraltrust.ai/blog/echo-chamber-context-poisoning-jailbreak; AI AGENT SECURITY FRAMEWORKS: B3 BENCHMARK: https://securitybrief.com.au/story/open-source-b3-framework-to-benchmark-ai-agent-security-unveiled; DIFFERENTIAL PRIVACY IN ENTERPRISE AI: https://arxiv.org/pdf/2501.18914; MULTI-AGENT DEFENCE PIPELINE AGAINST PROMPT INJECTION: https://arxiv.org/abs/2509.14285 #AI #AISafety #AISecurity #AISovereignty #AIGovernance #ResponsibleAI #TrustworthyAI #AIStressTest #Learning #History #Technology #Innovation

    ٣٦ من الدقائق
  11. ٠٦‏/١١‏/٢٠٢٥

    Invisible sources of contamination

    Over 171 years ago, a contaminated water pump in Victorian London killed 616 people in a month because the poison was invisible to inspection, undetectable by the science of the era and absolutely trusted by those who consumed it. New leading edge research from the UK AISI, Anthropic and the Alan Turing Institute demonstrates that language models remain vulnerable to persistent backdoors inserted via minimal poisoned data, challenging the assumption that larger training datasets dilute poisoning effects. In addressing Cholera, the rapid emergence of symptoms provided an advantage, enabling epidemiological responses - in contrast, data poisoning can remain dormant, undetectable and potentially active over time. Both demand a safety response that focuses on source control as well as detection, yet many organisations approach AI security as a post-deployment challenge. Is your organisation asking the right questions about data provenance before crisis forces the conversation? Profiled research: Data Poisoning Attack Research: https://arxiv.org/abs/2510.07192; AI Red Teaming and Adaptive Attacks Against Defences: https://arxiv.org/abs/2510.09023; Control-Theoretic Approaches to AI Guardrails: https://arxiv.org/abs/2510.13727; EU AI Act Implementation Framework: https://digital-strategy.ec.europa.eu/en/policies/european-approach-artificial-intelligence #AI #AISafety #AISecurity #AISovereignty #AIGovernance #ResponsibleAI #TrustworthyAI #AIStressTest #Learning #History #Technology #Innovation

    ٣٢ من الدقائق
  12. ٣٠‏/١٠‏/٢٠٢٥

    From nuclear testing moratoriums to AI safety thresholds

    Over 33 years ago, the US conducted its final nuclear test - transitioning to a science-based Stockpile Stewardship Program that maintains civilization-ending arsenals through simulations alone, never testing them again.​​ New AI safety research has found frontier models achieving 70% performance on complex software engineering tasks and demonstrating potential for weaponised capabilities - prompting developers to deploy unprecedented safety controls not because they've definitively crossed danger thresholds, but because they cannot rule out crossing them.​​ Whilst nuclear warheads and AI systems are very different - one frozen-in-time physics, the other evolving from under 10% to 70% capability in under two years - we can draw interesting parallels with regard to governing powerful technologies through computational assessment rather than direct testing when the risks of empirical validation become unacceptable.​​ Will your organisation wait for evidence of actual harm before implementing enhanced AI controls or will you trigger safeguards when capability thresholds are reached? Profiled research: International AI Safety Report 2025: First Key Update: https://doi.org/10.48550/arXiv.2510.13653; Detecting and Reducing Scheming in AI Models: https://openai.com/index/detecting-and-reducing-scheming-in-ai-models/; LLM Jailbreak Detection for (Almost) Free: https://doi.org/10.48550/arXiv.2509.14558; InvThink: Towards AI Safety via Inverse Reasoning: https://doi.org/10.48550/arXiv.2510.01569; AI Red-Teaming Design: Threat Models and Tools: https://cset.georgetown.edu/article/ai-red-teaming-design-threat-models-and-tools/

    ٣٤ من الدقائق

حول

The AI Stress Test explores new Enterprise AI research - including Frontier AI, Applied AI and Trusted AI developments - unpacking why it matters and what we can learn from historical parallels. **The AI Stress Test is currently on a brief hiatus while we reset the Podcast to an exciting new format - Looking forward to being back with you in March 2026!** #AI #EnterpriseAI #AIValue #FrontierAI #AppliedAI #TrustedAI #AIGovernance #AISafety #ResponsibleAI #AIStressTest #Learning #History #Technology #Innovation