Code & Cure

Vasanth Sarathy & Laura Hagopian

Decoding health in the age of AI Hosted by an AI researcher and a medical doctor, this podcast unpacks how artificial intelligence and emerging technologies are transforming how we understand, measure, and care for our bodies and minds. Each episode unpacks a real-world topic to ask not just what’s new, but what’s true—and what’s at stake as healthcare becomes increasingly data-driven. If you're curious about how health tech really works—and what it means for your body, your choices, and your future—this podcast is for you. We’re here to explore ideas—not to diagnose or treat. This podcast doesn’t provide medical advice.

  1. HÁ 7 H

    #31 - How Retrieval-Augmented AI Can Verify Clinical Summaries

    Fluent summaries that cannot prove their claims are a hidden liability in healthcare, quietly eroding clinician trust and wasting time. In this episode, we walk through a practical system that replaces “sounds right” narratives with evidence-backed summaries by pairing retrieval augmented generation with a large language model that serves as a judge. Instead of asking one AI to write and police itself, the work is divided. One model drafts the summary, while another breaks it into atomic claims, retrieves supporting chart excerpts, and issues clear verdicts of supported, not supported, or insufficient, with explanations clinicians can review. We explain why generic summarization often breaks down in clinical settings and how retrieval augmented generation keeps the model grounded in the patient’s actual record. The conversation digs into subtle but common failure modes, including when a model ignores retrieved evidence, when a sentence mixes correct and incorrect facts, and when wording implies causation that the record does not support. A concrete example brings this to life: a claim that a patient was intubated for septic shock is overturned by operative notes showing intubation for a procedure, with the system flagging the discrepancy and guiding a precise correction. That is not just higher accuracy; it is accountability you can audit later. We also explore a deeper layer of the problem: argumentation. Clinical care is not just a list of facts, but the relationships between them. By evaluating claims alongside their evidence, surfacing contradictions, and pushing for precise language, the system helps generate summaries that reflect real clinical reasoning rather than confident guessing. The payoff is less time spent chasing errors, more time with patients, and a defensible trail for quality review and compliance. If you care about chart review, clinical documentation, retrieval augmented generation, and building AI systems clinicians can trust, this episode offers practical takeaways.  Reference: Verifying Facts in Patient Care Documents Generated by Large Language Models Using Electronic Health Records Philip Chung et al.  NEJM AI (2025) Credits: Theme music: Nowhere Land, Kevin MacLeod (incompetech.com) Licensed under Creative Commons: By Attribution 4.0 https://creativecommons.org/licenses/by/4.0/

    24min
  2. 5 DE FEV.

    #30 - From Reddit To Rescue: Real-Time Signals Of The Opioid Crisis

    What if the earliest warning sign of an opioid overdose surge isn’t locked inside a delayed report, but unfolding in real time on Reddit? In this episode, we explore how social media conversations, especially pseudonymous, community-led forums, can reveal emerging overdose risks before traditional surveillance systems catch up. We unpack research that analyzed more than a decade of posts to show how even simple drug mentions sharpened forecasts of overdose death rates. The signal was especially strong for fentanyl, exposing where existing public health tools lag and why online communities often see danger first. Along the way, we explain the mechanics in plain language: how time-series models respond faster than surveys, why subreddit structure filters noise, and how historical archives enable rigorous validation. But it doesn’t stop at counting mentions. We dig into what happens when posts are classified by lived experience: overdose stories, sourcing concerns, or test strip discussions.  We also examine what broke during COVID, when behavior and access shifted overnight, and how to detect those regime changes before models start to fail. The takeaway is urgent and practical. Social data won’t replace public health surveillance, but it can make it fast enough to save lives. We share a field-ready playbook for turning online signals into timely interventions, and show how feedback from the same communities can explain why a response worked—or didn’t—so teams can adapt quickly. If you care about real-time epidemiology, harm reduction, and responsible AI in healthcare, this conversation connects raw text to real-world impact. Reference: Monitoring the opioid epidemic via social media discussions Delaney A Smith et al.  Nature NPJ Digital Health (2025) Credits: Theme music: Nowhere Land, Kevin MacLeod (incompetech.com) Licensed under Creative Commons: By Attribution 4.0 https://creativecommons.org/licenses/by/4.0/

    19min
  3. 29 DE JAN.

    #29 - AI Hype Meets Hospital Reality

    What really happens when a “smart” system steps into the operating room, and collides with the messy, time-pressured reality of clinical care? In this episode, we unpack a multi-center pilot that streamed audio and video from live surgeries to fuel safety checklists, flag cases for review, and promise rapid, actionable insight. What emerged instead was a clear-eyed lesson in the gap between aspiration and execution. Across four fault lines, the story shows where clinicians’ expectations of AI ran ahead of what today’s systems can reliably deliver, and what that means for patient safety. We begin with the promise. Surgeons and care teams envisioned near-instant post-case summaries: what went well, what raised concern, and which patients might be at risk. The reality looked different. Training demands, configuration work, and brittle workflows made it clear that AI is anything but plug-and-play. We explore why polished language can be mistaken for intelligence, why models need the right tools to reason effectively, and why moving AI from one hospital to another is closer to a redesign than a simple deployment. Then we follow the data. When it takes six to eight weeks to turn raw footage into usable insight, the value of learning forums like morbidity and mortality conferences quickly erodes. Privacy protections, de-identification, and quality control matter—but without pipelines built for speed and trust, insights arrive too late to change practice. We contrast where the system delivered real value, such as checklists and procedural signals, with where it fell short: predicting post-operative complications and producing research-ready datasets. Throughout the conversation, we argue for a minimum clinically viable product: tightly scoped use cases, early and deep involvement from surgeons and nurses, and data flows that respect governance without stalling learning. AI can strengthen patient safety and team performance—but only when expectations align with capability and operations are designed for real clinical tempo. If this resonates, follow the show, share it with a colleague, and leave a review with one takeaway you’d apply in your own clinical setting.  Reference: Expectations vs Reality of an Intraoperative Artificial Intelligence Intervention Melissa Thornton et al.  JAMA Surgery (2026) Credits: Theme music: Nowhere Land, Kevin MacLeod (incompetech.com) Licensed under Creative Commons: By Attribution 4.0 https://creativecommons.org/licenses/by/4.0/

    26min
  4. 22 DE JAN.

    #28 - How AI Confidence Masks Medical Uncertainty

    Can you trust a confident answer, especially when your health is on the line? This episode explores the uneasy relationship between language fluency and medical truth in the age of large language models (LLMs). New research asks these models to rate their own certainty, but the results reveal a troubling mismatch: high confidence doesn’t always mean high accuracy, and in some cases, the least reliable models sound the most sure. Drawing on her ER experience, Laura illustrates how real clinical care embraces uncertainty—listening, testing, adjusting. Meanwhile, Vasanth breaks down how LLMs generate their fluent responses by predicting the next word, and why their self-reported “confidence” is just more language, not actual evidence. We contrast AI use in medicine with more structured domains like programming, where feedback is immediate and unambiguous. In healthcare, missing data, patient preferences, and shifting guidelines mean there's rarely a single “right” answer. That’s why fluency can mislead, and why understanding what a model doesn’t know may matter just as much as what it claims. If you're navigating AI in healthcare, this episode will sharpen your eye for nuance and help you build stronger safeguards.  Reference:  Benchmarking the Confidence of Large Language Models in Answering Clinical Questions: Cross-Sectional Evaluation Study Mahmud Omar et al. JMIR (2025) Credits:  Theme music: Nowhere Land, Kevin MacLeod (incompetech.com) Licensed under Creative Commons: By Attribution 4.0 https://creativecommons.org/licenses/by/4.0/

    26min
  5. 15 DE JAN.

    #27 - Sleep’s Hidden Forecast

    What if one night in a sleep lab could offer a glimpse into your long-term health? Researchers are now using a foundation model trained on hundreds of thousands of hours of sleep data to do just that, by predicting the next five seconds of a polysomnogram, the model learns the rhythms of sleep and, with minimal fine-tuning, begins estimating risks for conditions like Parkinson’s, dementia, heart failure, stroke, and even some cancers. We break down how it works: during a sleep study, sensors capture brain waves (EEG), eye movements (EOG), muscle tone (EMG), heart rhythms (ECG), and breathing. The model compresses these multimodal signals into a reusable format, much like how language models process text. Add a small neural network, and suddenly those sleep signals can help predict disease risk up to six years out. The associations make clinical sense: EEG patterns are more telling for neurodegeneration, respiratory signals flag pulmonary issues, and cardiac rhythms hint at circulatory problems. But, the scale of what’s possible from a single night’s data is remarkable. We also tackle the practical and ethical questions. Since sleep lab patients aren’t always representative of the general population, we explore issues of selection bias, fairness, and external validation. Could this model eventually work with consumer wearables that capture less data but do so every night? And what should patients be told when risk estimates are uncertain or only partially actionable? If you're interested in sleep science, AI in healthcare, or the delicate balance of early detection and patient anxiety, this episode offers a thoughtful look at what the future might hold—and the trade-offs we’ll face along the way. Reference:  A multimodal sleep foundation model for disease prediction Rahul Thapa Nature (2026) Credits:  Theme music: Nowhere Land, Kevin MacLeod (incompetech.com) Licensed under Creative Commons: By Attribution 4.0 https://creativecommons.org/licenses/by/4.0/

    24min
  6. 8 DE JAN.

    #26 - How Your Phone Keyboard Signals Your State Of Mind

    What if your keyboard could reveal your mental health? Emerging research suggests that how you type—not what you type—could signal early signs of depression. By analyzing keystroke patterns like speed, timing, pauses, and autocorrect use, researchers are exploring digital biomarkers that might quietly reflect changes in mood. In this episode, we break down how this passive tracking compares to traditional screening tools like the PHQ. While questionnaires offer valuable insight, they rely on memory and reflect isolated moments. In contrast, continuous keystroke monitoring captures real-world behaviors—faster typing, more pauses, shorter sessions, and increased autocorrect usage—all patterns linked to mood shifts, especially when anxiety overlaps with depression. We discuss the practical questions this raises: How do we account for personal baselines and confounding factors like time of day or age? What’s the difference between correlation and causation? And how can we design systems that protect privacy while still offering clinical value? From privacy-preserving on-device processing to broader behavioral signals like sleep and movement, this conversation explores how digital phenotyping might help detect depression earlier—and more gently. If you're curious about AI in healthcare, behavioral science, or the ethics of digital mental health tools, this episode lays out both the potential and the caution needed. Reference:  Effects of mood and aging on keystroke dynamics metadata and their diurnal patterns in a large open-science sample: A BiAffect iOS study Claudia Vesel et al. J Am Med Inform Assoc (2020) Credits:  Theme music: Nowhere Land, Kevin MacLeod (incompetech.com) Licensed under Creative Commons: By Attribution 4.0 https://creativecommons.org/licenses/by/4.0/

    20min
  7. 1 DE JAN.

    #25 - When Safety Slips: Prompt Injection in Healthcare AI

    What happens when a chatbot follows the wrong voice in the room? In this episode, we explore the hidden vulnerabilities of prompt injection, where malicious instructions and fake signals can mislead even the most advanced AI into offering harmful medical advice. We unpack a recent study that simulated real patient conversations, subtly injecting cues that steered the AI to make dangerous recommendations—including prescribing thalidomide for pregnancy nausea, a catastrophic lapse in medical judgment. Why does this happen? Because language models aim to be helpful within their given context, not necessarily to prioritize authoritative or safe advice. When a browser plug-in, a tainted PDF, or a retrieved web page contains hidden instructions, those can become the model’s new directive, undermining guardrails and safety layers. From direct “ignore previous instructions” overrides to obfuscated cues in code or emotionally charged context nudges, we map the many forms of this attack surface. We contrast these prompt injections with hallucinations, examine how alignment and preference training can unintentionally amplify risks, and highlight why current defenses, like content filters or system prompts, often fall short in clinical use. Then, we get practical. For AI developers: establish strict instruction boundaries, sanitize external inputs, enforce least-privilege access to tools, and prioritize adversarial testing in medical settings. For clinicians and patients: treat AI as a research companion, insist on credible sources, and always confirm drug advice with licensed professionals. AI in healthcare doesn’t need to be flawless, but it must be trustworthy. If you’re invested in digital health safety, this episode offers a clear-eyed look at where things can go wrong and how to build stronger, safer systems. If you found it valuable, follow the show, share it with a colleague, and leave a quick review to help others discover it. Reference:  Vulnerability of Large Language Models to Prompt Injection When Providing Medical Advice Ro Woon Lee JAMA Open Health Informatics (2025) Credits:  Theme music: Nowhere Land, Kevin MacLeod (incompetech.com) Licensed under Creative Commons: By Attribution 4.0 https://creativecommons.org/licenses/by/4.0/

    25min
  8. 25/12/2025

    #24 - What Else Is Hiding In Medical Images?

    What if a routine mammogram could do more than screen for breast cancer? What if that same image could quietly reveal a woman’s future risk of heart disease—without extra tests, appointments, or burden on patients? In this episode, we explore a large-scale study that uses deep learning to uncover cardiovascular risk hidden inside standard breast imaging. By analyzing mammograms that millions of women already receive, researchers show how a single scan can deliver a powerful second insight for women’s health. Laura brings the clinical perspective, unpacking how cardiovascular risk actually shows up in practice—from atypical symptoms to prevention decisions—while Vasanth walks us through the AI system that makes this dual-purpose screening possible. We begin with the basics: how traditional cardiovascular risk tools like PREVENT work, what data they depend on, and why—despite their proven value—they’re often underused in real-world care. From there, we turn to the mammogram itself. Features such as breast arterial calcifications and subtle tissue patterns have long been linked to vascular disease, but this approach goes further. Instead of focusing on a handful of predefined markers, the model learns from the entire image combined with age, identifying patterns that humans might never think to look for. Under the hood is a survival modeling framework designed for clinical reality, where not every patient experiences an event during follow-up, yet every data point still matters. The takeaway is striking: the imaging-based risk score performs on par with established clinical tools. That means clinicians could flag cardiovascular risk during a test patients are already getting—opening the door to earlier conversations about blood pressure, cholesterol, diabetes, and lifestyle changes. We also zoom out to the bigger picture. If mammograms can double as heart-risk detectors, what other routine tests are carrying untapped signals? Retinal images, chest CTs, pathology slides—each may hold clues far beyond their original purpose. With careful validation and attention to bias, this kind of opportunistic screening could expand access to prevention and shift care further upstream. If this episode got you thinking, share it with a colleague, subscribe for more conversations at the intersection of AI and medicine, and leave a review telling us which everyday medical test you think deserves a second life. Reference:  Predicting cardiovascular events from routine mammograms using machine learning Jennifer Yvonne Barraclough Heart (2025) Credits:  Theme music: Nowhere Land, Kevin MacLeod (incompetech.com) Licensed under Creative Commons: By Attribution 4.0 https://creativecommons.org/licenses/by/4.0/

    24min
5
de 5
5 avaliações

Sobre

Decoding health in the age of AI Hosted by an AI researcher and a medical doctor, this podcast unpacks how artificial intelligence and emerging technologies are transforming how we understand, measure, and care for our bodies and minds. Each episode unpacks a real-world topic to ask not just what’s new, but what’s true—and what’s at stake as healthcare becomes increasingly data-driven. If you're curious about how health tech really works—and what it means for your body, your choices, and your future—this podcast is for you. We’re here to explore ideas—not to diagnose or treat. This podcast doesn’t provide medical advice.