694 episodes

Machine learning and artificial intelligence are dramatically changing the way businesses operate and people live. The TWIML AI Podcast brings the top minds and ideas from the world of ML and AI to a broad and influential community of ML/AI researchers, data scientists, engineers and tech-savvy business and IT leaders. Hosted by Sam Charrington, a sought after industry analyst, speaker, commentator and thought leader. Technologies covered include machine learning, artificial intelligence, deep learning, natural language processing, neural networks, analytics, computer science, data science and more.

The TWIML AI Podcast (formerly This Week in Machine Learning & Artificial Intelligence‪)‬ Sam Charrington

    • Technology
    • 4.7 • 394 Ratings

Machine learning and artificial intelligence are dramatically changing the way businesses operate and people live. The TWIML AI Podcast brings the top minds and ideas from the world of ML and AI to a broad and influential community of ML/AI researchers, data scientists, engineers and tech-savvy business and IT leaders. Hosted by Sam Charrington, a sought after industry analyst, speaker, commentator and thought leader. Technologies covered include machine learning, artificial intelligence, deep learning, natural language processing, neural networks, analytics, computer science, data science and more.

    Video as a Universal Interface for AI Reasoning with Sherry Yang

    Video as a Universal Interface for AI Reasoning with Sherry Yang

    Today we’re joined by Sherry Yang, senior research scientist at Google DeepMind and a PhD student at UC Berkeley. In this interview, we discuss her new paper, "Video as the New Language for Real-World Decision Making,” which explores how generative video models can play a role similar to language models as a way to solve tasks in the real world. Sherry draws the analogy between natural language as a unified representation of information and text prediction as a common task interface and demonstrates how video as a medium and generative video as a task exhibit similar properties. This formulation enables video generation models to play a variety of real-world roles as planners, agents, compute engines, and environment simulators. Finally, We explore UniSim, an interactive demo of Sherry's work and a preview of her vision for interacting with AI-generated environments.

    The complete show notes for this episode can be found at twimlai.com/go/676.

    • 49 min
    Assessing the Risks of Open AI Models with Sayash Kapoor

    Assessing the Risks of Open AI Models with Sayash Kapoor

    Today we’re joined by Sayash Kapoor, a Ph.D. student in the Department of Computer Science at Princeton University. Sayash walks us through his paper: "On the Societal Impact of Open Foundation Models.” We dig into the controversy around AI safety, the risks and benefits of releasing open model weights, and how we can establish common ground for assessing the threats posed by AI. We discuss the application of the framework presented in the paper to specific risks, such as the biosecurity risk of open LLMs, as well as the growing problem of "Non Consensual Intimate Imagery" using open diffusion models.

    The complete show notes for this episode can be found at twimlai.com/go/675.

    • 40 min
    OLMo: Everything You Need to Train an Open Source LLM with Akshita Bhagia

    OLMo: Everything You Need to Train an Open Source LLM with Akshita Bhagia

    Today we’re joined by Akshita Bhagia, a senior research engineer at the Allen Institute for AI. Akshita joins us to discuss OLMo, a new open source language model with 7 billion and 1 billion variants, but with a key difference compared to similar models offered by Meta, Mistral, and others. Namely, the fact that AI2 has also published the dataset and key tools used to train the model. In our chat with Akshita, we dig into the OLMo models and the various projects falling under the OLMo umbrella, including Dolma, an open three-trillion-token corpus for language model pretraining, and Paloma, a benchmark and tooling for evaluating language model performance across a variety of domains.

    The complete show notes for this episode can be found at twimlai.com/go/674.

    • 32 min
    Training Data Locality and Chain-of-Thought Reasoning in LLMs with Ben Prystawski

    Training Data Locality and Chain-of-Thought Reasoning in LLMs with Ben Prystawski

    Today we’re joined by Ben Prystawski, a PhD student in the Department of Psychology at Stanford University working at the intersection of cognitive science and machine learning. Our conversation centers on Ben’s recent paper, “Why think step by step? Reasoning emerges from the locality of experience,” which he recently presented at NeurIPS 2023. In this conversation, we start out exploring basic questions about LLM reasoning, including whether it exists, how we can define it, and how techniques like chain-of-thought reasoning appear to strengthen it. We then dig into the details of Ben’s paper, which aims to understand why thinking step-by-step is effective and demonstrates that local structure is the key property of LLM training data that enables it.

    The complete show notes for this episode can be found at twimlai.com/go/673.

    • 25 min
    Reasoning Over Complex Documents with DocLLM with Armineh Nourbakhsh

    Reasoning Over Complex Documents with DocLLM with Armineh Nourbakhsh

    Today we're joined by Armineh Nourbakhsh of JP Morgan AI Research to discuss the development and capabilities of DocLLM, a layout-aware large language model for multimodal document understanding. Armineh provides a historical overview of the challenges of document AI and an introduction to the DocLLM model. Armineh explains how this model, distinct from both traditional LLMs and document AI models, incorporates both textual semantics and spatial layout in processing enterprise documents like reports and complex contracts. We dig into her team’s approach to training DocLLM, their choice of a generative model as opposed to an encoder-based approach, the datasets they used to build the model, their approach to incorporating layout information, and the various ways they evaluated the model’s performance.

    The complete show notes for this episode can be found at twimlai.com/go/672.

    • 45 min
    Are Emergent Behaviors in LLMs an Illusion? with Sanmi Koyejo

    Are Emergent Behaviors in LLMs an Illusion? with Sanmi Koyejo

    Today we’re joined by Sanmi Koyejo, assistant professor at Stanford University, to continue our NeurIPS 2024 series. In our conversation, Sanmi discusses his two recent award-winning papers. First, we dive into his paper, “Are Emergent Abilities of Large Language Models a Mirage?”. We discuss the different ways LLMs are evaluated and the excitement surrounding their“emergent abilities” such as the ability to perform arithmetic Sanmi describes how evaluating model performance using nonlinear metrics can lead to the illusion that the model is rapidly gaining new capabilities, whereas linear metrics show smooth improvement as expected, casting doubt on the significance of emergence. We continue on to his next paper, “DecodingTrust: A Comprehensive Assessment of Trustworthiness in GPT Models,” discussing the methodology it describes for evaluating concerns such as the toxicity, privacy, fairness, and robustness of LLMs.

    The complete show notes for this episode can be found at twimlai.com/go/671.

    • 1 hr 5 min

Customer Reviews

4.7 out of 5
394 Ratings

394 Ratings

Ashish US ,

The best aiml podcast

I am following this podcast since 2017, my AI learning direction comes this podcast. Thankyou Sam !

mfnyai14 ,

Best pod for ai/ml practioners

Consistently great content - this is the podcast that actual ai/ml practitioners

atxfitgal ,

Clear and educational

I’m a total noob to these topics but an otherwise smart person, and this podcast strikes a perfect balance between technical and accessible. Looking forward to binging and following. Well done!

Top Podcasts In Technology

All-In with Chamath, Jason, Sacks & Friedberg
All-In Podcast, LLC
Lex Fridman Podcast
Lex Fridman
Hard Fork
The New York Times
No Priors: Artificial Intelligence | Machine Learning | Technology | Startups
Conviction | Pod People
TED Radio Hour
NPR
Acquired
Ben Gilbert and David Rosenthal

You Might Also Like

Practical AI: Machine Learning, Data Science
Changelog Media
Latent Space: The AI Engineer Podcast — Practitioners talking LLMs, CodeGen, Agents, Multimodality, AI UX, GPU Infra and al
Alessio + swyx
Data Skeptic
Kyle Polich
The AI Podcast
NVIDIA
Super Data Science: ML & AI Podcast with Jon Krohn
Jon Krohn
Machine Learning Street Talk (MLST)
Machine Learning Street Talk (MLST)