The Media Copilot

The Media Copilot

Hosted by journalist Pete Pachal, The Media Copilot is a weekly conversation with smart people on how AI is changing media, journalism, and the news.

  1. 1D AGO

    The Real Battle in AI Isn’t Capability. It’s Trust.

    Why the future of generative media may hinge on who owns the data and who gets paid for it. Generative AI can now create high-quality images and videos in seconds. But as the technology accelerates, a more fundamental question is emerging: Can AI-generated media ever be trusted at scale? In this episode of The Media Copilot, Pete Pachal speaks with Dr. Yair Adato, founder and CEO of Bria, about a growing divide in the AI ecosystem. On one side are models trained on vast, scraped datasets. On the other are systems built around licensed data, attribution, and control. At stake is not just quality, but ownership, accountability, and the future economics of creativity. Why This Matters The generative AI boom has largely focused on what these models can do. Less attention has been paid to how they are built and who benefits. As brands, media companies, and enterprises begin to integrate AI into real workflows, concerns around copyright, likeness rights, deepfakes, and data ownership are no longer theoretical. They are operational risks. This conversation reframes the debate: The future of AI may depend less on better models and more on building systems that businesses can actually trust. What We Cover  • Why “brand-safe” AI is becoming a business requirement, not a feature • The case for licensed data and a new attribution-driven data economy • How generative AI could reshape ownership and compensation for creators • Why visual AI presents higher stakes than text models • The limits of current models and the push toward greater control and transparency • How enterprises are integrating AI into real production workflows • The tension between automation and creativity in media and storytelling • Why AI will handle the “average” and humans will still define what is exceptional Notable Insight “This is by far the most advanced technology humanity has created,” says Dr.  Adato. “…and it took six years, not fifty.” About the 👤 Guest   LinkedIn:  linkedin.com/in/yair-adato-4936b236  Americans for Ben-Gurion University feature https://americansforbgu.org/generative-ai/ Bria (official site) https://bria.ai Bria AI LinkedIn: https://www.linkedin.com/company/briaai  Bria AI Instagram: https://www.instagram.com/bria.ai Bria AI Facebook: https://www.facebook.com/BriaAI About the show: To explore more conversations like this and see what’s new, visit the freshly updated Media Copilot website at mediacopilot.ai. You’ll find new episodes, expanded resources, and tools designed for journalists, communicators, and media leaders navigating the fast-changing world of AI. It’s the home base for everything Media Copilot and it’s just getting started. Enjoyed this episode? Subscribe to The Media Copilot on Substack, Apple Podcasts, Spotify, or your favorite app. On YouTube?  Tap the Like button and Subscribe to the YouTube channel. For more AI tools and resources built for media professionals, visit MediaCopilot.ai.Produced by Pete Pachal and Executive Producer Michele MussoEdited by the Musso Media Team Music: “Favorite” by Alexander Nakarada, licensed under CC BY 4.0All rights reserved. © AnyWho Media 2026

    37 min
  2. MAR 26

    Can You Certify Good AI use? This Organization Thinks So

    As AI reshapes journalism and media, Richard Murphy of the Alliance for Audited Media explains why the industry needs actual standards. AI is no longer experimental in media. It is operational. From drafting articles to generating images to influencing distribution, artificial intelligence is now embedded across the entire content pipeline in many organizations. But as adoption accelerates, trust is breaking down just as fast. In this episode of The Media Copilot, Pete Pachal talks with Richard Murphy, CEO of the Alliance for Audited Media, to unpack a growing industry response: ethical AI certification. Murphy explains how publishers, advertisers, and audiences are all asking the same question in different ways:  How do we know what is real, who created it, and whether we can trust it? The answer, at least in part, may lie in standards. Drawing from AAM’s newly developed framework, Murphy walks through the pillars of responsible AI use, from transparency and disclosure to human oversight and data protection. The goal is not to slow innovation, but to create guardrails that keep media credible in an era where AI can generate anything. Why This Matters Media has always relied on trust as its currency. AI is testing that foundation. When audiences cannot tell whether content is human-created, AI-assisted, or fully synthetic, credibility becomes fragile. At the same time, advertisers and partners are demanding proof that what they are funding or distributing meets ethical standards. This is where certification enters the picture. Ethical AI frameworks are quickly becoming more than best practice. They are positioning themselves as a competitive advantage, a compliance strategy, and potentially a defense against future regulation. The bigger shift is this: AI is not just changing how content is created.  It is redefining what accountability looks like in media. What we cover What “ethical AI certification” actually means in practiceThe 8 pillars of responsible AI use in media organizationsWhy disclosure is moving from optional to essentialThe difference between AI-assisted vs fully AI-generated contentWhere most trust failures are happening todayWhy self-regulation may be the industry’s best shot before government interventionHow AI is impacting not just content creation, but distribution and business modelsThe growing role of advertisers, partners, and audiences in demanding transparency About the 👤 Guest   LinkedIn (Personal Profile):  https://www.linkedin.com/in/rmurphy01 AAM Leadership Bio: https://auditedmedia.com/about/leadership Alliance for Audited Media): https://auditedmedia.com Digital Content Next (Articles): https://digitalcontentnext.org/blog/author/richmurphy/ About the show: To explore more conversations like this and see what’s new, visit the freshly updated Media Copilot website at mediacopilot.ai. You’ll find new episodes, expanded resources, and tools designed for journalists, communicators, and media leaders navigating the fast-changing world of AI. It’s the home base for everything Media Copilot and it’s just getting started. Enjoyed this episode? Subscribe to The Media Copilot on Substack, Apple Podcasts, Spotify, or your favorite app. On YouTube?  Tap the Like button and Subscribe to the YouTube channel. For more AI tools and resources built for media professionals, visit MediaCopilot.ai. Produced by Pete Pachal and Executive Producer Michele MussoEdited by the Musso Media Team . Music: “Favorite” by Alexander Nakarada, licensed under CC BY 4.0 All rights reserved. © AnyWho Media 2026

    46 min
  3. MAR 19

    When AI changes discovery, who still gets paid?

    AI is reshaping how people search, shop, and consume information and that shift is starting to challenge the business models that have supported media for decades. In this episode of The Media Copilot, host Pete Pachal sits down with Colin Jeavons, Founder and Chairman of Nomix Group, to explore what happens when AI becomes the middle layer between publishers and their audiences. From the collapse of traditional ad economics to the rising value of trust, this conversation breaks down how discovery is evolving, why some publishers may struggle to adapt, and where new opportunities are emerging across commerce, subscriptions, and AI-driven experiences. What we cover • How AI is changing search, discovery, and media economics • Why CPM-based advertising is under pressure • The growing importance of trust in content • The future of subscriptions and micropayments • How commerce and AI shopping may evolve • What publishers need to rethink right now Takeaways The old web rewarded volume. The next era may reward credibility. In Jeavons’ view, AI is speeding up a market correction that was already underway. Publishers built around commodity content and low value ad impressions face increasing risk. But organizations that create trusted reporting, specialized expertise, or high intent commerce content may still have a path forward. The future, he suggests, will not be defined by whether AI destroys publishing. It will be defined by which publishers learn how to operate in a world where attention is filtered through intelligent systems, trust carries a premium, and audiences are willing to pay for what feels indispensable. About the 👤 Guest   🔗 Colin Jeavons  LinkedIn: https://www.linkedin.com/in/colinjeavons/ 🔗 Nomix Group Website: https://nomix.group LinkedIn: https://www.linkedin.com/company/nomix-group/ About the show: To explore more conversations like this and see what’s new, visit the freshly updated Media Copilot website at mediacopilot.ai. You’ll find new episodes, expanded resources, and tools designed for journalists, communicators, and media leaders navigating the fast-changing world of AI. It’s the home base for everything Media Copilot and it’s just getting started. Enjoyed this episode? Subscribe to The Media Copilot on Substack, Apple Podcasts, Spotify, or your favorite app. On YouTube?  Tap the Like button and Subscribe to the YouTube channel. For more AI tools and resources built for media professionals, visit MediaCopilot.ai. Produced by Pete Pachal and Executive Producer Michele MussoEdited by the Musso Media Team  Music: “Favorite” by Alexander Nakarada, licensed under CC BY 4.0 All rights reserved. © AnyWho Media 2026

    50 min
  4. MAR 12

    Search is Changing Fast. Is Your Brand Ready for the Answer Engine Era?

    Search is changing fast, and AI is at the center of it. Search is no longer just about blue links and ranking on Google. More and more, people are getting their answers directly from AI tools like ChatGPT, Google AI Overviews, and other answer engines that summarize information, pull citations, and decide what gets surfaced in real time. That means visibility is changing, and so is the value of content. In this episode of The Media Copilot, Pete Pachal speaks with Josh Blyskal, who leads answer engine optimization research at Profound, a company focused on tracking how brands appear inside AI generated answers. Their conversation explores what answer engine optimization really means, how it differs from traditional SEO, and why specificity, utility, and structure now matter more than ever. What We Cover  • The shift from traditional search results to AI generated answers • What answer engine optimization (AEO) actually means • How AI tools break prompts into “fan out” searches • Why specificity and structured content matter more than ever • The role of citations and consensus in AI responses • How platforms like ChatGPT and Google AI Overviews choose sources • Why Reddit and user generated content still influence AI answers • The growing tension between AI discovery and publisher business models • Opportunities and risks for media organizations in the AI search era About the 👤 Guest Josh Blyskal • Website: https://www.joshblyskal.com • LinkedIn: https://www.linkedin.com/in/joshua-blyskal • X: https://twitter.com/joshblyskal • Speaker Deck: https://speakerdeck.com/joshbly Learn More About Profound • https://www.tryprofound.com About the show: To explore more conversations like this and see what’s new, visit the freshly updated Media Copilot website at mediacopilot.ai. You’ll find new episodes, expanded resources, and tools designed for journalists, communicators, and media leaders navigating the fast-changing world of AI. It’s the home base for everything Media Copilot and it’s just getting started. Enjoyed this episode? Subscribe to The Media Copilot on Substack, Apple Podcasts, Spotify, or your favorite app. On YouTube?  Tap the Like button and Subscribe to the YouTube channel. For more AI tools and resources built for media professionals, visit MediaCopilot.ai. Produced by Pete Pachal and Executive Producer Michele MussoEdited by the Musso Media Team  Music: “Favorite” by Alexander Nakarada, licensed under CC BY 4.0 All rights reserved. © AnyWho Media 2026

    47 min
  5. MAR 5

    Building the Newsroom AI Playbook Without Turning Journalism into Slop

    AI is rapidly becoming part of how news is produced, distributed, and discovered. But what does that actually look like inside a newsroom? In this episode of The Media Copilot, host Pete Pachal speaks with Gina Chua, Executive Editor at Large at Semafor and Executive Director of the Tow-Knight Center for Journalism Futures at the CUNY Graduate School of Journalism. Chua shares how Semafor is experimenting with practical AI tools that support journalists in everyday workflows. These include tools for copy editing and proofreading, systems that suggest relevant datasets for charts while a reporter is writing, and tools that help surface related reporting across different outlets and languages. The conversation also explores how newsrooms can organize large volumes of information. At Semafor, interview transcripts from events and panels are integrated into internal systems so reporters can quickly search conversations, locate quotes, and review context directly. Chua emphasizes that these tools are designed to assist newsroom work rather than replace editorial judgment. She also offers a useful way to think about large language models: they are built to work with language, not to verify facts. When used carefully with known text sources, they can help summarize, organize, and analyze information. Beyond newsroom workflows, the discussion turns to the broader shift happening in how people access information. AI tools, chatbots, and automated summaries are increasingly becoming a gateway to news, which raises important questions about trust, verification, and the future role of journalism. This episode looks at how reporters, editors, and media organizations are adapting as AI becomes part of the information ecosystem. What we cover • How Semafor is experimenting with AI tools inside the newsroom • Using transcripts and Slack to search interviews and discussions • Why language models are useful for handling text but not verifying facts • The role of human review in newsroom publishing decisions • How AI interfaces are changing the way audiences find news TIMESTAMPS: 00:00 – Intro: Journalism in the AI Era 02:15 – Gina’s Background & Semafor’s Model 06:00 – How Newsrooms Are Using AI Today 10:00 – Trust in a Synthetic World 14:00 – Transparency & Disclosure 18:30 – AI Tools Inside Reporting 23:00 – The Risk of Information Overload 27:30 – Reinventing the News Business 32:00 – Where AI Helps Most 36:30 – The Future of Journalism About the 👤 Guest GINA CHUA LinkedIn 👉 https://www.linkedin.com/in/ginachua X (Twitter) 👉 https://x.com/GinaSKChua Instagram 👉 https://www.instagram.com/gina_chua_nyc  Personal Website / Writing 👉 https://ginachua.me  Author Page (Semafor) 👉 https://www.semafor.com/author/gina-chua —- About the show: To explore more conversations like this and see what’s new, visit the freshly updated Media Copilot website at mediacopilot.ai. You’ll find new episodes, expanded resources, and tools designed for journalists, communicators, and media leaders navigating the fast-changing world of AI. It’s the home base for everything Media Copilot and it’s just getting started. Produced by Pete Pachal and Executive Producer Michele MussoEdited by the Musso Media Team  Music: “Favorite” by Alexander Nakarada, licensed under CC BY 4.0 All rights reserved. © AnyWho Media 2026

    51 min
  6. FEB 26

    She’s Building the AI Agent That Could Replace Your News Feed

    What if instead of scrolling headlines, you had a personal intelligence agent that understood what matters to you and delivered only signal, not noise? In this episode of The Media Copilot, Pete Pachal talks with Eva Cicinyte, co-founder and CEO of Gnomi, an AI-powered real-time news agent designed to synthesize global information into actionable insight. The goal isn’t summaries or more feeds. It’s context. Eva explains how her experience in political data analytics shaped her mission to make high quality understanding accessible to everyone, not just institutions with research teams. Gnomi pulls from global sources, social platforms, video, audio, and financial data to deliver personalized intelligence in real time. The platform’s new Finance Mode can even analyze live earnings calls as they happen, potentially surfacing market signals before headlines move prices.🔍 In this conversation  • Why Gnomi is built as an “intelligence layer,” not a news app • How AI agents could replace search and traditional feeds • The danger of engagement driven AI systems • Multilingual analysis and global perspective gaps • Using social and video data to detect emerging signals • Real time market insights from live earnings calls • The future of journalism in an AI first world • Ads, subscriptions, and the economics of AI tools If you care about the future of news, AI, finance, or how people will stay informed in the coming decade, this episode is a must watch. 00:00 – Intro: Why AI Agents Matter Now  Big-picture framing of the agent shift. 02:10 – Eva’s Background & Building Gnomi  How she entered the agent space and what problem they’re solving. 05:40 – What Actually Is an AI Agent?  Clear distinction between chatbots and agents. 09:15 – From Answers to Action  How agents move from generating text to executing workflows. 13:50 – Designing Guardrails & Trust  Why autonomy requires control and reliability. 18:20 – Real-World Use Cases  Where agents are already creating leverage. 22:45 – AI in the Workflow Stack  Replacing apps and orchestrating tools. 27:30 – Human + AI Collaboration  Why agents amplify people instead of replacing them. 32:10 – Infrastructure: Memory, Context & Systems  What makes agents actually autonomous. 37:00 – Competitive Advantage in the Agent Era  How companies should think about adoption. 41:30 – The Future of the Agent Economy  Where this is all headed next. About the 👤 Guest  LinkedIn 👉 https://www.linkedin.com/in/eva-cicinyte-1447161b2 Instagram (Personal) 👉 https://www.instagram.com/evapariscicinyte Official Website 👉 https://www.gnomi.com LinkedIn (Company Page) 👉 https://www.linkedin.com/company/gnomi Instagram (Company)👉 https://www.instagram.com/gnomi.app  About the show: To explore more conversations like this and see what’s new, visit the freshly updated Media Copilot website at mediacopilot.ai. You’ll find new episodes, expanded resources, and tools designed for journalists, communicators, and media leaders navigating the fast-changing world of AI. It’s the home base for everything Media Copilot and it’s just getting started. Enjoyed this episode? Subscribe to The Media Copilot on Substack, Apple Podcasts, Spotify, or your favorite app. On YouTube?  Tap the Like button and Subscribe to the YouTube channel. For more AI tools and resources built for media professionals, visit MediaCopilot.ai.Produced by Pete Pachal and Executive Producer Michele MussoEdited by the Musso Media Team Music: “Favorite” by Alexander Nakarada, licensed under CC BY 4.0All rights reserved. © AnyWho Media 2026

    43 min
  7. FEB 19

    Fake News at Machine Speed: Inside AI’s Impact on Media Trust

    Poynter’s Alex Mahadevan explains how newsrooms can use AI without losing the fundamentals of verification, context, and accountability. By The Copilot AI is already embedded in how people discover and consume news, from search to chat interfaces to automated summaries. So the question is no longer whether journalism will be shaped by AI. It’s how newsrooms maintain trust while experimenting responsibly. In this episode of The Media Copilot podcast, Pete Pachal sits down with Alex Mahadevan, Director of MediaWise and a faculty member at Poynter, to unpack what media literacy looks like now that anyone can generate convincing content at scale. Alex shares how his background in data and local journalism shaped his approach to tools, why public-facing AI ethics policies matter, and what it will take for news organizations to bring audiences along for the next phase of the information ecosystem. Why this matters Trust is the core product. AI can either widen the trust gap with errors and low-quality content, or help rebuild credibility through transparency, better products, and clearer communication about how journalism is made. This conversation gets practical about what responsible AI use looks like, where disclosures help and where they can unintentionally slow innovation, and why the newsroom AI divide is becoming a real competitive advantage for organizations that adapt. What we cover • Alex’s journey into journalism and the global mission of MediaWise • How AI is reshaping misinformation, trust, and newsroom transparency • Practical uses of chatbots, coding agents, and AI workflows • The widening divide between AI enthusiasts and skeptics in newsrooms • Ethics, job concerns, and gray areas around AI-assisted writing • What the future of news may look like beyond traditional articles About the 👤 Guest  🔗Alex Mahadevan  🔗Poynter / MediaWise  🔗MediaWise About the show: To explore more conversations like this and see what’s new, visit the freshly updated Media Copilot website at mediacopilot.ai. You’ll find new episodes, expanded resources, and tools designed for journalists, communicators, and media leaders navigating the fast-changing world of AI. It’s the home base for everything Media Copilot and it’s just getting started. Enjoyed this episode? Subscribe to The Media Copilot on Substack, Apple Podcasts, Spotify, or your favorite app. On YouTube?  Tap the Like button and Subscribe to the YouTube channel. For more AI tools and resources built for media professionals, visit MediaCopilot.ai. Produced by Pete Pachal and Executive Producer Michele MussoEdited by the Musso Media Team  Music: “Favorite” by Alexander Nakarada, licensed under CC BY 4.0 All rights reserved. © AnyWho Media 2026

    51 min
  8. FEB 5

    AI, and copyright: How media can decide between litigation or negotiation

    Lawsuits set public rules; contracts set private ones. A media attorney on how leverage, timing, and context decide the path. In this episode, Pete Pachal sits down with corporate and transactional attorney Jason Henderson, a streaming and licensing specialist who also happens to be a creative with real skin in the game. Jason breaks down why the popular “AI learns like humans” analogy only goes so far, how fair use really works in court, and why the future will be shaped less by courtroom theory and more by deal structures. The key parts of those deals that are often overlooked: indemnification and who actually bears the risk when things go sideways. From The New York Times and Perplexity headlines to the practical mechanics of licensing training data, this conversation gets grounded fast. Jason explains what matters most to media companies, what smaller publishers should watch, and why agentic browsing and attribution are shaping up to be the next pressure point. Why this matters: Media is facing a new kind of competition. Not always a stolen article, but a substituted experience. When AI tools summarize, synthesize, and answer in real time, the legal question is not only “Was it copied?” It is also “Does it replace the market for the original?” Jason outlines how courts evaluate that, why “transformative” is both the key term and the messiest one, and why the industry is drifting toward partnerships and licensing frameworks even as litigation continues. At the same time, the next wave is not just training bots or search bots. It is agents that behave like users and may be harder to block or even detect. The more AI becomes the interface to the web, the more urgent it becomes for publishers to understand the business and legal stakes. Key Takeaways Fair use is not a blanket shield. Courts look at purpose, transformation, and market impact, and the facts matter. Legitimate acquisition matters. Even if a use might be transformative, piracy can change the legal posture dramatically. Media’s biggest fear is substitution. Summaries and AI answers can erode subscriptions, traffic, and trust, even without verbatim copying. Deals are becoming more specific. Expect narrower permissions and more constraints on how data can be used for training or product features. Risk is moving through contracts. Indemnification is common, but it is only as strong as the indemnifier’s balance sheet and insurance. Attribution is the missing bridge. A clear “this came from” pathway could reduce conflict and rebuild value for original publishers. Agentic browsing will raise the temperature. When AI acts as a user proxy, blocking and enforcement become harder, and the business questions get sharper. 👤 Guest 🔗Jason Henderson    🔗Senior Attorney, JWL International    🔗Founder, Castle Bridge Media      🔗Co-host, Castle of Horror podcast (horror movie coverage) About the show: To explore more conversations like this and see what’s new, visit the freshly updated Media Copilot website at mediacopilot.ai. You’ll find new episodes, expanded resources, and tools designed for journalists, communicators, and media leaders navigating the fast-changing world of AI. It’s the home base for everything Media Copilot and it’s just getting started. Enjoyed this episode? Subscribe to The Media Copilot on Substack, Apple Podcasts, Spotify, or your favorite app. On YouTube?  Tap the Like button and Subscribe to the channel.  For more AI tools and resources built for media professionals, visit MediaCopilot.ai. Produced by Pete Pachal and Executive Producer Michele MussoEdited by the Musso Media Team  Music: “Favorite” by Alexander Nakarada, licensed under CC BY 4.0 All rights reserved. © AnyWho Media 2026

    51 min

Ratings & Reviews

5
out of 5
5 Ratings

About

Hosted by journalist Pete Pachal, The Media Copilot is a weekly conversation with smart people on how AI is changing media, journalism, and the news.

You Might Also Like