AI lab by information labs

information labs

AI lab podcast, "decrypting" expert analysis to understand Artificial Intelligence from a policy making point of view.

  1. AI lab TL;DR | Anna Mills and Nate Angell - The Mirage of Machine Intelligence

    MAY 26 · BONUS

    AI lab TL;DR | Anna Mills and Nate Angell - The Mirage of Machine Intelligence

    🔍 In this TL;DR episode, Anna and Nate unpack why calling AI outputs “hallucinations” misses the mark—and introduce “AI Mirage” as a sharper, more accurate metaphor. From scoring alternative terms to sparking social media debates, they show how language shapes our assumptions, trust, and agency in the age of generative AI. The takeaway: choosing the right words is a hopeful act of shaping our AI future. 📌 TL;DR Highlights ⏲️[00:00] Intro ⏲️[00:42] Q1-What’s wrong with the term “AI hallucination” — and how does “mirage” help? ⏲️[05:30] Q2-Why did “mirage” stand out among 80+ alternatives? ⏲️[10:30] Q3-How should this shift in language impact educators, journalists, or policymakers? ⏲️[10:10] Wrap-up & Outro 💭 Q1 - What’s wrong with the term “AI hallucination” — and how does “mirage” help? 🗣️ "There's no reason to think that AI is experiencing something, that it has a belief about what's real or what's not." (Anna) 🗣️ "It anthropomorphizes AI, and it also misleads us to think that this might be a technically fixable problem—as a person might take medication for mental illness—that maybe AI could be induced not to hallucinate." (Anna) 🗣️ "I did come up with my own criteria, which included: not implying that AI has intent or consciousness, implying that outputs don't match reality in some way, showing a connection to the patterns in the training data ideally, but also showing that AI can go beyond training data." (Anna) 🗣️ "The words used to describe different technologies can sometimes steer people in directions in relation to them that aren’t really beneficial." (Nate) 🗣️ "Just like how a desert produces a mirage under certain circumstances... It’s the same with AI. There’s a system at play... that can produce a certain situation, which can then be perceived by an observer as possibly misleading, inaccurate, or counterfactual." (Nate) 💭 Q2 - Why did “mirage” stand out among 80+ alternatives? 🗣️ "I actually went through and rated each term numerically on each of those criteria and did kind of a simple averaging of that to see which terms scored the highest." (Anna) 🗣️ "We decided that it was misleading to say 'Data Mirage,' because people would think the problem was in the data... and that’s not the case. So we ditched the 'data' part and just landed on 'AI Mirage'." (Anna) 🗣️ "We kind of realized, as we were discussing 'Mirage,' how important it was that it centered human judgment—and that wasn’t initially one of the criteria." (Anna) 🗣️ "Even when we know how it works and we know it’s wrong, sometimes there’s still that temptation... to say, 'Wow, I think it really nailed it this time.'" (Anna) 🗣️ "We really wanted to encourage this ongoing interrogation of the metaphors we use and the language we use, and how they're affecting our relationship with AI." (Anna) 💭 Q3 - How should this shift in language impact educators, journalists, or policymakers? 🗣️ "How do we build systems and train ourselves to think about how we want to interact with them, stay in control, and still be the ones making judgments and choices?" (Anna) 🗣️ "We are participating in shaping that future, and it’s not over. We don’t have to just capitulate and accept the term that’s used. We don’t have to accept someone’s vision of what AGI is going to be in five years. We’re all shaping this." (Anna) 🗣️ "In a way, it doesn’t really matter what term you end up with—just asking the question of whether 'hallucination' is a useful or accurate term can spark a really interesting and valuable discussion." (Nate) 🗣️ "There are many systemic issues we should be thinking about with AI. But I also believe in the power of the damning—of the words we use to talk about it—as being an important factor in all that." (Nate) 🗣️ "It’s useful for us as humans to have different words for those outputs we deem unexpected,  incorrect, or counterfactual. It helps us to talk about when an AI mirages rather than dumping all its outputs into one big undifferentiated basket." (Nate) 📌 About Our Guests 🎙️ Anna Mills | College of Marin 🌐 Anna Mills linkedin.com/in/anna-mills-oer  🎙️ Nate Angell | Nudgital 🌐 Nate Angell linkedin.com/in/nateangell  🌐 Article | Are We Tripping? The Mirage of AI Hallucinations https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5127162  Anna is a college writing instructor and a leading advocate for AI literacy in education, building on her combined teaching experience and technical knowledge.  Nate is the founder of Nudgital, a company that builds sustainability and growth at the intersection of communications, community, technology, and strategy.   #AI #ArtificialIntelligence #GenerativeAI

    21 min
  2. AI lab TL;DR | Emmie Hine - Can Europe Lead the Open-Source AI Race?

    MAY 12 · BONUS

    AI lab TL;DR | Emmie Hine - Can Europe Lead the Open-Source AI Race?

    🔍 In this TL;DR episode, Emmie Hine (Yale Digital Ethics Center) makes the case for Europe’s leadership in open-source AI—thanks to strong infrastructure, multilingual data, and regulatory clarity. With six key policy recommendations, the message is clear: trust and transparency can make EU models globally competitive. 📌 TL;DR Highlights ⏲️[00:00] Intro ⏲️[00:43] Q1-What advantages make the EU a strong contender in open-source AI compared to the US and China? ⏲️[03:08] Q2-How do EU regulations and initiatives enhance model trustworthiness and feasibility? ⏲️[05:43] Q3-What policy recommendations does the paper offer for EU AI governance and deployment? ⏲️[10:10] Wrap-up & Outro 💭 Q1 - What advantages make the EU a strong contender in open-source AI compared to the US and China? 🗣️ “The EU’s primary advantage is actually in regulatory leadership and trustworthiness.” 🗣️ “There’s access to a lot of multilingual data, which is really great and important for creating multilingual LLMs.” 🗣️ “As DeepSeek showed, big leaps are very possible.” 💭 Q2 - How do EU regulations and initiatives enhance model trustworthiness and feasibility? 🗣️ “Its framework of risk-based classification and how it encourages ethics by design—I think that’s really, really important for guiding responsible development and deployment of foundation models.” 🗣️ “The AI Office is really putting an emphasis on multi-stakeholder collaboration... bringing in civil society, academia, and industry.” 🗣️ “The Code of Practice will provide clarity on training data and copyright as well.” 💭 Q3 - Why focus on regulating specific AI apps instead of AI overall? 🗣️ “Establishing an EU-wide open-source foundation model governance framework... would combat open-washing, where companies say, ‘Oh yeah, this is open source,’ but it’s not really.” 🗣️ “A certification and benchmarking system to evaluate open models for security, reliability, ethics, and performance... would help boost user trust and also international competitiveness.” 🗣️ “Expanding funding programs—especially for SMEs and startups—can help them put out more competitive models while encouraging multilingual capabilities and ethical development.” 🗣️ “Ideally, this investment will ensure that these facilities are going to be energy efficient to help combat the climate impacts of models.” 🗣️ “There was definitely not a universal understanding of open-source technology in general, and specifically around open-source AI... So promoting digital literacy and responsible AI usage is going to be really, really crucial.” 📌 About Our Guest 🎙️ Emmie Hine | Yale Digital Ethics Center 🌐 Article | Open-Source Foundation Models Made in the EU: Why it is a Good Idea https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5191372  🌐Newsletter | The Ethical Reckoner https://ethicalreckoner.substack.com/  🌐 Emmie Hine linkedin.com/in/emmiehine Emmie Hine is a Research Associate at the Yale Digital Ethics Center and a PhD candidate in Law, Science, and Technology at the University of Bologna and KU Leuven. Her research focuses on the ethics and governance of emerging technologies in different geopolitical contexts.  #AI #ArtificialIntelligence #GenerativeAI

    11 min
  3. AI lab TL;DR | Milton Mueller - Why Regulating AI Misses the Point

    APR 21 · BONUS

    AI lab TL;DR | Milton Mueller - Why Regulating AI Misses the Point

    🔍 In this TL;DR episode, Milton Mueller (the Georgia Institute of Technology School of Public Policy) argues that what we call “AI” is really just part of a broader digital ecosystem. Instead of vague, top-down AI regulation, he calls for context-specific rules that address actual uses—like facial recognition or medical diagnostics—rather than the underlying technology. The key message: regulate behavior and applications, not “AI” as an abstract concept. 📌 TL;DR Highlights ⏲️[00:00] Intro ⏲️[00:47] Q1-Why do you see AI as “distributed computing,” not a new tech? ⏲️[04:05] Q2-Why aren’t issues like misinformation and bias new to AI? ⏲️[09:29] Q3-Why focus on regulating specific AI apps instead of AI overall? ⏲️[17:02] Wrap-up & Outro 💭 Q1 - Why do you see AI as “distributed computing,” not a new tech? 🗣️ "Distributed computing is a revolutionary technology... it started around 1945, 1950, and it's been progressing and evolving ever since." 🗣️ "What people are calling AI is really a digital ecosystem, which is a cybernetic system that consists of computing devices, networks, data, and software." 🗣️ "When people talk about governing AI, it's not just about AI... you're going to have to control everything about ICT, everything about digital ICT." 🗣️ "If indeed you're talking about some way of controlling in advance the capabilities to do machine learning applications, you really are committing yourself to controlling every element of that digital ecosystem." 💭 Q2 - Why aren’t issues like misinformation and bias new to AI? 🗣️ "People say AI is a problem because of its capability to distribute misinformation, but if you go back to 1995, we were already noticing that some of the things you find on the internet are not true." 🗣️ "In the early days of the internet, people thought that just because something was online, it had to be credible or true." 🗣️ "What was giving you these search results... was an early form of artificial intelligence. You had robots searching the internet and scanning websites according to a program." 🗣️ "What’s interesting is that AI-driven search engines were already giving biased results back in the '90s... search engine bias was something people were complaining about." 🗣️ "You are interacting with an artificial intelligence mechanism that is distributed and globalized, and you're trying to control the results or objecting to the results in ways that are very similar to what we're hearing now about AI." 💭 Q3 - Why focus on regulating specific AI apps instead of AI overall? 🗣️ "The problems posed by a specific AI application would require different kinds of responses depending on the nature of the problem." 🗣️ "Can you solve a problem like that by regulating facial recognition technology? And the answer is no. What you do is you regulate police practice." 🗣️ "You have to be more specific. It's not about regulating AI in a general way; it's about regulating specific applications in their specific context." 🗣️ "Whenever you hear somebody say 'AI is causing this problem,' just substitute the word computing, and you'll see how it makes just as much sense." 🗣️ "The idea that you can regulate technology to prevent it from doing any harm in advance is an idea that, as a historian of technology, just doesn't make any sense." 📌 About Our Guest 🎙️ Milton Mueller | the Georgia Institute of Technology School of Public Policy 🌐 Article | It's just distributed computing: Rethinking AI governance https://www.sciencedirect.com/science/article/pii/S030859612500014X  🌐 Milton Mueller https://www.linkedin.com/in/miltonmueller/  Milton Mueller is Professor at the Georgia Institute of Technology School of Public Policy. He specialises in the political economy of information and communication and is the author of seven books and numerous journal articles. His work informs not only public policy but also science and technology studies, law, economics, communications, and international studies. #AI #ArtificialIntelligence #GenerativeAI

    18 min
  4. AI lab TL;DR | Kevin Frazier - How Smarter Copyright Law Can Unlock Fairer AI

    APR 7 · BONUS

    AI lab TL;DR | Kevin Frazier - How Smarter Copyright Law Can Unlock Fairer AI

    🔍 In this TL;DR episode, Kevin Frazier (University of Texas at Austin school of Law) outlines a proposal to realign U.S. copyright law with its original goal of spreading knowledge. The discussion introduces three key reforms—an AI training presumption, research safe harbors, and data commons—to help innovators access data more easily. By reducing legal ambiguity, the proposals aim to support responsible AI development and level the playing field for startups and researchers.  📌 TL;DR Highlights ⏲️[00:00] Intro ⏲️[00:40] Q1-How do copyright laws limit AI’s training data and knowledge diffusion? ⏲️[04:05] Q2-How does the original intent of the Intellectual Property Clause conflict with current copyright rules? ⏲️[10:35] Q3-What reforms could align copyright with promoting science? ⏲️[15:36] Wrap-up & Outro 💭 Q1 - How do copyright laws limit AI’s training data and knowledge diffusion? 🗣️ "Copyright law was designed to promote knowledge creation. Now it functions as a bottleneck in that knowledge ecosystem." 🗣️ "Today's AI systems face a paradoxical constraint: they require vast oceans of human-created content to learn, and yet our copyright framework increasingly cordons off those essential waters." 🗣️ "Early publishers faced guild restrictions and royal monopolies that limited the dissemination of knowledge. Today's AI developers are navigating a similarly restrictive landscape through the barriers of copyright law." 🗣️ "When there's ambiguity in the law, that hinders innovation. It leads to litigation and poses a real threat to startups trying to determine whether they can use copyrighted information for training." 💭 Q2 - How does the original intent of the Intellectual Property Clause conflict with current copyright rules?  🗣️ "The IP clause starts off with a mandate that Congress spread knowledge. If something isn’t promoting the progress of science, then it can’t be interpreted as constitutional." 🗣️ "Copyright began as a 14-year term. Now it's expanded to more than 70 years — a huge restriction on the ability to spread knowledge." 🗣️ "The founders hated monopolies. They’d seen how royal prerogatives were used to quash innovation — and tried to create a better system for incentivizing knowledge." 🗣️ "AI tools are unparalleled in their ability to create knowledge. The question now is: can we spread that knowledge?" 🗣️ "The Constitution’s goal wasn’t just to reward creators — it was to spread science and useful arts as far and wide as possible." 💭 Q3 - What reforms could align copyright with promoting science? 🗣️ "We need a clear statutory presumption that using works for machine learning constitutes fair use — that sort of clarity is essential for startups and research institutions to compete." 🗣️ "Without robust datasets, the positive use cases of AI — from public health breakthroughs to AI tutors for differently-abled students — simply aren’t possible." 🗣️ "Imagine if all that data your smartwatch gathers went toward training AI models tailored to the public good — that’s the promise of data commons." 🗣️ "AI is like fire: it can spread and fuel incredible progress — but if we smother it with fire extinguishers too soon, only the biggest players will be able to benefit." 🗣️ "We must make sure this isn’t a world where only OpenAI, Anthropic, and Google build the models — we need a future with many options and many positive use cases of AI." 📌 About Our Guest 🎙️ Kevin Frazier | the University of Texas at Austin School of Law 🌐 Article | Progress Interrupted: The Constitutional Crisis in Copyright Law https://jolt.law.harvard.edu/digest/progress-interrupted-the-constitutional-crisis-in-copyright-law  🌐 Kevin Frazier https://www.linkedin.com/in/kevin-frazier-51811737/     Kevin Frazier is an AI Innovation and Law Fellow at the Austin School of Law of the University of Texas. He is also a Contributing Editor at Lawfare, a non-profit publication and he developed the first open-source Law and AI syllabus. #AI #ArtificialIntelligence #GenerativeAI

    17 min
  5. AI lab TL;DR | Paul Keller - A Vocabulary for Opting Out of AI Training and TDM

    MAR 24 · BONUS

    AI lab TL;DR | Paul Keller - A Vocabulary for Opting Out of AI Training and TDM

    🔍 In this TL;DR episode, Paul Keller (The Open Future Foundation) outlines a proposal for a common opt-out vocabulary to improve how EU copyright rules apply to AI training. The discussion introduces three clear use cases—TDM, AI training, and generative AI training—to help rights holders express their preferences more precisely. By standardizing terminology across the value chain, the proposal aims to bring legal clarity, promote interoperability, and support responsible AI development. 📌 TL;DR Highlights ⏲️[00:00] Intro ⏲️[00:41] Q1-Why is this vocabulary needed for AI training opt-outs? ⏲️[04:17] Q2-How does it help creators, AI developers, and policymakers and what are some of the concepts? ⏲️[11:55] Q3-What are its limitations, and how could it evolve? ⏲️[14:35] Wrap-up & Outro 💭 Q1 - Why is this vocabulary needed for AI training opt-outs? 🗣️ "At the core of the EU copyright framework is... the TDM exceptions – the exceptions for text and data mining that were introduced in the 2019 Copyright Directive." 🗣️ "It ensures that rights holders have some level of control over their works, and it makes sure that the majority of publicly available works are available to innovate on top of, to build new things." 🗣️ "The purpose of such a vocabulary is to provide a common language for expressing rights reservations and opt-outs that are understood in the same way along the entire value chain." 🗣️ "This vocabulary proposal is the outcome of discussions that we had with many stakeholders, including rights holders, AI companies, policymakers, academics, and public interest technologists." 💭 Q2 - How does it help creators, AI developers, and policymakers and what are some of the concepts?  🗣️ "At the very core, the idea of vocabulary is that you have some common understanding of language... that terms you use mean the same to other people that you deal with." 🗣️ "We offer these three use cases for people to target their opt-outs from... like sort of the Russian dolls: the wide TDM category that is AI training, and in that is generative AI training." 🗣️ "If all of these technologies sort of use the same definition of what they are opting out, it becomes interoperable and it becomes also relatively simple to understand on the rights holder side." 💭 Q3 - What are its limitations, and how could it evolve? 🗣️ "The biggest limitation is... we need to see if this lands in reality and stakeholders start working with this." 🗣️ "These information intermediaries... essentially convey the information from rights holders to model providers—then it has a chance to become something that structures this field." 🗣️ "It is designed as a sort of very simple, relatively flexible approach that makes it expandable." 📌 About Our Guest 🎙️ Paul Keller | The Open Future Foundation 🌐 Article | A Vocabulary for opting out of AI training and other forms of TDM https://openfuture.eu/wp-content/uploads/2025/03/250307_Vocabulary_for_opting_out_of_AI_training_and_other_forms_of_TDM.pdf  🌐 Paul Keller https://www.linkedin.com/in/paulkeller/        Paul Keller is the co-Founder and Director of Policy at the Open Future Foundation, a European nonprofit organization. He has extensive experience as a media activist, open policy advocate and systems architect striving to improve access to knowledge and culture. #AI #ArtificialIntelligence #GenerativeAI

    15 min
  6. AI lab TL;DR |  João Pedro Quintais - Untangling AI Copyright and Data Mining in EU Compliance

    MAR 3 · BONUS

    AI lab TL;DR | João Pedro Quintais - Untangling AI Copyright and Data Mining in EU Compliance

    🔍 In this TL;DR episode, João Quintais (Institute for Information Law) explains the interaction between the AI Act and EU copyright law, focusing on text and data mining (TDM). He unpacks key issues like lawful access, opt-out mechanisms, and transparency obligations for AI developers. João explores challenges such as extraterritoriality and trade secrets, offering insights into how voluntary codes of practice and contractual diligence could help align AI innovation with EU copyright rules. #AI #ArtificialIntelligence #GenerativeAI 📌 TL;DR Highlights ⏲️[00:00] Intro ⏲️[00:51] Q1-What are the key interactions between the AI Act and EU copyright law, particularly concerning text and data mining (TDM) practices? ⏲️[7:12] Q2-How do you see the transparency obligations of the AI Act shaping the relationship between AI developers and rightsholders? ⏲️[15:22] Q3-Is the extraterritorial reference of Recital 106 of the AI Act enforceable? ⏲️[24:35] Wrap-up & Outro 💭 Q1 - What are the key interactions between the AI Act and EU copyright law, particularly concerning text and data mining (TDM) practices 🗣️ "The AI Act links to Article 4 by obliging general-purpose AI (GPAI) model providers to identify and respect rights reservation mechanisms and disclose a sufficiently detailed summary about the training data used." 🗣️ "The Copyright Directive of 2019 introduces exceptions and limitations for text and data mining (TDM), with Article 3 aimed at research and Article 4 applying broadly but with additional requirements like rights reservation or opt-out mechanisms." 🗣️ "The concept of TDM is so broad that it applies to activities involved in pre-training and training AI models, impacting entities across the value chain, not just model providers." 🗣️ "Entities like Common Crawl or LAION that perform upstream activities like web scraping are not directly regulated by the AI Act but are part of the broader TDM definition under the Copyright Directive." 🗣️ "One debated requirement is the rights reservation or opt-out mechanism for publicly accessible online content." 💭 Q2 - How do you see the transparency obligations of the AI Act shaping the relationship between AI developers and rightsholders? 🗣️ "The transparency provision in Article 53(1)(d) requires GPAI model providers to make publicly available a sufficiently detailed summary of training data, balancing interests like copyright, privacy, and fair competition." 🗣️ "If the summary is too vague, it becomes meaningless; if too detailed, it might infringe on trade secrets—so finding a balance is critical." 🗣️ "The usefulness of the training data summary lies in clarifying whether TDM exception requirements, such as lawful access and respect for opt-outs, have been met." 🗣️ "A significant challenge is ensuring compliance when data sets are obtained from upstream providers not subject to the AI Act, raising questions about responsibility and enforcement." 🗣️ "The Act balances interests, acknowledging the impossibility of listing all copyrighted works used in training due to territorial fragmentation and low originality thresholds." 💭 Q3 - Is the extraterritorial reference of Recital 106 of the AI Act enforceable? 🗣️ "Recital 106 aims to prevent regulatory arbitrage by requiring compliance with EU copyright standards, even for AI models trained outside the EU." 🗣️ "The principle of territoriality in copyright law conflicts with the extraterritorial implications of the AI Act, as copyright rules are typically governed by the location of the activity." 🗣️ "Using contractual obligations and voluntary meta-regulation, such as commitments from upstream providers, offers a more consistent way to enforce compliance than extending the law extraterritorially." 🗣️ "The Act's compliance incentives might still push GPAI providers to align with EU standards to avoid severe sanctions, even if extraterritorial enforcement remains uncertain." 🗣️ "Some suggest contractual obligations or meta-regulation as more practical solutions to ensure upstream compliance with EU law." 📌 About Our Guest 🎙️  João Pedro Quintais | Assistant Professor, Institute for Information Law (IViR) 🌐 Article | Generative AI, Copyright, and the AI Act https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4912701  🌐 Joao Pedro Quintais https://www.ivir.nl/employee/quintais/  Dr João Pedro Quintais is Assistant Professor at the University of Amsterdam’s Law School, in the Institute for Information Law (IViR). João notably studies how intellectual property law applies to new technologies and the implications of copyright law and its enforcement by algorithms on the rights and freedoms of Internet users, on the remuneration of creators, and on technological development. João is also Co-Managing Editor of the widely read Kluwer Copyright Blog and has published extensively in the area of information law.

    25 min
  7. AI lab TL;DR | Anna Tumadóttir - Rethinking Creator Consent in the Age of AI

    FEB 10 · BONUS

    AI lab TL;DR | Anna Tumadóttir - Rethinking Creator Consent in the Age of AI

    🔍 In this TL;DR episode, Anna Tumadóttir (Creative Commons) discusses how the evolution of creator consent and AI has reshaped perspectives on openness, highlighting the challenges of balancing creator choice with the risks of misuse. Examines the limitations of blunt opt-out mechanisms like those in the EU AI Act, the implications for marginalized communities and open access, and explores the need for nuanced preference signals to preserve openness while respecting creators' intentions in the age of generative AI. 📌 TL;DR Highlights ⏲️[00:00] Intro ⏲️[00:38] Q1-How has the conversation around creator consent and AI evolved in the past few years? ⏲️[05:42] Q2-How has the tension between openness and creator choice played out so far, and what lessons can be learned from this tension? ⏲️[11:02] Q3-can we ensure that marginalized or underrepresented communities maintain agency over their contributions to the commons? ⏲️[19:03] Wrap-up & Outro 💭 Q1 - How has the conversation around creator consent and AI evolved in the past few years? 🗣️ "When AI became mainstream, most creators couldn’t have anticipated how their works might one day be used by machines—some uses align with their intentions, but others do not." 🗣️ "Individuals want more choice over how their work is used, and without tailored options, some are considering paywalls or not publishing at all." 🗣️ "The EU AI Act’s opt-out mechanism is a blunt instrument—it’s just a 'no,' not a nuanced reflection of creators’ varied preferences." 🗣️ "Creators may object to large companies using their works for AI training but be fine with nonprofits or research-focused uses, showing the need for more nuanced tools." 🗣️ "We’re focusing on developing 'preference signals'—mechanisms that let creators communicate specific preferences for how their work is used in AI models." 💭 Q2 - How has the tension between openness and creator choice played out so far, and what lessons can be learned from this tension? 🗣️ "Scientists and researchers who traditionally embraced open access are now reconsidering, fearing that commercial AI providers are exploiting their work." 🗣️ "Creators who once freely shared their work under CC licenses are now hesitant, either because they misunderstand AI training risks or feel exposed." 🗣️ "The worst outcome of this tension is less openness overall—creators retreating behind paywalls or choosing not to publish at all." 🗣️ "A perception persists in the open-source AI community that CC-licensed works are 'safe' to use, but creators’ motivations for sharing openly years ago don’t always align with today’s AI landscape." 🗣️ "To preserve openness while respecting creator intentions, we need mechanisms that enable a 'no unless' approach—minimizing restrictions while maximizing use." 💭 Q3 - can we ensure that marginalized or underrepresented communities maintain agency over their contributions to the commons? 🗣️ "Generative AI amplifies existing inequalities because it demands infrastructure like electricity, internet, and computing power—resources many regions lack." 🗣️ "Even if everyone had equal internet access, a one-size-fits-all approach to technology wouldn’t work due to local contexts and different needs." 🗣️ "Traditional knowledge should be exempt from broad data mining rights, allowing communities to explicitly give or revoke permissions for its use in AI training." 🗣️ "We need public AI infrastructures that ensure diversity and regional perspectives while maintaining communities’ agency over their contributions." 🗣️ "To prevent lopsided development, policies must go beyond tools like preference signals and address broader governance and societal frameworks." 📌 About Our Guest 🎙️ Anna Tumadóttir | Creative Commons 🌐 Article | Questions for Consideration on AI & the Commons https://creativecommons.org/2024/07/24/preferencesignals/  🌐 Anna Tumadóttir  https://creativecommons.org/person/annacreativecommons-org/ Anna is the CEO of Creative Commons, an international nonprofit organization that empowers people to grow and sustain the thriving commons of shared knowledge and culture. #AI #ArtificialIntelligence #GenerativeAI

    20 min
  8. AI lab TL;DR | Carys J. Craig - The Copyright Trap and AI Policy

    JAN 27 · BONUS

    AI lab TL;DR | Carys J. Craig - The Copyright Trap and AI Policy

    🔍 In this TL;DR episode, Carys J Craig (Osgoode Professional Development) explains the "copyright trap" in AI regulation, where relying on copyright favors corporate interests over creativity. She challenges misconceptions about copying and property rights, showing how this approach harms innovation and access. Carys offers alternative ways to protect human creativity without falling into this trap. 📌 TL;DR Highlights ⏲️[00:00] Intro ⏲️[00:46] Q1-What is the "Copyright Trap," and why could it harm AI and creativity? ⏲️[10:05] Q2-Can you explain the three routes that lead into the copyright trap and their relevance to AI? ⏲️[22:08] Q3-What alternatives should policymakers consider to protect creators and manage AI? ⏲️[28:45] Wrap-up & Outro 💭 Q1 - What is the "Copyright Trap," and why could it harm AI and creativity? 🗣️ “To turn to copyright law is to turn to really a false friend. The idea that copyright is going to be our friend, is going to help us in this situation,(...) it's likely to do more harm than good." 🗣️ “We are imagining increasingly in these policy debates that copyright and protection of copyright owners will be a kind of counterweight to corporate power and to the sort of extractive logics of Big Tech and AI development. I think that that is misguided. And in fact, we're playing into the interests of both the entertainment industries and big tech ” 🗣️ "When we run into the copyright trap, this sort of conviction that copyright is going to be the right regulatory tool, we are sort of defining how this technology is going to evolve in a way that I think will backfire and will actually undermine the political objectives of those who are pointing to the inequities and the unfairness behind the technology and the way that it's being developed.” 🗣️ "AI industry, big tech industry and the creative industry stakeholders are all, I think, perfectly happy to approach these larger policy questions through the sort of logic of copyright, sort of proprietary logic of ownership, control, exchange in the free market, licencing structures that we're already seeing taking hold" 🗣️ "What we're going to see, I think, if we run into the copyright trap is that certainly smaller developers, but really everyone will be training the technology on incomplete data sets, the data sets that reflect the sort of big packaged data products that have been exchanged for value between the main market actors. So that's going to lessen the quality really of what's going in generally by making it more exclusive and less inclusive." 💭 Q2 - Can you explain the three routes that lead into the copyright trap and their relevance to AI? 🗣️ ""The first route that I identify is what's sometimes called the if-value-then-right fallacy. So that's the assumption that if something has value, then there should be or must be some right over it.“ 🗣️ "Because something has value, whether economic or social, doesn't mean we should turn it into property that can be owned and controlled through these exclusive rights that we find in copyright law."  🗣️ "The second route that I identify is a sort of obsession with copying and the idea that copying is inherently just a wrongful activity. (...) The reality is that there's nothing inherently wrongful about copying. And in fact, this is how we learn. This is how we create. 🗣️ "One of the clearest routes into the copyright trap is saying, well, you know, you have to make copies of texts in order to train AI. So of course, copyright is implicated. And of course, we have to prevent that from happening without permission.. (...) But our obsession with the individuated sort of discrete copies of works behind the scenes is now an anachronism that we really need to let go.” 🗣️ "Using the figure of the artist as a reason to expand copyright control, and assuming that that's going to magically turn into lining the pockets of artists and creators seems to me to be a fallacy and a route into the copyright trap." 💭 Q3 - Why is output-based remuneration better for creators, AI developers, and society? 🗣️ "The health of our cultural environment (..) [should be] the biggest concern and not simply or only protecting creators as a separate class of sort of professional actors." 🗣️ "I think what we could do is shift our copyright policy focus to protecting and encouraging human authorship by refusing to protect AI generated outputs. 🗣️ "If the outputs of generative AI are substantially similar to works on which the AI was trained, then those are infringing outputs and copyright law will apply to them such that to distribute those infringing copies would produce liability under the system as it currently exists.“ 🗣️ "There are privacy experts who might be much better placed to say how should we curate or ensure that we regulate the data on which the machines are trained and I would be supportive of those kinds of interventions at the input stage. 🗣️ “Copyright seems like a tempting way to do it but that's not what it does. And so maybe rather than some of the big collective licencing solutions that are being imagined in this context, we'd be better off thinking about tax solutions, where we properly tax big tech and then we use that tax in a way that actually supports the things that we as a society care about, including funding culture and the arts." 📌 About Our Guest 🎙️ Carys J Craig | Osgoode Hall Law School 🌐 Article | The AI-Copyright Trap https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4905118  🌐 Carys J Craig https://www.osgoode.yorku.ca/faculty-and-staff/craig-carys-j/   Carys is the Academic Director of the Osgoode Professional Development LLM Program in Intellectual Property Law, and recently served as Osgoode’s Associate Dean. A recipient of multiple teaching awards, Carys researches and publishes widely on intellectual property law and policy, with an emphasis on authorship, users’ rights and the public domain. #AI #ArtificialIntelligence #GenerativeAI

    29 min

About

AI lab podcast, "decrypting" expert analysis to understand Artificial Intelligence from a policy making point of view.