Data Science at Home

Francesco Gadaleta
Data Science at Home

Artificial Intelligence, algorithms and tech tales that are shaping the world. Hype not included.

  1. 2 DAYS AGO

    Humans vs. Bots: Are You Talking to a Machine Right Now? (Ep. 273)

    In this episode of Data Science at Home, host Francesco Gadaleta dives deep into the evolving world of AI-generated content detection with experts Souradip Chakraborty, Ph.D. grad student at the University of Maryland, and Amrit Singh Bedi, CS faculty at the University of Central Florida.  Together, they explore the growing importance of distinguishing human-written from AI-generated text, discussing real-world examples from social media to news. How reliable are current detection tools like DetectGPT? What are the ethical and technical challenges ahead as AI continues to advance? And is the balance between innovation and regulation tipping in the right direction?    Tune in for insights on the future of AI text detection and the broader implications for media, academia, and policy.   Chapters    00:00 - Intro  00:23 - Guests: Souradip Chakraborty and Amrit Singh Bedi  01:25 - Distinguish Text Generation By AI  04:33 - Research on Safety and Alignment of Generative Model  06:01 - Tools to Detect Generated AI Text   11:28 - Water Marking 18:27 - Challenges in Detecting Large Documents Generated by AI  23:34 - Number of Tokens  26:22 - Adversarial Attack 29:01 - True Positive and False Positive of Detectors  31:01 - Limit of Technologies  41:01 - Future of AI Detection Techniques  46:04 - Closing Thought   Subscribe to our new YouTube channel https://www.youtube.com/@DataScienceatHome

    50 min
  2. 13 NOV

    AI vs. The Planet: The Energy Crisis Behind the Chatbot Boom (Ep. 271)

    In this episode of Data Science at Home, we dive into the hidden costs of AI’s rapid growth — specifically, its massive energy consumption. With tools like ChatGPT reaching 200 million weekly active users, the environmental impact of AI is becoming impossible to ignore. Each query, every training session, and every breakthrough come with a price in kilowatt-hours, raising questions about AI’s sustainability.   Join us, as we uncovers the staggering figures behind AI's energy demands and explores practical solutions for the future. From efficiency-focused algorithms and specialized hardware to decentralized learning, this episode examines how we can balance AI’s advancements with our planet's limits. Discover what steps we can take to harness the power of AI responsibly!   Check our new YouTube channel at https://www.youtube.com/@DataScienceatHome   Chapters 00:00 - Intro 01:25 - Findings on Summary Statics 05:15 - Energy Required To Querry On GPT 07:20 - Energy Efficiency In BlockChain 10:41 - Efficicy Focused Algorithm 14:02 - Hardware Optimization 17:31 - Decentralized Learning 18:38 - Edge Computing with Local Inference 19:46 - Distributed Architectures 21:46 - Outro     #AIandEnergy #AIEnergyConsumption #SustainableAI #AIandEnvironment #DataScience #EfficientAI #DecentralizedLearning #GreenTech #EnergyEfficiency #MachineLearning #FutureOfAI #EcoFriendlyAI #FrancescoFrag #DataScienceAtHome #ResponsibleAI #EnvironmentalImpact

    22 min
  3. 6 NOV

    Love, Loss, and Algorithms: The Dangerous Realism of AI (Ep. 270)

    Subscribe to our new channel https://www.youtube.com/@DataScienceatHome   In this episode of Data Science at Home, we confront a tragic story highlighting the ethical and emotional complexities of AI technology. A U.S. teenager recently took his own life after developing a deep emotional attachment to an AI chatbot emulating a character from Game of Thrones. This devastating event has sparked urgent discussions on the mental health risks, ethical responsibilities, and potential regulations surrounding AI chatbots, especially as they become increasingly lifelike.   🎙️ Topics Covered: AI & Emotional Attachment: How hyper-realistic AI chatbots can foster intense emotional bonds with users, especially vulnerable groups like adolescents. Mental Health Risks: The potential for AI to unintentionally contribute to mental health issues, and the challenges of diagnosing such impacts. Ethical & Legal Accountability: How companies like Character AI are being held accountable and the ethical questions raised by emotionally persuasive AI.   🚨 Analogies Explored: From VR to CGI and deepfakes, we discuss how hyper-realism in AI parallels other immersive technologies and why its emotional impact can be particularly disorienting and even harmful.   🛠️ Possible Mitigations: We cover potential solutions like age verification, content monitoring, transparency in AI design, and ethical audits that could mitigate some of the risks involved with hyper-realistic AI interactions. 👀 Key Takeaways: As AI becomes more realistic, it brings both immense potential and serious responsibility. Join us as we dive into the ethical landscape of AI—analyzing how we can ensure this technology enriches human lives without crossing lines that could harm us emotionally and psychologically. Stay curious, stay critical, and make sure to subscribe for more no-nonsense tech talk!   Chapters 00:00 - Intro 02:21 - Emotions In Artificial Intelligence 04:00 - Unregulated Influence and Misleading Interaction 06:32 - Overwhelming Realism In AI 10:54 - Virtual Reality 13:25 - Hyper-Realistic CGI Movies 15:38 - Deep Fake Technology 18:11 - Regulations To Mitigate AI Risks 22:50 - Conclusion   #AI#ArtificialIntelligence#MentalHealth#AIEthics#podcast#AIRegulation#EmotionalAI#HyperRealisticAI#TechTalk#AIChatbots#Deepfakes#VirtualReality#TechEthics#DataScience#AIDiscussion #StayCuriousStayCritical

    24 min

About

Artificial Intelligence, algorithms and tech tales that are shaping the world. Hype not included.

You Might Also Like

To listen to explicit episodes, sign in.

Stay up to date with this show

Sign in or sign up to follow shows, save episodes and get the latest updates.

Select a country or region

Africa, Middle East, and India

Asia Pacific

Europe

Latin America and the Caribbean

The United States and Canada