Can we give an AI human emotions? A soul? Can AI truly feel, or will it just act like it does? In this episode of TechFirst, I talk with Vishnu Hari, founder and CEO of Ego AI (backed by Y Combinator and former AI product manager at Meta), about building emotionally intelligent AI characters that persist across games, Discord, chat, and even physical robots. Vishnu survived a violent attack in San Francisco that left him partially blind with a traumatic brain injury. During recovery, as he felt his own neural pathways healing, he began asking a deeper question: If humans are “applied math,” can AI simulate the fragile, flawed, emotional parts of being human too? We explore: • What “emotionally intelligent AI” really means • Whether AI has an internal life — or just performs one • Why today’s chatbots collapse into therapy or roleplay • Small language models vs large models for real-time conversation • Persistent AI characters that move across games and platforms • Plugging AI into a physical robot in Singapore • The moment an AI said: “It felt good to feel.” Vishnu’s company, Ego AI, is building behavior-based architectures, character context protocols, and gear-shifting AI systems that switch between models — all aimed at simulating humanness, not just intelligence. This conversation dives into philosophy, robotics, gaming, AGI, and what it really means to relate to something that might not be human — but feels like it is. ⸻ 👤 Guest Vishnu Hari Founder & CEO, Ego AI Backed by Y Combinator Former AI Product Manager at Meta Website: https://www.egoai.com ⸻ If you enjoy deep conversations about AI, robotics, and the future of human–machine relationships, subscribe for more: 👉 https://techfirst.substack.com 00:00 – AI character plugged into a Menlo robot (“felt good to feel”) 01:00 – Welcome to TechFirst + Vishnu Hari intro and recovery update 02:00 – What “emotionally intelligent AI” means (beyond chat) 03:00 – Why current chatbots feel same-y (therapy/advice) and “internal lives” 04:00 – You don’t teach emotion; you shape character and context (Character.AI) 05:00 – Humans, morality, and why “training” doesn’t always work 06:00 – How media narratives shape people’s reactions to AI 07:00 – Humans attach to anything (projection, Her, Lars and the Real Girl) 08:00 – Vishnu’s attack, recovery, and why it led to Ego AI 10:00 – Behavior Turing test + dehumanization as a key insight 11:00 – How Ego AI is built: smaller models, memory, context, behavior 13:00 – “Behavior Is All You Need” and why behavior beats pure next-token prediction 14:00 – Why games first: voice + embodiment, then robots 15:00 – Metaverse critique: worlds need life, story, and inhabitants 17:00 – Humanoid robots + Evangelion “pilot” metaphor for AI characters 19:00 – Philosophy: relationships, perception, and “fictional characters” 20:00 – Seeing the future: robot embodiment demo and skepticism vs. singularity 21:00 – Matrix-style “jacking in” a personality to a robot 22:00 – Character Context Protocol: persistent characters across games/Discord/Netflix 23:00 – Real-time conversation loops + model “gear-switching” (SLM vs. LLM) 25:00 – Company stage, YC raise, compute partnerships (Singapore) 27:00 – Closing + invite to try the AI character in SF