34 avsnitt

Technical discussions with deep learning researchers who study how to build intelligence. Made for researchers, by researchers.

Generally Intelligent Kanjun Qiu

    • Teknologi

Technical discussions with deep learning researchers who study how to build intelligence. Made for researchers, by researchers.

    Episode 34: Seth Lazar, Australian National University: On legitimate power, moral nuance, and the political philosophy of AI

    Episode 34: Seth Lazar, Australian National University: On legitimate power, moral nuance, and the political philosophy of AI

    Seth Lazar is a professor of philosophy at the Australian National University, where he leads the Machine Intelligence and Normative Theory (MINT) Lab. His unique perspective bridges moral and political philosophy with AI, introducing much-needed rigor to the question of what will make for a good and just AI future.

    Generally Intelligent is a podcast by Imbue where we interview researchers about their behind-the-scenes ideas, opinions, and intuitions that are hard to share in papers and talks.

    About ImbueImbue is an independent research company developing AI agents that mirror the fundamentals of human-like intelligence and that can learn to safely solve problems in the real world. We started Imbue because we believe that software with human-level intelligence will have a transformative impact on the world. We’re dedicated to ensuring that that impact is a positive one.We have enough funding to freely pursue our research goals over the next decade, and our backers include Y Combinator, researchers from OpenAI, Astera Institute, and a number of private individuals who care about effective altruism and scientific research.Our research is focused on agents for digital environments (ex: browser, desktop, documents), using RL, large language models, and self supervised learning. We’re excited about opportunities to use simulated data, network architecture search, and good theoretical understanding of deep learning to make progress on these problems. We take a focused, engineering-driven approach to research.

    Website: https://imbue.com/LinkedIn: https://www.linkedin.com/company/imbue-ai/Twitter: @imbue_ai

    • 1 tim. 55 min
    Episode 33: Tri Dao, Stanford: On FlashAttention and sparsity, quantization, and efficient inference

    Episode 33: Tri Dao, Stanford: On FlashAttention and sparsity, quantization, and efficient inference

    Tri Dao is a PhD student at Stanford, co-advised by Stefano Ermon and Chris Re. He’ll be joining Princeton as an assistant professor next year. He works at the intersection of machine learning and systems, currently focused on efficient training and long-range context.



    About Generally Intelligent 

    We started Generally Intelligent because we believe that software with human-level intelligence will have a transformative impact on the world. We’re dedicated to ensuring that that impact is a positive one.  

    We have enough funding to freely pursue our research goals over the next decade, and our backers include Y Combinator, researchers from OpenAI, Astera Institute, and a number of private individuals who care about effective altruism and scientific research.  

    Our research is focused on agents for digital environments (ex: browser, desktop, documents), using RL, large language models, and self supervised learning. We’re excited about opportunities to use simulated data, network architecture search, and good theoretical understanding of deep learning to make progress on these problems. We take a focused, engineering-driven approach to research.  



    Learn more about us

    Website: https://generallyintelligent.com/

    LinkedIn: linkedin.com/company/generallyintelligent/ 

    Twitter: @genintelligent

    • 1 tim. 20 min
    Episode 32: Jamie Simon, UC Berkeley: On theoretical principles for how neural networks learn and generalize

    Episode 32: Jamie Simon, UC Berkeley: On theoretical principles for how neural networks learn and generalize

    Jamie Simon is a 4th year Ph.D. student at UC Berkeley advised by Mike DeWeese, and also a Research Fellow with us at Generally Intelligent. He uses tools from theoretical physics to build fundamental understanding of deep neural networks so they can be designed from first-principles. In this episode, we discuss reverse engineering kernels, the conservation of learnability during training, infinite-width neural networks, and much more.


    About Generally Intelligent 

    We started Generally Intelligent because we believe that software with human-level intelligence will have a transformative impact on the world. We’re dedicated to ensuring that that impact is a positive one.  

    We have enough funding to freely pursue our research goals over the next decade, and our backers include Y Combinator, researchers from OpenAI, Astera Institute, and a number of private individuals who care about effective altruism and scientific research.  

    Our research is focused on agents for digital environments (ex: browser, desktop, documents), using RL, large language models, and self supervised learning. We’re excited about opportunities to use simulated data, network architecture search, and good theoretical understanding of deep learning to make progress on these problems. We take a focused, engineering-driven approach to research.  



    Learn more about us

    Website: https://generallyintelligent.com/

    LinkedIn: linkedin.com/company/generallyintelligent/ 

    Twitter: @genintelligent

    • 1 tim. 1 min.
    Episode 31: Bill Thompson, UC Berkeley, on how cultural evolution shapes knowledge acquisition

    Episode 31: Bill Thompson, UC Berkeley, on how cultural evolution shapes knowledge acquisition

    Bill Thompson is a cognitive scientist and an assistant professor at UC Berkeley. He runs an experimental cognition laboratory where he and his students conduct research on human language and cognition using large-scale behavioral experiments, computational modeling, and machine learning. In this episode, we explore the impact of cultural evolution on human knowledge acquisition, how pure biological evolution can lead to slow adaptation and overfitting, and much more.



    About Generally Intelligent 

    We started Generally Intelligent because we believe that software with human-level intelligence will have a transformative impact on the world. We’re dedicated to ensuring that that impact is a positive one.  

    We have enough funding to freely pursue our research goals over the next decade, and our backers include Y Combinator, researchers from OpenAI, Astera Institute, and a number of private individuals who care about effective altruism and scientific research.  

    Our research is focused on agents for digital environments (ex: browser, desktop, documents), using RL, large language models, and self supervised learning. We’re excited about opportunities to use simulated data, network architecture search, and good theoretical understanding of deep learning to make progress on these problems. We take a focused, engineering-driven approach to research.  



    Learn more about us

    Website: https://generallyintelligent.com/

    LinkedIn: linkedin.com/company/generallyintelligent/ 

    Twitter: @genintelligent

    • 1 tim. 15 min
    Episode 30: Ben Eysenbach, CMU, on designing simpler and more principled RL algorithms

    Episode 30: Ben Eysenbach, CMU, on designing simpler and more principled RL algorithms

    Ben Eysenbach is a PhD student from CMU and a student researcher at Google Brain. He is co-advised by Sergey Levine and Ruslan Salakhutdinov and his research focuses on developing RL algorithms that get state-of-the-art performance while being more simple, scalable, and robust. Recent problems he’s tackled include long horizon reasoning, exploration, and representation learning. In this episode, we discuss designing simpler and more principled RL algorithms, and much more.


    About Generally Intelligent 

    We started Generally Intelligent because we believe that software with human-level intelligence will have a transformative impact on the world. We’re dedicated to ensuring that that impact is a positive one.  

    We have enough funding to freely pursue our research goals over the next decade, and our backers include Y Combinator, researchers from OpenAI, Astera Institute, and a number of private individuals who care about effective altruism and scientific research.  

    Our research is focused on agents for digital environments (ex: browser, desktop, documents), using RL, large language models, and self supervised learning. We’re excited about opportunities to use simulated data, network architecture search, and good theoretical understanding of deep learning to make progress on these problems. We take a focused, engineering-driven approach to research.  



    Learn more about us

    Website: https://generallyintelligent.com/

    LinkedIn: linkedin.com/company/generallyintelligent/ 

    Twitter: @genintelligent

    • 1 tim. 45 min
    Episode 29: Jim Fan, NVIDIA, on foundation models for embodied agents, scaling data, and why prompt engineering will become irrelevant

    Episode 29: Jim Fan, NVIDIA, on foundation models for embodied agents, scaling data, and why prompt engineering will become irrelevant

    Jim Fan is a research scientist at NVIDIA and got his PhD at Stanford under Fei-Fei Li. Jim is interested in building generally capable autonomous agents, and he recently published MineDojo, a massively multiscale benchmarking suite built on Minecraft, which was an Outstanding Paper at NeurIPS. In this episode, we discuss the foundation models for embodied agents, scaling data, and why prompt engineering will become irrelevant.  



    About Generally Intelligent 

    We started Generally Intelligent because we believe that software with human-level intelligence will have a transformative impact on the world. We’re dedicated to ensuring that that impact is a positive one.  

    We have enough funding to freely pursue our research goals over the next decade, and our backers include Y Combinator, researchers from OpenAI, Astera Institute, and a number of private individuals who care about effective altruism and scientific research.  

    Our research is focused on agents for digital environments (ex: browser, desktop, documents), using RL, large language models, and self supervised learning. We’re excited about opportunities to use simulated data, network architecture search, and good theoretical understanding of deep learning to make progress on these problems. We take a focused, engineering-driven approach to research.  



    Learn more about us

    Website: https://generallyintelligent.com/

    LinkedIn: linkedin.com/company/generallyintelligent/ 

    Twitter: @genintelligent

    • 1 tim. 26 min

Mest populära poddar inom Teknologi

Lex Fridman Podcast
Lex Fridman
SvD Tech brief
Svenska Dagbladet
Internetpionjärerna
Tele2
Elbilsveckan
Peter Esse & Christoffer Gullin
Acquired
Ben Gilbert and David Rosenthal
Darknet Diaries
Jack Rhysider

Du kanske också gillar

Latent Space: The AI Engineer Podcast — Practitioners talking LLMs, CodeGen, Agents, Multimodality, AI UX, GPU Infra and al
Alessio + swyx
Machine Learning Street Talk (MLST)
Machine Learning Street Talk (MLST)
The TWIML AI Podcast (formerly This Week in Machine Learning & Artificial Intelligence)
Sam Charrington
No Priors: Artificial Intelligence | Technology | Startups
Conviction | Pod People
Dwarkesh Podcast
Dwarkesh Patel
The AI Breakdown: Daily Artificial Intelligence News and Discussions
Nathaniel Whittemore