想知道AI如何像开“诸葛亮会”一样解决难题,又为何连小学生的乘法都搞不定吗?本期节目,几篇最新的论文将带我们一窥AI大脑的内部运作:看它如何上演“分身思考”的头脑风暴,如何被我们的“偏见”变得无聊。更重要的是,我将告诉你一个解锁它隐藏创造力的简单“咒语”,并揭示为何在训练AI时,不能只看“平均分”。
00:00:29 让AI更聪明的秘密:不是想得更长,而是想得更巧
00:06:28 分身思考:AI的大脑里如何开一场头脑风暴
00:11:13 为什么聪明的AI,竟学不会小学生的乘法?
00:18:03 为什么AI越来越无聊?换个问法,解锁它的隐藏技能
00:22:36 AI训练揭秘:你真的懂“平均”吗?
本期介绍的几篇论文:
[LG] Rethinking Thinking Tokens: LLMs as Improvement Operators
[Meta Superintelligence Labs & Anthropic]
https://arxiv.org/abs/2510.01123
---
[LG] Thoughtbubbles: an Unsupervised Method for Parallel Thinking in Latent Space
[Stanford University]
https://arxiv.org/abs/2510.00219
---
[LG] Why Can't Transformers Learn Multiplication? Reverse-Engineering Reveals Long-Range Dependency Pitfalls
[University of Chicago & MIT & University of Waterloo]
https://arxiv.org/abs/2510.00184
---
[CL] Verbalized Sampling: How to Mitigate Mode Collapse and Unlock LLM Diversity
[Northeastern University & Stanford University]
https://arxiv.org/abs/2510.01171
[LG] Per-example gradients: a new frontier for understanding and improving optimizers
[Google Deepmind]
https://arxiv.org/abs/2510.00236
---
Информация
- Подкаст
- ЧастотаЕжедневно
- Опубликовано3 октября 2025 г. в 00:31 UTC
- Длительность29 мин.
- ОграниченияБез ненормативной лексики