AI可可AI生活

[人人能懂] 从并行思考、结构化学习到认知解密

想知道AI如何像开“诸葛亮会”一样解决难题,又为何连小学生的乘法都搞不定吗?本期节目,几篇最新的论文将带我们一窥AI大脑的内部运作:看它如何上演“分身思考”的头脑风暴,如何被我们的“偏见”变得无聊。更重要的是,我将告诉你一个解锁它隐藏创造力的简单“咒语”,并揭示为何在训练AI时,不能只看“平均分”。

00:00:29 让AI更聪明的秘密:不是想得更长,而是想得更巧

00:06:28 分身思考:AI的大脑里如何开一场头脑风暴

00:11:13 为什么聪明的AI,竟学不会小学生的乘法?

00:18:03 为什么AI越来越无聊?换个问法,解锁它的隐藏技能

00:22:36 AI训练揭秘:你真的懂“平均”吗?

本期介绍的几篇论文:

[LG] Rethinking Thinking Tokens: LLMs as Improvement Operators  

[Meta Superintelligence Labs & Anthropic]  

https://arxiv.org/abs/2510.01123  

---

[LG] Thoughtbubbles: an Unsupervised Method for Parallel Thinking in Latent Space  

[Stanford University]  

https://arxiv.org/abs/2510.00219  

---

[LG] Why Can't Transformers Learn Multiplication? Reverse-Engineering Reveals Long-Range Dependency Pitfalls  

[University of Chicago & MIT & University of Waterloo]  

https://arxiv.org/abs/2510.00184  

---

[CL] Verbalized Sampling: How to Mitigate Mode Collapse and Unlock LLM Diversity  

[Northeastern University & Stanford University]  

https://arxiv.org/abs/2510.01171  

[LG] Per-example gradients: a new frontier for understanding and improving optimizers  

[Google Deepmind]  

https://arxiv.org/abs/2510.00236  

---