LessWrong (Curated & Popular)

LessWrong

Audio narrations of LessWrong posts. Includes all curated posts and all posts with 125+ karma.If you'd like more, subscribe to the “Lesswrong (30+ karma)” feed.

  1. 1H AGO

    "Product Alignment is not Superintelligence Alignment (and we need the latter to survive)" by plex

    tl;dr: progress on making Claude friendly[1] is not the same as progress on making it safe to build godlike superintelligence. solving the former does not imply we get a good future.[2] please track the difference. The term Alignment was coined[3] to point to the technical problem of understanding how to build minds such that if they were to become strongly and generally superhuman, things would go well. It has been increasingly adopted by frontier AI labs and much of the rest of the AI safety community to mean a much easier challenge, something like "having AIs that are empirically doing approximately what you ask them to do".[4] If it's possible to use an intent-aligned product to build a research system which discovers a new paradigm and breaks your guardrails, then it is not Aligned in the original sense. If you can use your intent aligned system to write code which jailbreaks other LLMs and enables them to do dangerous ML research, it is also not Aligned in the original sense. Conflating progress on product alignment with progress on superintelligence alignment seems to be lulling much of the AI safety community into a false sense of security. Why is Superintelligence [...] --- Outline: (01:18) Why is Superintelligence Alignment less prominent? (02:21) Why do we need Superintelligence Alignment to survive? The original text contained 10 footnotes which were omitted from this narration. --- First published: March 31st, 2026 Source: https://www.lesswrong.com/posts/mrwYCNocXCP2hrWt8/product-alignment-is-not-superintelligence-alignment-and-we --- Narrated by TYPE III AUDIO.

    4 min
  2. 2D AGO

    "Some things I noticed while LARPing as a grantmaker" by Zach Stein-Perlman

    Written to a new grantmaker. Most value comes from finding/creating projects many times your bar, rather than discriminating between opportunities around your bar. If you find/create a new opportunity to donate $1M at 10x your bar (and cause it to get $1M, which would otherwise be donated to a 1x thing), you generate $9M of value (at your bar).[1] If you cause a $1M at 1.5x opportunity to get funded or a $1M at 0.5x opportunity to not get funded, you generate $500K of value. The former is 18 times as good. You should probably be like I do research to figure out what projects should exist, then make them exist rather than I evaluate the applications that come to me. That said, most great ideas come from your network, not from your personal brainstorming. In some buckets, the low-hanging fruit will be plucked. In others, nobody's on the ball and amazing opportunities get dropped. If you're working in a high-value bucket where nobody's on the ball, tons of alpha is on the table. (Assuming enough donors or grantmakers will listen to you to fund your best stuff.) I talk about "10x opportunities" and "1x opportunities" for simplicity here. It [...] The original text contained 1 footnote which was omitted from this narration. --- First published: March 23rd, 2026 Source: https://www.lesswrong.com/posts/CzoiqGzpShprcv2Jd/some-things-i-noticed-while-larping-as-a-grantmaker --- Narrated by TYPE III AUDIO.

    12 min
  3. MAR 25

    "The Case for Low-Competence ASI Failure Scenarios" by Ihor Kendiukhov

    I think the community underinvests in the exploration of extremely-low-competence AGI/ASI failure modes and explain why. Humanity's Response to the AGI Threat May Be Extremely Incompetent There is a sufficient level of civilizational insanity overall and a nice empirical track record in the field of AI itself which is eloquent about its safety culure. For example: At OpenAI, a refactoring bug flipped the sign of the reward signal in a model. Because labelers had been instructed to give very low ratings to sexually explicit text, the bug pushed the model into generating maximally explicit content across all prompts. The team noticed only after the training run had completed, because they were asleep. The director of alignment at Meta's Superintelligence Labs connected an OpenClaw agent to her real email, at which point it began deleting messages despite her attempts to stop it, and she ended up running to her computer to manually halt the process. An internal AI agent at Meta posted an answer publicly without approval; another employee acted on the inaccurate advice, triggering a severe security incident that temporarily allowed employees to access sensitive data they were not authorized to view. AWS acknowledged that [...] --- Outline: (00:19) Humanitys Response to the AGI Threat May Be Extremely Incompetent (02:26) Many Existing Scenarios and Case Studies Assume (Relatively) High Competence (04:31) Dumb Ways to Die (07:31) Undignified AGI Disaster Scenarios Deserve More Careful Treatment (10:43) Why This Might Be Useful --- First published: March 19th, 2026 Source: https://www.lesswrong.com/posts/t9LAhjoBnpQBa8Bbw/the-case-for-low-competence-asi-failure-scenarios --- Narrated by TYPE III AUDIO.

    12 min

Ratings & Reviews

4.8
out of 5
12 Ratings

About

Audio narrations of LessWrong posts. Includes all curated posts and all posts with 125+ karma.If you'd like more, subscribe to the “Lesswrong (30+ karma)” feed.

You Might Also Like