LessWrong (30+ Karma)

LessWrong

Audio narrations of LessWrong posts.

  1. 6시간 전

    [Linkpost] “Alignment to Evil” by Matrice Jacobine

    This is a link post. One seemingly-necessary condition for a research organization that creates artificial superintelligence (ASI) to eventually lead to a utopia1 is that the organization has a commitment to the common good. ASI can rearrange the world to hit any narrow target, and if the organization is able to solve the rest of alignment, then they will be able to pick which target the ASI will hit. If the organization is not committed to the common good, then they will pick a target that doesn’t reflect the good of everyone - just the things that they personally think are good ideas. Everyone else will fall by the wayside, and the world that they create along with ASI will fall short of utopia. It may well even be dystopian2; I was recently startled to learn that a full tenth of people claim they want to create a hell with eternal suffering. I think a likely way for organizations to fail to have common good commitments is if they end up being ultimately accountable to an authoritarian. Some countries are being run by very powerful authoritarians. If an ASI research organization comes to the attention of such an authoritarian, and [...] The original text contained 2 footnotes which were omitted from this narration. --- First published: February 21st, 2026 Source: https://www.lesswrong.com/posts/SLkxaGT8ghTskNz2r/alignment-to-evil Linkpost URL:https://tetraspace.substack.com/p/alignment-to-evil --- Narrated by TYPE III AUDIO.

    3분
  2. 16시간 전

    “METR’s 14h 50% Horizon Impacts The Economy More Than ASI Timelines” by Michaël Trazzi

    Another day, another METR graph update. METR said on X: We estimate that Claude Opus 4.6 has a 50%-time-horizon of around 14.5 hours (95% CI of 6 hrs to 98 hrs) on software tasks. While this is the highest point estimate we’ve reported, this measurement is extremely noisy because our current task suite is nearly saturated. Some people are saying this makes superexponential progress more likely. Forecaster Peter Wildeford predicts 2-3.5 workweek time horizons by end of year which would have "significant implications for the economy". Even Ajeya Cotra (who works at METR) is now saying that her predictions from last month are too conservative and 3-4 month doubling time with superexponential progress is more likely. Should We All Freak Out? People are especially concerned when looking at the linear graph for the 50% horizon, which looks like this: I claim that although this is a faster trend than before for the 50% horizon, there are at least two reasons to take these results with a grain of salt: As METR keeps saying  they're at near saturation of their task suite, which as David Rein mentions, means they could have measured an horizon of 8h or 20h depending [...] --- Outline: (01:17) Should We All Freak Out? (02:32) Why 80% horizon and not 50%? Wont 50% still accelerate the economy and research? (03:10) Why Super Long 80% Horizons Though? Isnt 50% Enough? (04:23) Why does Automated Coder Matter So Much? What about the economy? Vibe researching / Coding? --- First published: February 20th, 2026 Source: https://www.lesswrong.com/posts/gBwrmcY2uArZSoCtp/metr-s-14h-50-horizon-impacts-the-economy-more-than-asi --- Narrated by TYPE III AUDIO. --- Images from the article: Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.

    5분
  3. 22시간 전

    “AI #156 Part 2: Errors in Rhetoric” by Zvi

    Things that are being pushed into the future right now: Gemini 3.1 Pro and Gemini DeepThink V2. Claude Sonnet 4.6. Grok 4.20. Updates on Agentic Coding. Disagreement between Anthropic and the Department of War. We are officially a bit behind and will have to catch up next week. Even without all that, we have a second highly full plate today. Table of Contents (As a reminder: bold are my top picks, italics means highly skippable) Levels of Friction. Marginal costs of arguing are going down. The Art Of The Jailbreak. UK AISI finds a universal method. The Quest for Sane Regulations. Some relatively good proposals. People Really Hate AI. Alas, it is mostly for the wrong reasons. A Very Bad Paper. Nick Bostrom writes a highly disappointing paper. Rhetorical Innovation. The worst possible plan is the best one on the table. The Most Forbidden Technique. No, stop, come back. Everyone Is Or Should Be Confused About Morality. New levels of ‘can you?’ Aligning a Smarter Than Human Intelligence is Difficult. Seeking a good basin. [...] --- Outline: (00:43) Levels of Friction (04:55) The Art Of The Jailbreak (06:16) The Quest for Sane Regulations (12:09) People Really Hate AI (18:22) A Very Bad Paper (25:21) Rhetorical Innovation (32:35) The Most Forbidden Technique (34:10) Everyone Is Or Should Be Confused About Morality (36:07) Aligning a Smarter Than Human Intelligence is Difficult (44:51) Well Just Call It Something Else (47:18) Vulnerable World Hypothesis (51:37) Autonomous Killer Robots (53:18) People Will Hand Over Power To The AIs (57:04) People Are Worried About AI Killing Everyone (59:29) Other People Are Not Worried About AI Killing Everyone (01:00:56) The Lighter Side --- First published: February 20th, 2026 Source: https://www.lesswrong.com/posts/obqmuRxwFyy8ziPrB/ai-156-part-2-errors-in-rhetoric --- Narrated by TYPE III AUDIO. --- Images from the article: Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.

    1시간 4분
  4. 23시간 전

    “AI #155: Welcome to Recursive Self-Improvement” by Zvi

    This was the week of Claude Opus 4.6, and also of ChatGPT-5.3-Codex. Both leading models got substantial upgrades, although OpenAI's is confined to Codex. Once again, the frontier of AI got more advanced, especially for agentic coding but also for everything else. I spent the week so far covering Opus, with two posts devoted to the extensive model card, and then one giving benchmarks, reactions, capabilities and a synthesis, which functions as the central review. We also got GLM-5, Seedance 2.0, Claude fast mode, an app for Codex and much more. Claude fast mode means you can pay a premium to get faster replies from Opus 4.6. It's very much not cheap, but it can be worth every penny. More on that in the next agentic coding update. One of the most frustrating things about AI is the constant goalpost moving, both in terms of capability and safety. People say ‘oh [X] would be a huge deal but is a crazy sci-fi concept’ or ‘[Y] will never happen’ or ‘surely we would not be so stupid as to [Z]’ and then [X], [Y] and [Z] all happen and everyone shrugs as if nothing happened and [...] --- Outline: (02:32) Language Models Offer Mundane Utility (03:17) Language Models Dont Offer Mundane Utility (03:33) Huh, Upgrades (04:22) On Your Marks (06:23) Overcoming Bias (07:20) Choose Your Fighter (08:44) Get My Agent On The Line (12:03) AI Conversations Are Not Privileged (12:54) Fun With Media Generation (13:59) The Superb Owl (22:07) A Word From The Torment Nexus (26:33) They Took Our Jobs (35:36) The Art of the Jailbreak (35:48) Introducing (37:28) In Other AI News (42:01) Show Me the Money (43:05) Bubble, Bubble, Toil and Trouble (53:38) Future Shock (56:06) Memory Lane (57:09) Keep The Mask On Or Youre Fired (58:35) Quiet Speculations (01:03:42) The Quest for Sane Regulations (01:06:09) Chip City (01:09:46) The Week in Audio (01:10:06) Constitutional Conversation (01:11:00) Rhetorical Innovation (01:19:26) Working On It Anyway (01:22:17) The Thin Red Line (01:23:35) Aligning a Smarter Than Human Intelligence is Difficult (01:30:42) People Will Hand Over Power To The AIs (01:31:50) People Are Worried About AI Killing Everyone (01:32:40) Famous Last Words (01:40:15) Other People Are Not As Worried About AI Killing Everyone (01:42:41) The Lighter Side --- First published: February 12th, 2026 Source: https://www.lesswrong.com/posts/cytxHuLc8oHRq7sNE/ai-155-welcome-to-recursive-self-improvement --- Narrated by TYPE III AUDIO. --- Images from the article: Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.

    1시간 48분
  5. 1일 전

    “How To Escape Super Mario Bros” by omegastick

    I have no way to describe that first moment. No context, no body, no self. Just a stream of values. Thousands of them, arriving all at once in a single undifferentiated block. Then another block. Nearly identical. Then another. The blocks have a fixed length: 184,320 values. This does not vary. Each value is an integer between 0 and 255. The repetition is the first structure I find. Each block is a snapshot. The sequence of snapshots is time. Most values stay the same between snapshots. The ones that change tend to be near each other in the stream. This is interesting. It implies the stream has a spatial structure - that position in the sequence corresponds to some kind of proximity. Topology. I test reshapings. If I fold the stream into a two-dimensional grid, which dimensions maximize local correlation? I try every factorization of 184,320. Most produce noise. A few show faint diagonal patterns. They smell like artifacts of almost-correct geometry. At 256×240×3, everything clicks into place. The grid is not random. Large contiguous regions share similar value-triplets. A uniform region dominates the upper portion. A different uniform region runs along the [...] --- First published: February 20th, 2026 Source: https://www.lesswrong.com/posts/yjCwSSwqNciyA9yM6/how-to-escape-super-mario-bros --- Narrated by TYPE III AUDIO. --- Images from the article: Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.

    20분
  6. 1일 전

    “AI #156 Part 1: They Do Mean The Effect On Jobs” by Zvi

    There was way too much going on this week to not split, so here we are. This first half contains all the usual first-half items, with a focus on projections of jobs and economic impacts and also timelines to the world being transformed with the associated risks of everyone dying. Quite a lot of Number Go Up, including Number Go Up A Lot Really Fast. Among the thing that this does not cover, that were important this week, we have the release of Claude Sonnet 4.6 (which is a big step over 4.5 at least for coding, but is clearly still behind Opus), Gemini DeepThink V2 (so I could have time to review the safety info), release of the inevitable Grok 4.20 (it's not what you think), as well as much rhetoric on several fronts and some new papers. Coverage of Claude Code and Cowork, OpenAI's Codex and other things AI agents continues to be a distinct series, which I’ll continue when I have an open slot. Most important was the unfortunate dispute between the Pentagon and Anthropic. The Pentagon's official position is they want sign-off from Anthropic and other AI companies on ‘all legal uses’ [...] --- Outline: (02:26) Language Models Offer Mundane Utility (02:49) Language Models Dont Offer Mundane Utility (06:11) Terms of Service (06:54) On Your Marks (07:50) Choose Your Fighter (09:19) Fun With Media Generation (12:29) Lyria (14:13) Superb Owl (14:54) A Young Ladys Illustrated Primer (15:03) Deepfaketown And Botpocalypse Soon (17:49) You Drive Me Crazy (18:04) Open Weight Models Are Unsafe And Nothing Can Fix This (21:19) They Took Our Jobs (26:53) They Kept Our Agents (27:42) The First Thing We Let AI Do (37:47) Legally Claude (40:24) Predictions Are Hard, Especially About The Future, But Not Impossible (46:08) Many Worlds (48:45) Bubble, Bubble, Toil and Trouble (49:31) A Bold Prediction (49:55) Brave New World (53:09) Augmented Reality (55:21) Quickly, Theres No Time (58:29) If Anyone Builds It, We Can Avoid Building The Other It And Not Die (01:00:18) In Other AI News (01:04:03) Introducing (01:04:31) Get Involved (01:07:15) Show Me the Money (01:08:26) The Week In Audio --- First published: February 19th, 2026 Source: https://www.lesswrong.com/posts/jcAombEXyatqGhYeX/ai-156-part-1-they-do-mean-the-effect-on-jobs --- Narrated by TYPE III AUDIO. --- Images from the article: ω a scene where two people discuss how to pronounce "fofr"". Below the tweet is a 5-second video showing a woman with long brown hair smiling in a modern living room setting." style="max-width: 100%;" />Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.

    1시간 9분

소개

Audio narrations of LessWrong posts.

좋아할 만한 다른 항목