LessWrong posts by zvi

zvi

Audio narrations of LessWrong posts by zvi

  1. -23 Ч

    “AI #156 Part 1: They Do Mean The Effect On Jobs” by Zvi

    There was way too much going on this week to not split, so here we are. This first half contains all the usual first-half items, with a focus on projections of jobs and economic impacts and also timelines to the world being transformed with the associated risks of everyone dying. Quite a lot of Number Go Up, including Number Go Up A Lot Really Fast. Among the thing that this does not cover, that were important this week, we have the release of Claude Sonnet 4.6 (which is a big step over 4.5 at least for coding, but is clearly still behind Opus), Gemini DeepThink V2 (so I could have time to review the safety info), release of the inevitable Grok 4.20 (it's not what you think), as well as much rhetoric on several fronts and some new papers. Coverage of Claude Code and Cowork, OpenAI's Codex and other things AI agents continues to be a distinct series, which I’ll continue when I have an open slot. Most important was the unfortunate dispute between the Pentagon and Anthropic. The Pentagon's official position is they want sign-off from Anthropic and other AI companies on ‘all legal uses’ [...] --- Outline: (02:26) Language Models Offer Mundane Utility (02:49) Language Models Dont Offer Mundane Utility (06:11) Terms of Service (06:54) On Your Marks (07:50) Choose Your Fighter (09:19) Fun With Media Generation (12:29) Lyria (14:13) Superb Owl (14:54) A Young Ladys Illustrated Primer (15:03) Deepfaketown And Botpocalypse Soon (17:49) You Drive Me Crazy (18:04) Open Weight Models Are Unsafe And Nothing Can Fix This (21:19) They Took Our Jobs (26:53) They Kept Our Agents (27:42) The First Thing We Let AI Do (37:47) Legally Claude (40:24) Predictions Are Hard, Especially About The Future, But Not Impossible (46:08) Many Worlds (48:45) Bubble, Bubble, Toil and Trouble (49:31) A Bold Prediction (49:55) Brave New World (53:09) Augmented Reality (55:21) Quickly, Theres No Time (58:29) If Anyone Builds It, We Can Avoid Building The Other It And Not Die (01:00:18) In Other AI News (01:04:03) Introducing (01:04:31) Get Involved (01:07:15) Show Me the Money (01:08:26) The Week In Audio --- First published: February 19th, 2026 Source: https://www.lesswrong.com/posts/jcAombEXyatqGhYeX/ai-156-part-1-they-do-mean-the-effect-on-jobs --- Narrated by TYPE III AUDIO. --- Images from the article: ω a scene where two people discuss how to pronounce "fofr"". Below the tweet is a 5-second video showing a woman with long brown hair smiling in a modern living room setting." style="max-width: 100%;" />Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.

    1 ч. 9 мин.
  2. -1 ДН.

    “Monthly Roundup #39: February 2026” by Zvi

    There really is a lot going on these days. I held off posting this because I was trying to see if I could write a net helpful post about the current situation involving Anthropic and the Pentagon. Anthropic very much wants to help DoW defend our country and make us strong. It is clear there have been some large misunderstandings here about how LLMs work. They are not ordinary tools like spreadsheets that automatically do whatever the user asks, nor would it be safe to make them so, nor do they predictably adhere to written rule sets or take instructions from their CEO in a crisis. And they are probabilistic. You do not and cannot get absolute guarantees. The only way to know if an AI model will do what you need in a crisis is something you needed to be do regardless of potential refusals, and which is also what you must do with human soldiers, which is to run the simulations and mock battles and drills and tests that tell you if the model can do and is willing to do the job. If there are irreconcilable differences and the military contract needs [...] --- Outline: (02:09) Bad News (04:03) Government Working (15:11) The Epstein Files (17:17) RIP Scott Adams (19:08) News You Cant Use But Click On Anyway (20:22) Were Putting Together A Team (20:47) You Cant Retire, I Quit (23:41) Jones Act Watch (33:24) Variously Effective Altruism (34:41) They Took Our Jobs And Now I Can Relax (35:39) While I Cannot Condone This (37:32) Good News, Everyone (39:19) Use Your One Time (42:34) Hands Off My Phone (45:14) Fun Theory (46:46) Good Advice (48:31) For Your Entertainment (51:07) Plur1bus (52:14) Gamers Gonna Game Game Game Game Game (57:58) Sports Go Sports (01:02:49) The Revolution of Retroactive Rising Expectations (01:04:15) I Was Promised Spying Cars (01:05:13) Prediction Market Madness (01:09:02) The Lighter Side --- First published: February 18th, 2026 Source: https://www.lesswrong.com/posts/3QPoEGfzHaywGDWKr/monthly-roundup-39-february-2026 --- Narrated by TYPE III AUDIO. --- Images from the article: Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.

    1 ч. 17 мин.
  3. -2 ДН.

    “On Dwarkesh Patel’s 2026 Podcast With Elon Musk and Other Recent Elon Musk Things” by Zvi

    Some podcasts are self-recommending on the ‘yep, I’m going to be breaking this one down’ level. This was one of those. So here we go. As usual for podcast posts, the baseline bullet points describe key points made, and then the nested statements are my commentary. Some points are dropped. If I am quoting directly I use quote marks, otherwise assume paraphrases. Normally I keep everything to numbered lists, but in several cases here it was more of a ‘he didn’t just say what I think he did did he’ and I needed extensive quotes. In addition to the podcast, there were some discussions around safety, or the lack thereof, at xAI, and Elon Musk went on what one can only describe as megatilt, including going hard after Anthropic's Amanda Askell. I will include that as a postscript. I will not include recent developments regarding Twitter, since that didn’t come up in the interview. I lead with a discussion of bounded distrust and how to epistemically consider Elon Musk, since that will be important throughout including in the postscript. What are the key takeaways? Elon Musk is more confused than [...] --- Outline: (02:56) Bounded Distrust (05:12) IN SPACE (09:56) The AI Will Follow You To Mars (22:32) xAI Business Plans (25:54) Optimus Prime (27:04) Beating China (30:02) SpaceX and How To Run a Company Elon Style (33:17) DOGE (35:29) TeraFab IN SPACE (35:47) Postscript: Safety Third at xAI (40:15) Elon Serves Back Saying That Which Is Not (42:51) Elons Army (43:55) Children Are Our Future (48:11) Where Do We Go From Here --- First published: February 17th, 2026 Source: https://www.lesswrong.com/posts/5yidbWsdWjNzWzLWZ/on-dwarkesh-patel-s-2026-podcast-with-elon-musk-and-other --- Narrated by TYPE III AUDIO. --- Images from the article: Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.

    49 мин.
  4. -3 ДН.

    “On Dwarkesh Patel’s 2026 Podcast With Dario Amodei” by Zvi

    Some podcasts are self-recommending on the ‘yep, I’m going to be breaking this one down’ level. This was very clearly one of those. So here we go. As usual for podcast posts, the baseline bullet points describe key points made, and then the nested statements are my commentary. Some points are dropped. If I am quoting directly I use quote marks, otherwise assume paraphrases. What are the main takeaways? Dario mostly stands by his predictions of extremely rapid advances in AI capabilities, both in coding and in general, and in expecting the ‘geniuses in a data center’ to show up within a few years, possibly even this year. Anthropic's actions do not seem to fully reflect this optimism, but also when things are growing on a 10x per year exponential if you overextend you die, so being somewhat conservative with investment is necessary unless you are prepared to fully burn your boats. Dario reiterated his stances on China, export controls, democracy, AI policy. The interview downplayed catastrophic and existential risk, including relative to other risks, although it was mentioned and Dario remains concerned. There was essentially no talk about alignment [...] --- Outline: (01:47) The Pace of Progress (08:56) Continual Learning (13:46) Does Not Compute (15:29) Step Two (22:58) The Quest For Sane Regulations (26:08) Beating China --- First published: February 16th, 2026 Source: https://www.lesswrong.com/posts/jWCy6owAmqLv5BB8q/on-dwarkesh-patel-s-2026-podcast-with-dario-amodei --- Narrated by TYPE III AUDIO. --- Images from the article: Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.

    29 мин.
  5. -6 ДН.

    “ChatGPT-5.3-Codex Is Also Good At Coding” by Zvi

    OpenAI is back with a new Codex model, released the same day as Claude Opus 4.6. The headline pitch is it combines the coding skills of GPT-5.2-Codex with the general knowledge and skills of other models, along with extra speed and improvements in the Codex harness, so that it can now handle your full stack agentic needs. We also got the Codex app for Mac, which is getting positive reactions, and quickly picked up a million downloads. CPT-5.3-Codex is only available inside Codex. It is not in the API. As usual, Anthropic's release was understated, basically a ‘here's Opus 4.6, a 212-page system card and a lot of benchmarks, it's a good model, sir, so have fun.’ Whereas OpenAI gave us a lot less words and a lot less benchmarks, while claiming their model was definitely the best. OpenAI: GPT-5.3-Codex is the most capable agentic coding model to date, combining the frontier coding performance of GPT-5.2-Codex with the reasoning and professional knowledge capabilities of GPT-5.2. This enables it to take on long-running tasks that involve research, tool use, and complex execution. Much like a colleague, you can steer and interact with GPT-5.3-Codex while [...] --- Outline: (01:50) The Overall Picture (03:00) Quickly, Theres No Time (04:15) System Card (04:49) AI Box Experiment (05:22) Maybe Cool It With Rm (07:02) Preparedness Framework (11:14) Glass Houses (12:16) OpenAI Appears To Have Violated SB 53 In a Meaningful Way (14:29) Safeguards They Did Implement (16:55) Misalignment Risks and Internal Deployment (18:38) The Official Pitch (24:28) Inception (26:12) Turn The Beat Around (27:35) Codex Does Cool Things (29:33) Positive Reactions (38:03) Negative Reactions (40:43) Codex of Ultimate Vibing --- First published: February 13th, 2026 Source: https://www.lesswrong.com/posts/CCDRjL7NZtNGtGheY/chatgpt-5-3-codex-is-also-good-at-coding --- Narrated by TYPE III AUDIO. --- Images from the article: Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.

    42 мин.
  6. 11 ФЕВР.

    “Claude Opus 4.6 Escalates Things Quickly” by Zvi

    Life comes at you increasingly fast. Two months after Claude Opus 4.5 we get a substantial upgrade in Claude Opus 4.6. The same day, we got GPT-5.3-Codex. That used to be something we’d call remarkably fast. It's probably the new normal, until things get even faster than that. Welcome to recursive self-improvement. Before those releases, I was using Claude Opus 4.5 and Claude Code for essentially everything interesting, and only using GPT-5.2 and Gemini to fill in the gaps or for narrow specific uses. GPT-5.3-Codex is restricted to Codex, so this means that for other purposes Anthropic and Claude have only extended the lead. This is the first time in a while that a model got upgraded while it was still my clear daily driver. Claude also pulled out several other advances to their ecosystem, including fast mode, and expanding Cowork to Windows, while OpenAI gave us an app for Codex. For fully agentic coding, GPT-5.3-Codex and Claude Opus 4.6 both look like substantial upgrades. Both sides claim they’re better, as you would expect. If you’re serious about your coding and have hard problems, you should try out both, and see what combination works [...] --- Outline: (01:55) On Your Marks (17:35) Official Pitches (17:56) It Compiles (21:42) It Exploits (22:45) It Lets You Catch Them All (23:16) It Does Not Get Eaten By A Grue (24:10) It Is Overeager (25:24) It Builds Things (27:58) Pro Mode (28:24) Reactions (28:36) Positive Reactions (42:12) Negative Reactions (50:40) Personality Changes (56:28) On Writing (59:11) They Banned Prefilling (01:00:27) A Note On System Cards In General (01:01:34) Listen All Yall Its Sabotage (01:05:00) The Codex of Competition (01:06:22) The Niche of Gemini (01:07:55) Choose Your Fighter (01:12:17) Accelerando --- First published: February 11th, 2026 Source: https://www.lesswrong.com/posts/5JNjHNn3DyxaGbv8B/claude-opus-4-6-escalates-things-quickly --- Narrated by TYPE III AUDIO. --- Images from the article: Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.

    1 ч. 14 мин.
  7. 10 ФЕВР.

    “Claude Opus 4.6: System Card Part 2: Frontier Alignment” by Zvi

    Coverage of Claude Opus 4.6 started yesterday with the mundane alignment and model welfare sections of the model card. Today covers the kinds of safety I think matter most: Sabotage, deception, situational awareness, outside red teaming and most importantly the frontier, catastrophic and existential risks. I think it was correct to release Opus 4.6 as an ASL-3 model, but the process Anthropic uses is breaking down, and it not on track to reliably get the right answer on Opus 5. Tomorrow I’ll cover benchmarks, reactions and the holistic takeaways and practical implications. I’m still taking it all in, but it seems clear to me that Claude Opus 4.6 is the best model out there and should be your daily driver, with or without Claude Code, on most non-coding tasks, but it is not without its weaknesses, in particular in writing and falling into generating more ‘AI slop’ style prose than Claude Opus 4.5. For coding tasks, I presume that Opus 4.6 with Claude Code is the play, especially with Agent Teams and fast mode available, and I’m using it myself, but Codex with GPT-5.3-Codex-Max is also a strong model and a viable alternative, and a fully [...] --- Outline: (01:32) Sabotage, Deception and Evaluation Integrity (03:42) Sandbagging On Dangerous Capability Evaluations (06:01) Situational Awareness (07:33) Inhibiting Evaluation Awareness (6.5) (09:06) Self-Preference (10:24) UK AISI Testing (11:40) Apollo Research Testing (14:24) Responsible Scaling Policy Evaluations (15:45) CBRN (mostly Biology) (18:43) Autonomy (26:40) Autonomy Benchmarks (29:53) Cyber (31:27) Ship It Anyway (33:40) You Are Not Ready --- First published: February 10th, 2026 Source: https://www.lesswrong.com/posts/togCQtFtfdF23xGNS/claude-opus-4-6-system-card-part-2-frontier-alignment --- Narrated by TYPE III AUDIO. --- Images from the article: Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.

    37 мин.
  8. 9 ФЕВР.

    “Claude Opus 4.6: System Card Part 1: Mundane Alignment and Model Welfare” by Zvi

    Claude Opus 4.6 is here. It was built with and mostly evaluated by Claude. Their headline pitch includes: 1M token context window (in beta) with State of the art retrieval performance. Improved abilities on a range of everyday work tasks. Model is improved. State of the art on some evaluations, including Terminal-Bench 2.0, HLE and a very strong lead in GDPval-AA. Claude Code now has an experimental feature called Agent Teams. Claude Code with Opus 4.6 has a new fast (but actually expensive) mode. Upgrades to Claude in Excel and the release of Claude in PowerPoint. Other notes: Price remains $5/$25, the same as Opus 4.5, unless you go ultra fast. There is now a configurable ‘effort’ parameter with four settings. Refusals for harmless requests with rich context are down to 0.04%. Data sources are ‘all of the above,’ including the web crawler (that they insist won’t cross CAPTCHAs or password protected pages) and other public data, various non-public data sources, data from customers who opt-in to that and internally generated data. They use ‘several’ data filtering methods. Thinking mode gives better [...] --- Outline: (03:45) A Three Act Play (04:57) Safety Not Guaranteed (10:53) Pliny Can Still Jailbreak Everything (12:48) Transparency Is Good: The 212-Page System Card (13:53) Mostly Harmless (17:45) Mostly Honest (19:01) Agentic Safety (20:27) Prompt Injection (23:07) Key Alignment Findings (33:48) Behavioral Evidence (6.2) (38:40) Reward Hacking and 'Overly Agentic Actions' (40:37) Metrics (6.2.5.2) (42:40) All I Did It All For The GUI (43:58) Case Studies and Targeted Evaluations Of Behaviors (6.3) (44:19) Misrepresenting Tool Results (45:09) Unexpected Language Switching (46:12) The Ghost of Jones Foods (47:54) Loss of Style Points (48:54) White Box Model Diffing (49:13) Model Welfare --- First published: February 9th, 2026 Source: https://www.lesswrong.com/posts/sWsSncqMLKyGZA9Ar/claude-opus-4-6-system-card-part-1-mundane-alignment-and --- Narrated by TYPE III AUDIO. --- Images from the article: Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.

    56 мин.

Оценки и отзывы

5
из 5
Оценок: 2

Об этом подкасте

Audio narrations of LessWrong posts by zvi

Вам может также понравиться