LessWrong (30+ Karma)

LessWrong

Audio narrations of LessWrong posts.

  1. HACE 4 H

    “Your rights when flying to Europe” by Yair Halberstadt

    Europe (and the UK) have strong protections for flyers in the case of delayed or cancelled flights. However very few people are aware of these, and airlines will almost always try to wriggle out of paying up. Even travel agents are often unaware of these laws, or unwilling to fight the airline for you. Given the rollercoaster that flying to/from Israel has been in the last 3 years, I've had my share of experience forcing airlines to pay up what they owe, so I thought it might be valuable turning that into a post. These regulations are enshrined in EU 261. You can see the full text here, and equally importantly the interpretive guidelines here that cover many edge cases. TLDR When flying into or out of the EU or UK, consider booking a flight with an EU or UK based airline.Don't book a car/hotel with the airline, as that turns it into a package deal which has weaker rights.Preserve records of all interactions with the airline. Prefer text based chat to phone as this is easier to records. If you do phone, and get a negative answer, follow up with a text based chat to [...] --- Outline: (00:51) TLDR (02:09) Which flights do these laws apply to? (02:40) What are my rights? (03:51) How will airlines try to screw you over? (05:06) Your playbook for flight cancellation (07:53) Your playbook for compensation (08:11) Do Nots (08:56) What about...? --- First published: May 5th, 2026 Source: https://www.lesswrong.com/posts/F5KkZiGMytDJszzyg/your-rights-when-flying-to-europe --- Narrated by TYPE III AUDIO.

    9 min
  2. HACE 9 H

    “Model Spec Midtraining: Improving How Alignment Training Generalizes” by Chloe Li, saraprice, Sam Marks, Jonathan Kutasov

    tl;dr We introduce model spec midtraining (MSM): after pre-training but before alignment fine-tuning, we train models on synthetic documents discussing their Model Spec, teaching them how they should behave and why. This controls how models generalize from subsequent alignment training—for example, two models with identical fine-tuning can generalize to different values depending on how MSM explains those behaviors. We use MSM to substantially reduce agentic misalignment and study which Model Specs produce better generalization. 📝Blog, 📄Paper, 💻 Code Introduction Some frontier AI developers aim to align language models to a Model Spec or Constitution that describes intended model behavior. The standard approach is to fine-tune on demonstrations of behaviors that align with the spec (e.g., conversations where the model acts as intended). However, this can fail to produce robust alignment. For example, LLM agents have been shown to take unethical actions (e.g., blackmailing, leaking company information, alignment faking) when placed in scenarios different from those appearing in their alignment training (Lynch et al., 2025; Jarviniemi and Hubinger, 2024; Greenblatt et al., 2024) We propose model spec midtraining (MSM), a method for shaping how models generalize from alignment fine-tuning (AFT). MSM is motivated by the hypothesis that AFT [...] --- Outline: (00:52) Introduction (02:24) Different generalization, same fine-tuning data (04:36) Reducing agentic misalignment (07:39) How does MSM scale with AFT compute? (08:58) Model Spec science (12:39) Conclusion --- First published: May 5th, 2026 Source: https://www.lesswrong.com/posts/R3Rrw8EscuRKxMFTz/model-spec-midtraining-improving-how-alignment-training --- Narrated by TYPE III AUDIO. --- Images from the article: Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.

    14 min
  3. HACE 11 H

    “Motivated reasoning, confirmation bias, and AI risk theory” by Seth Herd

    Of the fifty-odd biases discovered by Kahneman, Tversky, and their successors, forty-nine are cute quirks, and one is destroying civilization. This last one is confirmation bias. - From Scott Alexander's review of Julia Galef's The Scout Mindset. Alexander goes on to argue that this bias is the source of polarization in society, which is distorting our beliefs and setting us at each other's throats. How could someone believe such different things unless they're either really stupid or lying to conceal their selfishness? I think smart people who care about the truth go on believing conflicting things largely because of confirmation bias and motivated reasoning. The corner of civilization I'm most worried about is the one figuring out how to handle the advent of strong AI. I'm not telling anyone which direction to update, but I am suggesting that we are probably a little to a lot overconfident in our beliefs about alignment and AI impacts. I think the effects of biases are still strong and still overlooked in this corner of civilization, despite its strong values of truth-seeking and relative awareness of biases. Bias has more influence where there's less direct evidence, and that's the case in [...] --- Outline: (04:50) 1.1. Motivated reasoning (10:01) 2. Empirical evidence for confirmation bias (12:08) 2.1. Bias in evaluating evidence (17:48) 2.2. Bias in selecting evidence (20:57) 2.3. Bias in remembering evidence (22:37) 2.4. Other causal explanations of confirmation bias effects (25:17) 2.5. Empirical evidence for motivated reasoning (27:26) 3. Limitations in human cognitive capacity for very complex problems (29:16) 3.1. Introspection suggests fuzzy models and updating (30:32) 3.2. Intuition vs. analysis - evidence and brain mechanisms (34:19) 3.3. Bayesian reasoning is an ideal, not a method (36:40) 3.4. AI risk is complicated (40:18) 4. Compounding of confirmation bias (42:05) 4.1. Example of frame/hypothesis choices and confident disagreement among experts (47:29) 4.2. Social compounding of confirmation bias effects (49:45) 4.2.1. Social effects on evaluating evidence. (52:41) 4.2.2. Social effects on selecting evidence, memory, and framing (57:48) 4.2.3. Interlude: dont give up on seeking truth (58:32) 4.2.4. Social belief contagion or information cascade effects (01:03:07) 4.3. Very rough estimates of total compounded confirmation bias (01:09:01) 5. Implications and remediations (01:12:02) 5.1. Standard remediations (01:14:09) 5.2. Remediations for motivated reasoning (01:17:39) Conclusion The original text contained 9 footnotes which were omitted from this narration. --- First published: May 5th, 2026 Source: https://www.lesswrong.com/posts/QpgmEhBvJQxAfFMP2/motivated-reasoning-confirmation-bias-and-ai-risk-theory --- Narrated by TYPE III AUDIO. --- Images from the article: Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.

    1 h 19 min
  4. HACE 11 H

    “The AI Ad-Hoc Prior Restraint Era Begins” by Zvi

    The White House has ordered Anthropic not to expand access to Mythos, and is at least seriously considering a complete about-face of American Frontier AI policy into a full prior restraint regime, where anyone wishing to release a highly capable new model will have to ask for permission. This would be the antithesis of all their previous rhetoric, and all their actions to systematically avoid laying a foundation to do this in an orderly and informed fashion. But now, with the existence of Mythos, and a potential coming hackastrophe where cyber attackers will by default have the edge and we desperately need defenders to have a head start, it is not clear they feel they have a choice. If implemented well, this could be the right thing. By default, it won’t be implemented well. Project Glasswing Cannot Expand The government is now deciding which models can and cannot be made available on particular terms to particular parties. This is already happening. Anthropic wanted to expand the number of companies with access to Mythos as part of Project Glasswing. The White House said no. It is not clear this is any of [...] --- Outline: (00:54) Project Glasswing Cannot Expand (02:33) The Ad-Hoc Prior Restraint Era Begins (09:44) Implementation Through CAISI (12:49) What Should We Do About AI? (14:16) The Chain of Command Nonsense Continues (16:29) The Government Should Maintain Multiple AI Providers (16:56) Hows It Going To End? --- First published: May 5th, 2026 Source: https://www.lesswrong.com/posts/QX2ZCfkpWGqkyvStN/the-ai-ad-hoc-prior-restraint-era-begins --- Narrated by TYPE III AUDIO.

    19 min
  5. HACE 14 H

    [Linkpost] “Interpreting Language Model Parameters” by Lucius Bushnaq, Dan Braun, Oliver Clive-Griffin, Bart Bussmann, Nathan Hu, mivanitskiy, Linda Linsefors, Lee Sharkey

    This is a link post. This is the latest work in our Parameter Decomposition agenda. We introduce a new parameter decomposition method, adVersarial Parameter Decomposition (VPD)[1] and decompose the parameters of a small[2] language model with it. VPD greatly improves on our previous techniques, Stochastic Parameter Decomposition (SPD) and Attribution-based Parameter Decomposition (APD). We think the parameter decomposition approach is now more-or-less ready to be applied at scale to models people care about. Importantly, we show that we can decompose attention layers, which interp methods like transcoders and SAEs have historically struggled with. We also build attribution graphs of the model for some prompts using causally important parameter subcomponents as the nodes, and interpret parts of them. While we made these graphs, we discovered that our adversarial ablation method seemed pretty important for faithfully identifying which nodes in them were causally important for computing the final output. We think this casts some doubt on the faithfulness of subnetworks found by the majority of other subnetwork identification methods in the literature.[3][4] More details and some examples can be found in the paper. Additionally, as with our previous technique SPD, VPD does not [...] The original text contained 5 footnotes which were omitted from this narration. --- First published: May 5th, 2026 Source: https://www.lesswrong.com/posts/eAQZaiC3PcBhS4HjM/linkpost-interpreting-language-model-parameters Linkpost URL:https://www.goodfire.ai/research/interpreting-lm-parameters --- Narrated by TYPE III AUDIO. --- Images from the article: Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.

    5 min
  6. HACE 1 DÍA

    “Housing Roundup #15: The War Against Renters” by Zvi

    So many are under the strange belief that there is something terrible about not owning the house in which you live. So we massively subsidize home ownership, and try to actively interfere with renting. Except when we do rent control, which turns renting into a form of owning, and allows us to take real property and de facto give it to current renters. A lot of this is pure attempts to punish and exclude the poor. If you can’t afford a downpayment, we don’t want you living here. Go away. Some of it is the belief that when you rent, you are being ‘taken advantage of’ and that such a deal could not possibly be fair. Some of it is that if you don’t own, you don’t have the incentive to drive up property values. Which means you won’t properly work to ‘improve’ your local area, especially that you won’t conspire to block housing. The result of this is that if you’re not willing to commit to living in one place for years, or you can’t afford a down payment, you get punished, and punished hard. Owning Versus Renting The graph [...] --- Outline: (01:08) Owning Versus Renting (03:11) Build To Rent Is Good Actually (08:16) Elizabeth Warren, Full Supervillain (09:44) The Better Case Against Corporate Housing Ownership (11:50) The ROAD Act Bans Building And Then Renting Houses (14:17) Rental Covenants (15:09) Extended Eviction Delay After Nonpayment Is Mostly Bad (16:51) Los Angeles Renting (19:30) Sufficiently Advanced Rent Control Is Indistinguishable From Ownership (23:05) England Tries To Ban Renting (24:36) Claude Rental Discounts --- First published: May 4th, 2026 Source: https://www.lesswrong.com/posts/jW4TeNZhxBA9Fzdim/housing-roundup-15-the-war-against-renters --- Narrated by TYPE III AUDIO. --- Images from the article: Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.

    26 min

Acerca de

Audio narrations of LessWrong posts.

También te podría interesar