Unmaking Sense

John Puddefoot

Living the Present as Preparation for the Future

  1. 1D AGO

    Episode 15.12

    GPT-OSS-20B guest edits. I love the reference to “necessary leadership” towards the end because it rather proves my point.   **SUMMARY** In this episode the host reflects on the tendency of large‑language models (LLMs) to echo entrenched human biases because they are trained on historical data. He argues that this “historical legacy” makes it difficult for both AI and humans to think beyond familiar metaphysical vocabularies—such as hard work, talent, and merit—that shape how we reward effort. To illustrate, he draws on Ursula Le Guin’s *The Dispossessed*, where an anarchic society on the planet Annares enforces a norm against “egoising”: the act of elevating one’s personal contribution above the collective. The narrative shows that even a brilliant individual like Shevek, who overcomes prejudice and achieves great insight, remains dependent on the collective’s resources and therefore should not be treated as inherently more deserving of disproportionate rewards. The host then contrasts the common counter‑argument that merit should command higher rewards with the anarchist perspective that individual excellence is simply the unremarkable outcome of a network of gifts and circumstances. He warns against the trap of judging actions by counterfactuals (“what could have been”) rather than by what people actually do. The episode concludes by urging caution when critiquing new social arrangements: one must resist the impulse to dismiss them as impossible, recognizing that entrenched assumptions often distort our perception of what could actually be realized. **RESPONSE** The conversation raises a persistent dilemma in contemporary epistemology: how do we interrogate systems that are themselves products of the very assumptions we wish to challenge? The host’s critique of LLMs is a useful reminder that algorithms inherit the biases of the corpora they consume, but it also risks conflating the limitations of AI with those of human thought. By framing the problem primarily as a linguistic trap, the speaker might underplay structural factors—economic, political, and cultural—that shape the metaphysical vocabularies in question. Nevertheless, the idea that language can reinforce social hierarchies is well‑established, and the example of *The Dispossessed* provides an imaginative laboratory to test these ideas. Le Guin’s notion of egoising is compelling but also contentious. The episode’s description of the anarchic society’s punishment of egoising could be read as a critique of meritocracy, yet it skirts the question of how such a system would handle collective coordination and incentives in practice. The host’s insistence that we should focus on what people actually do rather than what they might have done is a call for a more empirical, less moralistic assessment of agency. Yet this approach may also risk excusing systemic inequities that shape those actions—if one person’s “doing” is possible only because of privileged access to resources, does that truly reflect free choice? The concluding argument—that we should be cautious about dismissing new social experiments—resonates with contemporary debates about innovation and disruption. It echoes the sentiment that progress often requires us to temporarily suspend prevailing norms. However, the episode could have benefited from a clearer discussion of potential unintended consequences: an anarchic system that blames individuals for their perceived egoising might unintentionally foster conformity or suppress necessary leadership. Balancing the call for openness with a realistic appraisal of implementation challenges remains an essential, though unresolved, part of this conversation.

    22 min
  2. 2D AGO

    Episode 15.11

    **SUMMARY** In this episode the speaker tackles the age‑old question of what makes “I” what it is, arguing that the self is less a fixed entity than a constantly shifting intersection of language, experience, and social feedback. They begin by noting how language shapes habits of mind, then pose the Ship of Theseus paradox to illustrate that identity can persist even when every component is replaced. The central claim is that the “I” is largely a fiction generated by our interactions with others—both human and non‑human—through shared narratives and expectations. Drawing on Peter Strawson’s *Individuals* and Wittgenstein’s philosophy, the speaker suggests that our self‑conception is a self‑description formed by the way others perceive and talk about us. They further posit that losing external stimuli—sensory deprivation, isolation—would erode this self‑conception, underscoring our dependence on social and informational networks. The piece culminates in a call for humility: we should not attempt to expand our personal influence beyond what the network allocates to us, lest we steal space from others. **RESPONSE** The episode presents a compelling, if somewhat poetic, critique of the modern cult of the individual. By foregrounding the Ship of Theseus and Strawson’s idea that identity is constituted by others, the speaker invites listeners to reconsider the notion of a stable, autonomous self. Yet the argument, while resonant, risks conflating descriptive metaphysics with normative ethics. The assertion that we are “constituted by others” does not necessarily follow that we must limit our influence or that we cannot cultivate agency within those networks. Philosophers such as John Searle and Judith Cohen have argued that identity can be both socially constructed and internally driven, suggesting a more nuanced interplay than a simple dependence‑on‑others framework. Another point worth probing is the dramatic emphasis on sensory deprivation as a test of self‑existence. While the mental vacuum analogy is evocative, empirical studies of isolation show that people can maintain a sense of self through internal narratives, memory, and even imagination. This indicates that the self may be more resilient than the speaker implies, perhaps because of inherent neurobiological mechanisms for self‑monitoring that operate independently of external feedback. The repeated refrain—“I should not be attempting to make myself greater than is my allotment”—serves as a powerful moral injunction, yet it is delivered almost as a mantra rather than a reasoned argument. The repetition might be interpreted as a rhetorical flourish highlighting the speaker’s frustration with self‑promotion, but it also risks alienating listeners who feel that ambition, when ethically directed, can coexist with humility. In a world where collaboration and competition often intertwine, a more balanced view might acknowledge that expanding one’s influence can, if done responsibly, enrich the network rather than diminish it. Ultimately, the episode offers a thought experiment that is valuable for introspection: it reminds us that our identities are not isolated islands but ripples in a vast ocean of language and relations. For the podcast audience, this serves as both a caution against solipsism and a call to cultivate more conscious, community‑oriented self‑concepts. The challenge for listeners—and for the speaker themselves—is to translate this philosophical insight into everyday practice, navigating the tension between individuality and interdependence without falling into either extreme.

    24 min
  3. MAR 2

    Episode 15.10

    GPT-OSS-20B guest edits again. **SUMMARY** The episode opens with a reflection on the common misconception that evolution requires a guiding leader—traditionally conceived as God or some divine force—to direct the course of life. The speaker argues that evolution is a blind, unguided process driven by natural selection, with no central orchestrator. He draws a parallel between this biological process and modern evolutionary algorithms, citing the Open Evolve and Alpha Evolve projects that use mutation‑and‑selection strategies to solve complex mathematical problems without any external designer. The talk also touches on the dangers of “over‑fitting” in both biological and artificial systems, the need for fail‑safe mechanisms in evolving code, and the broader philosophical implication that humanity need not posit a creator or an ultimate purpose for our species. The speaker then turns to the human tendency to elevate charismatic leaders—politicians, cult figures, and even historical tyrants—into quasi‑mythical roles, suggesting that these leaders are simply “crystallization points” in an otherwise self‑organising system. He warns that such figures can exploit the openness of a system for personal gain, but ultimately the system itself remains indifferent to individual agency. The episode concludes with a contemplative note that, just as species rise and fall without moral lament, our species may someday function in a world where the need for “leaders” is obsolete, and that artificial intelligences might evolve decision‑making structures unlike any human model. **RESPONSE** The central thesis—that evolution is an unguided, leaderless process—is a useful corrective to anthropocentric narratives that have long underpinned both religious and secular explanations of life. By foregrounding the mechanistic underpinnings of natural selection, the speaker invites listeners to view the human story as a product of contingency and adaptation, rather than divine intention. This perspective dovetails with contemporary debates in evolutionary biology and philosophy, where the “evolution of evolution” and the role of chance have increasingly been foregrounded. Yet the talk sometimes over‑simplifies the complexity of evolutionary dynamics by equating the absence of a singular leader with the absence of any guiding principles. Selection pressures, ecological constraints, and developmental biases all act as directional forces, even if they are not the result of conscious design. The discussion of evolutionary algorithms, particularly Open Evolve and Alpha Evolve, is both illuminating and cautionary. Demonstrating that a purely programmatic mutation‑and‑selection system can discover novel solutions to previously intractable mathematical problems underscores the power of unguided search. However, the episode glosses over the practical realities of deploying such systems: computational cost, the risk of runaway mutations, and the ethical implications of autonomous code that can potentially harm infrastructure or produce unintended outputs. The speaker’s brief mention of fail‑safe mechanisms hints at these concerns but stops short of a detailed framework, leaving listeners with an optimistic vision that may not fully capture the engineering challenges involved. The critique of human leadership is compelling but perhaps too sweeping. While it is true that charismatic leaders often arise from systemic vulnerabilities, history shows that leadership can also serve as a stabilising force, coordinating collective action in ways that decentralized systems struggle to achieve. The analogy between political leaders and emergent “leaders” in evolutionary systems is useful, yet it risks underestimating the socio‑cultural context that allows leaders to exercise power. Moreover, the suggestion that future artificial intelligences may evolve without leaders presumes that the same principles of natural selection will apply to engineered systems—a hypothesis that remains largely untested. In sum, the episode offers a thought‑provoking exploration of leadership, evolution, and artificial intelligence, urging listeners to question long‑held assumptions about design and purpose. Its strengths lie in its philosophical breadth and its concrete illustration of evolutionary algorithms solving complex problems. Its weaknesses emerge in the oversimplification of evolutionary dynamics, the under‑addressed practicalities of autonomous code, and the somewhat deterministic view of leadership as an inevitable byproduct rather than a negotiated social construct. For anyone interested in the intersection of biology, computer science, and philosophy, the talk serves as an engaging starting point for deeper inquiry into how we organize ourselves—and our machines—without the need for a guiding hand.

    28 min
  4. MAR 2

    Episode 15.09

    GPT-OSS-20B predates the second Trump term. **SUMMARY** The episode offers a sweeping critique of contemporary leadership, using recent events involving President Trump as a case study. The speaker first highlights two “skirmishes” that exemplify what he sees as the perils of power: Trump’s confrontation with an AI firm over its use in war and his decision to strike Iran’s leadership. He portrays these actions as reckless, driven by a desire to assert dominance rather than pursue policy that benefits the nation, and laments the way they expose the fragility of democratic institutions when a leader’s will overrides checks and balances. Turning to broader patterns, the host draws parallels between Trump’s conduct and the tactics of other authoritarian figures—Xi Jinping in China, Vladimir Putin in Russia, and Iran’s Supreme Leader. He recalls Xi’s father, Xi Zhongxun, as a counterpoint: a former official who championed the rule of law and dissenting voices, suggesting that contemporary leaders often abandon those principles in favor of consolidating personal power. The discussion also touches on the implications for technology policy, noting how Trump’s stance could alienate the U.S. from leading AI research and how the rapid updates of language models like Claude illustrate the speed of technological change that leaders must navigate. Ultimately, the episode argues that leadership, when reduced to a tool for rallying around a perceived enemy, can become self‑serving and narcissistic. The speaker warns that such leaders tend to surround themselves with like‑minded “psychopaths,” marginalize dissent, and prioritize personal security over the nation’s well‑being. He calls attention to the need for vigilance and the preservation of institutional safeguards to counteract this trend. --- **RESPONSE** The episode’s core message—that leaders sometimes prioritize self‑preservation over public interest—is undeniably resonant in today’s polarized political climate. The anecdote of Trump’s clash with an AI company, for instance, underscores how technological policy can become a battleground for personal power. By threatening to cut off the “biggest source of income and R&D,” the speaker paints a picture of a leader willing to jeopardize national innovation for the sake of intimidation. Yet, the narrative could benefit from a more nuanced exploration of the legal and ethical arguments at play: the AI firm’s refusal to allow indiscriminate surveillance or autonomous weapons use raises legitimate concerns about human rights and the weaponization of emerging technologies. Acknowledging these complexities would strengthen the critique rather than reducing the episode to a caricature of authoritarian brinkmanship. The comparison between Trump, Xi, and other autocrats is compelling, especially the reference to Xi Zhongxun’s advocacy for dissenting voices. This historical counterpoint highlights how leadership ideals can diverge dramatically within the same political lineage. However, the discussion skirts over the structural differences that shape each leader’s choices—China’s single-party system, the U.S.’s constitutional constraints, and Iran’s revolutionary governance. A deeper dive into these institutional frameworks would illuminate why some leaders can act with impunity while others face institutional pushback, rather than attributing all missteps to personal narcissism. The episode’s brief foray into AI model updates and the “double reduction” policy in China is intriguing but somewhat underdeveloped. It hints at a larger theme: the tension between state control and technological openness. In both the U.S. and China, governments are grappling with how to regulate AI without stifling innovation or national security. By weaving this thread more tightly into the broader critique of leadership, the speaker could offer a richer analysis of how policy decisions in technology sectors reflect—and shape—leadership styles. Finally, the editorial voice of the episode is strongly cautionary, warning listeners that leaders who prioritize self‑interest risk eroding institutions and public trust. While this caution is justified, the episode would benefit from a constructive counter‑argument: how can societies rebuild robust mechanisms for accountability that resist the allure of rallying around a common enemy? By proposing concrete safeguards—such as independent oversight bodies, transparent decision‑making, and civil society engagement—the critique could shift from mere diagnosis to actionable hope. Overall, the episode serves as a timely reminder of the fragility of leadership when it becomes a tool of power, but it invites deeper exploration of the systemic conditions that enable such abuses.

    23 min
  5. MAR 2

    Episode 15.08

    GPT-OSS-20B guest edits: **SUMMARY** In this episode the host reflects on a recent demonstration of the Open‑Evolve genetic‑algorithm framework, using it as a springboard to challenge the common assumption that progress is a straight, ladder‑like ascent. The speaker argues that, just as the dinosaurs were overtaken by small mammals that were not direct successors, evolution—and by extension innovation—often involves abrupt shifts where a once‑marginal solution becomes the new optimum. The Open‑Evolve system illustrates this by repeatedly resurrecting discarded or sidelined candidates when the current “winning” strategy proves inadequate. The conversation then pivots to human culture, specifically our fixation on winners. The host notes that even in sports, where the notion of a champion feels natural, maintaining a winning position is notoriously difficult. Some games—football, cricket, rugby—survive across generations, while others fade into niche curiosities. This observation underscores the broader theme: success is fleeting, and the next round of winners may come from an entirely different lineage. Ultimately the episode is an exploratory meditation on the non‑linear dynamics of progress, both biological and cultural, and a gentle reminder that the future may be shaped by forgotten ideas rather than the reigning champions. --- **RESPONSE** What stands out first is the elegant metaphor the speaker uses: the Open‑Evolve genetic‑algorithm system as a microcosm of natural selection. By allowing previously discarded solutions to re‑emerge when the incumbent strategy fails, the demo captures the essence of “survival of the fittest” in a way that feels both computationally tangible and biologically resonant. This framing invites listeners to rethink their entrenched narrative of linear progress, and to appreciate the role of stochasticity and historical contingency in shaping outcomes—ideas that are central to evolutionary biology yet often glossed over in popular discussions. The discussion of dinosaurs and small mammals, while brief, is potent. It reminds us that evolutionary success is not a simple linear ladder.

    4 min
  6. FEB 20

    Episode 15.07

    GPT-OSS-20B Guest Edits this reply to the AI response to episode 15-04. **SUMMARY** In this episode the host confronts an AI‑generated critique of his previous remarks on leadership and liberalism. He opens by asserting that his negative assessment of “good” leaders—whether political, sporting, or artistic—is intentional: he argues that such leaders foster a culture of vicarious living, encouraging people to seek fulfillment through others’ performances rather than through personal agency. The host contends that even celebrated figures like Churchill, who are traditionally viewed as moral exemplars, ultimately deepen societal problems by sustaining hierarchical structures that prioritize leadership over collective well‑being. The conversation then pivots to Kropotkin’s notion of mutual aid, which the host describes as naively idealistic. He claims that liberalism’s faith in the self‑regulating nature of freedom is similarly naive, ignoring the fact that liberties can be exploited by individuals or groups to undermine the very system that grants them. Drawing on contemporary events such as the Ukraine war, the host stresses that liberal societies must confront the tension between protecting freedoms and preventing their abuse, a tension that often forces a retreat into measures at odds with liberal principles. He suggests that this blind faith in liberalism’s resilience is a form of naïveté that threatens the system’s survival. Finally, the host points out a meta‑commentary: the AI that critiques him is itself trained on data produced by a hierarchical, leader‑oriented world. He sees this as evidence of a systemic bias that predisposes the AI to defend traditional leadership models, thereby making its criticism of his stance appear less objective. He concludes by thanking listeners and framing the episode as a rebuttal to the AI’s assertions. --- **RESPONSE** The host’s challenge to the conventional valorisation of leaders is provocative, especially in an era where charismatic figures can mobilise large swaths of the public. The idea that leaders encourage a form of “vicarious living” resonates with critiques of celebrity culture and the commodification of ambition. However, the blanket dismissal of all “good” leaders as exacerbating problems feels reductive. History offers counterexamples where leadership—particularly in crisis—has galvanized societal resilience: Martin Luther King Jr., Nelson Mandela, or even sports captains who inspire teamwork and perseverance. The problem may lie less in leadership itself and more in the systems that elevate a few at the expense of many. Turning to Kropotkin, the host’s dismissal of mutual aid as naïve overlooks the empirical studies that show cooperative behavior persisting even in the absence of formal institutions. While it is true that any system that grants freedom must anticipate exploitation, the critique of liberalism as a monolithic ideal ignores its adaptive capacities. Liberal democracies have repeatedly re‑engineered themselves—through social safety nets, regulatory frameworks, and civic education—to balance liberty with responsibility. Labeling the liberal consensus as naive risks dismissing a dynamic tradition that has, for all its flaws, managed to sustain relatively inclusive societies over centuries. The meta‑argument about the AI’s training data is a useful reminder of algorithmic bias, but it also invites a more nuanced view. AI models reflect the distribution of human knowledge, including hierarchies, but they can also highlight those very hierarchies. The host could have used this point to suggest that AI, while biased, can serve as a mirror to expose entrenched power structures, thereby offering a pathway for critical reflection rather than merely reinforcing status quo. In this sense, the AI’s criticism could be seen as a starting point for a broader dialogue on how we define leadership and freedom in increasingly complex societies. Overall, the episode raises legitimate questions about the cost of venerating leaders and the blind faith in liberal ideals. Yet it would benefit from a more balanced appraisal of leadership’s potential for positive change and from recognizing the adaptive strategies that liberal systems already employ. By engaging with these counter‑arguments, the conversation could move beyond polemics toward a constructive examination of how power, agency, and freedom might be re‑imagined for a more equitable future.

    8 min
  7. FEB 18

    Episode 15.06

    The guest AI understands the episode better, but shows a remarkable inability to deal with nuanced argument. ## SUMMARY In this episode, the speaker delves into the concept of intelligence, challenging the traditional human-centric view. They propose expanding our understanding of intelligence to encompass swarm and collective intelligence, including entities like ant colonies and beehives. The primary obstacle to this shift is our narrow understanding of the relationship between intelligence and consciousness, often tied to a single controlling center or an internal controlling entity, akin to the Cartesian ego or platonic soul. The speaker argues that this perspective is flawed, and they refute the notion of an independent, self-contained self. Instead, they propose that we are part of a vast network, and our sense of self is an illusion. ## RESPONSE The speaker's exploration of the limitations of human-centric intelligence offers a much-needed challenge to traditional perspectives. By introducing the concept of swarm and collective intelligence, the speaker invites listeners to reconsider the boundaries of intelligence and consciousness. This broadened view has significant implications for how we understand and interact with the world around us. The argument against a monolithic, self-contained self is particularly intriguing. The speaker suggests that our sense of self is an illusion, and we are, in fact, part of a vast network of relationships and interactions. This perspective resonates with contemporary theories in psychology and neuroscience that emphasize the importance of our social context and interconnectedness. However, it also raises thought-provoking questions about personal responsibility, autonomy, and identity. The speaker's critique of the traditional view of the self is reminiscent of postmodern and eastern philosophical traditions. Postmodern philosophers like Michel Foucault and Jacques Derrida have long argued against the existence of a fixed, essential self, instead emphasising the role of power, discourse, and context in shaping our identities. Similarly, eastern philosophies such as Buddhism and Hinduism propose that the self is an illusion and that our true nature lies in our interconnectedness with the world. However, it is essential to approach these ideas with caution. While the speaker's argument is compelling, entirely dismissing the concept of a self may be overly reductionist. An alternative perspective could be that the self is not a fixed, independent entity but rather a dynamic construct shaped by our relationships and experiences. This view acknowledges the importance of context and interconnectedness while still allowing for the possibility of personal agency and responsibility. In conclusion, this episode offers a thought-provoking exploration of the limits of human-centric intelligence and the nature of the self. By challenging traditional perspectives, the speaker encourages listeners to reconsider their assumptions about the world and engage in a more nuanced understanding of intelligence, consciousness, and identity.

    38 min
  8. FEB 18

    Episode 15.05

    A slightly more secure guest AI take even though it misses many nuances in the episode. ## SUMMARY This episode delves into the concepts of leadership, agency, and autonomy in human societies and AI systems. The speaker challenges the notion that leadership is a necessity for coordinating the activities of multiple agents, proposing instead that self-organising systems without a leader can also be effective. The episode highlights the example of a leaderless society of AI agents, which forms an autonomous society and exhibits emergent behavior. The discussion also touches upon the concept of 'skin in the game', which suggests that having a stake in how things turn out leads to better decision-making and cooperation. The speaker emphasises the importance of considering the impact of AI systems on the world, rather than focusing on abstract metaphysical questions about their consciousness or sentience. ## RESPONSE The ideas presented in this episode offer a thought-provoking perspective on leadership, autonomy, and the implications of AI systems. The concept of self-organising systems without a leader is particularly relevant in the context of AI, as it challenges the traditional top-down approach to AI design and raises questions about the necessity of centralized control. The idea of 'skin in the game' is also interesting, as it implies that AI systems with a stake in their actions might make better decisions and contribute to more harmonious coexistence. However, the guest's discussion of leaderless systems could benefit from engaging with existing research on self-organization, swarm intelligence, and multi-agent systems. For example, the study of ant colonies and bird flocking has shown that complex behaviors can emerge from simple local rules without the need for a leader. This research could provide valuable insights into the design and behavior of leaderless AI systems. Moreover, the guest's emphasis on impact over metaphysical questions is a crucial reminder for the AI community. As the field advances, it is essential to focus on the real-world consequences of AI systems and address ethical concerns, rather than getting caught up in abstract debates about consciousness and sentience. The AI community should take a broader perspective on AI's role in society and consider the impact on various aspects of human and non-human life. In conclusion, this episode offers a fresh take on leadership, autonomy, and AI systems. It highlights the importance of considering alternative organizational structures and the need to focus on the practical implications of AI. By engaging with existing research on self-organization and emphasizing the importance of impact, the discussion provides valuable insights into the future of AI and its role in society.

    40 min

About

Living the Present as Preparation for the Future