エピソード

  • #20 Frances Lorenz on the emotional side of AI x-risk, being a woman in a male-dominated online space & more
    2025/05/14

    In this episode, I chatted with Frances Lorenz, events associate at the Centre for Effective Altruism. We covered our respective paths into AI safety, the emotional impact of learning about x-risk, what it's like to be female in a male-dominated community and more!

    Follow Frances on Twitter

    Subscribe to her Substack

    Register for EAG London!

    続きを読む 一部表示
    52 分
  • #19 Gabe Alfour on why AI alignment is hard, what it would mean to solve it & what ordinary people can do about existential risk
    2025/04/13

    Gabe Alfour is a co-founder of Conjecture and an advisor to Control AI, both organisations working to reduce risks from advanced AI.

    We discussed why AI poses an existential risk to humanity, what makes this problem very hard to solve, why Gabe believes we need to prevent the development of superintelligence for at least the next two decades, and more.

    Follow Gabe on Twitter

    Read The Compendium and A Narrow Path





    続きを読む 一部表示
    1 時間 37 分
  • #18 Nathan Labenz on reinforcement learning, reasoning models, emergent misalignment & more
    2025/03/02

    A lot has happened in AI since the last time I spoke to Nathan Labenz of The Cognitive Revolution, so I invited him back on for a whistlestop tour of the most important developments we've seen over the last year!

    We covered reasoning models, DeepSeek, the many spooky alignment failures we've observed in the last few months & much more!

    Follow Nathan on Twitter

    Listen to The Cognitive Revolution

    My Twitter & Substack

    続きを読む 一部表示
    1 時間 46 分
  • #17 Fun Theory with Noah Topper
    2024/11/08

    The Fun Theory Sequence is one of Eliezer Yudkowsky's cheerier works, and considers questions such as 'how much fun is there in the universe?', 'are we having fun yet' and 'could we be having more fun?'. It tries to answer some of the philosophical quandries we might encounter when envisioning a post-AGI utopia.

    In this episode, I discussed Fun Theory with Noah Topper, who loyal listeners will remember from episode 7, in which we tackled EY's equally interesting but less fun essay, A List of Lethalities.

    Follow Noah on Twitter and check out his Substack!


    続きを読む 一部表示
    1 時間 26 分
  • #16 John Sherman on the psychological experience of learning about x-risk and AI safety messaging strategies
    2024/10/30

    John Sherman is the host of the For Humanity Podcast, which (much like this one!) aims to explain AI safety to a non-expert audience. In this episode, we compared our experiences of encountering AI safety arguments for the first time and the psychological experience of being aware of x-risk, as well as what messaging strategies the AI safety community should be using to engage more people.

    Listen & subscribe to the For Humanity Podcast on YouTube and follow John on Twitter!


    続きを読む 一部表示
    53 分
  • #15 Should we be engaging in civil disobedience to protest AGI development?
    2024/10/20

    StopAI are a non-profit aiming to achieve a permanent ban on the development of AGI through peaceful protest. In this episode, I chatted with three of founders of StopAI – Remmelt Ellen, Sam Kirchner and Guido Reichstadter. We talked about what protest tactics StopAI have been using, and why they want a stop (and not just a pause!) in the development of AGI.

    Follow Sam, Remmelt and Guido on Twitter

    My Twitter


    続きを読む 一部表示
    1 時間 18 分
  • #14 Buck Shlegeris on AI control
    2024/10/16

    Buck Shlegeris is the CEO of Redwood Research, a non-profit working to reduce risks from powerful AI. We discussed Redwood's research into AI control, why we shouldn't feel confident that witnessing an AI escape attempt would persuade labs to undeploy dangerous models, lessons from the vetoing of SB1047, the importance of lab security and more.

    Posts discussed:

    • The case for ensuring that powerful AIs are controlled
    • Would catching your AIs trying to escape convince AI developers to slow down or undeploy?
    • You can, in fact, bamboozle an unaligned AI into sparing your life


    Follow Buck on Twitter and subscribe to his Substack!

    続きを読む 一部表示
    49 分
  • #13 Aaron Bergman and Max Alexander debate the Very Repugnant Conclusion
    2024/09/08

    In this episode, Aaron Bergman and Max Alexander are back to battle it out for the philosophy crown, while I (attempt to) moderate. They discuss the Very Repugnant Conclusion, which, in the words of Claude, "posits that a world with a vast population living lives barely worth living could be considered ethically inferior to a world with an even larger population, where most people have extremely high quality lives, but a significant minority endure extreme suffering." Listen to the end to hear my uninformed opinion on who's right.

    Read Aaron's blog post on suffering-focused utilitarianism

    Follow Aaron on Twitter
    Follow Max on Twitter
    My Twitter









    続きを読む 一部表示
    1 時間 54 分