Podcast Episodes
Back to Search“The Case Against AI Control Research” by johnswentworth
The AI Control Agenda, in its own words:
… we argue that AI labs should ensure that powerful AIs are controlled. That is, labs should make sure that t…
1 year, 1 month ago
“Don’t ignore bad vibes you get from people” by Kaj_Sotala
I think a lot of people have heard so much about internalized prejudice and bias that they think they should ignore any bad vibes they get about a pe…
1 year, 1 month ago
“[Fiction] [Comic] Effective Altruism and Rationality meet at a Secular Solstice afterparty” by tandem
(Both characters are fictional, loosely inspired by various traits from various real people. Be careful about combining kratom and alcohol.)
The orig…
1 year, 1 month ago
“Building AI Research Fleets” by bgold, Jesse Hoogland
From AI scientist to AI research fleet
Research automation is here (1, 2, 3). We saw it coming and planned ahead, which puts us ahead of most (4, 5, …
1 year, 1 month ago
“What Is The Alignment Problem?” by johnswentworth
So we want to align future AGIs. Ultimately we’d like to align them to human values, but in the shorter term we might start with other targets, like …
1 year, 1 month ago
“Applying traditional economic thinking to AGI: a trilemma” by Steven Byrnes
Traditional economics thinking has two strong principles, each based on abundant historical data:
Principle (A): No “lump of labor”: If human populat…
1 year, 1 month ago
“Passages I Highlighted in The Letters of J.R.R.Tolkien” by Ivan Vendrov
All quotes, unless otherwise marked, are Tolkien's words as printed in The Letters of J.R.R.Tolkien: Revised and Expanded Edition. All emphases mine.…
1 year, 1 month ago
“Parkinson’s Law and the Ideology of Statistics” by Benquo
The anonymous review of The Anti-Politics Machine published on Astral Codex X focuses on a case study of a World Bank intervention in Lesotho, and te…
1 year, 1 month ago
“Capital Ownership Will Not Prevent Human Disempowerment” by beren
Crossposted from my personal blog. I was inspired to cross-post this here given the discussion that this post on the role of capital in an AI future …
1 year, 1 month ago
“Activation space interpretability may be doomed” by bilalchughtai, Lucius Bushnaq
TL;DR: There may be a fundamental problem with interpretability work that attempts to understand neural networks by decomposing their individual acti…
1 year, 1 month ago