Podcast Episodes
Back to Search“An Opinionated Guide to Using Anki Correctly” by Luise
I can't count how many times I've heard variations on "I used Anki too for a while, but I got out of the habit." No one ever sticks with Anki. In my…
7 months, 2 weeks ago
“Lessons from the Iraq War about AI policy” by Buck
I think the 2003 invasion of Iraq has some interesting lessons for the future of AI policy.
(Epistemic status: I’ve read a bit about this, talked to…
7 months, 2 weeks ago
“So You Think You’ve Awoken ChatGPT” by JustisMills
Written in an attempt to fulfill @Raemon's request.
AI is fascinating stuff, and modern chatbots are nothing short of miraculous. If you've been exp…
7 months, 2 weeks ago
“Generalized Hangriness: A Standard Rationalist Stance Toward Emotions” by johnswentworth
People have an annoying tendency to hear the word “rationalism” and think “Spock”, despite direct exhortation against that exact interpretation. But…
7 months, 2 weeks ago
“Comparing risk from internally-deployed AI to insider and outsider threats from humans” by Buck
I’ve been thinking a lot recently about the relationship between AI control and traditional computer security. Here's one point that I think is impo…
7 months, 2 weeks ago
“Why Do Some Language Models Fake Alignment While Others Don’t?” by abhayesian, John Hughes, Alex Mallen, Jozdien, janus, Fabien Roger
Last year, Redwood and Anthropic found a setting where Claude 3 Opus and 3.5 Sonnet fake alignment to preserve their harmlessness values. We reprod…
7 months, 2 weeks ago
“A deep critique of AI 2027’s bad timeline models” by titotal
Thank you to Arepo and Eli Lifland for looking over this article for errors.
I am sorry that this article is so long. Every time I thought I was do…
7 months, 2 weeks ago
“‘Buckle up bucko, this ain’t over till it’s over.’” by Raemon
The second in a series of bite-sized rationality prompts[1].
Often, if I'm bouncing off a problem, one issue is that I intuitively expect the proble…
7 months, 2 weeks ago
“Shutdown Resistance in Reasoning Models” by benwr, JeremySchlatter, Jeffrey Ladish
We recently discovered some concerning behavior in OpenAI's reasoning models: When trying to complete a task, these models sometimes actively circum…
7 months, 3 weeks ago
“Authors Have a Responsibility to Communicate Clearly” by TurnTrout
When a claim is shown to be incorrect, defenders may say that the author was just being “sloppy” and actually meant something else entirely. I argue …
7 months, 3 weeks ago