Podcast Episodes
Back to Search"Cryonics and Regret" by MvB
This post is not about arguments in favor of or against cryonics. I would just like to share a particular emotional response of mine as the topic bec…
2 years, 8 months ago
"Unifying Bargaining Notions (2/2)" by Diffractor
Alright, time for the payoff, unifying everything discussed in the previous post. This post is a lot more mathematically dense, you might want to dig…
2 years, 10 months ago
"The ants and the grasshopper" by Richard Ngo
Inspired by Aesop, Soren Kierkegaard, Robin Hanson, sadoeuphemist and Ben Hoffman.
One winter a grasshopper, starving and frail, approaches a colony o…
2 years, 10 months ago
"Steering GPT-2-XL by adding an activation vector" by TurnTrout et al.
Summary: We demonstrate a new scalable way of interacting with language models: adding certain activation vectors into forward passes. Essentially, w…
2 years, 10 months ago
"An artificially structured argument for expecting AGI ruin" by Rob Bensinger
Philosopher David Chalmers asked: "Is there a canonical source for "the argument for AGI ruin" somewhere, preferably laid out as an explicit argument…
2 years, 11 months ago
"How much do you believe your results?" by Eric Neyman
You are the director of a giant government research program that’s conducting randomized controlled trials (RCTs) on two thousand health intervention…
2 years, 11 months ago
"Mental Health and the Alignment Problem: A Compilation of Resources (updated April 2023)" by Chris Scammell & DivineMango
This is a post about mental health and disposition in relation to the alignment problem. It compiles a number of resources that address how to mainta…
2 years, 11 months ago
"On AutoGPT" by Zvi
The primary talk of the AI world recently is about AI agents (whether or not it includes the question of whether we can’t help but notice we are all …
2 years, 11 months ago
"GPTs are Predictors, not Imitators" by Eliezer Yudkowsky
(Related text posted to Twitter; this version is edited and has a more advanced final section.)
Imagine yourself in a box, trying to predict the next …
3 years ago
"Discussion with Nate Soares on a key alignment difficulty" by Holden Karnofsky
https://www.lesswrong.com/posts/iy2o4nQj9DnQD7Yhj/discussion-with-nate-soares-on-a-key-alignment-difficulty
Crossposted from the AI Alignment Forum. M…
3 years ago