Conversation

Quote Tweet
Replying to @DavidDuvenaud @aidangomezzz and @AndrewCritchCA
Paul Christiano has written some relatively banal doom scenarios, in which even governance that improves human lives and power along all measurable axes is paired with so much manipulation and deception that humans are effectively marginalized: alignmentforum.org/posts/HBxe6wdj
1
Quote Tweet
Replying to @aidangomezzz
Some detailed scenarios of how we could lose influence have been written by @AndrewCritchCA , in which power is delegated to AI gradually because it’s effective, and anyone pushing back is marginalized: alignmentforum.org/posts/LpM3EAak
1
Quote Tweet
Replying to @aidangomezzz
This one is long but should definitely be on the list: cold-takes.com/without-specif The DeepMind safety team also put together a good literature review trying to summarize different threat models: alignmentforum.org/posts/wnnkD6P2
1
Quote Tweet
Replying to @aidangomezzz
Joe's is the best full breakdown of the argument I'm aware of arxiv.org/abs/2206.13353 (shorter version I haven't read: joecarlsmith.com/2023/03/22/exi)
1