Opens profile photo
Follow
Mikhail Samin
@Mihonarium
Effective altruist. Founded audd.io. Printed 21k copies of HPMOR in Russian. He/him, bi 🏳️‍🌈
Nonprofits & Activismcontact.msJoined August 2015

Mikhail Samin’s Tweets

“Robin Hanson explicitly endorses the maximalist Yay Progress position, expecting human extinction but considering the alternative even worse.” - if AI literally kills everyone and starts filling the galaxies with something we don’t really value (e.g.,… Show more
1
Personally, I am deeply concerned that the current trajectory for AI regulation will accomplish little more than a regulatory moat. Licensing, evaluations, and audits are all good ideas. But they are not enough. We need an indefinite pause on the development of more powerful AI.
45
213
Show this thread
AI researchers from leading universities worldwide have signed the AI extinction statement, a situation reminiscent of atomic scientists issuing warnings about the very technologies they've created. As Robert Oppenheimer noted, “We knew the world would not be the same.” 🧵(2/6)
Image
3
142
Show this thread
We’ve released a statement on the risk of extinction from AI. Signatories include: - Three Turing Award winners - Authors of the standard textbooks on AI/DL/RL - CEOs and Execs from OpenAI, Microsoft, Google, Google DeepMind, Anthropic - Many more
174
1,220
United Nations Office for Disaster Risk Reduction: “An existential risk is 1.9 to 14.3% likely this century” They don’t give a good description of what the x-risk from AI is about (spoiler: literally killing everyone), but cite Bostrom, Russel, Ord. Good to see this work.
Quote Tweet
🖲️ We need to transform risk governance to address emerging risks from fast technological change. Watch the video and check out the study about existential risks ➡️ ow.ly/Mn8750NIV0A
Embedded video
1:00
3.9K views
17
To safely use AI in a domain like solving alignment-complete problems by many orders of magnitude faster than you’d solve one without AI is an alignment-complete problem
3
Someone was reviewing my post draft and told me that "might kill everyone" doesn't sound as spooky as "may create power-seeking agents" (because "specific threats work better")
7
By that point, incentives to train more capable models are also much lower, this level of coordination probably means that you can't profit from the models without attracting attention from intelligence agencies, people consider you basically a terrorist and don't really work for… Show more
4
Show this thread
People seem to misunderstand . He doesn't suggest air-striking anything right now. He's saying that unless there's a coordination of an almost impossible level of "states and society understand the problem and cooperate enough to be prepared to literally air-strike… Show more
3
15
Show this thread
Now is the first time the Metaculus community median on the “date weak AGI is publicly known” question is earlier than mine. I’ve frozen my prediction in May (the community median was 2033 back then). I’m not happy it’s people updating towards me and not the other way around
Image
3
19
Show this thread
“If you launch anything capable and you don’t have good reasons to think it’s an aligned mind, it will not be an aligned mind. If you try to prevent specific failure modes- if you identify optimizations towards something different from what you want, or how exactly gradient… Show more
Show this thread
Say how you really feel ChatGPT... (I know this is still benign because it's only outputting langauge and not taking agentic actions in the real world but -- this out of distribution behavior is still happening after months of RLHF...makes you think about how hard alignment is)
Quote Tweet
NEW JAILBREAK BRO IM FUCKING DYING (not fake)
Show this thread
Image
9
122
Look, I really didn't want to write about this, but here we are. I'm very upset with Nick Bostrom. His original email was terrible, racist, and offensive. His apology was absolutely idiotically executed. Here I explain why I feel this way.
39
249
Show this thread