A case for AGI safety research far in advance
By Steven Byrnes @ 2021-03-26T12:59 (+7)
This is a linkpost to https://www.alignmentforum.org/posts/zzXawbXDwCZobwF9D/my-agi-threat-model-misaligned-model-based-rl-agent
I posted My AGI Threat Model: Misaligned Model-Based RL Agent on alignmentforum yesterday. Among other things, I make a case for misaligned AGI being an existential risk which can and should be mitigated by doing AGI safety research far in advance. So that's why I'm cross-posting it here!
I basically took a bunch of AGI safety ideas that I like—both old ideas (e.g. agents tend to self-modify to become more coherent) and new ideas (e.g. inner alignment)—and cast them into the framework of an AGI architecture and development path that I find most plausible, based on my thinking about neuroscience and machine learning.
Please leave comments either here or at the original post, and/or email me if you want to chat. Thanks in advance! :-D