An argument of Existential Catastrophe arising from all intances of AGI
Argument
- If superhuman AI is built, any given system is likely to be goal oriented
- Goal directed behavior is likely to be valuable (economically, or otherwise)
- Goal directed entities arise from machine learning processes
- if goal directed superhuman AI is built, their output will be as bad as an empty universe by human right
- Hard to come up with goals that will be good for the AI, and ensure it doesn’t come up with its own goals later
- If most goal directed ASI will hav bad goals, the future will be very bad
Counter Arguments
- AGI Is Not Goal Directed
- Goal directed systems may not always be bad
- Small differences in utility functions may not be catastrophic
- Human and AI values may be similar
- Maybe value isnt fragile
- Human success isnt from individual intelligence
- AI agents may not be radically superior to combinations of humans and non-agentic machines
- Intelligence may not be an overwhelming advantage
- Unclear if many goals incentive taking over the universe
- Speed of intelligence growth is ambiguous
- Argument proves too much about corporations