Coordinative AI development hypothetical

https://arbital.com/p/4j

by Eliezer Yudkowsky Apr 8 2015 updated Dec 16 2015

What would safe AI development look like if we didn't have to worry about anything else?


A simplified/easier hypothetical form of the [ known algorithm nonrecursive] path within the Value achievement dilemma. Suppose there was an effective world government with effective monitoring of all computers; or that for whatever other imaginary reason rogue AI development projects were simply not a problem. What would the ideal research trajectory for that world look like?

Usefulness: