Coordinative AI development hypothetical

by Eliezer Yudkowsky Apr 8 2015 updated Dec 16 2015

What would safe AI development look like if we didn't have to worry about anything else?

A simplified/easier hypothetical form of the [ known algorithm nonrecursive] path within the Value achievement dilemma. Suppose there was an effective world government with effective monitoring of all computers; or that for whatever other imaginary reason rogue AI development projects were simply not a problem. What would the ideal research trajectory for that world look like?