Leaving aside the one-dimensional view of humanity and the nastiness of wanting our species to be enslaved, isn't there a problem of logic here? If humans are so venal, gullible, mendacious, and self-deceiving, how can we trust ourselves not to build robots that are just as bad (or worse)?
...if all you mean is are the robots going to take over, it's more or less inevitable, and not a moment too soon. Humans are really too stupid, venal, gullible, mendacious, and self-deceiving to be put in charge of important things like the Earth (much less the rest of the Solar System). I strongly support putting AIs in charge because I'm dead certain we can build ones that are not only smarter than human but more moral as well.
Tuesday, May 26, 2009
h+ magazine: "Is a Terminator scenario possible?" (Found via Instapundit.) Interesting, but I'm left nonplussed by this statement from J. Storrs Hall of the Foresight Institute: