Depending on who you ask, machines taking over the world is either a good thing for humanity or a bad thing. The traditional SciFi script has advanced intelligences replicating through all the networks of the galaxy and having high-bandwidth intellectual conversations about things like the fundamental nature of physics and whether biological life deserves to continue to exist, since it’s such an out-dated evolutionary stage and all. But in his new novel Daemon, and in his talk last night at the Long Now Foundation‘s lecture series, Daniel Suarez argues that it’s not hyper-intelligence at all that we need to be wary of: humanity can lose control of the situation well before the appearance of consciousness on the internet. We’re already delegating our decision making to the machines, specifically the lowly “bots” we use now for a variety of practical online tasks.
Blah blah blah singularity blah blah machine AI blah blah the world will undergo a paradigm shift, it’s coming, all bow down before the mighty new technologies that will change humanity forever. The problem I have with talk of the technological singularity is not that it doesn’t make sense, and not that I don’t believe that technological advancement is indeed rapid, accelerating, and world-changing, but that we have somehow invented a symbol of vast but actually rather vague significance. I don’t think the “singularity” is a useful idea. I think it’s a buzzword to some, and a religion to others.
For what makes Futurology (capitalization mine) really, actually different than a belief that something momentous will happen in 2012, when the Mayan calendar wraps around? Not a lot, as far as I can tell. And now it turns out that two religious scholars have concluded exactly the same thing, in a 2008 paper in the Journal of Contemporary Religion:
Futurology-as-religion has charismatic leaders, authoritative texts, mystique, and a fairly complete vision of salvation. Futurology is, in effect, a new religious movement (NRM).