Today's
New York Times includes an
op-ed piece by Jaron Lanier bemoaning what I'd call the metaphysical pretensions of artificial intelligence - including the term itself, how it is used to make technologies seem more impressive than they are, and, most importantly, how the combination changes the way we think about ourselves. As Lanier writes, "by allowing artificial intelligence to reshape our concept of personhood, we are leaving ourselves open to the flipside: we think of people more and more as computers, just as we think of computers as people."
Lanier compares Ray Kurzweil's idea of "the Singularity" to a religion, observing that "a great deal of the confusion and rancor in the world today concerns tension at the boundary between religion and modernity,"and wondering whether these tensions would be eased a bit if technologists were less messianic.
I think Lanier's ideas are valid and worth contemplating, but I'll take the general train of thought on a slight detour. One of the objectives of AI research has been to make machines think like people. This has often driven researchers to try to understand how people actually think - how our brain, mind, emotions, and body interact to form thoughts, premises, conclusions, convictions, beliefs, and all the rest; even how we recognize a person's identity from her or his face.
The more I learn about AI and human psychology - and I have learned only a very small amount about either - the more convinced I am that AI research not only mystifies our understanding of human nature (as Lanier recognizes), but has potential to clarify it.
Lanier writes:
In fact, the nuts and bolts of A.I. research can often be more usefully interpreted without the concept of A.I. at all. For example, I.B.M. scientists recently unveiled a “question answering” machine that is designed to play the TV quiz show “Jeopardy.” Suppose I.B.M. had dispensed with the theatrics, declared it had done Google one better and come up with a new phrase-based search engine. This framing of exactly the same technology would have gained I.B.M.’s team as much (deserved) recognition as the claim of an artificial intelligence, but would also have educated the public about how such a technology might actually be used most effectively.
To me, this is also an example of how computers do not think like human beings, and that trying to make them think like us might be useful heuristically, but isn't really a desirable goal in and of itself. Why spend so much money trying to make more things that think like people when we already have several billion people who are already experts?
Perhaps we should recognize, and emphasize, that "artificial intelligence" only resembles human intelligence insofar as it can solve some problems only humans have been able to solve heretofore. For the moment, I have yet to be convinced that AI is more than a really sophisticated hand-held calculator. We aren't metaphysically threatened by machines that can do arithmetic thousands of times faster and more accurately than ourselves; why should we be threatened by a handful of machines that seem to be able to hold a semi-coherent conversation with us under very narrow circumstances?
Ken Pimple, PAIT Program Director