Saw the movie Ex
Machina. The outside shots, filmed in Valldalen,
Norway, are are simply gorgeous. Good flick and provoked some
ruminating (avoiding plot details).
There seems no a
priori reason to suppose that machine intelligence cannot reach
the point of passing the Turing
test. A complex enough programed machine able to “learn”
from extracting patterns from massive data and using them to interact
with humans should be able to “exhibit intelligent behavior
equivalent to, or indistinguishable from, that of a human.” One
can imagine such a machine as pictured in the movie.
But what does the
Turing test really test. An “artificial intelligence” might be
able to interpret and respond to the full range of human behavior and
simulate the same. It might be able to “read” a conscious human
better than an actual human might by picking up on subtle physical
manifestations (as stored in its memory). With a large enough data
base behind it and a multitude of “learned” behaviors it might
convince a human that it was indeed intelligent and even self-aware.
But would it be? Would the ability to simulate human behavior
completely enough to appear human actually be human or entail
consciousness? If programed with a sub-routine causing it to seek to
persist (i.e., resist termination), would it be a self seeking
self-preservation? Would programing allowing it to read human
emotions and respond “appropriately” with simulated emotion mean
it actually felt such emotions?
Would a machine
intelligence able to simulate human behavior and emotions actually be
able to love, hate, feel empathy and act with an awareness of itself
and, perhaps more importantly, of an Other? Or might there still be
something missing?
Smoked a cigar on my
favorite bench while considering all this and watched some ants going
about their business. Ants are extremely complex biological machines
acting and reacting within their environment with purpose and an
overall drive to self-perpetuate (both as individuals and as a
collective). They may be conscious even if not self aware. Or is a
certain basic self-awareness something that goes with being alive?
Would even a very complex machine ever be alive even if very
“intelligent?”
My guess is that
machine intelligence – even if very complex and advanced and
equipped with a self-referential sub-program allowing algorithmic
analysis of itself – would not be conscious or alive. Thus not
capable of emotion and therefore what we might call coldly
rational. Is this why Bill
Gates, Stephen
Hawking and others are concerned about AI?