Robert Lawrence Kuhn is the creator, writer and host of "Closer to Truth," a public television series and online resource that features the world's leading thinkers exploring humanity's deepest questions. Kuhn is co-editor with John Leslie, of "The Mystery of Existence: Why Is There Anything at All?" (Wiley-Blackwell, 2013). This article is based on "Closer to Truth" interviews produced and directed by Peter Getzels and streamed at www.closertotruth.com. Kuhn contributed this article to Live Science's Expert Voices: Op-Ed & Insights.
According to techno-futurists, the exponential development of technology in general and artificial intelligence (“AI”) in particular — including the complete digital replication of human brains — will radically transform humanity via two revolutions. The first is the "singularity," when artificial intelligence will redesign itself recursively and progressively, such that AI will become vastly more powerful than human intelligence ("superstrong AI"). The second revolution will be "virtual immortality," when the fullness of our mental selves can be uploaded perfectly to nonbiological media (such as silicon chips), and our mental selves will live on beyond the demise of our fleshy, physical bodies.
AI singularity and virtual immortality would mark a startling, transhuman world that techno-futurists envision as inevitable and perhaps just over the horizon. They do not question whether their vision can be actualized; they only debate when will it occur, with estimates ranging from 10 to 100 years. [Artificial Intelligence: Friendly or Frightening? ]
I'm not so sure. Actually, I'm a skeptic — not because I doubt the science, but because I challenge the philosophical foundation of the claims. Consciousness is the elephant in the room, and most techno-futurists do not see it. Whatever consciousness may be, it affects the nature of the AI singularity and determines whether virtual immortality is even possible.
It is an open question, post-singularity, whether superstrong AI without inner awareness would be in all respects just as powerful as superstrong AI with inner awareness, and in no respects deficient? In other words, are there kinds of cognition that, in principle or of necessity, require true consciousness? For assessing the AI singularity, the question of consciousness is profound .
What is consciousness?
Consciousness is a main theme of "Closer To Truth," and among the subtopics I discuss with scientists and philosophers on the program is the classic "mind-body problem" — what is the relationship between the mental thoughts in our minds and the physical brains in our heads? What is the deep cause of consciousness? (All quotes that follow are from "Closer To Truth.")
NYU Philosopher David Chalmers famously described the "hard problem" of consciousness: "Why does it feel like something inside? Why is all our brain processing — vast neural circuits and computational mechanisms — accompanied by conscious experience? Why do we have this amazing inner movie going on in our minds? I don't think the hard problem of consciousness can be solved purely in terms of neuroscience."
"Qualia" are the core of the mind-body-problem. "Qualia are the raw sensations of experience," Chalmers said. "I see colors — reds, greens, blues — and they feel a certain way to me. I see a red rose; I hear a clarinet; I smell mothballs. All of these feel a certain way to me. You must experience them to know what they're like. You could provide a perfect, complete map of my brain [down to elementary particles] — what's going on when I see, hear, smell — but if I haven't seen, heard, smelled for myself, that brain map is not going to tell me about the quality of seeing red, hearing a clarinet, smelling mothballs. You must experience it." [A Brief History of Artificial Intelligence ]
Can a computer be conscious?
To Berkeley philosopher John Searle, computer programs can never have a mind or be conscious in the human sense, even if they give rise to equivalent behaviors and interactions with the external world. (In Searle's "Chinese Room" argument, a person inside a closed space can use a rule book to match Chinese characters with English words and thus appear to understand Chinese, when, in fact, she does not.) But, I asked Searle, "Will it ever be possible, with hyperadvanced technology, for nonbiological intelligences to be conscious in the same sense that we are conscious? Can computers have 'inner experience'?"
"It's like the question, 'Can a machine artificially pump blood as the heart does?'" Searle responded. "Sure it can — we have artificial hearts. So if we can know exactly how the brain causes consciousness, down to its finest details, I don't see any obstacle, in principle, to building a conscious machine. That is, if you knew what was causally sufficient to produce consciousness in human beings and if you could have that [mechanism] in another system, then you would produce consciousness in that other system. Note that you don't need neurons to have consciousness. It's like saying you don't need feathers in order to fly. But to build a flying machine, you do need sufficient causal power to overcome the force of gravity."
"The one mistake we must avoid," Searle cautioned, "is supposing that if you simulate it, you duplicate it. A deep mistake embedded in our popular culture is that simulation is equivalent to duplication. But of course it isn't. A perfect simulation of the brain — say, on a computer — would be no more conscious than a perfect simulation of a rainstorm would make us all wet."
To robotics entrepreneur (and MIT professor emeritus) Rodney Brooks, "there's no reason we couldn't have a conscious machine made from silicon." Brooks' view is a natural consequence of his beliefs that the universe is mechanistic and that consciousness, which seems special, is an illusion. He claims that, because the external behaviors of a human, animal or even a robot can be similar, we "fool ourselves" into thinking "our internal feelings are so unique."
Can we ever really assess consciousness?
"I don't know if you're conscious. You don't know if I'm conscious," said Princeton University neuroscientist Michael Graziano. "But we have a kind of gut certainty about it. This is because an assumption of consciousness is an attribution, a social attribution. And when a robot acts like it's conscious and can talk about its own awareness, and when we interact with it, we will inevitably have that social perception, that gut feeling, that the robot is conscious.
"But can you really ever know if there's 'anybody home' internally, if there is any inner experience?" he continued. "All we do is compute a construct of awareness."
Warren Brown, a psychologist at Fuller Theological Seminary and a member of UCLA's Brain Research Institute, stressed "embodied cognition, embodied consciousness," in that "biology is the richest substrate for embodying consciousness." But he didn't rule out that consciousness "might be embodied in something nonbiological." On the other hand, Brown speculated that "consciousness may be a particular kind of organization of the world that just cannot be replicated in a nonbiological system."
Neuroscientist Christof Koch, president and chief scientific officer of the Allen Institute for Brain Science, disagrees. "I am a functionalist when it comes to consciousness," he said. "As long as we can reproduce the [same kind of] relevant relationships among all the relevant neurons in the brain, I think we will have recreated consciousness. The difficult part is, what do we mean by 'relevant relationships'? Does it mean we have to reproduce the individual motions of all the molecules? Unlikely. It's more likely that we have to recreate all the [relevant relationships of the relevant] synapses and the wiring ("connectome") of the brain in a different medium, like a computer. If we can do all of this at the right level, this software construct would be conscious."
I asked Koch if he'd be "comfortable" with nonbiological consciousness.
"Why should I not be?" he responded. "Consciousness doesn't require any magical ingredient."
Radical visions of consciousness
A new theory of consciousness — developed by Giulio Tononi, a neuroscientist and psychiatrist at the University of Wisconsin (and supported by Koch) — is based on "integrated information" such that distinct conscious experiences are represented by distinct structures in a heretofore unknown kind of space. "Integrated information theory means that you need a very special kind of mechanism organized in a special kind of way to experience consciousness," Tononi said. "A conscious experience is a maximally reduced conceptual structure in a space called 'qualia space.' Think of it as a shape. But not an ordinary shape — a shape seen from the inside."
Tononi stressed that simulation is "not the real thing." To be truly conscious, he said, an entity must be "of a certain kind that can constrain its past and future — and certainly a simulation is not of that kind."
Regarding the promise of brain replication to achieve virtual immortality, Tononi is not convinced. According to his theory of integrated information, "what would most likely happen is, you would create a perfect 'zombie' — somebody who acts exactly like you, somebody whom other people would mistake for you, but you wouldn't be there."
Inventor and futurist extraordinaire Ray Kurzweil believes that "we will get to a point where computers will evidence the rich array of emotionally subtle types of behavior that we see in human beings; they will be very intelligent, and they will claim to be conscious. They will act in ways that are conscious; they will talk about their own consciousness and argue about it just the way you and I do. And so the philosophical debate will be whether or not they really are conscious — and they will be participating in the debate."
Kurzweil argues that assessing the consciousness of other [possible] minds is not a scientific question. "We can talk scientifically about the neurological correlates of consciousness, but fundamentally, consciousness is this subjective experience that only I can experience. I should only talk about it in first-person terms (although I've been sufficiently socialized to accept other people's consciousness). There's really no way to measure the conscious experiences of another entity."
"But I would accept that these nonbiological intelligences are conscious," Kurzweil concluded. "And that'll be convenient, because if I don't, they'll get mad at me."
I offer two conjectures. The first is that the nature of the AI singularity differs profoundly in the case where it is literally conscious, with humanlike inner awareness, from the case where it is not literally conscious — even though in both cases superstrong AI is vastly more intelligent than humans and by all accounts appears to be equally conscious. This difference between being conscious and appearing conscious would become even more fundamental if, by some objective, absolute standard, humanlike inner awareness conveys some kind of intrinsic worthiness to entities manifesting it. I agree with the claim that after superstrong AI exceeds some threshold, science could never, even in principle, distinguish actual inner awareness from apparent inner awareness. But I do not agree with what usually follows: that this everlasting uncertainty about inner awareness and conscious experience in other entities (nonbiological or biological) makes the question irrelevant. I think the question maximally relevant.
My second conjecture, of which I am more sure, is that unless humanlike inner awareness can be wholly recreated, even synthesized, by physical manipulations alone, uploading one's neural patterns and pathways, however complete, could never preserve the original, first-person mental self (the private "I"), and virtual immortality would be impossible. (While it certainly seems feasible, given escalating technologies, to embed hyperadvanced hardware in brains — repairing damage, augmenting functions, transmitting thoughts — such enhancements, no matter how advanced, would not mean reproducing or transferring first-person consciousness. They wouldn't even be in the same category.)
Both conjectures turn on the deep cause of consciousness.
Alternative causes of consciousness
Through my conversations (and decades of night-musings), I have arrived at five alternative causes of consciousness (there may be others). Traditionally, the choice is between physicalism/materialism (No. 1 below) and dualism (No. 4), but the other three possibilities deserve consideration.
- Consciousness is entirely physical, solely the product of physical brain, which, at its deepest levels, comprises the fields and particles of fundamental physics. This is "physicalism" or "materialism," and it is overwhelmingly the prevailing theory of scientists. To many materialists, the utter physicality of consciousness is more an assumed premise than a derived conclusion.
- Consciousness is an independent, nonreducible feature of physical reality that exists in addition to the fields and particles of fundamental physics. This may take the form of a new, independent (fifth?) physical force or of a radically new organization of reality (e.g., ‘qualia space’ as postulated by integrated information theory).
- Consciousness is a nonreducible feature of each and every physical field and particle of fundamental physics. Everything that exists has a kind of "proto-consciousness," which, in certain aggregates and under certain conditions, can generate human-level inner awareness. This is "panpsychism," one of the oldest theories in philosophy of mind (going back to pre-modern animistic religions and the ancient Greeks). Panpsychism, in various forms, is an idea being revived by some contemporary philosophers in response to the seemingly intractable "hard problem" of consciousness.
- Consciousness requires a radically separate, nonphysical substance that is independent of a physical brain, such that reality consists of two radically disparate parts — physical and nonphysical substances, divisions, dimensions or planes of existence. This is "dualism." While human consciousness requires both a physical brain and this non-physical substance (somehow working together), following the death of the body and the dissolution of the brain, this nonphysical substance of or by itself could maintain some kind of conscious existence. (Though this nonphysical substance is traditionally called a “soul” — a term that carries heavy theological implications — a soul is not at all the only kind of thing that such a nonphysical substance could be.)
- Consciousness is ultimate reality; the only thing that's really real is consciousness — everything, including the entire physical world, is derived from an all-encompassing "cosmic consciousness." Each individual instance of consciousness — human, animal, robotic or otherwise — is a part of this cosmic consciousness. Eastern religions, in general, espouse this kind of view. (See Deepak Chopra for contemporary arguments that ultimate reality is consciousness.)
Will superstrong AI be conscious?
I'm not going to evaluate each competing cause of consciousness. (That would require a course, not a column.) Rather, for each cause, I'll speculate whether nonbiological intelligences with superstrong AI (following the AI singularity) could be conscious and possess inner awareness.
- If consciousness is entirely physical, then it would be almost certainly true that nonbiological intelligences with superstrong AI would have the same kind of inner awareness that we do. Moreover, as AI would rush past the singularity and become ineffably more sophisticated than the human brain, it would likely express forms of consciousness higher than we today could even imagine.
- If consciousness is an independent, nonreducible feature of physical reality, then it would remain an open question whether nonbiological intelligences could ever experience true inner awareness. (It would depend on the deep nature of the consciousness-causing feature and whether this feature could be manipulated by technology.)
- If consciousness is a nonreducible property of each and every elementary physical field and particle (panpsychism), then it would seem likely that nonbiological intelligences with superstrong AI could experience true inner awareness (because consciousness would be an intrinsic part of the fabric of physical reality).
- If consciousness is a radically separate, nonphysical substance not causally determined by the physical world (dualism), then it would seem impossible that superstrong AI (alone), no matter how advanced, could ever experience true inner awareness.
- If consciousness is ultimate reality (cosmic consciousness), then anything could be (or is) conscious (whatever that may mean), including nonbiological intelligences.
Remember, in each of these cases, no one could detect, using any conceivable scientific test, whether the nonbiological intelligences with superstrong AI had the inner awareness of true consciousness.
In all aspects of behavior and communications, these nonbiological intelligences would seem to be equal to (or superior to) humans. But if these nonbiological intelligences did not, in fact, have the felt sense of inner experience, they would be "zombies" ("philosophical zombies" to be precise), externally identical to conscious beings, but blank inside.
And this dichotomy elicits (a bit circularly) our probative question: whether true conscious experience and inner awareness in nonbiological intelligences would bring about distinctive, richer cognitions (however subtle and undetectable) or represent higher forms of absolute, universal value (however anthropomorphic this may seem).
Is virtual immortality possible?
Now, what about virtual immortality — digitizing and uploading the fullness of one's first-person mental self (the "I") from wet, mushy, physical brains that die and decay to new, more permanent (nonbiological) media or substrates? Could this actually work?
Again, the possibilities for virtual immortality relate to each of the alternative causes of consciousness.
1. If consciousness is entirely physical, then our first-person mental self would be uploadable, and some kind of virtual immortality would be attainable. The technology might take hundreds or thousands of years — not decades, as techno-optimists believe — but barring human-wide catastrophe, it would happen.
2. If consciousness is an independent, nonreducible feature of physical reality, then it would be possible that our first-person mental self could be uploadable — though less clearly than in No. 1 above, because not knowing what this consciousness-causing feature would be, we could not know whether it could be manipulated by technology, no matter how advanced. But because consciousness would still be physical, efficacious manipulation and successful uploading would seem possible.
3. If consciousness is a nonreducible feature of each and every elementary physical field and particle (panpsychism), then it would seem probable that our first-person mental self would be uploadable, because there would probably be regularities in the way particles would need to be aggregated to produce consciousness, and if regularities, then advanced technologies could learn to control them.
4. If consciousness is a radically separate, nonphysical substance (dualism), then it would seem impossible to upload our first-person mental self by digitally replicating the brain, because a necessary cause of our consciousness, this nonphysical component, would be absent.
5. If consciousness is ultimate reality, then consciousness would exist of itself, without any physical prerequisites. But would the unique digital pattern of a complete physical brain (derived, in this case, from consciousness) favor a specific segment of the cosmic consciousness (i.e., our unique first-person mental self)? It's not clear, in this extreme case, that uploading would make much difference (or much sense).
In trying to distinguish these alternatives, I am troubled by a simple observation. Assume that a perfect digital replication of my brain does, in fact, generate human-level consciousness (surely alternative 1, possibly 2, probably 3, not 4, 5 doesn’t matter). This would mean that my first-person self and personal awareness could be uploaded to a new medium (nonbiological or even, for that matter, a new biological body). But if "I" can be replicated once, then I can be replicated twice; and if twice, then an unlimited number of times.
So, what happens to my first-person inner awareness? What happens to my "I"?
Assume I do the digital replication procedure and it works perfectly — say, five times.
Where is my first-person inner awareness located? Where am I?
Each of the five replicas would state with unabashed certainty that he is "Robert Kuhn," and no one could dispute them. (For simplicity of the argument, physical appearances of the clones are neutralized.) Inhabiting my original body, I would also claim to be the real “me,” but I could not prove my priority.
I'll frame the question more precisely. Comparing my inner awareness from right before to right after the replications, will I feel or sense differently? Here are four obvious possibilities, with their implications:
- I do not sense any difference in my first-person awareness. This would mean that the five replicates are like super-identical twins — they are independent conscious entities, such that each begins instantly to diverge from the others. This would imply that consciousness is the local expression or manifestation of a set of physical factors or patterns. (An alternative explanation would be that the replicates are zombies, with no inner awareness — a charge, of course, they will deny and denounce.)
- My first-person awareness suddenly has six parts — my original and the five replicates in different locations — and they all somehow merge or blur together into a single conscious frame, the six conscious entities fusing into a single composite (if not coherent) "picture." In this way, the unified effect of my six conscious centers would be like the "binding problem" on steroids. (The binding problem in psychology asks how do our separate sense modalities like sight and sound come together such that our normal conscious experience feels singular and smooth, not built up from discrete, disparate elements). This would mean that consciousness has some kind of overarching presence or a kind of supra-physical structure.
- My personal first-person awareness shifts from one conscious entity to another, or fragments, or fractionates. These states are logically (if remotely) possible, but only, I think, if consciousness would be an imperfect, incomplete emanation of evolution, devoid of fundamental grounding.
- My personal first-person awareness disappears upon replication, although each of the six (original plus five) claims to be the original and really believes it. (This, too, would make consciousness even more mysterious.)
Suppose, after the replicates are made, the original (me) is destroyed. What then? Almost certainly my first-person awareness would vanish, although each of the five replicates would assert indignantly that he is the real "Robert Kuhn" and would advise, perhaps smugly, not to fret over the deceased and discarded original.
At some time in the future, assuming that the deep cause of consciousness permits this, the technology will be ready. If I were around, would I submit? I might, because I'm confident that 1 (above) is true and 2, 3 and 4 are false, and that the replication procedure would not affect my first-person mental self one whit. (So I sure wouldn't let them destroy the original.)
Bottom line, for me for now: The AI singularity and virtual immortality must confront the deep cause of consciousness.
Follow all of the Expert Voices issues and debates — and become part of the discussion — on Facebook, Twitter and Google+. The views expressed are those of the author and do not necessarily reflect the views of the publisher. This version of the article was originally published on Live Science.