The Meta-Problem of Consciousness | Professor David Chalmers | Talks at Google

Video Statistics and Information

Video
Captions Word Cloud
Reddit Comments
Captions
[MUSIC PLAYING] DAVID CHALMERS: Thanks for coming out. It's good to be here. As Eric said, I am a philosopher thinking about consciousness. Coming from a background in the sciences and math, it always struck me that the most interesting and hardest unsolved problem in the sciences was the problem of consciousness. And way back 25 years ago when I was in grad school, it seemed to be the best way to come at this from a big picture perspective was to go into philosophy and think about the foundational issues that arise in thinking about consciousness from any number of different angles, including the angles of neuroscience and psychology and AI. In this talk, I'm going to present a slightly different perspective on the problem after laying out some background, the perspective of what I call the meta-problem of consciousness. I always liked the idea that you approach a problem by stepping one level up, taking the metaperspective. I love this quote, "Anything you can do, I can do meta." I have no idea what the origins was. I like the fact this is attributed to Rudolf Carnap, one of my favorite philosophers. But anyone who knows Carnap's work, it's completely implausible he would ever say anything so frivolous. It's also being attributed to my thesis advisor, Doug Hofstadter, author of "Godel, Escher, Bach" and a big fan of the metaperspective. But he assures me he never said it either. But the metaperspective on anything is stepping up a level. The meta-problem, as I think about it, is it's called the meta-problem because it's a problem about a problem. A metatheory is a theory about a theory. Meta-problem is a problem about a problem. In particular, it's the problem of explaining why we think there is a problem about consciousness. So there's a first-order problem, the problem of consciousness. Today, I'm going to focus on a problem about it. But I'll start by introducing the first-order problem itself. The first-order problem is what we call the hard problem of consciousness. It's the problem of explaining why and how physical processes should give rise to conscious experience. You've got all of these neurons firing in your brain, bringing about all kinds of sophisticated behavior. We can get it to be [INAUDIBLE] explaining our various responses, but there's this big question about how it feels from the first person point of view. That's the subjective experience. I like this illustration of the hard problem of consciousness. It seems to show someone's hair catching fire, but I guess it's a metaphorical illustration of the subjective perspective. So the hard problem is concerned with what philosophers call phenomenal consciousness. The word consciousness is ambiguous 1,000 ways. But phenomenal consciousness is what it's like to be a subject from the first person point of view. So a system is phenomenally conscious if there's something it's like to be it. A mental state is phenomenally conscious if there's something it's like to be in it. So the thought is there are some systems-- so there's something it's like to be that system. There's something it's like to be me. I presume there's something it's like to be you. But presumably, there's nothing it's like to be this lectern. As far as we know, the lectern does not have a first person perspective. This phrase was made famous by my colleague, Tom Nagel at NYU, who back in 1974 wrote an article called "What Is It Like To Be A Bat?". And the general idea was, well, it's very hard to know what it's like to be a bat from the third person point of view, just looking at it as a human who has different kinds of experience. But presumably, very plausibly, there is something it's like to be a bat. The bat is conscious. It's having subjective experiences, just of a kind very different from ours. In human subjective experience, consciousness divides into any number of different kinds or aspects, like different tracks of the inner movie of consciousness. We have visual experiences like the experience of, say, these colors, blue and red and green from the first person point of view and of depth. There are sensory experiences like the experience of my voice, experiences of taste and smell. They're experiences of your body. Feeling pain or orgasms or hunger or a tickle or something, they all have some distinctive first person quality. Mental images like recalled visual images, emotional experiences like a experience of happiness or anger. And indeed, we all seem to have this stream of a current thought or at the very least, we're kind of chattering away to ourselves and reflecting and deciding. All of these are aspects of subjective experience, things we experience from the first person point of view. And I think these subjective experiences are, at least on the face of it, data for the science of consciousness to explain. These are just facts about us that we're having these subjective experiences. If we ignore them, we're ignoring the data. So if you catalog the data that, say, the science of consciousness needs to explain, there are certainly facts about our behavior and how we respond in situations. There are facts about how our brain is working. There are also facts about how subjective experiences, and on the face of it, they're data. And it's these data that pose what I call the hard problem of consciousness. But this gets contrasted with the easy problems, the so-called easy problems of consciousness, which are the problems of explaining behavioral and cognitive functions. Objective things you can measure from the third person point of view typically tied to behavior. Perceptual discrimination of a stimulus, I can discriminate two different things in my environment. I can say, that's red, and that's green. I can integrate the information about the color and the shape. I can use it to control my behavior. Walk towards the red one rather than the green one. I can report it, say that's red, and so on. Those are all data too for science to explain. But we've got a bead on how to explain though they don't seem to pose as big a problem. Why? We explain those easy problems by finding a mechanism, typically a neural or computational mechanism that performs the relevant function to explain how it is that I get to say there's a red thing over there or walk towards it. Well, you find the mechanisms involving perceptual processes and action processes in my brain that leads to that behavior. Find the right mechanism that performs the function you've explained what needs to be explained with the easy problems of consciousness. But for the hard problem, for subjective experience, it's just not clear that this standard method works. It looks like explaining all that behavior still leaves open a further question. Why does all that give you subjective experience? Explain the reacting, the responding, the controlling, the reporting, and so on. It still leaves open the question, why is all that accompanied by subjective experience. Why doesn't it go on in the dark without consciousness, so to speak? There seems to be what the philosopher Joe Levine has called a gap here, an explanatory gap, between physical processes and subjective experience. At least our standard kinds of explanation, which work really well for the easy problems of behavior and so on, don't obviously give you a connection to the subjective aspects of experience. And there's been a vast amount of discussion of these things over-- I mean, well, for centuries, really. But it's been particularly active in recent decades, philosophers, scientists, all kinds of different views. Philosophically, you can divide approaches to the hard problem into at least two classes. One is an approach on which consciousness is taken to be somehow irreducible and primitive. We can't explain it in more basic physical terms, so we take it as a kind of primitive. And that might lead to dualist theories of consciousness where consciousness is somehow separate from and interacts with the brain. Recently very popular has been the class of panpsychist theories of consciousness. I know Galen Strawson was here a while back talking. He very much favors panpsychist theories where consciousness is something basic in the universe underlying matter. And indeed, there are idealist theories where consciousness underlies the whole universe. So these are all extremely speculative but interesting views that I've explored myself. There are also a reductionist theories of consciousness from functionalist approaches, where consciousness is just basically taken to be a giant algorithm or computation, biological approaches to consciousness-- my colleague Ned Block was here, I know, talking about neurobiology-based approaches, where it's not the algorithm that matters, but the biology it's implemented in-- and indeed, the kind of quantum approaches that people like Roger Penrose and Stuart Hameroff have made famous. I think there's interesting things to say about all of these approaches. I think that right now, at least, most of the reductionist approaches leave a gap. But the non-reductionist approaches have other problems in seeing how it all works. Today, I'm going to take a different kind of approach, this approach through the meta-problem. One way to motivate this is to-- I often get asked, well, you're a philosopher. It's fine. You get to think about these things like the hard problem of consciousness. How can I, as a scientist or an engineer or an AI researcher-- how can I do something to contribute, to help get this at this hard problem of consciousness? Is this just a problem for philosophy? For me to work on it as a AI researcher, I need something I can operationalize, something I can work with and try to program. And as it stands, it's just not clear how to do that with the hard problem. If you're a neuroscientist, there are some things you can do. You can work with humans and look at their brains and look for the neural correlates of consciousness, the bits of the brain that go along with being conscious. Because at least with humans, we can take as a plausible background assumption that the system is conscious. For AI, we can't even do that. We don't know which AI systems we're working with that are conscious. We need some operational criteria. In AI, we mostly work on modeling things like behavior and objective functioning. For consciousness, those are the easy problems. So how does someone coming from this perspective make a connection to the hard problem of consciousness? Well, one approach is to work on certain problems among the easy problems of behavior that shed particular light on the hard problem. And that's going to be the approach that I look at today. So the key idea here is there are certain behavioral functions that seem to have a particularly close relation to the hard problem of consciousness. In particular, we say things about consciousness. We make what philosophers call phenomenal reports, verbal reports of conscious experiences. So I'll say things like, I'm conscious, I'm feeling pain right now, and so on. Maybe the consciousness and the pain are subjective experiences. But the reports, the utterances, I am conscious, well that's a bit of behavior. In principle, explaining those is among the easy problems. It's objectively measurable response. We can find a mechanism in the brain that produces it. And among our phenomenal reports, there's the special class we can call the problem reports, reports expressing our sense that consciousness poses a problem. Now admittedly, not everyone makes these reports. But they seem to be fairly widespread, especially among philosophers and scientists thinking about these things. But furthermore, it's a sense that it's fairly easy to find a very wide class of people who think about consciousness. People say things like, there is a problem of consciousness, a hard problem. On the face of it, explaining behavior doesn't explain consciousness. Consciousness seems non-physical. How would you ever explain the subjective experience of red and so on? It's an objective fact about us-- at least about some of us-- that we make those reports. And that's a fact about human behavior. So the meta-problem of consciousness then, at a second approximation, is roughly the problem of explaining these problem reports, explaining, you might say, the conviction that we're conscious and that consciousness is puzzling. And what's nice about this is that although the hard problem is this airy fairy problem about subjective experience that's hard to pin down, this is a puzzle ultimately about behavior. So this is an easy problem, one that ought to be open to those standard methods of explanation in the cognitive and brain sciences. So there's a research program. There's a research program here. So I like to think of the meta-problem as something we could play that role. I talked about earlier, if you're an AI researcher thinking about this, the meta-problem is an easy problem, a problem about behavior, that's closely tied to the hard problem. So it's something we might be able to make some progress on using standard methods of thinking about algorithms and computations or thinking about brain processes and behavior while still shedding some light, at least indirectly, on the hard problem. It's more tractable than the hard problem. But solving it ought to shed light on the hard problem. And today, I'm just going to kind of lay out the research program and talk about some ways in which it might potentially shed some light. This is interesting to a philosopher because it looks like an instance of what people sometimes call genealogical analysis. It goes back to Friedrich Nietzsche on the genealogy of morals. Instead of thinking about what's good or bad, let's look at where our sense of good or bad came from, the genealogy of it all in evolution or in culture or in religion. And people think a genealogical approach to God, instead of thinking about does God exist or not, let's look at where our belief in God came from. Maybe there's some evolutionary reason for why people believe in God. This often leads, not always, but often leads to a kind of debunking of our beliefs about those domains. Explain why we believe in God in evolutionary terms, no need for the God hypothesis anymore. Explain how moral beliefs and evolutionary terms, maybe no need to take morality quite so seriously. So some people, at least, are inclined to take an approach like this with consciousness too. If you think about the meta-problem explaining our beliefs about consciousness, that might ultimately debunk our beliefs about consciousness. This leads to a philosophical view, which has recently attracted a lot of interest, a philosophical view called illusionism, which is the view that consciousness itself is an illusion. Or maybe that the problem of consciousness is an illusion. Explain the illusion, and we dissolve the problem. I take that in terms of the meta-problem, that view roughly comes to solve the meta-problem. It will dissolve the hard problem. Explain why it is that we say all these things about consciousness, why we say, I am conscious, why we say, consciousness is puzzling. If you can explain all that in algorithmic terms, then you'll remove the underlying problem because you'll have explained why we're puzzled in the first place. Actually, walking over here today, I noticed that just a couple of blocks away, we have the Museum of Illusions, so I'm going to check that out later on. But if illusionism is right, added to all those perceptual illusions is going to be the problem of consciousness itself. It's roughly an illusion thrown up by having a weird kind of self model with a certain kind of algorithm that attributes to ourselves special properties that we don't have. So one line on the meta-problem is the illusionist line. Solve the meta-problem, you'll get to treat consciousness as an illusion. That's actually a view that has many antecedents in the history of philosophy, one way or another. Even Immanuel Kant and his great critique of pure reason had a section where he talked about the self or the soul as a transcendental illusion. We seem to have this indivisible soul. But that's the kind of illusion thrown out by our cognitive processes. The Australian philosophers, Ullin Place and David Armstrong, had versions of this that I might touch on a bit later. Daniel Dennett, a leading reductionist thinker about consciousness has been pushing for the last couple of decades the idea that consciousness involves a certain kind of user illusion. And most recently, the British philosopher, Keith Frankish, has been really pushing illusionism as a theory of consciousness. He has a book centering around a paper by Keith Frankish on illusionism as a theory of consciousness that I recommend to you. So one way to go with the meta-problem is the direction of illusionism. But one nice thing about-- many people find illusionism completely unbelievable. They find, how could it be that consciousness is an illusion? Look, we just have these subjective experiences. It's a data about our nature. And I confess, I've got some sympathy with that reaction. So I'm not an illusionist myself. I'm a realist about consciousness in the philosopher's sense, where a realist about something is someone who believes that thing is real. I think consciousness is real. I think it's not an illusion. I think that solving the meta-problem does not dissolve the hard problem. But the nice thing about the meta-problem is you can proceed on it-- to some extent, at least in initial neutrality-- on that question, is consciousness real or is it an illusion. It's a basic problem about our objective functioning in these reports. What explains those? There's a neutral research program here that both realists, illusionists, people of all kinds of different views of consciousness can explain. And then we can come back and look at the philosophical consequences. So I'm not an illusionist. I think consciousness is real. I've got to say, I do feel the temptation of illusionism. I find it really intriguing and in some ways attractive view. It's just fundamentally unbelievable. Nevertheless, I think that the meta-problem should be a tractable problem. Solving it, at the very least, will shed much light on the hard problem of consciousness even if it doesn't solve it. If you can explain our conviction that we're conscious, somehow the source, the roots of our conviction that we are conscious, must have something to do with consciousness especially if consciousness is real. So I think it's very much a good research program for people to explain. So then I'll move on now to just outlining the research program a little bit more and then talk a bit about potential solutions and on impact on theories of consciousness before wrapping up with a little bit more about illusionism. So this meta-problem, which I've been pushing recently, opens up a tractable empirical research program for everyone, reductionists, non-reductionists, illusionists, non-illusionists. We can try to solve it and then think about the philosophical consequences. Now what is the meta-problem? Well, the way I'm going to put it is it's the problem of topic-neutrally explaining problem intuitions or else explaining why that can't be done. And I'll unpack all the pieces of that right now. First, starting with problem intuitions. What are problem intuitions? Well, there are the things we say. There are things we think I say. Consciousness seems irreducible. I might think consciousness is irreducible. People might be disposed, have a tendency to say or think those things. Problem intuitions all take to be roughly, that tendency. We have dispositions to say and think certain things about consciousness. What are the core problem intuitions? Well, I think they break down into a number of different kinds. There is the intuition that consciousness is non-physical. We might think of that as a metaphysical intuition about the nature of consciousness. There are intuitions about explanation. Consciousness is hard to explain, explaining behavior doesn't explain consciousness. There are intuitions about knowledge of consciousness. Some of you may know the famous thought experiment of Mary in the black and white room who knows all about the objective nature of color vision and so on, but still doesn't know what it's like to see red. She sees red for the first time. She learns something new. That's an intuition about knowledge of consciousness. There are what philosophers call modal intuitions about what's possible or imaginable. One famous case is the case of a zombie, a creature who is physically identical to you and me but not conscious. Or maybe an AI system, which is functionally identical to you and me, but not conscious. That at least seems conceivable to many people. So this is the philosophical zombie. Unlike the zombies and movies, which have weird behaviors and go after brains and so on, the philosophical zombie is a creature that seems, at least behaviorally, may be physically like a normal human, but doesn't have any conscious experiences. All the physical states, none of the mental states. And it seems to many people that's at least conceivable. We're not zombies. I don't think anyone here is a zombie-- I hope. But nonetheless, it seems that we can make sense of the idea. And one way to pose the hard problem is, why are we not zombies. So this imagined ability of zombies is one of the intuitions that gets the problem going. And then you can go on and catalog more and more intuitions about the distribution of conscious, maybe the intuition that robots won't be conscious. That's an optional one, I think. Or consciousness matters morally in certain ways, and the list goes on. So I think there is an interdisciplinary research program here of working on those intuitions about consciousness and trying to explain them. Experimental psychology and experimental philosophy-- a newly active area-- can study people's intuitions about consciousness. We can work on models of these things, computational models or neurobiological models, of these intuitions and reports. And indeed, I think there's a lot of room for philosophical analysis. And there's just starting to be a program of people doing these things in all these fields. I mean, it is an empirical question, how widely these intuitions are shared. You might be sitting there thinking, come on, I don't have of these intuitions. Maybe this is just you. My sense is-- from the psychological study to date-- it seems that some of these intuitions about consciousness are at least very widely shared, at least as dispositions or intuitions, although they are often overridden on reflection. But the current data on this is somewhat limited. Although there is a lot of empirical work on intuitions about the mind concerning things like belief, like when do kids get the idea that your beliefs about the world can be false, concerning the way your self persists through time-- could you exist after the death of your body-- where consciousness is concerned, there's work on the distribution of consciousness. Could a robot be conscious? Could a group be conscious? Here's a book by Paul Bloom, "Decartes' Baby" that catalogs a lot of this interesting work, making the case that many children are intuitive dualists. Thinks they're naturally inclined to think there's something non-physical about the mind. So far, most of this work has not been so much on these core problem intuitions about consciousness, but there's work developing in this direction. Sara Gottlieb and Tania Lombrozo have a very recent article called "Can Science Explain The Human Mind" on people's judgments about when various mental phenomena are hard to explain. And they seem to find that yes, subjective experience and things to which people have privileged first person access seem to pose the problem big time. So there's the beginning of a research program here. I think there's room for a lot more. The topic neutrality part-- when I say we're looking for a topic neutral explanation of problem intuitions, that's roughly to say an explanation that doesn't mention consciousness itself. It's put in neutral terms. It's neutral on the existence of consciousness. The most obvious one would be something like an algorithmic explanation. Now here is the algorithm the brain is executing that generates our conviction that we're conscious and our reports about consciousness. There may be some time between an algorithm and consciousness, but to specify the algorithm, you don't need to make claims about consciousness. So the algorithmic version of the meta-problem is roughly find the algorithm that generates our problem intuition. So that's, I think, in principle a research program that maybe an AI researchers in combination with psychologists-- the psychologist could help isolate data about the way that the human beings are doing it, how these things are generated in humans. And the AI researcher can try and see about implementing that algorithm in machines and see what results. And I'll talk about a little bit of research in this direction in just a moment. OK now I want to say something about potential solutions to the problem. Like I said, this is a big research program. I don't claim to have the solution to the meta-problem. I've got some ideas, but I'm not going to try and lay out a major solution. So here are a few things, which I think might be part of a solution to the problem, many of which have got antecedents here and there in scientific and philosophical discussion. Some promising ideas include retrospective models, phenomenal concepts, introspective opacity, the sense of acquaintance. Let me just say something about a few of these. One starting idea that almost anyone is going to have here is somehow models of ourselves are playing a central role here. Human beings have models of the world, naive physics, naive psychology, models of other people, and so on. We also have models of ourselves. It makes sense for us to have models of ourselves and our own mental processes. This is something that the psychologist Michael Graziano has written a lot on. We have internal models of our own cognitive processes, including those tied to consciousness. And somehow something about our introspective models explains our sense, A, that we are conscious and B, that this is distinctively problematic. And I think anyone thinking about the meta-problem, this has got to be at least the first step. We have these introspective models. If you were an illusionist, they'll be false models. If you're a realist, they needn't be false models. But at the very least, these introspective models are involved, which is fine. But the devil's in the details. How do they work to generate this problem? A number of philosophers have argued we have special concepts of consciousness, introspective concepts of these special subjective states. People call these phenomenal concepts, concepts of phenomenal consciousness. And one thing that's special is these concepts are somehow independent of our physical concepts. They explain we've got one set of physical concepts for modeling the external world. We've got one set of introspective concepts from modeling our own mind. And these concepts, just by virtue of the way they're designed, are somewhat independent of each other. And that partly explains why consciousness seems to be independent of the physical world intuitively. So maybe that independence of phenomenal concepts could go some distance to explaining our problem reports. So I think there's got to be something to this as well. At the same time, I don't this goes nearly far enough because we have concepts of many aspects of the mind, not just of the subjective experiential past but things we believe and things we desire. And so when I believe that Paris is the capital of France, that's part of my internal self model. But that doesn't seem to generate the hard problem in nearly the same way in which the experience of red does. So a lot more needs to be said about what's going on in cases like having the experience of red and having the sense that that generates a gap. So it doesn't generalize to everything about the mind. Some people have thought that what we might call introspective opacity plays a role, that when we introspect what's going on in our minds, we don't have access to the underlying physical states. We don't see the neurons in our brains. We don't see that consciousness is physical. So we see it as non-physical. Most recently, the physicist Max Tegmark has argued in this direction, saying somehow consciousness is substrate-independent. We don't see the substrate. So then we think maybe it can float free of the substrate. Armstrong made an analogy with the case of someone in a circus where-- the headless person illusion where someone's there with a veil across their head, and you don't see their head. So you see them as having no head. Here is a 19th century booth at a circus, so-called headless woman. There's a veil over her head. You don't see the head so somehow, it looks-- at least for a moment-- like the person doesn't have a head. So Armstrong says maybe that's how it is with consciousness. You don't see it as physical, so you see it as non-physical. But still the question comes up, how do we make this inference. There's something that's special goes on in cases like color and taste and so on. The color experience seems to attribute primitive properties to objects like redness, greenness, and so on, when, in fact, in the external world at the very least, they have complex reducible properties. Somehow, our internal models of color treat colors like red and green as if they are primitive things. It turns out to be useful to have these models of things. We treat certain things as primitive, even though they're reducible. And it sure seems that when we experience colors, we experience greenness as a primitive quality even though it may be a very, very complex reducible property. That's something about our model of colors. The philosopher Wolfgang Schwartz tried to make an analogy with sensor variables in image processing. You've got some visual senses and a camera or something you need to process the image. Well, you've got some sensor variables to represent the sensory inputs that the various sensors are getting. And you might treat them as a primitive dimension because that's the most useful way to treat them. You don't treat them as certain amounts of lights or photons firing. You don't need to know about that. You use these sensor variables and treat them as a primitive dimension. And all that will play into a model of these things as primitive, maybe taking that idea and extending it to introspection. These conscious states are somehow like sensor variables in our model of the mind. And somehow, these internal models give us the sense of being acquainted with primitive concrete qualities and of our awareness of them. This is still just laying out. I don't think this is still yet actually explaining a whole lot. But it's laying out-- it's narrowing down what it is that we need to explain to solve the meta-problem. But just to put the pieces together, here's a little summary. One thing I like about this summary is you can read it in either an illusionist tone of voice, as an account of the illusion of consciousness-- so this is how false introspective models work-- or in a realist tone of voice, as an account of our true correct models of consciousness. But we can set it out in a way which is neutral on the two and then try and figure out later whether these models are correct, as the realist says, or incorrect, as the illusionist says. We have introspective models deploying introspective concepts of our internal states that are largely independent of our physical concepts. These concepts are introspectively opaque, not revealing any of the underlying mechanisms. Our perceptual models perceptually attribute primitive perceptual qualities to the world. And our introspective models attribute primitive mental relations to those qualities. These models produce the sense of acquaintance, both with those qualities and with our awareness of those qualities. Like I said, this is not a solution to the meta-problem, but it's trying, at least, to pin down some parts of the roots of those intuitions and to narrow down what needs to be explained. To go further, you want, I think, to test these explanations, both with psychological studies to see if this is plausibly what's going on in humans-- this is the kind of thing which is the basis of our intuitions-- and computational models to see if, for example, we could program this kind of thing into an AI system and see if it can generate somehow qualitatively similar reports and intuitions. You might think that last thing is a bit far fetched right now, but I know of at least one instance of this research program, which has been put into play by Luke Muehlhauser and [INAUDIBLE] two researchers at Open Philanthropy very interested in AI and consciousness. They actually built-- they took some ideas about the meta-problem from something I'd written about it and from something that the philosopher Francois Kammerer had written about it. A couple of basic ideas about where problem intuitions might come from. And they tried to build them into a computational model. They built a little software agent, which had certain axioms about colors and how they work. There's the red and there's green and certain axioms about their own subjective experiences of colors. And then they combined it with a little theorem prover. And they saw what did this little software agent come up with. And it came up with claims like, hey, well, my experiences of color are distinct from any physical state, and so on. OK they cut a few corners. This is not a yet truly a convincing sophisticated model of everything going on in the human mind. But it shows that there's a research program here of trying to find the algorithmic basis of these states. And I think as more sophisticated models develop, we might be able to use these to kind of provide a way in for AI researchers in thinking about this topic. Of course, there is the question, you model all this stuff better and better in a machine, then is the machine actually going to be conscious or is it just going to have found self models that replicate what's going on in humans. So some people have proposed an artificial consciousness test. Aaron Sloman, Susan Schneider, Ed Turner have suggested somehow that if a machine seems to be puzzled about consciousness in roughly the ways that we are, maybe that's actually a sign that it's conscious. So if a machine actually looks to us as if it's puzzled by consciousness, is that a sign of consciousness? These people-- this is suggested as a kind of Turing test for machine consciousness. Find machines which are conscious like we are. Of course, the opposing point of view is going to be no, the machine is not actually conscious. It's just like machine that studied up for the Turing test by reading the talk like a human book. It's like, damn, do I really need to convince those humans that I'm conscious by replicating all those ill-conceived confusions about consciousness. Well I guess I can do it if I need to. Anyway, I'm not going to settle this question here. But I do think that if we somehow find machines being puzzled, it won't surprise me that once we actually have serious AI systems, which engagement in natural language and modeling of themselves and the world, they might well find themselves saying things like, yeah, I know in principle I'm just a set of silicon circuits, but I feel like so much more. I think that might tell us something about consciousness. Let me just say a little bit about theories of consciousness. I do think a solution to the meta-problem and a solution to the hard problem ought to be closely connected. The illusionist has solved the meta-problem. You'll dissolve the hard problem. But even if you're not an illusionist about consciousness, there ought to be some link. So here's a thesis. Whatever explains consciousness should also partly explain our judgments, now reports about consciousness. The rationale here is it would just be very strange if these things were independent, if the basis of consciousness played no role in our judgments about consciousness. So they can use this as a way of evaluating or testing theories of consciousness. For theory of consciousness says mechanism M is the basis of consciousness, that M should also partly explain our judgments about consciousness. Whatever the basis is ought to explain the reports. And you can use this. You can bring this to bear on various extant theories of consciousness. Here's one famous current theory of consciousness, integrated information theory developed by Giulio Tononi and colleagues at the University of Wisconsin. Tononi says the basis of consciousness is integrated information, a certain kind of integration of information for which to and he has a measure that he calls phi. Basically, when your phi is high enough, you get consciousness. A consciousness is high phi, and there's a mathematical definition. But I won't go into it here. But it's a really interesting theory. So here's a-- basically it analyzes a network property of systems of units. And it's got a informational measure called phi that's supposed to go with consciousness. Question, if integrated information is the basis of consciousness, It ought to explain problem reports, at least in principle. Challenge, how does that work? And it's at least far from obvious to me how integrated information will explain the problem reports. It seems pretty dissociated from them. On Tononi's view, you can have simulations of systems with high phi that have zero phi. They'll go about making exactly the same reports but without consciousness at all. So phi is at least somewhat dissociable. You get systems with very high phi, but no tendency to report. Maybe that's less worrying. Anyway, here's a challenge for this theory, for other theories. Explain, not just how high phi gives you consciousness, but how it plays a central role in the algorithms that generate problem reports. Something similar goes for many other theories, biological theories, quantum theories, global workspace, and so on. But let me just wrap up by saying something about the issue of illusionism that I was talking about near the start. Again, you might be inclined to think that this approach through the meta-problem tends, at least very naturally, to lead to illusionism. And I think it can be-- it certainly provides, I think, some motivation for illusionism, the view that consciousness doesn't exist, we just think it does. On this view, again, a solution to the meta-problem dissolves the hard problem. So here's one way of putting the case for illusion. If there is a solution to the meta-problem, then there is an explanation of our beliefs about consciousness that's independent of consciousness. There's an algorithm that explains our beliefs about consciousness. It doesn't mention consciousness. Arguably, it could be in place without consciousness. Arguably, that kind of explanation could debunk our beliefs about consciousness the same way that perhaps explaining beliefs about God in evolutionary terms might debunk belief in God. It certainly doesn't prove that God doesn't exist. You might think that if you can explain our beliefs in terms of evolution, it somehow removes the justification or the rational basis for those beliefs. So something like that, I think, can be applied to consciousness too. And there's a lot to be said about analyzing the extent to which this might debunk the beliefs. On the other hand, the case against illusionism is very, very strong for many people. And the underlying worry is that some of the illusionism is completely unbelievable. It's just a manifest fact about ourselves that we have conscious experience, we experience red, we feel pain, and so on. To deny those things is to deny the data. No, the dielectric here is complicated. The illusionist will come back and say, yes, but I can explain why illusionism is unbelievable. These models we have, these self models of consciousness, are so strong that they were just wired into us by evolution. They're not models we can get rid of. So my view predicts that my view is unbelievable. And the question is, what-- the dialectical situation is complex and interesting. But maybe I could just wrap up with two expressions of absurdity on either side of this question, the illusionist and the anti-illusionist, both finding absurdity in the other person's views. Here's Galen Strawson, who was here. Galen's view is very much that illusionism is totally absurd. In fact, he thinks it's the most absurd view that anyone has ever held. There occurred in the 20th century, the most remarkable episode in the whole history of ideas, the whole history of human thought, a number of thinkers denied the existence of something we know with certainty to exist, consciousness. He thinks this is just a sign of incredible philosophical pathology. Here's the rationalist philosopher, Eliezer Yudkowsky, and something he wrote a few years ago on zombies and consciousness and the epiphenomenalist view that consciousness plays no causal role, where he was engaging some stuff I wrote a couple of decades ago. He said, "this zombie argument"-- the idea we can imagine zombies physically like us but without consciousness-- "may be a candidate for the most deranged idea in all of philosophy. The causally closed cognitive system of trauma's internal narrative is malfunctioning, in a way, not by necessity but just in our own universe miraculously happens to be correct." And here he is expressing this debunking idea that on this view, there's an algorithm that generates these intuitions about consciousness. And that's all physical. And there's also this further layer of non-physical stuff. And just by massive coincidence, the physical algorithm is a correct model of the non-physical stuff. That's a form of debunking here. It would take a miracle for this view to be correct. So I think both of these views are onto-- these objections are onto something. And to make progress on this on either side, we need to find a way of getting past these absurdities. You might say, well, there's middle ground between very strong illusionism and very strong epiphenomenalism. It tends to slide back to the same problems. Other forms of illusionism, weaker forms don't help much with the hard problem. Other forms of realism are still subject to this. It takes a miracle for this view to be correct critique. So I think to get beyond absurdity here, both sides need to do something more. The illusionist needs to do more to explain how having a mind could be like this, even though it's not at all the way that it seems. They need to find some way to recapture the data. Realists need to explain how it is that these meta-problem processes are not completely independent of consciousness. Realists need to explain how meta-problem processes, the ones that generate these intuitions and reports and convictions about consciousness, are essentially grounded in consciousness even if it's possible somehow for them or conceivable for them to occur without consciousness. Anyway so that's just to lay out a research program. I think a solution to the meta-problem that meets these ambitions might just possibly solve the hard problem of consciousness or at the very least shed significant light on it. In the meantime, the meta-problem is a potentially tractable research project for everyone, and might I recommend to all of you. Thanks. [APPLAUSE] AUDIENCE: Yes, I just want to say I think it's very interesting, this concept of, we have these collection of mental models and that this collection of mental models is consciousness, basically. Consciousness defines the collection of these mental models that we have. And the problem with consciousness is that we don't understand the physical phenomenon that causes these mental models or that stimulates these mental models. So we just have this belief that it's ephemeral or not real or something like that. And if you take that view, then what's interesting is that you could simulate these mental models like robot could simulate these mental models. And you could simulate consciousness as well. And even if the underlying physical phenomena that fuels these mental models is different-- robots have different sensors, et cetera-- you could still get the same consciousness effect in both cases. DAVID CHALMERS: Yeah, I think that's right. Or at the very least, it looks like you ought to be able to get the same models, at least, in a robot. If the models themselves are something algorithmic and ought to be, you ought to be able to design a robot that has, at the very least, let's say, isomorphic models and some sense that is conscious. Of course, it's a further question-- at least by my [? lights-- ?] whether then the robot will be conscious. And that was the question I alluded to in talking about the artificial consciousness test. But you might think that would at least be very good evidence that the robot is conscious. If it's got a model of consciousness just like ours, it seems very plausible there ought to be a very strong link between having a model like that and being conscious. I think probably something like Ned Block-- who was here arguing against machine consciousness-- would say, no, no, the model is not enough. The model has to be built of the right stuff. So it's gotta be built of biology. And so on. But at by my [? light, ?] I think if I had found an AI system that had a very serious version of our model of consciousness, I'd take that as a very good reason to believe it's conscious. AUDIENCE: In the IIT theory, is there a estimate or plausible estimate for what the value of phi is for people and for other systems? DAVID CHALMERS: Basically, no. It's extremely hard to measure in systems of any size at all. Because the way it's defined, it involves taking a sum over every possible partition of a system. It turns out A, it's hard to measure in the brain because you've got to involve the causal dependencies set between different units on neurons. But even for a pure algorithmic system, you've got a neural network laid out in front of you. It's computationally intractable to measure the fire of one of those once I get to bigger than 15 units or so. So Tononi would like to say this is an empirical theory and in principle empirically testable. But there's the in principle. It's extremely difficult to measure phi. Some people, Scott Aaronson the computer scientist, has tried to put forward counterexamples to the theory, which were basically very, very simple systems like matrix multipliers that multiply two large matrices. Turn out to have enormous phi, phi as big as you like if the matrices are big enough and therefore by Tononi's theory will not just be conscious, but as conscious as a human being. And Aaronson put this forward as a reductio ad absurdum of the IIT theory. I think Tononi basically bit the bullet and said, oh yeah, those matrix multipliers are actually having some high degree of consciousness. So I think IIT is probably at least missing a few pieces if it's going to be developed. But it's a research program too. AUDIENCE: You mentioned belief as an example of something where there's another mental quality, but people don't seem to have the same sense that it is very hard to explain. In fact, it almost seems too easy where people-- like a belief about something sort of feels like just how things are. But you have to reflect on a belief to notice it as a belief. Do you think there is also or has there been research related to this question into why is that different? It seems like another angle of attack on this problem. It's just like, why doesn't this generate the same hard problem. DAVID CHALMERS: Yeah. In terms-- I'm not sure if there's been research from the perspective of the meta-problem or of theory of mind. Certainly, people have thought, in their own right, what is the difference in belief and experience that makes them so different. This goes way back to David Hume, a philosopher a few centuries ago who said basically, perception is vivid. Impression of impressions and ideas. And impressions like experiencing colors are vivid in force and vivacity, and ideas are merely a faint copy or something. But that's just the first order. And then there are contemporary versions of this kind of thing, far more sophisticated ways of saying a similar thing. But yeah, you could, in principle, explore that through the meta-problem. Why does it seem to us that perception is so much more vivid? What about our models of the mind makes perception seem so much more vivid than belief and makes beliefs seem structural and empty whereas perception is so full of light? But no, I don't know of work on that from the meta-problem perspective. Like I said, there's not that much work on these introspective models directly. There is work on theory of mind about beliefs. Tends to be about models of other people. It may be there's something I could dig through in my literature on belief that says something about that. It's a good place to push. AUDIENCE: Thanks. AUDIENCE: I wanted to bring up Kurt Godel. You mentioned your advisor wrote "Godel, Escher, Bach". There's something that seems very like Godel-- Godelian or whatever about this whole discussion in that-- so Godel showed that given a set of axioms in mathematics, that it would either be consistent or complete but not both. And it seems like when Daniel Dennett-- Daniel Dennett seems to have a set of axioms where he cannot construct consciousness from them. He seems to be very much in this consistent camp, like he wants to have a consistent framework but is OK with the incompleteness. And I wonder if a similar approach could be taken with consciousness where we could, in fact, prove that consciousness is independent of Daniel Dennett's set of axioms, the same way they proved-- after Godel, they proved the Continuum Hypothesis was independent of ZF set theory, and then they added the axiom of choice, made it ZFC set theory. So I wonder if we could show that in Daniel Dennett's world, we are essentially zombies or we are either zombies or not. It doesn't matter. Either statement could be true. And then find what is the minimum axiom that has to be added to Dennett's axioms in order to make consciousness true. DAVID CHALMERS: Interesting. I thought for a moment this was going to go in a different direction and you were going to say Dennett is consistent but incomplete. AUDIENCE: Yes. DAVID CHALMERS: He doesn't have consciousness in his picture. I'm complete. I've got consciousness-- AUDIENCE: Yes. DAVID CHALMERS: --but inconsistent. That's why I say all these crazy theory. AUDIENCE: Right, yeah. DAVID CHALMERS: And you're faced with the choice of not having consciousness and being incomplete or having consciousness and somehow getting this hard problem and being forced into, at least, puzzles and paradoxes. But the way you put it was friendlier to me. Yeah, certainly, Doug Hofstadter himself has written a lot on analogies between the Godelian paradoxes and the MindBody problem. And he thinks always our self models are always doomed to be incomplete in the Godelian way. He thinks that that might be somehow part of the explanation of our puzzlement, at least about consciousness. Someone like Roger Penrose, of course, takes this much more seriously literally. He thinks that the computational aspects of computational systems are always going to be limited in the Godel way. He thinks human beings are not so limited. He thinks we've got mathematical capacities to prove theorems, to see the truth of certain mathematical claims that no formal system could ever have. So he thinks that we somehow go beyond the incomplete Godelian-- I don't know if he actually thinks we're complete, but at least we're not incomplete in the way that finite computational systems are incomplete. And furthermore, he thinks that extra thing that humans have is tied to consciousness. I never quite saw how that last step goes, even if we didn't have these special non-algorithmic capacities to see the truth of mathematical theorems, how would that be tied to consciousness. But at the very least, there are structural analogies to be drawn between those two cases, about incompleteness of certain theories. How literally we should take the analogies, I'd have to think about it. AUDIENCE: Has there been some consideration that the problem of understanding consciousness inherently must be difficult because we address the problem using consciousness? I'm reminded of the halting problem in computer science where we say that in the general case, a program cannot be written to tell whether another program will halt because what if you ran it on itself. It can't be broad enough to include its own execution. So I wonder if there is a similar corollary in consciousness where we use consciousness to think about consciousness and so therefore, we may not have enough equipment there to be able to unpack it. DAVID CHALMERS: Yeah, it's tricky. People say it's like, you use a ruler. To measure a ruler-- well, I can do this ruler to measure many other things. But it can't measure itself. It's not [INAUDIBLE]. Well, on the other hand, you can measure one ruler using another ruler. Maybe you can measure one consciousness using another. The brain-- [INAUDIBLE] the brain can't study the brain. But the brain actually does a pretty good job of studying the brain. There are some self-referential paradoxes there. And I think that, again, is at the heart of Hofstadter's approach. But I think we'd have to look for very, very specific conditions under which systems can't study themselves. I did always like the idea that if the mind was simple enough that we could understand it, we would be too simple to understand the mind. So maybe something like that could be true of consciousness. On the other hand, I actually think that if you start thinking that consciousness can go along with very simple systems, I think at the very least, we ought to be able to study consciousness in other systems simpler than ourselves. And boy, if I could solve the hard problem even in dogs, I'd be satisfied. Yeah? AUDIENCE: Hey, so I have a question about how the meta-problem research program might proceed, sort of related to the last question. So certainly things we believe about our own consciousness, even if we all say them, probably some of them are false. Our brain has a tendency to hide what reality is like. If you look at visual perception, there's what's called lightness constancy. Our brain subtracts out the lighting in the environments so we actually see more reliably what the color of objects are. Like these viral examples of the black and gold dress is an example of this. And when you're presented with an explanation of it, it's like, huh? My brain does that? It's not something we have access to. Or Yani Laurel-- DAVID CHALMERS: Laurel Yani, yeah. AUDIENCE: --illusion is another one where when you hear the explanation, the scientists that understand it, our own introspection doesn't include that. So how do you proceed with trying to get at what consciousness really is versus what our whatever simplified or distorted view might be? DAVID CHALMERS: Yeah, I think well one view here would be that we never have access to the mechanisms that generate consciousness, but we still have access to the conscious states themselves. Actually, Karl Lashley said this decades ago. He said no process of the brain is ever conscious. The processes that get you to the states are never conscious. The states they get you to are conscious. So take your experience of the dress. For me, it was white and gold. So I knew that. Each of us was certain that I am-- I was certain that I was experiencing white and gold. Maybe you were certainly you were experiencing blue and black. AUDIENCE: I forget which it was. All I remember is I was right. [LAUGHTER] DAVID CHALMERS: You were sure that, yeah, those idiots can't be looking at this right. I think the natural way to describe this, at least, is that each of us was certain what kind of conscious experience we were having, but what we had no idea about was the mechanisms by which we got there. So the mechanisms are completely opaque. But the states themselves were at least prima facie transparent. I think that would be the standard view. Even a realist about consciousness could go with that. They'd say we know what conscious states where we know what those conscious states are. We don't know the processes by which they are generated. The illusionist, I think, wants to go much further and say, well it seems to you that you know what conscious state you're having. It seem to you that you're experiencing yellow and gold. Sorry, yellow and white, whatever it was. Gold and white. AUDIENCE: Black and gold is what I remember. DAVID CHALMERS: No, black and blue, I think, and-- AUDIENCE: Blue? DAVID CHALMERS: --gold and white. It seems to you you're experiencing gold and white. But, in fact, that too is just something thrown up by another model. The yellow gold was a perceptual model. Then there was an introspective model that said you are experiencing gold and white when maybe, in fact, you're just a zombie. Or who knows what's actually going on in your conscious state. So the illusionist view, I think, has to somehow take this further and say, not just the processes that generate the conscious states, but maybe the conscious states themselves are somehow opaque to us. AUDIENCE: It feels like some discussion of generality of a problem is missing from this discussion. The matrix multiplier example of having high phi is still-- it's not a general thing. Is there someone exploring the space, the intersection of generality and complexity that leads to consciousness as an emergent behavior? DAVID CHALMERS: When you say generality, there's the idea that a theory should be general, that it should apply to every system. You mean mechanisms? AUDIENCE: Generality of the agent. If I can write an arbitrarily complex program to play tic-tac-toe and all it will ever be able to do is play tic-tac-toe, it has no outputs to express anything else. DAVID CHALMERS: Yeah. So general in the sense of AGI, artificial general intelligence. Some aspects of consciousness seem to be domain general like for example, maybe insofar as belief and reasoning is conscious. Those are domain general. But much of perception doesn't seem especially domain general. Right? Color is very domain. Taste is very domain specific. So it's still conscious. AUDIENCE: But if my agent can't express problem statements, like if I don't give it an output by which it can express problem statements, you can never come to a conclusion about its consciousness. DAVID CHALMERS: I'd like to distinguish intelligence and consciousness and even be able to-- even natural language and being able to address a problem statement and analyze a problem, that's already a very advanced form of intelligence. I think it is very plausible that a mouse has got some kind of consciousness, even though it's got no ability to address problem statements, and many of its capacities may be very specialized. It's still much more general than a simple neural network that can only do one thing. A mouse can do many things. But I'm not sure that I see an essential-- I certainly see a connection between intelligence and generality. We want to say somehow a high degree of generality is required for high intelligence. I'm not sure there's the same connection for consciousness. I think consciousness can be extremely domain-specific as a taste and maybe vision are. Or it can be domain-general. So maybe those two across cut each other a bit. AUDIENCE: So it seems to me like the meta-problem as it's formulated implies some amount of separation or epiphenomenalism between consciousness and brain states. And one thing that I think underlies a lot of people's motivation to do science is that it has causal import. Like predicting behaviors is clearly a functionally useful thing to do, and if you can predict all of behavior without having to explain consciousness, their motivation for explaining consciousness sort of evaporates and it feels like, yeah, well, what's the point of even thinking about that because it's just not going to do anything for me. What do you say to someone when they say that to you? DAVID CHALMERS: What is the thing that they said to me again? AUDIENCE: That maybe consciousness exists, maybe it doesn't. But if I can explain all of human behavior and all of the behavior of the world in general without recourse to such concepts, then I've done everything that there is that's useful, like explaining consciousness isn't a useful thing to do. And thus, I'm not interested in this, and it may-- DAVID CHALMERS: I see. AUDIENCE: --as well not be real. DAVID CHALMERS: I think epiphenomenalism could be true. I certainly don't have any commitment to it, though. It's quite possible that consciousness has a role to play in generating behavior that we don't yet understand. And maybe thinking hard about the meta-problem can help us get clearer on those roles. I think if you've got any sympathy to panpsychism, maybe consciousness is intimately involved with how physical processes get going in the first place. And there are people who want to pursue interactionist ideas where consciousness interacts with the brain. Or if you're a reductionist, consciousness may be just a matter of the right algorithm. In all those views, consciousness may have some role to play. But just say it turns out that you can explain all of behavior, including these problems, without bringing in consciousness. Does that mean that consciousness is not something we should care about and not something that matters? I don't think that would follow. Maybe it wouldn't matter for certain engineering purposes, say you want to build a useful system. But at least in my view, consciousness is really the only thing that matters. It's the thing that makes life worth living. It's what gives our lives meaning and value and so on. So it might turn out that consciousness is not that useful for explaining other stuff. But if it's the source of intrinsic significance in the world, then understanding consciousness would still be absolutely essential to understanding ourselves. Furthermore, if it comes to developing other systems like AI systems or dealing with non-human animals and so on, we absolutely want to know. We need to know whether they're conscious because if they're conscious, they presumably have moral status. If they can suffer, then it's very bad to mistreat them. If they're not conscious, then you might-- I think it's very plausible-- treat non-conscious systems, we can treat how we like. And it doesn't really matter morally. So the question of whether an AI system is conscious or not is going to be absolutely vital for how we interact with it and how we build our society. That's not a question of engineering usefulness. It's a question of connecting with our most fundamental values. AUDIENCE: Yeah, I completely agree. I just-- I haven't found that formulation to be very convincing to others necessarily. AUDIENCE: Hi, thanks so much for coming and chatting with us today. I'm really interested in some of your earlier work, the extended mind [? distributed ?] cognition. And you're at a company speaking with a bunch of people who do an incredibly cognitively demanding task. DAVID CHALMERS: Yeah. AUDIENCE: Most of the literature that I've read on this topic uses relatively simple examples saying like it's difficult to think just inside your head on these relatively simple things. And if you take a look at the programs that we build, on a mundane day-to-day basis, they're millions of lines long. I've read people in the past say something like the Boeing 777 was the most complicated thing that human beings have ever made, and I think most of us would look at that and say, we got that beat. The things that large internet companies do, the size, the complexity of that is staggering. And yet if we close our eyes, everyone in here is going to say, I'm going to have difficulty writing a 10 line program in my head. So I've just sort of, as an open, I'd be very interested in hearing your thoughts about how the activity of programming connects to the extended mind ideas. DAVID CHALMERS: Yeah, so this, I guess, is a reference to something that I got started in about 20 years ago with my colleague, Andy Clark. We wrote an article called "The Extended Mind" about how processes in the mind can extend outside the brain when we become coupled to our tools. And actually, our central example back then in the mid-90s was a notebook, someone writing stuff in a notebook. And even then, we knew about the internet, and we had some internet examples. I guess this company didn't exist yet in '95. But now, of course, our minds have just become more and more extended. And smartphones came along a few years later, and everyone is coupled very, very closely to their phones and their other devices that couple of them very, very closely to the internet. Now it's suddenly the case that a whole lot of my memory is now offloaded onto the servers of your company somewhere or other, whether it's in the mail systems or navigation mapping systems or other systems. Yeah, most of my navigation has been offloaded to maps. And much of my memory [INAUDIBLE] has been offloaded. Well, maybe that's in my phone. But other bits of my memory are offloaded into my file system on some cloud service. So certainly, yeah, vast amounts of my mind are now existing in the cloud. And if I were somehow to lose access to those completely, then I'd lose an awful lot of my capacities. So I think now we are now extending into the cloud thanks to you guys and others. The question's specifically about programming. Programming is a kind of active interaction with our devices. I mean, I think programming is something that takes a little bit longer. It's a longer timescale so the core cases of the extended mind involve automatic use of our devices, which are always ready to hand. We can use them to get information, to act in the moment, which is the kind of thing that the brain does. So insofar as programming is a slower process-- and I remember from my programming days, all the endless hours of debugging and so on-- then it's at least going to be a slower timescale for the extended mind. But still, Feynman talked about writing this way. Someone looked at Feynman's work and a bunch of notes he had about a physics problem he was thinking about. And someone said to him, oh it's nice you have this record of your work. And Feynman said, that's not a record of my work. That's the work. That is the thinking. I was writing it down and so on. I think, at least my recollection from my programming days, is that when you're actually writing a program, it's not like you just do a bunch of thinking and then code your thoughts. The programming is to some very considerable extent your thinking. So is that the sort of thing you're-- AUDIENCE: Yes, absolutely. [INTERPOSING VOICES] If we, I think as people that program, start to reflect on what we do, and very few of us actually-- if you're the tech lead of a system, maybe you've got it in your head. But you would agree that most of the people on the team who have come more recently only have a chunk of it in their head. And yet, they're somehow still able to contribute. DAVID CHALMERS: Oh, yeah. This is now distributed cognition. The extended mind, the extended cognition starts with an individual and then extends their capacities out using their tools or their devices or even other people. So maybe my partner serves as my memory, but it's still centered on an individual. But then there's the closely related case of distributed cognition, where you have a team of people who are doing something and making joint decisions and carrying out joint actions in an absolutely seamless way. And I take it at a company like this, there are going to be any number of instances of distributed cognition. I don't know whether the company as a whole has one giant Google mind or maybe there's just a near infinite number of separate Google minds for all the individual teams and divisions. But I think probably some anthropologist has already done a definitive analysis of distributed cognition in this company. But if they haven't, they certainly need to. AUDIENCE: Thank you. [APPLAUSE]
Info
Channel: Talks at Google
Views: 165,604
Rating: undefined out of 5
Keywords: talks at google, ted talks, inspirational talks, educational talks, The Meta-Problem of Consciousness, Professor David Chalmers, David Chalmers, Consciousness, human behavior
Id: OsYUWtLQBS0
Channel Id: undefined
Length: 71min 11sec (4271 seconds)
Published: Tue Apr 02 2019
Related Videos
Note
Please note that this website is currently a work in progress! Lots of interesting data and statistics to come.