字幕列表 影片播放 列印英文字幕 [APPLAUSE] - Two out of the three fundamental mysteries about our place in the universe have already been resolved. The first is literally about our place in the universe. Many years ago Copernicus told us that we were not at its centre, that we were just a tiny dot suspended in the abyss. This is an image of the earth taken from the probe Voyager 1 as it was leaving the solar system from about six billion kilometres away. All of human history, all of the history of life on Earth, has taken place on that pale blue dot. The second mystery, Darwin then revealed that we humans are just one branch, or one twig, of a beautifully rich and delicate evolutionary tree. And that much of the machinery of life is shared even with the lowliest of our fellow creatures. The third mystery is that of consciousness, our inner universe. Now earlier this year, for the third time in my life, I ceased to exist. As the propofol anaesthetic flowed from the cannula in my wrist into my bloodstream and then into my brain, there was a falling apart. A blackness. An absence. And then, I was back. Drowsy and disoriented, but definitely there. And when you wake from a deep sleep, you might be confused what time it is, especially in flying somewhere, but you'll know that some time has passed. There seems to be some basic continuity between your consciousness then, and your consciousness now. But coming around from a general anaesthetic, it could have been five minutes. It could have been five hours. It could have been five days, or five years. I was simply not there. A premonition of the oblivion of death. And general anaesthesia doesn't just work on your brain. It doesn't just work on your mind. It works on your consciousness. By altering the delicate electrochemical circuitry inside your head, the basic ground state of what it is to be is temporarily abolished. And in this process lies one of the greatest remaining mysteries in science and philosophy. How does consciousness happen? Why is life in the first person? It is going away, and coming back. The modern incarnation of this problem is usually traced to Descartes, who in the 17th century distinguished between matter stuff, res extensa, the stuff that these desks are made of, that clothes are made of. But also the brains and bodies and made of, material stuff. And res cogitans, the stuff of thought, of feelings. The stuff of consciousness. And in making this distinction, he gave rise to the now infamous mind/body problem, and life has never been simple ever since. But Descartes actually generated even more mischief with his doctrine of the beast machine, which I'm going to mention now, because it anticipates where I'm going to end up as the bell rings when I finish in an hour. Before Descartes, people commonly believed in something called the great chain of being, with rocks and plants at one end, and other non-human animals, a bit higher up than humans, and then angels and gods at the very top. And this great scale of being was also a scale of moral virtue, so that humans had more moral virtues than animals and plants, and then rocks and so on. Now Descartes, in making this division between mind and matter, argued that only humans had minds, and therefore moral status, while other animals didn't have minds. They were merely physiological machines, or beast machines, morally equivalent to plants, and to rocks. And in this view, the physiological mechanisms that give rise to the property of being alive were not relevant to the presence of mind or consciousness. Now I'm going to propose, at the end of this talk, the opposite. That our conscious sense of self arises because of, and not in spite of, the fact that we, too, are beast machines. So to get there, let's return to the apparent mystery of consciousness. Now as recently as 1989, which is quite a while ago, but not that long ago, Stuart Sutherland, who was founding professor of experimental psychology at my university of Sussex, had this to say. "Consciousness is a fascinating but elusive phenomenon. It is impossible to specify what it is, what it does, or why it evolved. Nothing worth reading has been written on it." [LAUGHTER] It's quite a pessimistic point of view. And that may have been true then. I don't think it was true then, but in any case things have changed a lot since. And more or less, about the time that Sutherland made these remarks, we can see the birth, or the rebirth, of the study of consciousness within the neurosciences. And a good landmark is this paper by Francis Crick and Christof Koch, published in 1990. And they start their paper by saying that it is remarkable that most of the work in cognitive sciences, and the neurosciences, makes no reference to consciousness or awareness at all. And then they go on to propose their own theory of what the neural correlates of consciousness are. What it is in the brain that is responsible for being conscious. And since then, over the last 25 years, there's been first a trickle, and now a deluge of research on the brain basis of conscious experience. Some of this work is being carried out in my laboratory, the Sackler Centre, the consciousness science, which was founded six years ago with Hugo Critchley, my co-director. And there are now even specialised academic journals, The Neuroscience of Consciousness, which I started last year with Oxford University Press. And this is a real change of the tide. When I started out more than 20 years ago, it was thought to be a very-- it was thought to be career suicide to want to study consciousness, scientifically. And it may still be, we don't know. Let's see. So while the brain basis of consciousness is still a mystery, it is, in some sense, an accessible mystery. And the author, Mark Haddon, put this very nicely, I think. He said the raw material of consciousness is not on the other side of the universe. It didn't happen 14 billion years ago. And it's not squirrelled away deep inside an atom. The raw material of consciousness is right here, inside your head, and you can hold the brain in your hands. But the brain won't deliver its secrets very easily. What's extraordinary about the brain is not so much the number of neurons, though there are about 90 billion. It's not even the number of connections, though there are so many that if you counted one every second, it would take you about three million years to finish counting. What's truly extraordinary are the patterns of connectivity, which to a large extent, are still not known, but within which are inscribed everything that makes you, you. The challenge is then this, at least the the way I see it. How can the structure and dynamics of the brain, in connection with the body and the environment, account for the subjective phenomenological properties of consciousness? And considering things this way, we come up against what the philosopher David Chalmers has often called the hard problem of consciousness. And the idea is this. There is an easy problem. The easy problem is to understand how the combined operations of the brain and the body give rise to perception, to cognition, to thinking, to learning, to behaviour. How the brain works, in other words. The hard problem is to understand why and how any of this should have anything to do with consciousness at all. Why aren't we just robots, or philosophical zombies, without any in a universe? Now there's a tempting intuition here, which is that, even if we solve the hard problem, even if we solve the easy problem, the hard problem would still remain as mysterious as it seems now. But this just seems wrong-headed to me. It may not be necessary to explain why consciousness exists at all, in order to make progress in understanding its material basis. And this for me, is the real problem of consciousness; how to account for its various properties in terms of biological mechanisms without pretending that it doesn't exist at all, as you do if you solve the easy problem, and without trying to account for why it's parts of the universe in the first place, which is the hard problem. And in the history of science, we've been somewhere similar before. It's hard to say if it's exactly the same situation. But in our understanding of life, eminent biochemists of the time found it entirely mysterious how biological mechanisms could give rise to the property of being alive. And there were proposed of things, like elan vital and essence vital, and all sorts of other stuff. And although we don't yet understand everything about life, this initial sense of mystery about life has, to a large extent, dissolved as biologists have just got on with the business of understanding the properties of living systems in terms of mechanisms. An important part of this story was the realisation that life is not just one thing, but rather a constellation of partially dependent, partially separable, processes, like metabolism, homeostasis, and reproduction. In the same way, to make progress on the real problem of consciousness, it can be useful to distinguish different aspects or dimensions of what it is to be conscious. The space of possible minds, if you like. And one simple classification is into conscious level, which is the property of being conscious at all. For example, the difference between being in a dreamless sleep, or under general anaesthesia, and being awake and conscious as you are now. And the conscious content, when you are conscious, you're conscious of something. The myriad of sights, sounds, smells, emotions, feelings, and beliefs that populate your inner universe at any one time. And one thing you are conscious of when you are conscious, is the experience, the specific experience, of being you, and this is conscious self. And it's the third dimension of consciousness. Now I don't claim these distinctions mark completely independent aspects of what it is to be conscious, but they're a pragmatically useful way of breaking down the problem a bit. So let's start with conscious level. What are the fundamental brain mechanisms that underlie our ability to be conscious at all? And we can think of this, at least a first approximation, as a scale from being completely unconscious, as if you were in a coma, or under general anaesthesia, to being awake, alert, and fully conscious as you are now. And there's various states in between being drowsy, being mildly sedated and so on. What's important is that, while being conscious and being awake often go together, this is not always the case. For instance, when you are dreaming you are asleep, but you are having conscious experiences. The conscious experience of your dreams. And on the other side of this diagram, there are pathological states, like the vegetative state, where physiologically you will go through sleep/wake cycles, but there is nobody at home. There is no consciousness happening. So what are the specific mechanisms that underlie being conscious and not simply being physiologically awake? Well there are a number of possibilities. Is it the number of neurons? Well actually, probably not. There are more neurons in your cerebellum, this bit at the back of your brain, than in the rest of your brain put together. In fact are about four times more neurons in your cerebellum than in the rest of your cortex. But if you have damage to your cerebellum, yeah, you'll have some problems with coordination and other things, some cognitive problems, but you won't lose consciousness. It's not just the number of neurons. Doesn't seem to be any particular region. In fact, there are regions that, if you suffer damage, you will permanently lose consciousness; in thalamina, the nuclei, and the thalamus deep inside the brain. But these seems to be more like on/off switches than actual generators of conscious experience. It's not even neural activity, at least not simple kinds of neural activity. Your brain is still highly active doing unconscious states, during the sleep. And even if your brain is highly synchronised, one of the first theories of consciousness was it depended on neurons firing in synchrony with each other. If your brain is too synchronised, you will lose consciousness, and this happens in states of absence epilepsy. What seems to be the case is that, being conscious it all, depends on how different brain regions talk to each other in specific ways. And this was some groundbreaking work by Marcello Massimini in Milan about 10 years ago. And what he did here, was he stimulated the cortex of the brain with a brief pulse of electromagnetic energy, using a technique called transcranial magnetic stimulation or TMS. And then he used EEG electroencephalography to listen to the brain's echos. A little bit like banging on the brain and listening to its electrical response. And what he noticed when you do this, and you can see on the left is asleep, and on the right is awake. And this is very much slowed down. When you stimulate the brain in a sleep condition, there is still a response. There's still an echo, but the echo stays very localised to the point of stimulation. It doesn't travel around very much, and it doesn't last very long. But when you stimulate a conscious brain, there's a spatial temporally complex response. This echo bounces around all over the cortex in very interesting ways. What's more, the complexity of this echo can be quantified. You can apply some simple algorithms to describe how complex, how rich, this pattern of interactivity is. This is also from the Milan group. And what they've done here is, they basically look at the echo as it moves around the brain. And they see the extent to which you could describe it, the minimum description length. How much can you compress the image of that echo? Much the same way that algorithms make compressed files from digital images in your phone. And they came up with an index called the perturbational complexity index. And what you find is, you now have a number that you can attach to how conscious you are. This is, I think, really intriguing, because it's a first step towards having an actual measurement of conscious level. This graph on the bottom shows this measure applied to a variety of conscious states, ranging from pathological conscious states, like the vegetative state, where there is no consciousness at all, all the way through locked in syndrome, and then healthy waking. And you can immediately see that techniques like this might already have clinical value in diagnosing potential for consciousness patients might have after severe brain injury. Now at Sussex, we are continuing work along these lines. We actually look, instead of bashing on the brain with a sharp pulse of energy, we want to see whether we can get something similar just by recording the spontaneous activity of the brain. So we look at spontaneous dynamics from, in this case, waking states and anaesthesia. This is work with my PhD students Michael Schartner and Adam Barrett. We measure its complexity, and indeed, we find that we can distinguish different levels of consciousness just by the spontaneous activity of the brain. In a way this isn't that surprising, because we know various things change. The balance of different frequencies of your brain activity changes when you lose consciousness. But this doesn't have to do with that. This is independent of that. There's something specific that is being detected by these changes in complexity. More recently we've applied the same measures to sleep, in this case taking advantage with colleagues in Milan of recordings taken from directly within the human cerebral cortex. These are implanted electrodes. And we see much the same story. If you compare where the two areas are, you compare the complexity of wakeful rest, and early non-rem sleep, where you are not dreaming very much. You see that complexity fools a great deal. What's interesting here is, if you compare wakeful rest to REM sleep, where people will often report dreams if you wake them up, the level of complexity is very much as it is during the wakeful state. There's something else going on here, which is that the complexity in the frontal part of the brain seems to be higher than in other parts of the brain. And that's something we still don't understand fully, yet. I just wanted to give you something hot off the press, so to speak, which is where you've also been applying these measures now to data taken from people under the influence of psychedelic drugs; psilocybin, ketamine, and LSD. And what we find, at least in our hands to start with here, is that the level of complexity actually increases as compared to the baseline state, which is not something we've seen before in any other application of these measures. So what's important about this way of looking at things is that, it's grounded in a theory that tries to explain why certain sorts of brain dynamics go along with being conscious. And put very simply, the idea is this-- and it goes back to Guilio Tononi and Gerald Edelman, people that I went to work with in America about 18 years ago-- the idea is very simple. Consciousness is extremely informative. Every conscious experience you have, or have had, or will have, is different from every other conscious experience you have had, are having, or will have. Even the experience of pure darkness rules out a vast repertoire of alternative possible experiences that you could have, or might have in the future. There's a huge amount of information for the organism in any conscious experience. At the same time every experience that you have is highly integrated. Every conscious scene is experienced, as all of a piece, is bound together. We don't experience colours and shapes separately in any way. It's conscious experiences at the level of phenomenology combine these properties. They are the one hand highly informative, composed of many different parts. On the other, bound together in a unified whole. And this motivates us to search for mathematical measures which have the same property, which are neither lacking in information. On the left, you see a system which is all connected together, so it can't enter very many different states. On the right is a system which is completely dissociated, so they can enter states, but it's not a single system. We want measures that track this middle ground of systems, that combine both integration and differentiation. And a number of these measures now exist. There are some equations here, which we can talk about later if you like, that try to target this middle ground. And time will tell whether, by applying these more phenomenologically grounded measures, we come up with even more precise practical measures of consciousness. Now why is this business of measurement important? And I want to make a general point here, which is that, if you're trying to naturalise a phenomenon which seems mysterious, the ability to measure it is usually one of the most important steps you can take. And we've seen numerous examples of this. The history of our understanding of heat and temperature is one very good example Here's an early thermometer from the 18th century, which used the expansion of air. But of course there are many problems in generating a reliable thermometer and a scale of temperature, if you don't already have fixed points. And if you don't know what heat is you get trapped in a kind of vicious circle that took a long time to break out of. But people did break out of this, and the development of the thermometers catalysed our understanding of heat from being something that flowed in and out of objects, to being something that was identical to a physical property. The mean molecular kinetic energy of molecules in a substance. And having that concept of heat now allows us to talk about temperature far beyond the realms of human experience. We can talk about the temperature on the surface of the sun, in a sense, the way we can talk about the temperature of interstellar space, close to absolute zero. None of these things make any sense and in our phenomenological experience of hot and cold. So this brings me to my first take-home message. Measurement is important, and consciousness, conscious level, depends on a complex balance of differentiation and integration in brain dynamics, reflecting the fact that conscious experiences themselves are both highly informative and always integrated. Now when we are conscious, we are conscious of something. So what are the brain mechanisms that determine the content of consciousness? And the hero for this part of the story is the German physicist and physiologist Hermann Von Helmholtz. And he proposed the idea that the brain is a kind of prediction machine. That what we see, hear, and feel are nothing other than the brain's best guess about the causes of sensory inputs. And the basic idea is, again, quite simple. The brain is locked inside its bony skull home, and has very indirect access to the external world. All it receives are ambiguous and noisy sensory signals, which are highly and directly related to this external world of objects, and so on, if there is an external world of objects out there at all. They know about that. Perception in this view is, by necessity, a process of inference in which the brain interprets these ambiguous and noisy sensory signals with respect to some prior expectations or beliefs about the way the world is. And this forms the brain's best guess of the causes of the sensory signals that are impacting our sensory surfaces all the time. What we see is the brain's best guess of what's out there. I want to give you a couple of examples that illustrate this process. It's quite easy to do, in a way. This first example is a well-known visual illusion called Edelstein's Checkerboard. Now here, you'll see two patches. You'll see patches A and B. And I hope to you they look to be different shades of grey. Do they? They look to be different shades of grey. Of course they are exactly the same shade of grey. I can illustrate that by putting an alternative image, and joining up those two patches. You'll see that's it's the same shade of grey. You may not believe me, so what I'll do is, I'll shift it along, and you'll see even more clearly. There are no sharp edges. It's the same shade of grey. What's going on here, of course, is that the brain is unconsciously applying its prior knowledge that a cast shadow dims the appearance of surfaces that it casts onto. So we therefore see the patch B as being lighter than it really is, in order to account for that effect. And this is of course an illustration of the success of the visual system, not its failure. The visual system is a very bad physical light metre, but that's not what it's supposed to do. It's supposed to, or one thing it's supposed to do, is to interpret the causes of sensory signals in terms of meaningful objects in the world. It's also an example of what we sometimes call cognitive impenetrability. Even if you know the patches are the same shade of grey, when I take that bar away, they again look different. Can't do much about that. The second example just shows you how quickly the brain can take in new prior information to change the nature of conscious perception. This is a so-called Mooney image. And if you haven't seen it before, hopefully what you will see here is just a passing of black and white splotches. Does everybody kind of get that? Black and white splotches? Some of you might have seen this before. And now what I'm going to do is fill it in, and you see something very different. What you'll see is a very meaningful scene here involving at least two objects, a woman, a hat and a horse. Now if you stare at this for a while, I won't leave it up for too long-- but if you just look at it for a little bit, and then I take that image away again, you should still be able to see the objects within that image. Now for me this is quite remarkable, because the sensory information hasn't changed at all. All that's changed are your brain's prior expectations about what that sensory data signifies. And this changes what you consciously see. Now this also works in the auditory domain. Here are two spectrograms. This is something called sine wave speech, and what you see here are two time frequency representations of speech sounds. The one on the top has all the sharp acoustical features that provide normal speech removed. A little bit like thresholding an image. And the bottom is something else. So I'm going to play the top first, and let's see what it sounds like. [STRANGE BEEPS AND NOISES] And now I'll play you something else. (BOTTOM SOUND - A MAN'S VOICE): Jazz and swing fans like fast music. - So I hope you all understood that piece of sage advice. And now if I play the original sound again-- [BEEPS AND WHISTLES THAT SOUND LIKE THE SENTENCE] - Yeah? This is exactly the same. Again, all this change is what we expect that sound to signify. [SAME SOUND PLAYED AGAIN] - One more time, just for luck. It's not just a bunch of noisy whistles, it's speech. Now this typical framework for thinking about these kinds of effects is Bayesian Inference. And this is a form of probabilistic reasoning, which is applicable in all sorts of domains, not just in neuroscience, in medical diagnosis, and all sorts of things, like finding lost submarines. But in neuroscience, we talk about the Bayesian brain. And it's a way of formalising Helmholtz's idea that perception is a form of best guessing. And the idea is that sensory signals and prior beliefs can be represented as probability distribution. So for instance, this yellow curve is the probability of something being the case, maybe that you've got a brief glimpse of an object moving to the right. The sensory data may say something different. It may have a probability that peaks at a different angle of movement. Maybe it's drifting in a different direction. And the optimal combination of the prior, and the likelihood, the yellow curve and the red curve, is this green curve, which we will call the posterior distribution. And that represents the best optimal combination of these two sorts of evidence. And the idea is, well that's what we perceive. Thinking about perception in this way does something rather strange to the way, classically in neuroscience, people have thought about perception. The classical view is that the brain processes sensory information in a bottom-up, or feed-forward direction. This is a picture of the visual system of the monkey, and the idea is that information comes in through the retina, then goes through the thalamus. It then goes to the back of the brain. And as the sensory signals percolate deeper and deeper and deeper into the brain, they encode or represent progressively more sophisticated features of objects. So you start out at early levels the visual cortex with response to luminance, and edges, and then higher up to objects, including other monkeys. What's important here is the perceptual heavy lifting is done by information flowing in this bottom-up or feed-forward direction. Now the Bayesian brain idea says something very different. It says that what's really important are the top-down or inside-out connections that flow from the centre of the brain back out. And we've known for a long time there's a large number, a very large number of these connections, and some descriptions more than flow the other way around. But the function has been rather mysterious. Thinking about the Bayesian brain gives us a nice way to interpret this. Which is that it's exactly these top-down or inside-out connections that convey predictions from high levels of the brain to lower levels, to lower levels, back out to the sensory surfaces. So these blue arrows convey the brain's predictions about the causes of sensory signals. And then what flows in the feed-forward or bottom-up direction, from the outside-in, that's just the prediction area, the difference between what the brain expects and what it gets at each level of description. So this is often called predictive coding, or predictive processing, in some formal frameworks. And the idea is that minimization of prediction error occurs across all levels of this hierarchy at the same time, and what we then perceive is the consequence of this joint minimization of prediction error. So you can think of perception as a sort of controlled hallucination, in which our perceptual predictions are being reined in at all points by sensory information from the world and the body. Now there's quite a lot of experiments that show that something like this is probably going on in the brain. These are a couple of examples. And since they're-- I was looking for the best example so they don't come from my lab at all. [LAUGHTER] This is from Lars Muckli in Glasgow. He's shown using advanced brain reading techniques, which I won't describe, that you can decode the context of what a person is seeing from parts of the visual cortex that isn't even receiving any input. And what's more, you can decode better when you decode from the top part of the cortex, which is supposed to receive predictions from higher levels. So that suggests there are predictions being fed back. And another study by Andre Bastos and Pascal Fries in Germany, they used a method called Granger causality, which is sensitive to information flow in systems. And they find that top-down signals and bottom-up signals are conveyed in different frequency bands in the cortex, which is what you'd predict from predictive coding. One last experiment which I find particularly interesting is an experiment from a Japanese group of Masanori Murayama. And they used to optigenetics, which is a way of using lights to selectively turn on or off neural circuits. And in this experiment they showed that by just deactivating top superficial levels of somatosensory cortex in a mouse brain, the part of the mouse brain that's sensitive to touch, they could affect how well that mouse was able to do tactile discriminations. Those top-down connections were coming from a motor cortex. So there's a lot of evidence that top-down connections in the brain are important for perception, is the basic message there. But what's rather strange, and what I'm going to tell you next is that all this stuff is all very good, but predictive processing is not a theory of consciousness. Nothing I've said has anything to do with consciousness, at all. It has to do-- it's a very general theory of how brains do what they do. How they do perception, how they do cognition, how they do action. So somewhat counter-intuitively, I think this is exactly why it's a great theory of consciousness. And the reason I think this is because it allows us to ask all sorts of questions about the real problem. About what it is, what happens in brains that underlies what you happen to be conscious of right now, without getting sucked into the metaphysical pluckhole of why you are conscious in the first place. In other words, it provides a powerful approach to looking for neural correlates of consciousness, those things in the brain that go along with being conscious. Because we can now take advantage of a very general theory of how brains do what they do, rather than just looking at this region or that region. So what does predictive processing, or the Bayesian brain say about consciousness, specifically? Well many years ago, some influential experiments revealed a very strong connection between top-down signalling and conscious contents. In this example by Alvaro Pascual-Leone and Vincent Walsh, what they did was they had people look at visual motion, examples of visual motion. And they used TMS, this interventional technique where you can zap the brain very briefly. I mentioned it before. But they used it here, specifically to interrupt the top-M down signalling that was evoked by this perception of visual motion. And the result was that, if you interrupted specifically the top-down feedback, you would abolish the conscious perception of visual motion, even if you left the bottom-up signalling intact. So that was an early key. Now, more recently, in our lab and in many other labs all over the place, we've been asking some other questions about the relationship between what you expect and what you consciously experience. One of the most basic questions you can ask is, do we consciously see what we expect to see? Or do we see what violates our expectations of what we expect? And a recent study from our group, led by Yair Pinto, used a method called continuous flash suppression to address this question. It's illustrated here. You see different images in the different eyes. In one eye you see this rapidly changing Mondrian pattern of squares. And in the other eye, you either see a face or a house. And they change contrast like this. So initially, the person would just see this random pattern, and then they'll see either a house or a face. And simply, you just ask them to expect to see-- you just tell them a face is more likely or a house is more likely. And what we find over a number of studies is that we see faces more quickly when that's what we're expecting to see. It may seem obvious, but it could be the other way around. At least in these studies, we see what we expect to see, not what violates our expectations. That's the data. And the same goes for houses. These kinds of studies support the idea that it's the top-down predictions that are really important for determining what we're conscious of. There's another experiment which I will just mention. We did pretty much the same thing. This is called motion induced blindness. If you are in a lab rather than in a lecture theatre, and you stare at this central point here, then this red dot might disappear from time to time. And what we did was after it disappeared, we changed its colour and we led people to expect the colour change to be one thing or another. And again, it reappeared more quickly if it changed colour in the way you were expecting. Again, I am confirming that once your expectations were validated, then that accelerated your conscious awareness of something in the world. Now that's just behavioural evidence. That's just asking people what they see and when they see it. We've also been interested in the brain mechanisms that underlie and shape how our expectations change, what we consciously see. And we've been particularly interested in something called the alpha rhythm. Now the alpha rhythm is an oscillation of about 10 hertz or 10 cycles per second. That's especially prominent in the visual cortex, across the back of the brain. In one study, led by in this case a PhD student, Maxine Sherman, with Ryota Kanai, in Sussex. What we did here, we manipulated people's expectations of what they we're likely to see. And it's a very boring experiment, this was. The only thing they could see was what we call Gabor patches. They're just very dim patches of lines that are blurry around the edges. But the visual system loves these kinds of things. They activate early visual cortex very, very well. And people were expecting either that a patch was there, or that it wasn't there, in different conditions. And while doing this we measured brain activity. And to cut a long story short, what we found was that there were certain phases, certain parts of the cycle, this 10 Hertz cycle, at which your expectation had a greater effect on what you said that you saw. So there was part of this cycle, as the alpha rhythm, there was part of it where your expectations dominated your perception. And there was another part of it which was the opposite, where your sensory signals were more important in determining what you saw at that point. So this suggests that this oscillation in the back of the brain is orchestrating this exchange of predictions and prediction errors. And that's the sort of cycle that might be the neural mechanism for conscious vision. And other theories about what the alpha rhythm is doing, there are many. One is that it's doing nothing, it's just the brain idling away, and I think this is at least a more interesting way to think about it. Another experiment we've done with another PhD student Asa-Chang, we showed people these very fast changing luminance sequences. And it turns out that your brains will learn to predict the specific changes in these sequences that change you very quickly. And the signature of this learning, again, happens to be in the alpha rhythm, and suggests that this oscillation has something important to do with how the brain learns and updates predictions about sensory signals. But we do not go around the world looking at Gabor patches or rapidly changing things like this. We go around the world looking at people and objects. And that's what our visual world is composed of. So can any of these ideas say anything about our everyday visual experience? And I think that's a very important challenge in neuroscience to cross. Get out of the lab and think about real world experiences. So we've been using virtual reality over the last few years to try to get at some of these ideas. This is an Oculus Rift, which is now available to buy, I think. And we've been using these to address some of these real world aspects of visual perception. And one of these real world aspects is called perceptual presence. And this is the observation that, in normal perception, objects really seem to be there, as opposed to being images of objects. And this is, of course, what Magritte plays with in his famous painting, The Treachery of Images. For instance, this is an object. I think it's there, and in some sense I can perceive the back of it, even though I can't see the back of it, even though the back of it's not giving me any sensory data, I perceive it as an object with a back. How does one explain that? Well one idea you can come up with within this Bayesian brain framework is that, the brain is not only predicting the possible causes of the sensory signals getting right here, right now. But it's also predicting how sensory signals would change were I to make particular actions. Were I to pick this object up and move it around, or just move my eyes from one place to another. There's a long paper. I wrote about that which I-- please don't read it. [LAUGHTER] - But that's the basic idea. And how do you test an idea like that? So we've been using some innovative virtual reality methods, or augmented reality methods, with my post-doc, Keisuke Suzuki. And what we do is, we have virtual objects, and these virtual objects, they either behave as a normal object would. They're all weird, unfamiliar objects. But they can either behave as a normal object would behave, so you can learn to predict what would happen. This one is weird. It always shows you the same face, a little bit like having the moon on a plate in front of you. And then there are other conditions where objects respond to your movements, but they do so in unreliable and strange ways. So the question is, what does the brain learn about these objects, and how do we experience them? Do we experience them as objects in different ways when they behave differently? And we're still doing those experiments. Another way we can use VR, is to investigate what happens in visual hallucinations of the kind experienced in psychosis, and in certain other more pharmacologically induced conditions. What we're doing here is, we're coupling immersive virtual reality, imagine you've got a headset strapped to your head again, with clever image processing that models the effects of overactive priors on visual perception to generate a highly immersive experience. This is Sussex campus, actually, but now it seems quite different than it did at lunchtime today. I'll tell you that. What we've done, we've recorded this panoramic video which we can feedback through VR headset. And we've processed this video through one of these Google deep dream algorithms you might have seen, that can generate sort of bowls of pasta that looks like animals. And this might seem like a lot of fun. It is fun, but there is a serious purpose here, because it allows us to model the effects, model very unusual forms of perception and how they might play out in different ways in the visual hierarchy. And understanding how visual hallucinations might happen, and how the wider effects they have on the mind, I think, is a very important part of studying visual perception. So that brings us to the second take-home message, which is that what we consciously see is the brain's best guess of the causes of its sensory input. Normal perception is a fantasy that is constrained by reality. Now before I move on to the last section, I want to pay tribute to an unlikely character in a talk about neuroscience, which is Ernst Gombrich. Ernst Gombrich was one of the foremost historians of art of the 20th century. And it turns out that Gombrich's approach to understanding art had a lot in common with ideas and the Bayesian brain. And more specifically, with the idea that perception is largely an act of imagination, or construction, on the part of the perceiver. And this is most apparent in his concept of the beholder's share, which emphasises that the viewer brings an awful lot to the table in the act of experiencing an artwork. So he had this to say in his 1960 book, Art and Illusion, "the artist gives the beholder 'more to do', he draws them into the magic circle of creation and allows him to experience something of the thrill of making which had once been the privilege of the artist." I think for me this is very powerful when looking at, especially, things like Impressionist art. And here, one way to think about this is, that the artist has reverse engineered the whole perceptual process, so that what's there are not the objects, the end points of perception, but rather the raw materials; the patterns of light that engage our perceptual machinery in doing its work. And for me this might be why paintings like this are particularly powerful. Now the final dimension of consciousness I want to talk about is conscious self. The fundamental experience of being someone. Being someone like you. There are many aspects to our experience of being a conscious self. There is the bodily self, the experience of being and identifying with a particular body. A bit of the world goes around with you in the world all the time. There's the perspectival self, the experience of seeing the world, or experiencing the world, from a particular first person perspective, usually somewhere in the body, but not always. There is the volitional self, the experience of intending to do things, and of making things happen in the world of agency. And these ideas are, of course, often associated with concepts of will. Then there's the narrative self. This is where-- only until now, we don't have to worry about the concept of I, but when we get to the narrative self, there is now and I. There is a continuity of self experience from hour to hour, from day to day, from month to month, and from year to year, that you associate a name with, and a particular set of autobiographical memories. And finally, there's a social self. The way I experience being me is partly dependent on the way I perceive you as perceiving me. I'm just going to talk, in the minutes remain, about the bodily self. This is something we're working on quite a lot in Sussex. The experience of identifying with, and owning, a particular body. And the basic idea I want to convey, is again, very simple. It's just that we should think of our experience of body ownership in the same way that we think about our experience of other things, as well. That is, it's the brain's best guess of the causes of body-related signals. And the brain is always making this inference. It's making its inference about what in the world is part of the body, and what is not part of the body. But it has access, in this case, to other sorts of sensory signals, not just visual signals, or tactile signals, but also proprioceptive signals. The orange arrows here. These inform the brain about the body's configuration and position in space. And then also, and often overlooked, are interoceptive signals. These are signals that originate from within the body, that tell the brain about the physiological state or condition of the inside of the internal physiological milieu. And you can think the idea is that our experience of embodied self-hood is the brain's best guess of the causes of all the signals put together. Yeah, that's just to emphasise interoception. An important part of this idea is that interoception, the sense of the body from within, should work along the same principles, the same Bayesian principles that we've been thinking about, vision and audition, previously. That is, our experience of the inside of our bodies is the brain's best guess of the causes of the signals that come from the inside of our bodies. So we can think of, again, top-down predictions carrying predictions about what the bodily state is like, and bottom-M up prediction errors that report the differences between what's going on and what the brain expects. So what is our experience of the inside of our bodies? Well, way back at the beginning of psychology, William James and Carl Langer proposed that emotions, emotional experience, was really about the brain's perception of changes in its physiological state, rather than perception of the outside world. So, in this classic example, seeing a bear does not in itself generate the experience of fear. Rather seeing the bear sets in train, a load of physiological changes preparing for fight and flight responses. And it's the perception of those bodily changes in the context with the bear being around that leads to our experience of fear. So the Bayesian perspective just generalises that idea, and says that emotional experience is the brain's best guess of the causes of interoceptive signals. And this fits very nicely with a lot of evidence. And this is just one study done by a Finnish group. And all they did here was, they had people report where on their bodies they felt various emotions to take place. And so you feel anxiety in one part of your body. You feel fear in another, and so on. So our experience of emotion does seem to be intrinsically embodied. Now another part of our experience of being a body is the body as a physical object in the world. And this might seem quite easy to take for granted, since our physical body is just always there. It goes around with us, it changes over the years, in unfortunate ways. But it's always there. But it would be a mistake to take our experience of body ownership for granted. And there are some classic experiments that demonstrate how malleable our experience of body ownership is. This is the famous rubber hand experiment. Probably some of you have seen this. What happens here is that a volunteer has their hand hidden under a table, and the fake hand is put on top of the table, and then both hands are simultaneously stroked with a paintbrush. And it turns out that just seeing a hand-like object where a hand might be, feeling touch, and then seeing the object being touched, is enough evidence that the brain's best guess becomes that fake hand is, in fact, part of my body. Sort of part of my body. This is what it looks like in practise. Here you can see the fake hands, focusing on it. There's the real hands, not focusing on it, simultaneous stroking-- and there are various ways you can test it. [AUDIENCE LAUGHTER] - I found in doing this, it works even better on children, by the way, if you do that. So that's interesting, because that's using visual and tactile signals to convince the brain that this object is part of its body. In my lab, we've been interested in whether these signals that come from inside the body also play a role. So we set up a virtual reality version of this rubber hand illusion, where people wear these goggles, and they see a virtual fake hand. And we also record their heartbeats. And now what we can do is, we can make the virtual hand flash either in time or out of time with their heart beat. And we asked the question, do people experience this virtual hand as more belonging to them when it's flashing in time, rather than out of time, with their heartbeat? And the answer is that it does. And this is just some data, basically that, bigger than that, which means that, indeed, they experience the hand as more their own. The way we measure that actually, is that first we can ask them. That's the easiest way. Then we can also ask them to point to where they think their hand really is, and we can see how far they drift from where their hand really is to where the virtual hand is. And that provides a more objective way of measuring the strength of the effect. Here's what it looks like in practise. Again if you can see this, that's the real hand. That's somebody's virtual hand. Again, imagine you're wearing a headset so you'll see this in 3-D. And you can just about see it pulsing to read and back, I hope. And you can also do some other things with these virtual reality rubber hands that you couldn't do with real rubber hands. For instance, you can map movements of the real hand to the virtual hand, so you can start to ask questions about the extent to which the virtual hand moves as I predict it to move. How much does that affect the extent to which I feel it to be part of my body? You can make it change colour. So you can have somebody embody a skin colour associated with a cultural out-group, and see if they become less racist as a result. And then my favourite is where you can change, actually, the size of the body. And that's coming up here. So here what we do is, we can have the hand telescope up and down in size. And again, this might seem like fun, and it is fun, but there is a serious purpose. There are various conditions. There's in fact, a condition called Alice in Wonderland Syndrome, where people report that all parts of their body are, indeed, telescoping up and down in size. And in a more subtle way, there are lots of body dysmorphias, of subtle misperceptions of body shape, which might be associated with eating disorders. And so these sorts of techniques allow us to approach, in a very fine grained way, how people might mis-perceive their own bodies. That brings me to the third take-home message about self. And with apologies to Descartes, the take home message is that, I predict myself, therefore I am. In the last nine minutes, before the bell rings, I want to go full circle and return to this Cartesian idea of the beast machine. To try to convince you that our experience of being a conscious self is intimately tied up with our beast machine nature. And to do this, I need to mention one final aspects of perceptual inference, which has a lot to do with Karl Friston, who's done a lot of work in the Bayesian brain UCL here in London. And if we think of the brain as being in the business of minimising prediction errors, this can be done either by updating our perceptual predictions, which is what I've been talking about so far. And this what Helmholtz said. Or we can minimise prediction errors by making actions. We can change what we predict, or we can make an action so that our predictions come true. You can change with sensory input, or you can change what you believe about your sensory input. One point of doing this is, that you can make actions, then, to find out more about the world that surrounds you. And this is what Helmholtz has in mind when he says that each movement we make, by which we alter the appearance of objects, should be thought of as an experiment designed to test whether we've understood correctly the invariant relations of the phenomena before us. Which Gregory, much later, said something similar when he talked about perception as hypothesis testing. The point of this is, that we make eye movements, and other kinds of movements, to understand what the world is like. That, in fact, there was is tomato there, for instance. But there's another way to think about active inference, which is that, when we minimise prediction error, what we're actually doing is controlling a sensory variable. We're preventing it from changing, because we're making our prediction about what it is come true. And this is the use of active inference, to control or regulate something, rather than to understand what the causes of that something are. And this brings a very different tradition to mind, which is 20th century cybernetics. And this is Ross Ashby, who was a pioneer of this way of thinking. And he, with Roger Conan right at the end of his life, wrote a paper. The title of the paper, was "Every good regulator of a system must be a model of that system". The idea here is, if you want to regulate something very precisely, then you need a good model of what effects that system. Now you could apply this idea to the external world, as well. When you try to catch a cricket ball, you are actually trying to control the level of the angle above the horizon. But it applies more naturally, I think, to the internal state of our body. So really, what matters about my internal physiological condition, I don't really need to know exactly what it's like inside my body, and care about it. But I need to control it, and my brain needs to regulate it. So this way of thinking about active inference applies more naturally to interoception. Think about it this way that. Having good predictive models are always useful, but we can have a pendulum that swings, on the one hand, towards control. We can use these predictive models for control, and that's more applied to the state of our internal body. Or we can swing the other way, and think about perception, understanding. You could think of the instrumental, epistemic ways of thinking about the role of action and perception. And this brings to mind-- I mentioned Karl Friston. He's come up with this thing called the free energy principle. And I can only nod to the vast body of work he's done here on this. With the slogan, which is that organisms, over the long run, maintain themselves in states in which they expect to be in, in virtue of having good predictive models about their own internal condition. So this takes us right back to Descartes, but in a very different way. As I said right at the beginning of this lecture, for Descartes our physiological reality was rather irrelevant to our minds, our rationality, our consciousness. This is a quote from a 1968 paper on his beast machine argument, "without minds to direct their bodily movements, animals must be regarded as unthinking, unfeeling machines that move like clockwork." Now I think if you try to think how this idea of our predictive models controlling our internal physiological states, and the resulting experiences that perceptual content that might give rise to, you can make the opposite case. And the opposite case would be that conscious self-hood emerges because of, and not in spite of, the fact that we are beast machines. And I think this is a deeply embodied view of consciousness and self, and it speaks to this fundamental link in continuity between what it is to be alive, what it is to have a mind, and what it is to be a conscious self. So I repeat, the third take-home message should make even more sense now. That I predict myself, therefore I am. And I am a conscious self because I'm a bag of self-fulfilling predictions about my own physiological persistence over time. Now why does any of this matter? It's a lot of interesting ideas, but why should we be interested in studying consciousness? Well it's a very interesting thing, I hope I've convinced you. But there are lots of practical reasons to be interested as well. There are between 20 and 60,000 patients in the UK alone, who are in disorders of consciousness. You are in the vegetative state, or in coma, or in some other severely abnormal state of consciousness. Having better measures of conscious level, as I described at the beginning, is going to really change again, and how we treat people like this. And of course, in psychiatry. Psychiatric disorders are increasing that prevalence across all modern societies, and it's estimated one in six of us, at any one time, are suffering from a psychiatric condition. And understanding the mechanisms that underlie conscious content and conscious self, because a lot of psychiatric disorders include disturbances of the way we experience our body, even though that might not be the most obvious symptom, can help us understand the mechanisms involved in psychiatric disorders, not just the symptoms. There are also some more general reasons for studying consciousness, which bring up some ethical questions. When does consciousness emerge in development on newborn babies conscious? Or does consciousness start even in the womb? Maybe different dimensions of consciousness emerge at different times. Are other animals conscious? Well I think it can make a very good case for mammals and primates, but what about the octopus? The octopus has more neurons in its arms than in its central brain. They're very smart creatures. Here, you have to ask the question not only, what is it like to be an octopus, but what is it like to be an octopus arm? And finally, with the rise of artificial intelligence, we should begin to ask questions about what would it take for a machine to have some kind of subjective experience. I don't think we're anywhere near that yet, but we should consider what science can tell us about its possibility, because that would raise some very, very tricky ethical questions. But, fundamentally, consciousness remains fascinating for me for the same reason that it's motivated people throughout the ages. I mean, Hippocrates, the founder of modern medicine, put it one way. He said, from the brain and from the brain alone arise our sorrows, our joys. And he also had his first view of psychiatry that madness comes from its moistness. And then Francis Crick, in the 1990s, who I mentioned in the beginning. He gave birth, if you like, to the modern neuroscience of consciousness. He said much the same thing in his astonishing hypothesis. But there is this mystery and wonder still, about how the biological machinery inside our heads gives rise to the rich inner universe that makes life worth living. And despite this mystery, modern science is making progress. I hope I've given you a flavour, even though we don't understand how consciousness happens, we can begin to understand its mechanisms. So we should not be afraid of naturalising consciousness. It's not a bad thing to understand its basis in the material world. As so often in science, with greater understanding comes a larger sense of wonder, and a greater realisation that we are part of, and not apart from, the rest of nature. [AUDIENCE APPLAUSE]
B1 中級 Anil Seth的《意識的神經科學》。 (The Neuroscience of Consciousness with Anil Seth) 61 4 Josh 發佈於 2021 年 01 月 14 日 更多分享 分享 收藏 回報 影片單字