字幕列表 影片播放 列印英文字幕 - [automated message] This UCSD TV program is presented by University of California Television. Like what you learn? Visit our website, or follow us on Facebook and Twitter to keep up with the latest programs. ♪ [music] ♪ - [male] We are the paradoxical ape, bi-pedal, naked, large brain, long the master of fire, tools, and language, but still trying to understand ourselves. Aware that death is inevitable, yet filled with optimism. We grow up slowly. We hand down knowledge. We empathize and deceive. We shape the future from our shared understanding of the past. CARTA brings together experts from diverse disciplines to exchange insights on who we are and how we got here. An exploration made possible by the generosity of humans like you. ♪ [music] ♪ - [Evelina] Thanks very much for having me here. So today, I will tell you about the language system and some of its properties that may bear on the question of how this system came about. I will begin by defining the scope of my inquiry, because people mean many different things by language. I will then present you some evidence, suggesting that our language system is highly specialized for language. Finally I will speculate a bit on the evolutionary origins of language. So linguistic input comes in through our ears or our eyes. Once it's been analyzed perceptually, we interpret it by linking the incoming representations toward our stored language knowledge. This knowledge, of course, is also used for generating utterances during language production. Once we've generated an utterance in our heads, we can send it down to our articulation system. The part that I focus on is this kind of high-level component of language processing. This cartoon shows you the approximate locations of the brain regions that support speech perception, shown in yellow, visual letter and word recognition in green. This region is known as the visual word-form area. Articulation in pink and high-level language processing in red. What differentiates the high-level language processing regions from the perceptual and articulation regions is that they're sensitive to the meaningfulness of the signal. So for example, the speech regions will respond just as strongly to foreign speech as they respond to meaningful speech in our own language. The visual word-form area responds just as much to a string of consonants as it does to real words. The articulation regions can be driven to a full extent by people producing meaningless sequences of syllables. But in contrast, the high-level language-processing system or network, or sometimes I refer to it just as the language network for short, cares deeply about whether the linguistics signal is meaningful or not. In fact, the easiest way to find this system in the brain is to contrast responses to meaningful language stimuli, like words, phrases, or sentences. Some control conditions like linguistically degraded stimuli. The contrast I use most frequently is between sentences and sequences of non-words. A key methodological innovation that laid the foundation for much of my work was the development of tools that enable us to define the regions of the language network, functionally at the individual subject level, using contrasts like these. Here, I'm showing you sample language regions in three individual brains. This so-called functional localization approach has two advantages. One, it circumvents the need to average brains together, which is what's done in the common approach, and it's a very difficult thing to do because brains are quite different across people. Instead, in this approach, we can just average the signals that we extract from these key regions of interest. The second advantage is that it allows us to establish a cumulative research enterprise, which I think we all agree is important in science, because comparing results across studies and labs is quite straightforward if we're confident that we're referring to the same brain regions across different studies. This is just hard or impossible to do in the traditional approach, which relies on very coarse anatomical landmarks, like the inferior frontal gyrus or the superior temporal sulcus, which span many centimeters of cortex and are just not at the right level of analysis. So what drives me and my work is the desire to understand the nature of our language knowledge and the computations that mediate language comprehension and production. However, these questions are hard, especially given the lack of animal models for language. So for now, I settle on more tractable questions. For example, one, what is the relationship between the language system and the rest of human cognition? Language didn't evolve, and it doesn't exist in isolation from other evolutionarily older systems, which include the memory and attention mechanisms, the visual and the motor systems, the system that supports social cognition, and so on. That means that we just can't study language as an isolated system. A lot of my research effort is aimed at trying to figure out how language fits with the rest of our mind and brain. The second question delves inside the language system, asking, "What does its internal architecture look like?" It encompasses questions like, "What are the component parts of the language system? And what is the division of labor among them, in space and time?" Of course, both of those questions ultimately should constrain the space of possibilities for how language actually works. So they place constraints on both the data structures that underlie language and the computations that are likely performed by the regions of the system. Today, I focus on the first of these questions. Okay. So now onto some evidence. So the relationship between language and the rest of the mind and brain has been long debated, and the literature actually is quite abundant with claims that language makes use of the same machinery that we use for performing other cognitive tasks, including arithmetic processing, various kinds of executive function tasks, perceiving music, perceiving actions, abstract conceptual processing, and so on. I will argue that these claims are not supported by the evidence. Two kinds of evidence speak to the relationship between language and other cognitive systems. There is brain-imaging studies, brain-imaging evidence, and investigations of patients with brain damage. In fMRI studies, we do something very simple. We find our regions that seem to respond a lot to language, and then we ask how do they respond to other various non-linguistic tasks. If they don't show much of a response, then we can conclude that these regions are not engaged during those tasks. In the patient studies, we can evaluate non-linguistic abilities in individuals who don't have a functioning language system. If they perform well, we can conclude that the language system is not necessary for performing those various non-linguistic tasks. So starting with the fMRI evidence, I will show you responses in the language regions to arithmetic, executive function tasks, and music perception today. So here are two sample regions, the regions of the inferior frontal cortex around Broca's area and the region in the posterior middle temporal gyrus, but the rest of the regions of this network look similar in their profiles. The region on the top is kind of in and around this region known as Broca's area, except I don't use that term because I don't think it's a very useful term. In black and gray, I show you the responses to the two localizer conditions, sentences and non-words. These are estimated in data that's not used for defining these regions. So we divide the data in half, use half of the data to define the regions and the other half to quantify their responses. I will now show you how these regions respond when people are asked to do some simple arithmetic additions, perform a series of executive function tasks, like for example, hold a set of locations in spacial memory, spacial locations in working memory, or perform this classic flanker task, or listen to various musical stimuli. For arithmetic and various executive tasks, we included a harder and an easier condition, because we wanted to make sure that we can identify regions that are classically associated with performing these tasks, which is typically done by contrasting a harder and an easier condition of a task. So I'll show you now, in different colors, responses to these various tasks, starting with the region on the lower part of the screen. So we find that this region doesn't respond during arithmetic processing, doesn't respond during working memory, doesn't respond during cognitive control tasks, and doesn't respond during music perception. Quite strikingly to me at the time, a very similar profile is observed around this region, which is smack in the middle of so-called Broca's area, which appears to be incredibly selective in its response for language. Know that it's not just the case that these, for example, demanding tasks fail to show a hard versus an easy difference. They respond pretty much at or below fixation baseline when people are engaged in these tasks. So that basically tells you that these language regions work as much when you're doing a bunch of math in your head or hold information in working memory, as what you're doing when you're looking at a blank screen. So they really do not care. So of course, to interpret the lack of the response in these language regions, you want to make sure that these tasks activate the brain somewhere else. Otherwise, you may have really bad tasks that you don't want to use. Indeed, they do. So here, I'll show you activations for the executive function tasks, but music also robustly activates the brain outside of the language system. So here are two sample regions, one in the right frontal cortex, one in the left parietal cortex. You see the profiles of response are quite different from the language regions. For each task, we see robust responses, but also a stronger response to the harder than the easier condition across these various domains. These regions turn out to be part of this bilateral frontal parietal network, which is known in the literature by many names, including the cognitive control network or the multiple demand system, the latter term advanced by John Duncan, who wanted to highlight the notion that these regions are driven by many different kinds of cognitive demands. So these regions appeared to be sensitive to effort across tasks, and their activity has been linked to a variety of goal-directed behaviors. Interestingly if you look at the responses of these regions to our language localizer conditions, we find exactly the opposite of what we find in the language regions. They respond less to sentences than sequences of non-words, presumably because processing sentences requires less effort, but clearly this highlights, again, the language and the cognitive control system are clearly functionally distinct. Moreover, damage to the regions of the multiple demand network has been shown to lead to decreases in fluid intelligence. So Alex Woolgar reported a strong relationship between the amount of tissue loss in frontal and parietal cortices and a measure of IQ. This is not true for tissue loss in the temporal lobes. It's quite striking. You can actually calculate for this many cubic centimeters of loss in the MD system, you lose so many IQ points. It's a strong, clear relationship. So this system is clearly an important part of the cognitive arsenal of humans because the ability to think flexibly and abstractly and to solve new problems are exactly . . . These are the kinds of abilities that IQ tests aim to measure, are considered kind of one of the hallmarks of human cognition. Okay. So as I mentioned, the complementary approach for addressing questions about language specificity and relationship to other mental functions is to examine cognitive abilities in individuals who lack a properly functioning language system. Most telling are cases of global aphasia. So this is a severe disorder which affects pretty much the entire front temporal language system, typically due to a large stroke in the middle cerebral artery and lead to profound deficits in comprehension and production. Rosemary Varley at UCL has been studying this population for a few years now. With her colleagues, she has shown that actually these patients seem to have preserved abilities across many, many domains. So she showed that they have in-tact arithmetic abilities. They can reason causally. They have good nonverbal social skills. They can navigate in the world. They can perceive music and so on and so forth. Of course, these findings are then consistent with the kind of picture that emerges in our work in fMRI. Let's consider another important non-linguistic capacity, which a lot of people often bring up when I tell them about this work. How about the ability to extract meaning from non-linguistic stimuli? Right? So given that our language regions are so sensitive to meaning, we can ask how much of that response is due to the activation of some kind of abstract, conceptual representation that language may elicit, rather than something more language-specific, a semantic representation type. So to ask these questions, we can look at how language regions respond to nonverbal, meaningful representations. In one study, we had people look at events like this or the sentence-level descriptions of them, and either we had them do kind of a high-level semantic judgment test, like decide whether the event is plausible, or do a very demanding perceptual control task. Basically what you find here is, again, the black and gray are responses to the localizer conditions. So in red, as you would expect, you find strong responses to this and to the condition where people see sentences and make semantic judgments on them. So what happens when people make semantic judgments on pictures? We find that some regions don't care at all about those conditions, and other regions show reliable responses, but they're much weaker than those elicited by the meaningful sentence condition. So could it be that some of our language regions are actually abstract semantic regions? Perhaps. But for now, keep in mind that the response to the sentence-meaning condition is twice stronger, and it is also possible that participants may be activating linguistic representations to some extent when they encounter meaningful visual stimuli. So to answer this question more definitively, we're turning to the patient evidence again. If parts of the language system are critical for processing meaning in non-linguistic representations, then aphasic individuals should have some difficulties with nonverbal semantics. First, I want to share a quote with you from Tom Lubbock, a former art critic at The Independent, who developed a tumor in the left temporal lobe which eventually killed him. As the tumor progressed, and he was losing his linguistic abilities, he was documenting his impressions of what it feels like to lose the capacity to express yourself using verbal means. So he wrote, "My language to describe things in the world is very small, limited. My thoughts, when I look at the world, are vast, limitless, and normal, same as they ever were. My experience of the world is not made less by lack of language, but is essentially unchanged." I think this quote quite powerfully highlights the separability of language and thought. So in work that I'm currently collaborating on with Rosemary Varley and Nancy Kanwisher, we are evaluating the global aphasics performance on a wide range of tasks, requiring you to process meaning in nonverbal stimuli. So for example, can they distinguish between real objects and novel objects that are matched for low-level visual properties? Can they make plausibility judgments for visual events? What about events where plausibility is conveyed simply by the prototypicality of the roles? So you can't do this task by simply inferring that a watering can doesn't appear next to an egg very frequently. Right? It seems like the data so far is suggesting that they indeed seem fine on all of these tasks, and they laugh just like we do when they see these pictures because they're sometimes a little funny. So they seem to process these just fine. So this suggests, to me, that these kinds of tasks can be performed without a functioning language system. So even if our language system stores some abstract conceptual knowledge in some parts of it, it tells me at least that that code must live somewhere else as well. So even if we lose our linguistic way to encode this information, we can have access to it elsewhere. So to conclude this part, fMRI in patients sudies converge suggesting that the front temporal language system is not engaged in and is not needed for non-linguistic cognition. Instead, it appears that these regions are highly specialized for interpreting and generating linguistics signals. So just a couple minutes on what this means. So given this highly selective response to language stimuli that we observe, can we make some guesses already about what these regions actually do? I think so. I think a plausible hypothesis is that this network houses our linguistic knowledge, including our knowledge of the sounds of the language, the words, the constraints on how sounds and words can combine with one another. Then essentially the process of language interpretation is finding matches between the pieces of the input that are getting into our language system and our previously stored representations. Language production is just selecting the relevant subset of the representations to then convey to our communication partner. This way . . . The form that this knowledge takes is a huge question in linguistic psychology and neuroscience. So one result I don't have time to discuss is that contra some claims, it doesn't seem to be the case that syntactic processing is localized to a particular part of this language system. It seems it's widely distributed across. Anywhere throughout the system, you find sensitivity to both word-level meanings and compositional aspects of language, which is much in line with all current linguistic theorizing, which doesn't draw a short boundary between the lexicon and grammar. So this way of thinking about the language system as a store of our language knowledge makes it pretty clear that the system is probably not innate. In fact, it must arise via experience with language as we accumulate this language store. It's also presumably dynamic, changing all the time as we get more and more linguistic input through our lifetimes. I assume that our language knowledge is plausibly acquired with domain general statistical learning mechanisms, just like much other knowledge. So what changed in our brains that allowed for the emergence of this system? So one thing that changed is that our association cortices expanded. So these are regions that are sensory and motor regions and include frontal, temporal, and parietal regions. Okay. So these people have noted for a long time. I think I'm kind of in the camp of people who think that our brains are not categorically different in any way. They're just scaled-up versions of other primate brains. I think there's quite good evidence for that. So how does the system emerge? So I think one thing that was different between us and chimps is that there is a protracted course to the brain development in humans. So between birth and adulthood, our brains increase threefold, compared to just twofold in chimps. It's a big difference. Basically this just makes us exceptionally susceptible to environmental influences, and we can soak stuff up from the environment very, very easily. So as our brains grow, we make more glial cells. We make more synapses. Our axions continue to grow and become myelinated, and it's basically tissue that's ready to soak up the regularity that we see in the world. Of course, it comes at a cost. That's why we have totally useless babies that can't do anything, but apparently somehow it was worth. . . The tradeoffs were worth it. Okay. So the conclusions. We have this system. It's highly selective in its responses. It presumably emerges over the course of our development and would enable probably some combination of the expansion of these association cortices, where we can store vast amounts of symbolic information and this protracted brain development, which makes us great learners early on. Thank you. [applause] ♪ [music] ♪ - [Rachel] So I want to start this story actually in the 1800s. So in 1800, a young physician named Itard decided to take on the task of teaching a young boy French. Turns out that this young boy, who they think was between 10 and 12 years old, he had no language. He'd been discovered running around in the woods, naked and unable to communicate. Itard thought that this was a very important task because he thought that civilization was based on the ability to empathize and also on language. So he tried valiantly, for two years, to teach this young boy named Victor . . . He named him Victor because over this two-year time span, the only language or sounds that this boy could make was the French 'er', which sounds like Victor. Hence, he had his name. But after two years, Victor was unable to speak any French and comprehended very little French and primarily communicated with objects. Then Itard wrote this up after two years. He wrote up his findings, and he said that he thought that a major reason why Victor didn't learn French was because he was simply too old. But of course, this was the 1800s, and he didn't speculate as to what it was about being 10 years old with no language that would prevent you from learning language if you had a daily tutor trying to teach you French. There's an enormous amount of irony in this particular story, because Itard was the house physician for the first school for the deaf in the entire world. Here, we have a picture of it. The first school for the deaf was begun in 1760 in Paris. At the time that Itard was teaching Victor, he was at the school with all of these children who used sign language and all of these teachers who used sign language. Nonetheless, it never occurred to him to try to teach Victor sign language, but we can't blame him because in the 1800s, sign language was not considered to be a language. In fact, that particular discovery and realization wouldn't happen for another century and a half. So we can imagine why he didn't teach Victor sign language. The question is: could Victor have actually learned sign language if Itard had tried to teach it to him? So I'm going to try to answer that question today through a series of studies. But before I talk about our studies, I want to talk about something that all of the speakers who have preceded me have talked about, which is that one thing that's . . . The defining characteristic of language is that it's highly structured, and it's highly structured at all of these multiple levels, so that speech sounds make up words. Words make up words and phonemes. These words are strung together in sentences with syntax, and the specific syntax helps us understand and produce very specific kinds of meanings. Now one aspect of this language structure is that humans have evolved to the point where children learn this structure naturally. Nobody has to teach them. Nobody has to have a tutor to sit with them for two years to teach them the structure of French or the structure of English or the structure of ASL. Simply by being around people who use the language, young children naturally acquire all of this multilevel and complicated structure. That is to say all children do this if, in fact, they can access the language around them, if they have normal hearing. They're born with normal hearing. They hear people talk. Before you know it, they're talking themselves. But if children are born profoundly deaf, they cannot hear the speech around them. We know that lip-reading is insufficient to learn language because most of the speech sounds are invisible. What happens to these children? If there's no sign language in the environment, they can't learn a visual form of language either. So it happens that there are large numbers of children who actually are like Victor, in the sense that they grow up without language, without learning a language, but they are like Victor-- they are unlike Victor, in that they weren't running around in the woods, nude and having a very harsh life. So how does the lack of language in childhood affect the ability to learn language? Or does it affect the ability to learn language? This has been the focus of studies that we have been doing for many years in our laboratory. Because we're using deaf children and sign language as a means to model language acquisition and its effect on the brain, I think it's appropriate that I talk a little bit about the kinds of stimuli we do and about American Sign Language. So first of all, you should know that American Sign Language, unlike many of the sign languages that have been discussed up to this point, is a very sophisticated language. It's evolved clearly over 200 years. We might even say that American Sign Language evolved with the development of the United States of America and spread as civilization went across, as white men went across the continent. So American Sign Language has a phonological system, a morphological system, syntax, and so forth. So for those of you who don't know sign . . . I know there are many people here who do know sign. I want to show you what some of this structure looks like. So I'm going to play you two video tapes. For those of you who don't know sign, I would like you to guess which one is syntactically structured. For those of you who do know sign, maybe you could keep the answer to yourselves. That's one. Here's two. So how many think two? How many think one? Okay. For those of you who think two, I captioned this. So this is really 'kon, dird, lun, blid, mackers, gancakes.' Number one is a fully formed sentence with a subordinate clause. The reason I'm showing you these two sentences is, for those of you who don't know the language, you can't perceive or parse this particular structure. This is what knowing a language is about. For those of you who know ASL and who know this language, you know that the second example had all of these signs which were non-signs, possible signs, but really just non-signs. This is part of what knowing language is about. How do people learn this particular structure? The question that we're interested in is: how does being a young child help people learn this particular structure? So we did a series of experiments. When we started this work, it wasn't even clear that age would make a difference in sign language acquisition. Sign language is gestural. Sign language is mimetic. Maybe anybody can learn sign language at any time in their lives. So in one experiment, what we did is we recruited a number of people. This is in Canada, who were born deaf and who used ASL. We asked them. We created an experiment where we had a set of sentences that varied in complexity, and we showed them these ASL sentences, and we asked them simply to point to a picture that reflected the meaning of the sentence that they saw. We were quite struck by our findings. What you see here is that deaf people who learned ASL from birth, from their parents, performed very, very well on this task, in contrast to deaf people who were adults, who've been signing for over 20 years performed at chance. So they had great difficulty understanding some of these basic sentences in ASL. So this suggests that there are age of acquisition effects for sign, as there are for spoken language. Everybody sort of . . . The word on the street is it's much harder to learn a language if you're an adult than you're a child. But what if it's something deeper than this? What if there's something about learning a language in childhood that sets up the ability to learn language, that creates the ability to learn language? So we did another experiment, also in Canada, but we decided in order to test this particular hypothesis, we should switch languages. So we're no longer testing ASL here. What we're testing here is English. We devised an experiment in English, where we had a set of sentences, and some of them were ungrammatical, and some of them were grammatical. This is a common kind of task that psycholinguists use. Notice here that the people who were born profoundly deaf and for whom ASL was a first language are near-native in English. So this is a second language. So learning a language early, even though it's in sign, helps people learn a second language. Notice also that they performed . . . Their performance was indistinguishable from normally-hearing people who had learned other languages at birth, German, Urdu, Spanish, and French. So there seems to be that there's something about learning a language early in life, regardless of whether it's sign language or spoken language, that actually helps people learn more language. It's not simply learning language when one is little. But as also part of this experiment, we tested a group of individuals who had been signing for 20 years and had gone through the educational system in Canada, who were born deaf. On this task, they performed at chance. On a grammaticality judgment task, it's either yes or no. So it's at chance. So we see that individuals who are deprived of language, who aren't able to learn language at a young age, perform poorly on their primary language, sign language, and they perform very poorly on a second language, which is ASL, and we see the reverse. So there's something really special going on here about learning language at an early age. What might this be? And might it be in the brain? So in another set of studies, what we did is take this population that we had been looking at, and we decided to neuro-image their language processing to see whether this might give us some clues as to differences between first and second-language learners and people who had language and people who did not have language. So in this study, also done in Canada, with colleagues at the Montreal Neurological Institute, we did fMRI. Maybe many of you have had MRIs. We showed the subjects sentences, like the sentences that you saw, and we asked them to make grammatical judgments on these sentences. We tested 22 people. They were all born profoundly deaf. They all used American Sign Language as a primary language. They had all gone through the educational system, but they ranged in the age at which they were first able to acquire language. This is all the way from birth up to age 14. So if the age at which you learn your first language doesn't make a difference, then we should expect the neuro-processing patterns of all of these individuals to be similar. If age of acquisition makes a difference, we should see different patterns in the brain. In fact, this is what we have. This is what we found. When we did the analysis, we found that there were seven regions in the brain, primarily in the left hemisphere. As you know now, the left hemisphere has areas that are responsible for language. One effect that we found was that in the language regions of the left hemisphere, the earlier the person learned their first language, the more activation we saw in the language hemisphere. However, the older the person was when they learned their first language, the less activation we got in the language areas of the brain. So if there's less activation, is there something else going on here? We actually found a second effect, which we were not expecting at all, which is in the back part of the brain, the posterior part of the brain, in visual processing. This particular effect was that the longer the person matured, the older the . . . Without language, the older they were when they learned language, we found greater activation, more neural resources being devoted to visual processing. So we see that here in this group of deaf signers, we have two complementary reciprocal effects of when a child learns his or her . . .when an individual learns their first language and what the brain seems to be doing in terms of processing that language. So that for people who learned language early in life, almost all of their neural resources are devoted to processing the meaning and structure of that language. For people who learned their first language later in life, more neural resources are devoted to just trying, perhaps, to figure out what the signal was. Was this a word? Was it glum? Or was it gleam? So we have this reciprocal relationship between perceptual processing and language processing. This particular pattern is not unique to deaf signers. There's work by Tim Brown and Shleger that showed that younger children often have more posterior activation than older children. There are also some clinical populations, such as autistic individuals, particularly those who have low . . . whose language skills are not well-developed, will often show more processing in the occipital lobe. So this is not a pattern that is unique to deafness. So then the next question we had is whether, in fact . . . How does language develop when an individual first starts to learn it when they're much older, for example, when they're a teen? We have been very fortunate to have followed five or six children in our laboratory who had no language until they were 13 to 14 years of age, for a variety of reasons. Two of these children are from the United States. These other children are from other countries. Actually this particular circumstance, while we might think of it as being very rare, is actually very common, particularly in underdeveloped countries. So the way in which we have observed or analyzed language acquisition is to use normal procedures that people use to study children's language acquisition. We get a lot of spontaneous data from them, and we analyze it. So one question we had is: if you're 13 years old, and you don't have a language system, will you develop language like a baby? Or will you do something else? Because you have a developed cognitive system. Will you jump in the middle of the task? How will this progress? To answer this question, we need to look a little bit at how normally-hearing children or deaf or hearing children develop language when they are exposed to it as a young age. The major hallmark of children's language acquisition is that they very quickly, as they're acquiring the grammar of their language, their sentences get longer and longer. The reasons their sentences get longer and longer is because they're learning all of these . . .the morphology, the syntax. As they say ideas, as they're expressing their ideas, they're better able to use grammar to express them. So these data show the average length of children's expressions. Two of these children are normally hearing and acquiring English, and two of these children are acquiring ASL. So we see that, in fact, the teens that we have been following show no increase in their language. They're able to learn language and put words together, but, in fact, we don't see an increase in their grammar. In the last study, we wanted to neuro-image these children. We wanted to see what are their brains doing with the language that they have. So we used magnetoencephalography, which is a different technique, which is complementary to the fMRI. What we did for this is we studied their vocabulary, and we made stimuli that we knew that they knew, words that were in their vocabulary, and that looked something like this. In the first instance, the picture matched the sign. In the second instance, it didn't. When that happens, the brain goes, "Uh-oh," and you get this N-400 response. That's what we were localizing in the brain for these children. Because we're using vocabulary that they have, we know that they knew these words. We asked them to press buttons while they were doing this task, and we knew that they were accurate. We didn't only test these children. We also tested control groups. So some of these control groups are deaf. Some are hearing. Some are first language learners, and some are second language learners. The first panel shows the response of a group of normally hearing adults doing this task while looking at pictures and listening to words. This is data that Katie Travis used also to look at children's development, neural development. The second panel, these are deaf adults who learned ASL from birth. You can see that their processing is very much like the hearing adults who are speaking English, primarily left hemisphere in the language areas, with some support or help from the right hemisphere. Actually these patterns are indistinguishable. Both the hearing adults and the deaf adults learned language from birth, even though it was in a different form. What's this last panel? These are college students who are normally hearing. They have been learning sign for about three years, which is about the same amount of time that our cases were learning language. So we see that responding on this task in speech in ASL, whether it's a first language or a second language, so long as the subject had language from birth, looks fairly similar. What about the cases? We were able to neuro-image two cases, and you can see that their neural processing patterns are quite different, and they look neither like second language learners, nor do they look like deaf adults. These children have been signing for three years, and they had no language before they started to learn how to sign. You can see primarily that there's a huge response in the right hemisphere, in the occipital and parietal areas. One of the subjects also shows some response in the language areas. We can see that even though they're acquiring language, they're doing it in a very different way, and their brains are responding very differently. So we see that, in fact, there are huge effects of language environment on both the development of language, but also how the brain processes language. So we see that it seems to be that the human language capacity, both understanding language and expressing language, but also the brain's ability to process language is very dependent upon the baby's brain being exposed to or immersed in language from birth. It's through this analyzing language and working on the data that it's being fed that, in fact, I think the neural networks of language are being created. So language is a skill that is not innate but emerges from the interaction of the child with the environment, through linguistic communication. That was probably the answer that Itard was looking for and might be the reason why Victor did not acquire language. Thank you. [applause] ♪ [music] ♪ - [Edward] Okay. So I wanted to first thank the organizers really for this kind invitation to join this cast of stellar thinkers about the biology and behavior of language. I actually want to switch the title a little bit to something more specific. I want to talk to you about organization, in particular a kind of organization that I refer to as a taxonomy. The organization that I'm actually really referring to is the organization of sound, in particular speech sounds and how those are actually processed in a very important part of the brain called the superior temporal gyrus, also known as Wernicke's area. The main focus of my lab is actually to understand the basic transformation that occurs when you have an acoustic stimulus and how it becomes transformed into phonetic units. In other words, basically, how do we go from the physical stimulus that enters our ears into one that's essentially a mental construct, one that's a linguistic one? To basically ask the simple question, what is the structure of that kind of information as it's processed in the brain? Now this actually turns out to be a very complex problem because it's one that actually arises from many levels of computation that occur in the ascending auditory system. As sounds actually come through the ears, they go up through at least seven different synaptic connections across many different parts of the brain, even bilaterally, to where they're actually processed in the non-primary auditory cortex in the superior temporal gyrus. What we know about this area from animal studies and non-human primates, for example, is this is an area that no longer is tuned to basic low-level sound features, like pure tones, pure frequencies, but in one that is actually tuned to very broad, complex sounds. There have been very nice work in fMRI that's actually demonstrated that this area is far more selective to complex sounds, like speech, over non-speech sounds. So the basic question is not really about where is this processing going on. The question is how. Okay. What is the structure of information in this transformation that's going on? In particular, what kind of linkages can we make between that physical stimulus and the internal one, which is a phonetic one? For me, I think it's really important to acknowledge some really important fundamental contributions that occur, that give us some insight and put them in a very important perspective. For me, I think one of the most important pieces of work that led to this work that I'm about to describe from our lab was actually 25 years ago, using an approach that's actually far more complicated and difficult to achieve than what we do in our own work. This is using single unit and single neuron recordings that were recorded from patients that were undergoing neurosurgical procedures for their clinical routine care. This very extremely rare but precious opportunity to actually record from certain brain areas while someone is actually recording-- listening to speech. These are from my close colleague and mentor, Dr. Ojamin, who's in the audience today. But why I think it's so important to acknowledge this work is a lot of the clues about what I'm about to describe were actually seen 25 years ago. This figure that's extracted from that paper, where they actually showed and could record from single brain cells, called neurons, in the superior temporal gyrus, that they were active and corresponding to very specific sounds. But if you actually look at where those sounds are, they're not exactly corresponding to the same exact, let's say, phonemes or the same exact sounds, but, in fact, they are corresponding to a class of sounds. This was an observation that was made in this paper. They thought perhaps this is some mention of phonetic category representation there. But it wasn't that clear, actually, and there were a lot of other really important observations that were made in that paper. Now from a linguistic perspective, in thinking about how, behaviorally, we organize this information in the brain, there's actually a wonderful way to approach it. It's not perfect, but a very wonderful way to think about how languages across the world actually share a similar and shared inventory of speech sounds, not all completely the same. Each language has a different number, but they highly overlap. The reason why they overlap is because they are produced by the same vocal tract. This is essentially like a periodic table of sound elements for human language and speech. So this table actually has two really important dimensions. The horizontal dimension is actually one that we call the place of articulation. It's referencing where in the vocal track these sounds are made. For example, bilabial sounds, the 'P' and the 'B' require you to actually have a transient occlusion at the lips, 'ba. ' You cannot make those sounds without that particular articulatory movement. Whereas some of the other sounds, like a 'D' or a 'T', a 'da' a 'ta' a 'da' we call alveolar because the front of the tongue tip is actually placed against the teeth. So these are actually referencing where occlusions are occurring in the vocal tract when we speak, and those actually correlate necessarily to very specific acoustic signatures. The other dimension is what we call the manner of articulation. So the manner of articulation is actually telling you a little bit more about not so much where, but how in the vocal tract the constrictions are made in order to produce those sounds. We have certain ones, like plosives, where you have complete closure of the vocal tract and then a transient release, other sounds where you have near-complete, like a fricative, like 'sha,' 'za,' those sounds that we call fricative. If you actually look at vowels, they actually have a similar structural organization. There actually is something that actually references where in the vocal tract, either the front, middle, or back, or the degree of open and closure. So for both consonants and vowels, there actually is a structure that we know about, linguistically and phonologically, about how these things are organized. I think the thing that interests me is that, like I referenced before, that this is something like a periodic table. There is something fundamental about these units to our ability to perceive speech. These phonological representations are not necessarily the ones that we think of as these letters that we call phonemes, but actually groups of phonemes that share something in common, what we call features. These are the members of small categories which combine to form the speech sounds of human language. This became very attractive to me as a model of something to look for in the brain because of . . . Essentially why it could be so important is that languages actually do not vary without limit, but they actually reflect some single or limited general pattern, which is actually rooted in both the physical and cognitive capacities of the human brain, and I would add the vocal tract. This is not a new kind of thinking, but it's one that has not been clearly elucidated in terms of its biological mechanisms. So in order for us to get this information, it requires a very special opportunity, the one where we can't actually record directly from the brain. In many ways, this is actually a lot more coarse than the kind of recordings that were done almost 25 years ago. These are ones from electrode sensors that are placed on the brain in order to localize seizures in patients that have epilepsy. In the seven to 10 days that they are usually waiting to be localized, we have a very, again, precious opportunity to actually have some of the participant, the patient volunteers, listen to natural, continuous speech and look at those neural responses on these electrode recordings to see how information is distributed in the superior temporal gyrus when they're listening to these sounds. This gives you a sense of actually what that neural activity pattern looks like. [audio sample] - We're going to slow down that sentence a lot here. - [audio sample] Ready tiger go to green five now. - So you can see that the information is being processed in a very precise, both spacial and temporal, manner in the brain. This is exactly the reason why this kind of information has been elusive, because we do not currently have a method that actually has both spacial and temporal resolution and, at the same time, covers all of these areas simultaneously. So it's, again, in the context of these rare opportunities with human patient volunteers that we can conduct this kind of research. So the natural question is . . . Of course, now that I've shown you that we can actually see a pattern in the brain, both that's temporally and spatially specific, what actually happens when we try to deconstruct some of those sound patterns from the brain? This just gives you an example, again, in the superior temporal gyrus, where those sounds are activating the brain. An example of the spectrogram for a given sentence, in this case, it's, "In what eyes there were." The last part of that figure basically shows you that pattern across different electrodes. It's not all happening in the same particular way. You have very specific evoked responses that actually occur at different parts of the superior temporal gyrus. I want to show you what happens when you look at just one of those electrodes. If you look at the neural response of that one particular electrode that's labeled e1, and you organize the neural response by different phonemes, okay, you can actually see, again, on the vertical access, starting with 'da,' 'ba,' 'ga,' 'ta,' 'ka.' You can see that this electrode . . . Those hundreds or thousands of neurons that are under this electrode are very selectively responsive to this set of sounds that we call plosives. It's not one phoneme, but a category, and they share this feature that we actually know, linguistically, to be called plosive. I can show you a series of other electrodes. Electrode two has a very different kind of sensitivity. It's showing you that it really likes those sounds 'sha,' 'za,' 'sa,' 'fa.' This is an electrode that is, again, not tuned to one phoneme, but actually tuned to the category of sibilant fricatives in linguistic jargon. We have another electrode, e3, that is selective to low-back vowels, these "ah" based ones. Another one that is a little bit more selective to high-fronted vowels, 'E.' Even another electrode, e5, that is corresponding to nasal sounds. So this is a very low-level description, but it's actually the first time we've ever seen in this kind of principled way, obtained through very precise spatial and temporal recordings, the ability to resolve phonetic feature selectivity at single electrodes in the human brain. Now this is not enough. We need to really address this issue of structure. That's one of the themes here. Are all of these things just equally distributed as features? In the original thinking about these things, you could have a binary list of features. It turns out that features, in and of themselves, actually have structure and have relationships with one another. So what we did, in order to look at that structure in the brain, we looked at hundreds of electrodes that were recorded over a dozen patients. Each one of those columns actually corresponds to one electrode and one particular superior temporal gyrus in someone's brain. Like I just showed you before, the vertical axis is actually how they're organized by different phonemes. What we did here was we used a statistical method called hierarchical clustering. What hierarchical clustering is used for is finding the patterns in this data. What the hierarchical clustering showed us and sorted this data was that, in fact, there is, indeed, structure in the brain's responses to human speech sounds, and it looks like this. So we've organized the hierarchical clustering as a function of a single electrode's, again, a single column's selectivity to different phonemes, but we've also organized this clustering as a population response across all of the electrodes and looking at that selectivity for different phonemes. So we have two different axes that we're actually looking at the brains large distributed response to speech sounds. We're using this method which is what we call unsupervised, meaning we're not telling it any linguistic information, or we're not organizing the data. We're just saying, "Tell us how the brain is organizing this information." What we see from this is that when we actually look at where this information is being organized, one of the biggest divisions between different parts of different kinds of selectivity in the brain are what we would call the difference between consonants and vowels or really, actually, between obstruents and continuants, in linguistic jargon. But within those different categories, you actually have sub-classification. So within the consonants, you actually have a subdivision between plosives and fricatives. Between the sonorants you actually have referencing for different positions of the tongue, low back, low front, high front, different classes of vowels and, in fact, nasal. So basically this is telling you that feature selectivity in the brain is actually hierarchically structured. The second thing is that instead of using phonemes in order to organize the responses, we actually use features. So as an example, that term dorsal actually refers to the tongue position when it's fairly back, like for 'G' 'K' sounds. You can see that when we organize things by features, you have a much cleaner delineation. The electrode responses seem to be much more tuned to phonetic features shown below, as they are, compared to when you plot them as phonemes. Okay. So this essentially disproves any idea that there is individual phoneme representation in the brain, at least not one that's locally encoded, but tells you that the brain is organized by its sensitivity to phonetic features. Now relating it to a phonetic feature is the first step, and it's one that's really important because it's referencing the one we know about from linguistics and the one that we know behaviorally. But how do we connect this to the physical stimulus that's actually coming through our ears? That's where we have to make a linkage to actually something about the sound properties. Are these things truly abstract features that are being picked up by the brain? Or actually, are they referencing specific sound properties? Basically the answer is the latter. It's that what we're actually seeing is sensitivity to particular spectral temporal features. In the top row, I am showing you basically . . . When we look at the average tuning curves, the frequency versus time, tuning curves for each one of those different classifications for plosive fricatives, they're very similar to the acoustic structure when you average those particular phonemes in the brain. So what this means is the tuning that we're seeing that's corresponding to phonetic features is, in fact, one that is tuned to high order acoustic spectral temporal ones. The brain is selecting specific kind of acoustic information and converting it into what we perceive as phonetic. In the interest of time, I'm going to sort of skip more in-depth information about vowels and plosives and how those are specifically encoded. But in summary, what we've found is that there's actually a multidimensional feature space, actually, for speech sounds in the human superior temporal gyrus. This feature space is organized in a way that actually shows hierarchical structure. The hierarchical structure is fairly strongly driven by the brain, in particular, this auditory cortex sensitivity to acoustic differences, which are most signified actually in the manner of articulation distinctions, linguistically. What's interesting about this is it actually does correlate quite well with some known perceptual behavior. So I would like to conclude there and acknowledge some of the really important people from my lab, postdoctoral fellow Nima Mesgarani who did most of this work with one of our graduate students, Connie Cheung. Thank you. [applause] ♪ [music] ♪
B1 中級 美國腔 CARTA:語言是如何進化的。大腦中的語言 (CARTA: How Language Evolves: Language in The Brain) 254 32 J.s. Chen 發佈於 2021 年 01 月 14 日 更多分享 分享 收藏 回報 影片單字