Scientists reaching consensus on how brain processes speech

May 26, 2009
Brain

Neuroscientists feel they are much closer to an accepted unified theory about how the brain processes speech and language, according to a scientist at Georgetown University Medical Center who first laid the concepts a decade ago and who has now published a review article confirming the theory.

In the June issue of , the investigator, Josef Rauschecker, PhD, and his co-author, Sophie Scott, PhD, a neuroscientist at University College, London, say that both human and non-human primate studies have confirmed that speech, one important facet of language, is processed in the along two parallel pathways, each of which run from lower- to higher-functioning neural regions.

These pathways are dubbed the "what" and "where" streams and are roughly analogous to how the brain processes sight, but are located in different regions, says Rauschecker, a professor in the department of physiology and biophysics and a member of the Georgetown Institute for Cognitive and Computational Sciences.

Both pathways begin with the processing of signals in the , located inside a deep fissure on the side of the brain underneath the temples - the so-called "temporal lobe." Information processed by the "what" pathway then flows forward along the outside of the temporal lobe, and the job of that pathway is to recognize complex auditory signals, which include communication sounds and their meaning (semantics). The "where" pathway is mostly in the parietal lobe, above the temporal lobe, and it processes spatial aspects of a sound - its location and its motion in space - but is also involved in providing feedback during the act of speaking.

Auditory perception - the processing and interpretation of sound information - is tied to anatomical structures; signals move from lower to higher brain regions, Rauschecker says. "Sound as a whole enters the ear canal and is first broken down into single tone frequencies, then higher-up neurons respond only to more complex sounds, including those used in the recognition of speech, as the neural representation of the sound moves through the various brain regions," he says.

Both human and nonhuman primate studies were examined in this review.

In humans, researchers use functional magnetic resonance imaging (fMRI) to "watch" activity move between in experiments testing speech "(re)cognition," Rauschecker says. In non-human primates, investigators use a technique known as single-cell recording, which can measure changes within a single neuron. To do this, anesthetized animals are equipped with microelectrodes that can pick up activity in pinpointed brain areas, a technique that can be used only rarely in human patients but provides much better resolution.

"In both species, we are using species-specific communication sounds for stimulation, such as speech in humans and rhesus-specific calls in rhesus monkeys," Rauschecker says. "We find that the structure of these communication sounds is similar across species."

What is so interesting to Rauschecker is that although speech and language are considered to be uniquely human abilities, the emerging picture of brain processing of language suggests "in evolution, language must have emerged from neural mechanisms at least partially available in animals," he says.

"Speech, or the early process of language, is well modeled by animal communication systems, and these studies now demonstrate that primate auditory cortex, across species, displays the same patterns of hierarchical structure, topographic mapping, and streams of functional processing," Rauschecker says. "There appears to be a conservation of certain processing pathways through evolution in humans and nonhuman primates."

While this research is basic science trying to solve fundamental questions about the brain, it may ultimately yield some valuable insights into disorders that involve problems in comprehending auditory signals, such as autism and schizophrenia, he says.

"Understanding speech is one of the major problems seen in autism, and a person with schizophrenia hears sounds that are just hallucinations," Rauschecker says. "Eventually, this area of research will lead us to better treatment for these issues.

"But mostly, we are fascinated by the fact that humans can make such exquisite sense of the slight variation in sound waves that reach our ears, and only lately have we been able to model how the brain knows how to attach meaning to these sounds in terms of communication." he says.

Source: Georgetown University Medical Center (news : web)

Explore further: Researchers have developed a computational framework for standardizing neuroscience data worldwide

add to favorites email to friend print save as pdf

Related Stories

Researchers produce 'neural fingerprint' of speech recognition

Nov 10, 2008

Scientists from Maastricht University (Netherlands) have developed a method to look into the brain of a person and read out who has spoken to him or her and what was said. With the help of neuroimaging and data mining techniques ...

Why can’t I learn a new language?

Jul 08, 2008

Adults, even the brightest ones, often struggle with learning new languages. Dr Nina Kazanina in the Department of Psychology at the University of Bristol explains why.

Researchers identify language feature unique to human brain

Mar 23, 2008

Researchers at the Yerkes National Primate Research Center, Emory University, have identified a language feature unique to the human brain that is shedding light on how human language evolved. The study marks the first use ...

Where the brain makes sense of speech

Dec 19, 2007

Researchers have identified regions of the brain where speech sounds are perceived as having abstract meaning, rather than as just a stream of sensory input. They said their identification of the regions demonstrates that ...

Recommended for you

Researchers unlock mystery of skin's sensory abilities

12 hours ago

Humans' ability to detect the direction of movement of stimuli in their sensory world is critical to survival. Much of this stimuli detection comes from sight and sound, but little is known about how the ...

Tackling neurotransmission precision

Dec 18, 2014

Behind all motor, sensory and memory functions, calcium ions are in the brain, making those functions possible. Yet neuroscientists do not entirely understand how fast calcium ions reach their targets inside ...

User comments : 0

Please sign in to add a comment. Registration is free, and takes less than a minute. Read more

Click here to reset your password.
Sign in to get notified via email when new comments are made.