Beauchamp: Difference between revisions

From OpenWetWare
Jump to navigationJump to search
No edit summary
(28 intermediate revisions by the same user not shown)
Line 4: Line 4:
|-valign="top"
|-valign="top"
|width=400px style="padding: 5px; background-color: #ffffff; border: 2px solid #CC6600;" |
|width=400px style="padding: 5px; background-color: #ffffff; border: 2px solid #CC6600;" |
====AFNI Short Course: October 4-8, 2010====
{{Click || image=Beauchampposter.jpg | link=Beauchamp:ShortCourse2010| width=170px | height=250px  }}
[[Beauchamp:ShortCourse2010|Click for information on the short course]]
<h3>Home</h3>
<h3>Home</h3>


The Beauchamp Lab studies the neural mechanisms for visual perception and multisensory integration in human subjects. Many complex tasks require us to integrate information from multiple modalities. For instance, when trying to understanding spoken language, we make use of both the auditory information in the heard speech and the visual information from the facial movements of the speaker. This multisensory integration is especially important under conditions in which one modality is degraded, such as a loud room. Understanding the neural mechanisms for multisensory integration under these conditions may help us treat patients with sensory deficits, such as the decreased visual and auditory sensitivity common in aging. To understand the neural basis of multisensory integration, the primary method used is blood-oxygen level dependent functional magnetic resonance imaging (BOLD fMRI). fMRI experiments are conducted using the research-dedicated 3 tesla scanner in the UT MRI Center adjacent to the lab. Because of the limited temporal and spatial resolution of fMRI, we often combine it with other methods. It is particularly useful to combine fMRI with electrical or magnetic stimulation of the brain to determine if a region of activity observed in fMRI is truly important for the cognitive operation of interest. Electrical recording from patients implanted with electrodes for the treatment of medically intractable epilepsy is also an important tool, because it allows direct recording of the activity of small populations of neurons, as opposed to the indirect nature of the BOLD signal. Anatomically, the primary focus of the lab is on the superior temporal sulcus, a brain area critical for multisensory integration and visual perception of complex movements.
The Beauchamp Lab studies the neural mechanisms for multisensory integration and visual perception in human subjects; anatomically, the primary focus of the lab is on the superior temporal sulcus, a brain area critical for both the integration of auditory, visual, and somatosensory information and for the perception of complex visual motion, such as mouth movements. Many everyday tasks require us to integrate information from multiple modalities, such as during conversation when we make use of both the auditory information we hear in spoken speech and the visual information from the facial movements of the talker. Multisensory integration is especially important under conditions in which one modality is degraded, such as in a noisy room. Even in healthy young adults, there is considerable variability in people's ability to integrate auditory and visual speech, but this difference in even more pronounced when other populations are examined. Very young children rely exclusively on auditory information to understand language, but in normal lifespan development visual speech plays an increasing role, sometimes becoming dominant as hearing declines with age. Other populations also show interesting differences: deaf children commonly use a cochlear implant to allow them to hear, but the early lack of auditory input sometimes prevents them from ever properly integrating auditory and visual speech. To understand the neural mechanisms of multisensory integration and visual perception, our primary method is blood-oxygen level dependent functional magnetic resonance imaging (BOLD fMRI). fMRI experiments are conducted using the research-dedicated 3 tesla scanner in the UT MRI Center adjacent to the lab. Because of the limitations of fMRI, we often combine it with other methods, including transcranial magnetic stimulation (TMS) and electrical stimulation and recording. Through these sophisticated studies, we hope to unlock one of nature's great mysteries: how the brain performs amazing computational feats, such as understanding speech, that allow us to make sense of the auditory and visual world around us. Every advance in deepening our knowledge of these processes is not only exciting for its own sake but will also help children and patients with language and perceptual difficulties.


[[Image:UTHealthLogo.jpg|120px|UT Health Logo]]   


You can reach us at:
You can reach us at: Michael.S.Beauchamp (at) uth.tmc.edu, Department of Neurobiology and Anatomy, University of Texas Medical School at Houston, 6431 Fannin Street, Suite G.550G, Houston, Texas 77030. Telephone (713) 500-5978.
  Department of Neurobiology and Anatomy
  University of Texas Medical School at Houston
  6431 Fannin Street, Suite G.550G
  Houston, Texas 77030

Revision as of 12:40, 7 February 2012

Brain picture
Beauchamp Lab



Home

The Beauchamp Lab studies the neural mechanisms for multisensory integration and visual perception in human subjects; anatomically, the primary focus of the lab is on the superior temporal sulcus, a brain area critical for both the integration of auditory, visual, and somatosensory information and for the perception of complex visual motion, such as mouth movements. Many everyday tasks require us to integrate information from multiple modalities, such as during conversation when we make use of both the auditory information we hear in spoken speech and the visual information from the facial movements of the talker. Multisensory integration is especially important under conditions in which one modality is degraded, such as in a noisy room. Even in healthy young adults, there is considerable variability in people's ability to integrate auditory and visual speech, but this difference in even more pronounced when other populations are examined. Very young children rely exclusively on auditory information to understand language, but in normal lifespan development visual speech plays an increasing role, sometimes becoming dominant as hearing declines with age. Other populations also show interesting differences: deaf children commonly use a cochlear implant to allow them to hear, but the early lack of auditory input sometimes prevents them from ever properly integrating auditory and visual speech. To understand the neural mechanisms of multisensory integration and visual perception, our primary method is blood-oxygen level dependent functional magnetic resonance imaging (BOLD fMRI). fMRI experiments are conducted using the research-dedicated 3 tesla scanner in the UT MRI Center adjacent to the lab. Because of the limitations of fMRI, we often combine it with other methods, including transcranial magnetic stimulation (TMS) and electrical stimulation and recording. Through these sophisticated studies, we hope to unlock one of nature's great mysteries: how the brain performs amazing computational feats, such as understanding speech, that allow us to make sense of the auditory and visual world around us. Every advance in deepening our knowledge of these processes is not only exciting for its own sake but will also help children and patients with language and perceptual difficulties.

UT Health Logo

You can reach us at: Michael.S.Beauchamp (at) uth.tmc.edu, Department of Neurobiology and Anatomy, University of Texas Medical School at Houston, 6431 Fannin Street, Suite G.550G, Houston, Texas 77030. Telephone (713) 500-5978.