Visual influences on auditory speech perception
Multisensory research has demonstrated that observing lip movements is critical for phoneme acquisition during development as well as when auditory information is degraded by environmental noise or auditory deficits. This line of research is supported by our current NIH K99/R00 grant support. Our lab’s model of multisensory speech perception proposes that visual information facilitates speech processing through multiple distinct anatomical and functional networks integrating different types of information across the senses, including (1) the use of visual contextual information associated with auditory speech signals (e.g., the gender and identity of the speaker), (2) the use of preparatory lip motion predictive of the timing of phonemic onsets, (3) the use of lipreading information, as well as (4) the use of low-level arousal mechanisms.
Auditory influences on visual processing
Our lab investigates the mechanisms through which auditory information facilitates visual processes, with particular interest in identifying what information is relayed between the senses (and how this process can differ based on task or the neural mechanism). For example, hearing a sound can improve your ability to localize, identify, and respond to a visual object that appears at the same time. Is this due to the sound increasing levels of arousal or attention, or by giving us two chances to respond to the object (referred to as intersensory redundancy), or does it provide additional statistically relevant information about the object that helps us tune visual processing? We have evidence that all three processes are at work depending on the task. Most importantly, these effects get bigger when sensory damage occurs, highlighting this as an important protective and compensatory mechanism in the brain.
Understanding altered sensory and cognitive functions in patients with intrinsic brain tumors
In collaboration with Shawn Hervey-Jumper’s lab at the University of California, San Francisco hospital, we are working to improve functional outcomes following brain surgery. This research program seeks to broaden cognitive mapping that occurs in the operating room (with the goal of minimizing cognitive deficits after surgery) beyond language and motor functions to other critical cognitive processes including sensory integration and attention. Using electrocorticography (ECoG), we are developing tools to allow the quick assessment of the functional responses from a broad range of cortical areas without requiring overt responses (two significant limitations of current cortical stimulation mapping approaches). As part of this program, we are longitudinally testing intrinsic brain tumor patients on a battery of cognitive and perceptual tests to study clinical outcomes as well as basic research questions. For example, where does brain damage most strongly affect cognitive processes such as listening, talking, or maintaining attention? Critically, if these areas are damaged, can they recover through neural plasticity?
Synesthesia is a neurological phenomenon in which individuals are born with additional links between their senses. Some experience black numbers and letters as having colors (e.g., 2 may look blue or have the feeling that it should be red), or sounds may elicit colors or tastes — synesthesia can theoretically link up any two senses. These experiences are relatively common (2-4% of the population have one or more forms) and possessing synesthesia is not associated with any disorders.
We are currently conducting research on the relationship between synesthesia and more typical multisensory experiences. If you believe you may have synesthesia and are interested in participating in research, please contact us at firstname.lastname@example.org