James W. Lewis, PhD – Associate Professor
1 Medical Center Drive
PO Box 9303
Morgantown, WV 26506
Physiology, Pharmacology, & Neuroscience; Rockefeller Neuroscience Institute
PhD, Neurobiology, California Institute of Technology, Pasadena, CA
Neuroimaging, Medical College of Wisconsin, Milwaukee, WI
Our neuroimaging research interests include:
- Systems-level visual, auditory, and multisensory processing using functional magnetic resonance imaging (fMRI).
- Cognitive models of how the brain is organized to make sense of the everyday sounds we hear.
- Sensori-motor processing networks in the context of cognitive models of object and action knowledge representations.
- Cortical plasticity and adaptations made by the brains of children and adults with autism spectrum disorder (ASD) to process real-world multisensory information.
- Studies of acoustic signal processing mechanisms by the human brain that may be amenable to advancing biologically inspired hearing aid algorithm design.
Our primary research projects use state-of-the-art 3 Tesla functional magnetic resonance imaging (fMRI) at the Center for Advanced Imaging at WVU. We use a full scale simulation-MRI scanner to train children to be comfortable with neuroimaging environments. Other projects, typically related to hearing perception, utilize evoked response potentials (ERPs) and basic electroencephalography (EEG) techniques.
Description of Research
Our research explores general principles of how the human brain processes auditory information, especially with regard to our ability to combine sound with vision and touch (“multisensory integration”).Thus, the goals of our current research, using fMRI, are threefold. First is to explore brain regions responsible for our ability to recognize and identify sounds (natural or environmental sounds, in contrast to speech sounds which tend to activate specialized language-related brain areas in humans). Second is to explore brain systems responsible for our spatial perception of sound, including localizing sounds in three-dimensional space and/or perceiving a moving sound-source. And the third is to examine where and how the information processed along the auditory pathways becomes integrated with visual and somatosensory/motor processing pathways in the brain, providing us with a unified percept of the “multisensory” objects we experience every day.
Our group investigates the general principles of how the human brain processes auditory information, including studies of multisensory perception and advancing models of cognition. We primarily use functional magnetic resonance imaging (fMRI), with the 3T scanner at the Center for Advanced Imaging (http://www.hsc.wvu.edu/cai/), but also use neurophysiological techniques such as collecting evoked response potentials (ERPs). One of our research goals is to understand how the human brain is organized to represent knowledge of sensory events, and be able to gain a sense of meaning behind what we see and hear. A global model of how “all natural sounds” are processed (by category) are depicted in the figure below from Brefczynski-Lewis & Lewis, (2017).
For instance, the rotating 3D model of the human brain (left cortical hemisphere) illustrating cortex that is more responsive to hearing and recognizing environmental sounds (Fig. 1, yellow) in contrast to hearing but not recognizing the same
sounds played backwards. N=24 participants, α<0.05). Below are some of the forward and backward-played sound pairs, which can be heard by clicking on the icons. Can you identify the backward-played sounds?
For over a century, human neuropsychological (brain lesion) studies have indicated that semantic knowledge of “living versus non-living things” are represented in distinct brain regions or networks. Our research has uniquely contributed to our understanding of these models from the perspective of hearing perception. In particular, different colored brain regions (Fig. 2) depict areas of human cerebral cortex (group-averaged data) that are preferentially activated by one category of action sound relative to each of the three other categories indicated. This included a four-fold dissociation of two sub-categories of “living” sound sources (Human versus Animal action sounds) and two sub-categories of “non-living” sound-sources (Mechanical and Environmental). These types of studies have advanced our understanding of grounded cognition models for how acoustic and object knowledge (and our ability to think about objects) are organized or mediated by the brain.
To determine more specific functions of various “activated” regions (colored areas on the brain models in Figure 3 above), one approach we use is to examine “extremes” in cortical organization for representing acoustic knowledge. For example, we found that strongly
listeners more strongly activated the left hemisphere when hearing and recognizing action sounds produced by hand tools (e.g. hammering, sawing, drawing). Red depicts brain regions preferentially activated when hearing and processing hand-manipulated tool sounds, while blue depicts regions more responsive to animal vocalizations. Green shows cortex activated while participants (20 in each group) made dominant hand movements as if manipulating a variety of different tools. Note that the inferior parietal lobule (“IPL”; yellow overlap region) was strongly activated only in the hemisphere opposite the dominant hand, being very responsive when recognizing tool-related sounds.
These findings fit well with clinical studies, in that lesions to the IPL in the hemisphere opposite the dominant hand can lead to a severe disruption in one’s conceptual understanding of how to appropriately use tools or common objects (termed “ideational apraxia”). For instance, when asked to manipulate objects of common use, such patients might produce a hammering gesture with a set of keys, or attempt to use scissors to write.
We have also been able to examine how visual experience might affect brain networks that encode knowledge representations of every-day natural sounds by examining the brain of congenitally blind individuals. For instance, we found that recognizable human action sounds (in contrast to unrecognizable, backward-played versions of those sounds) preferentially activated distinct networks in the blind versus the sighted (Fig. 4). This included the blind group activations in response to human action sounds included significantly different cortical networks
Another focus of our research involves studying “bottom-up” acoustic signal processing using fMRI. Using fMRI, we recently reported that parametric increases in harmonic content, quantified by a harmonics-to-noise ratio (HNR value), of both animal vocalizations and artificially created iterated rippled noises (IRNs) led to parametric increases in activation of auditory cortices. These HNR-sensitive foci (Fig. 3, Tier 2; green and blue regions) were situated between tonotopically organized primary auditory cortices (Tier 1, yellow) and regions preferential for human non-verbal vocalizations and speech (Tier 3, purple hues).
Research Assistant III
Chris has been involved in research at WVU for over a decade. He has managed a public
health research study looking at heart disease in women in the Ohio River Valley
region of West Virginia. His interests are in fMRI and resting state fMRI data
processing, data integrity, and student mentoring.
Neuroscience PhD Candidate
Research interests: Brain adaptations for sensory and cognitive processing in individuals
with autism spectrum disorder (ASD).
SURE Research Student
I am currently a sophomore at West Virginia University, majoring in Biomedical Engineering
with a Minor in Mathematics. Technological advances in the medical field are occurring
continuously, and biomedical engineers are at the forefront of these discoveries.
I have an intense passion for all facets of STEM, and I am extremely eager to apply
my education to real world applications.
Former Lab Members:
Psychology Graduate Student
WVU Undergraduate Student
Project title: “Resting-state MRI clinical-translational applications to bed-ridden
in-patient health recovery”
Summer Undergraduate Research Interns (SURIs):
William J. Talkington (2005, 2006)
PhD program at WVU, postdoctoral fellow at WVU
Audrey Jajosky (2006)
MD/PhD program at WVU
Lauren R. Engel (2007)
MD program in Colorado
Laura M. Skipper (2008)
PhD program at Temple University
Kristina M. Rapuano (2009)
PhD program at National Institutes of Health
Stephen Gray (2010)
PhD program at University of Chicago
Susannah Engdahl (2011)
PhD program at University of Michigan
Hayley N. Still (2013)
Undergraduate student at Lewis & Clark College, OR
Matthew Preda (2014)
Undergradaute student at Wheaton College
Magenta Silberman (2015)
Undergraduate student at Syracuse University
Gabriela Valencia (2016)
Undergraduate student at Loyola University Chicago
A list of pubished work in MyBibliography:
- Brefczynski-Lewis JA,
Lewis JW. (2017)
Auditory object perception: A neurobiological model and prospective review.
Neuropsychologia, 105:223-242. PMCID:
- Webster PJ, Skipper-Kallal LM, Frum CA, Still HN, Ward BD,
Divergent human cortical regions for processing distinct acoustic-semantic
categories of natural sounds: Animal action sounds vs. vocalizations.
Frontiers in Neuroscience (January 2017) 10:579.. doi: 10.3389/fnins.2016.00579
- Bauer CE, Brefczynski-Lewis JA, Marano G, Mandich MB, Stolin A, Martone P,
Lewis JW, Jaliparthi G, Raylmann RR, Majewski S.
Concept of an Upright Wearable Positron Emission Tomography (PET) Imager in
Brain & Behavior.
- Geangu E, Quadrelli E,
, Macchi Cassia V, & Turati, C. (2015).
By the sound of it. An ERP investigation of human action sound processing in
Developmental Cognitive Neuroscience
- Talkington WJ, Taglialatela JP,
Using naturalistic utterances to investigate cortical pathways for processing
communication sounds in humans and non-human primates
. Invited Review for
eangu E, Quadrelli E,
, Macchi Cassia V, & Turati, C. (2015).
By the sound of it. An ERP investigation of human action sound processing in
. Developmental Cognitive Neuroscience
Lewis JW, Talkington WJ, Tallaksen KC, Frum CA.
Auditory object salience: human cortical processing of non-biological action
sounds and their acoustic signal attributes.
Frontiers in Systems Neuroscience (2012 May). doi:10.3389/fnsys.2012.00027.
- Talkington WJ, Rapuano KM, Hitt L, Frum CA,
Humans mimicking animals: A cortical hierarchy for human vocal communication
J. Neuroscience. (June 2012) 32(23):8084-93.
- Lewis JW, Talkington WJ, Puce A, Engel LR, Frum C.
Cortical networks representing object categories and high-level attributes of
familiar real-world action sounds.
J. Cog. Neuroscience (2011) 23(8):2079-2101.
- Lewis JW, Frum C, Brefczynski-Lewis JA, Talkington WJ, Walker NA, Rapuano
KM, Kovach AL.
Cortical network differences in the sighted versus early blind for recognition
of acoustic events and situational relationships.
Human Brain Mapping (2011; Epub2 2010) (32):2241-55. doi:10.1002/hbm.21185.
- Lewis JW. (2010).
Audio-visual perception of everyday natural objects-hemodynamic studies in humans.
Invited book chapter in
Multisensory Object Perception of the Primate Brain. Eds. MJ Naumer, J Kaiser.
Springer, Oxford University Press. Chapter 10:155-190.
- Ortigue S, Bianchi-Demicheli F, Patel N, Frum C,
Neuroimaging of Love: fMRI Meta-Analysis Evidence toward New Perspectives in
J Sex Med. (2010) (11):3541-52. doi: 10.1111/j.1743-6109.2010.01999.x.
- Engel LR, Frum C, Puce A, Walker NA,
Different categories of living and non-living sound-sources activate distinct
- Lewis JW, Talkington WJ, Walker NA, Spirou GA, Jajosky A, Frum
C, Brefczynski-Lewis, JA. (2009)
Human cortical organization for processing vocalizations indicates representation
of harmonic structure as a signal attribute.
J. Neuroscience; 29(7):2283-2296.
- Huddleston WE,
Lewis JW, Phinney RE, DeYoe EA.
Auditory and visual attention-based apparent motion share functional parallels.
Journal of Cognitive Neuroscience (2008), 70(7):1207-1216.
- Brefczynski-Lewis JA, Datta R,
Lewis JW, DeYoe EA.
The topography of visuospatial attention as revealed by a novel visual field
Journal of Cognitive Neuroscience (2008).
Department of Physiology & Pharmacology, WVU:
For related fMRI studies of sensory perception, please visit the DeYoe Lab home page:
For information on generating cortical flat maps and maps of monkey cortex, please
visit the Van Essen home page:
Other cool links:
Visual illusion demos: