1 / 6

Neural processing of language in the absence of auditory input

Neural processing of language in the absence of auditory input. Naja Ferjan Ramirez In collaboration with Matt Leonard, Christina Torres, Marla Hatrak , Eric Halgren , & Rachel Maybery Mayberry Lab Multimodal Imaging Lab Cognitive Neuroscience 2013 Spring Retreat &

lorna
Download Presentation

Neural processing of language in the absence of auditory input

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Neural processing of language in the absence of auditory input Naja Ferjan Ramirez In collaboration with Matt Leonard, Christina Torres, Marla Hatrak, Eric Halgren, & Rachel Maybery Mayberry Lab Multimodal Imaging Lab Cognitive Neuroscience 2013 Spring Retreat & Kavli Institute for Brain and Mind Symposium

  2. Early stages = modality specific AUDITORY WORDS WRITTEN WORDS • Are the neural underpinnings of word meaning dependent on auditory language experience? • Congenitally deaf individuals receive no auditory input • Some acquire sign language as their native language from birth Later stages = modality independent? Left dominant lexico-semantic network Marinkovic et al. 2003 Is the meaning of ASL signs in deaf native signers processed like the meaning of auditory & written words in hearing people (i.e. left dominant, fronto-temporal areas)?

  3. Word Onset Anatomically constrained Magnetoencephalography Average waveform from single MEG sensor amplitude • Deaf native signers (n=11) watch ASL signs • Hearing English speakers (n=8) listen to English words ms 200 100 300 400 500 600 700

  4. Integrating MEG (temporal resolution)… … with MRI (spatial resolution) ms 200 100 300 400 500 600 700 Method: Dale et al, 2000

  5. Sensory processing: Modality specific Sign in native deaf 300-350 ms Speech in hearing 300-350 ms Lexico-semantic processing: Independent of modality through which early language is received Neural systems for high-level language processing are supramodal!

  6. Thank you! NSF BCS-0924539 NIH RO1DC012797

More Related