1 / 27

Cross-modal Prediction in Speech Perception

Cross-modal Prediction in Speech Perception. Carolina Sánchez, Agnès Alsius, James T. Enns & Salvador Soto-Faraco Multisensory Research Group Universitat Pompeu Fabra Barcelona. Auditory + visual performance. MSI enhancement. Background. Visual + Auditory. Improve Speech Perception.

pabla
Download Presentation

Cross-modal Prediction in Speech Perception

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Cross-modal Prediction in Speech Perception Carolina Sánchez, Agnès Alsius, James T. Enns & Salvador Soto-Faraco Multisensory Research Group Universitat Pompeu Fabra Barcelona

  2. Auditory + visual performance MSI enhancement Background Visual + Auditory Improve Speech Perception Multisensory Integration

  3. Background • Prediction within one sensory modality • Many levels of information processing • Phonological prediction “ This morning I went to the library and borrowed a … book” (De Long, 2005; Pickering, 20707) • Visual prediction: Visual search (Enns, 2008; Dambacher, 2009) • Sensorimotor prediction: forward model (Wolpert, 1997)

  4. Predictive coding Pickering, 2007

  5. Hypothesis • If there exists prediction within the same modality, and if predictive coding models can account for prediction at a phonological level, then … Predictive Coding could occur across different sensory modalities too.

  6. Indirect evidences of cross-modal transfer in speech ERPs • Amplitud reduction • Shortening latency /pa/ high visual saliency /ka/ short visual saliency time van Wassenhove’s , 2005

  7. Our study • Visual prediction • Auditory prediction • Visual-to-auditory cross-modal prediction • Auditory-to-visual cross-modal prediction

  8. speech non speech Context fragment Target fragment Visual prediction Task : AV Match vs. AV Mismatch With visual informative visual context Visual stream Auditory stream Without informative context V A

  9. * Reaction time Without informative context * With visual informative context msec match mismatch Results With previous context participants respond faster than without it. VISUAL PREDICTION

  10. speech non speech Context fragment Target fragment Auditory prediction Task : AV Match vs. AV Mismatch With auditory informative auditory context Visual stream Auditory stream Without informative context V A

  11. * Reaction time Without informative context With auditory informative context * msec mismatch match Results With previous context participants respond faster than without it. AUDITORY PREDICTION

  12. Visual vs. Auditory Rts Visual prediction Auditory prediction Rts Without informative context Without informative context With visual informative context With auditory informative context msec msec * * incongruent congruent congruent incongruent

  13. Conclusions • Visual prediction • Auditory prediction Is this prediction cross-modal?

  14. V A Match V V A A Mismatch V A Predictability of Vision-to-AuditionDesign of the experiment Unimodal continued Unimodal continued Visual stream Auditory stream Discontinued Discontinued Mismatch Cross-modal continued Match Mismatch

  15. V V A A Mismatch Mismatch V A Mismatch Predictability of Vision-to-AuditionStimuli Cross-modal continued Discontinued Unimodal continued

  16. * Results Reaction time Participants were faster in the cross-modal condition than in the completely incongruent one. msec VISUAL –TO-AUDITORY PREDICTION Cross-modal continued Unimodal continued Discontinued Visual Auditory

  17. V A Match Match V V A A Mismatch Mismatch V A Mismatch Predictability of Audition-to-Vision Design of the experiment Unimodal continued Unimodal continued Visual stream Auditory stream Discontinued Discontinued Cross-modal continued

  18. Results Reaction time We didn’t find any difference between the mismatch condicions NO AUDITORY-TO-VISUAL PREDICTION msec Cross-modal continued Unimodal continued Discontinued Visual Auditory

  19. Conclusions There is some kind of prediction from vision-to-auditory modality There is not any prediction from auditory-to-vision modality Does this prediction depend on the language?

  20. Results (L1) Spanish participants with spanish sentences Canadian participants with english sentences Reaction time Reaction time * * msec msec Cross-modal continued Cross-modal continued Unimodal continued Unimodal continued Discontinued Discontinued Visual Visual Auditory Auditory VISUAL –TO-AUDITORY PREDICTION IN NATIVE LANGUAGE

  21. Spanish participants with spanish sentences Results (L1) Canadian participants with english sentences Reaction time Reaction time msec msec Cross-modal continued Cross-modal continued Unimodal continued Unimodal continued Discontinued Discontinued Visual Visual Auditory Auditory No differences between the mismatch conditions No prediction from auditory-to-visual modality in native language

  22. Conclusions There is some kind of prediction from vision-to-auditory modality in L1 There is not any prediction from auditory-to-vision modality L1 What happens with an unknown language?

  23. Canadian participants with spanish sentences Reaction time msec NO VISUAL-TO-AUDITORY IN OTHER LANGUAGE Unknown language : visual to auditory Cross-modal continued Unimodal continued Discontinued Visual Auditory

  24. Unknown language: auditory to visual Spanish participants with english sentences Canadian participants with spanish sentences Reaction time Reaction time msec msec Cross-modal continued Cross-modal continued Unimodal continued Unimodal continued Discontinued Discontinued Visual Visual Auditory Auditory No differences between the mismatch conditions No prediction from auditory-to-visual modality in other language

  25. Conclusions • No visual-to-auditory cross-modal prediction in an unknown language… it seems that some level of knowledge about the articulatory phonetics of the language is required to obtain the advantage of the predictive coding • No auditory-to-visual cross-modal prediction

  26. General Conclusions • Unimodal prediction from visual to visual modality from auditory to auditory • L1: ASYMMETRY • Cross-modal prediction from visual-to-auditory modality • No cross-modal prediction from auditory-to-visual modality • Unknown language: previous knowledge of the language is neccesary to make the prediction • No cross-modal prediction from visual-to-auditory modality • No cross-modal prediction from auditory-to-visual modality

  27. Thanks to… • Agnès Alsius, Postdoc Queen’s University • Antonia Najas, MA/ Research Assistant Universitat Pompeu Fabra • Phil Jaekl, Postdoc Universitat Pompeu Fabra - All the people of the Vision Lab, UBC, Vancouver Thanksforyourattention!!

More Related