1 / 95

Within-category acoustic variation is signal, not noise:

Within-category acoustic variation is signal, not noise: A new look at Categorical Perception and the problem of covariance Bob McMurray University of Rochester. Michael Tanenhaus Richard Aslin Meghan Clayards Joe Toscano Dana Subik Julie Markant . David Gow *. Mass General Hospital.

maude
Download Presentation

Within-category acoustic variation is signal, not noise:

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Within-category acoustic variation is signal, not noise: A new look at Categorical Perception and the problem of covariance Bob McMurray University of Rochester

  2. Michael Tanenhaus Richard Aslin Meghan Clayards Joe Toscano Dana Subik Julie Markant David Gow* Mass General Hospital University of Rochester *Errors & Inaccuracies attributable here. Collaborators

  3. Overview 1) Speech perception and word recognition. 2) Gradient sensitivity to acoustic detail in lexicon? 3) How might such sensitivity be used? 4) Developmental thoughts and data. 5) Conclusions

  4. Acoustic Sublexical Units /la/ /ap/ /a/ /b/ /l/ /p/ Speech Perception: Mapping of acoustics to sublexical units. Lexicon Word Recognition: Sublexical to lexical mapping. Overview of word recognition. This work examines the interface between speech perception and spoken word recognition. To a first approximation these fields carve up the perceptual pie fairly independently. Speech Perception and Word Recognition

  5. The interface of speech and words... Speech Perception: By the end of processing, a noisy acoustic signal is carved up into discrete units. Word Recognition: Since the early processing tosses out this variability, it can (for the most part) be ignored. Therefore: how are these units mapped onto lexical items in real-time? Combining these approaches might yield a more efficient system. Speech Perception and Word Recognition

  6. Merging approaches Merging approaches: an example Embedded Words A purely phonemic representation creates ambiguity. shipping quietly / S I p I N kw AI t l i/ ship inquiry / S I p I N kw AI ®i/ Speech Perception and Word Recognition Q:How does the system know when to begin a new word? A coarse phonemic representation shouldyield extremely long recognition latencies (which we don’t see).

  7. Embedded Words shipping quietly / ShIpI N kw AI t l i/ ship inquiry / Sh I ˘ ph I N kw AI ® i/ Sensitivity to fine-grained (subphonemic) detail can disambiguate the phrases. Representation of vowel length is continuous not discrete (Marslen-Wilson & Warren, 1988; Salverda, Dahan & McQueen, in press). Continuous representation of phonetic features allows probabilistic constraint satisfaction. Speech Perception and Word Recognition

  8. We need to look at the signal… Not just embedded words... Embeddings not the only source of [phonemic] ambiguity. • Place assimilation: • Eight Babies vs. Ape Babies • If /t/ assimilated place from “Babies” this becomes ambiguous (phonemicially) Subphonemic cues could help. • Phonological Reduction: • Reduced vowels all sound like /´/ • Original vowel may leave subphonemic trace. Speech Perception and Word Recognition

  9. A modest proposal Key properties of a combined system include: 1) Continuous sensitivity to fine-grained detail in the signal. 2) Information is retained and used to improve word recognition. Speech Perception and Word Recognition • What if the perceptual system could use fine-grained acoustic/phonetic cues to • Anticipate upcoming material • Resolve prior ambiguity • Show sensitivity to temporal organization.

  10. 100 B Discrimination % /p/ ID (%/pa/) 0 P B VOT P • Sharp identification of speech sounds on a continuum • Discrimination poor within a phonetic category Assessing continuous sensitivity Categorical Perception Speech Perception and Word Recognition Despite context dependency, system can extract phonemes (et al) remarkably discretely.

  11. Categorical Perception • Suggests gradations in acoustic properties are discarded in favor of discrete units. • But: the strong form of CP has never been shown. Evidence against it comes from: Speech Perception and Word Recognition • Discrimination Tasks • Pisoni and Tash (1974) Pisoni & Lazarus (1974) • Carney, Widin & Viemeister (1977) • Training • Samuel (1977) Pisoni, Aslin, Perey & Hennessy (1982) • Goodness Ratings • Miller (1997) Massaro & Cohen (1983)

  12. Categorical Perception Despite this, the notion of speech perception as discarding variability remains implicit in much of the literature. When subjects respond phonologically, performance looks categorical… But much of this work is characterized by explicit, metalinguistic tasks. Off-line measures. Non-word stimuli. Speech Perception and Word Recognition

  13. Assessing Continuous Sensitivity These measures don’t assess lexical activation or its timecourse. But Spoken Word Recognition happens… VERY fast. Without much conscious awareness Bit-by-bit (all the information is not available at once) Need an on-line measure… Speech Perception and Word Recognition

  14. Lexical Sensitivity to Acoustic Detail Andruski, Burton & Blumstein (1994) Misiurski, Blumstein, Rissman & Berman (in press) 3 Voiceless Tokens Peel: VOT=80 Peel1/3: VOT=53 Peel2/3: VOT=27 No difference between Peel and Peel1/3 Speech Perception and Word Recognition Fully voiceless tokens prime semantic associates better than 2/3 voiced tokens. Peel -> Banana >>>> Peel2/3 -> Banana 2/3 voiced stimuli prime competitor better than unmodified. Peach -> Sand <<<< Peach2/3 -> Sand

  15. Lexical Sensitivity to Acoustic Detail • Andruski, Burton & Blumstein (1994) • Misiurski, Blumstein, Rissman & Berman (in press) • However: • What is the timecourse of sensitivity? • Significant effects with an ISI of 50ms. • Sometimes effects can be seen at 250ms. Speech Perception and Word Recognition

  16. 2/3 voiced stimuli were close to category boundary. • Difference between 2 items does not provide evidence for gradiency. Priming for Competitor Andruski et al (schematic) Gradient Sensitivity “Categorical” Perception 25 50 80 VOT (ms) Lexical Sensitivity to Acoustic Detail Andruski, Burton & Blumstein (1994) Misiurski, Blumstein, Rissman & Berman (in press) What is the extent of the acoustic sensitivity? Speech Perception and Word Recognition

  17. Lexical Sensitivity to Acoustic Detail Gradiency: A monotonic relationship between lexical activation and one or more acoustic cues such that subphonemic changes in the acoustic cue result in corresponding changes in lexical activation. Gradient sensitivity to acoustic detail In short: patterns of lexical activation become a way of storing (or at least indicating) acoustic gradation. Moreover, if information is preserved in patterns of lexical activation, it must be retained long enough to be used. Need a measure sensitive to both acoustic detail and detailed temporal dynamics of lexical activation.

  18. What kind of measure (acoustic detail)? Use speech continua—more stimulus levels yields a better picture of relationship. • KlattWorks: generate synthetic continua from natural speech. • 9-step VOT continuum (0-40 ms) • 6 pairs of words. • beach/peach bale/pale bear/pear • bump/pump bomb/palm butter/putter • 6 fillers. • Lamp Leg Lock Ladder Lip Leaf • Shark Shell Shoe Ship Sheep Shirt Gradient sensitivity to acoustic detail

  19. How do we measure this (lexical activation)? How do we tap on-line recognition? With an on-line task: Eye-movements Subject hear spoken language and manipulate objects in a visual world. Visual world includes set of objects with interesting linguistic properties. A beach, a peach and some unrelated items. Eye-movements are monitored throughout task. Gradient sensitivity to acoustic detail

  20. How do we measure this (lexical activation)? • Why use eye-movements and visual world paradigm? • Natural Task • Eye-movements generated very fast (within 200ms of first bit of information). • Lots of eye movements for each token. • Eye movements time-locked to speech. At any given time, an eye movement can only reflect information the subject has heard (by then). • Subjects aren’t aware of eye-movements. • Evidence that fixation probability maps onto lexical activation (Allopenna, Magnuson & Tanenhaus, 1998) Gradient sensitivity to acoustic detail

  21. IR Headtracker Emitters Head-Tracker Cam Monitor Head 2 Eye cameras Computers connected via Ethernet Subject Computer Eyetracker Computer Head Mounted Eye Tracking Gradient sensitivity to acoustic detail

  22. Head Mounted Eye Tracking Three cameras. One measures head position (from IR sensors on monitor) The other two record eye position. 250 Hz realtime stream of fixation coordinates. Gradient sensitivity to acoustic detail Parsed into Saccades, Fixations, Blinks, etc… Head movement compensation. Output in ~screen coordinates.

  23. Experiment 1: Methods A moment to view the items Gradient sensitivity to acoustic detail

  24. Experiment 1: Methods 500 ms later Gradient sensitivity to acoustic detail

  25. Experiment 1: Methods Bear Gradient sensitivity to acoustic detail Repeat 1080 times…

  26. 1 0.9 0.8 0.7 0.6 0.5 0.4 0.3 0.2 0.1 0 0 5 10 15 20 25 30 35 40 Experiment 1: Identification Results High agreement across subjects and items for category boundary proportion /p/ Gradient sensitivity to acoustic detail B VOT (ms) P By subject:17.25 +/- 1.33ms By item: 17.24 +/- 1.24ms

  27. Trials 1 2 3 4 5 200 ms Time Experiment 1: Fixation Analysis Gradient sensitivity to acoustic detail Target =Bear Competitor =Pear Unrelated =Lamp, Ship

  28. 0.9 0.8 VOT=0 Response= VOT=40 Response= 0.7 0.6 0.5 0.4 Fixation proportion 0.3 0.2 0.1 0 0 400 800 1200 1600 2000 0 400 800 1200 1600 Time (ms) More looks to competitor than unrelated items Experiment 1: Fixation Results Gradient sensitivity to acoustic detail

  29. 1 0.9 0.8 ID Function after filtering Actual Data 0.7 0.6 0.5 0.4 0.3 0.2 0.1 0 0 5 10 15 20 25 30 35 40 Yields a “perfect” categorization function. Experiment 1: Fixation Analysis Gradient sensitivity to acoustic detail proportion /p/ B VOT (ms) P Trials with low-frequency response excluded.

  30. target Fixation proportion Fixation proportion time time Experiment 1: Fixation Analysis • Given that • the subject heard bear • clicked on “bear”… How often was the Subject looking at the “pear”? Categorical Results Gradient Effect Gradient sensitivity to acoustic detail target target competitor competitor competitor competitor

  31. 20 ms 25 ms 30 ms 10 ms 15 ms 35 ms 40 ms 0.16 0.14 0.12 0.1 0.08 0.06 0.04 0.02 0 0 400 800 1200 1600 0 400 800 1200 1600 2000 Experiment 1: Fixation Analysis Response= Response= VOT VOT 0 ms 5 ms Competitor Fixations Gradient sensitivity to acoustic detail Time since word onset (ms) Smaller effect on the amplitude of activation—more effect on the duration: Competitors stay active longer as VOT approaches the category boundary.

  32. 0.08 0.07 0.06 0.05 0.04 0.03 0.02 0 5 10 15 20 25 30 35 40 Area under the curve: Clear effects of VOT B: p=.026* P: p<.001*** Linear Trend B: p=.032* P: p<.001*** Experiment 1: Fixation Analysis Response= Response= Looks to Competitor Fixations Gradient sensitivity to acoustic detail Looks to Category Boundary VOT (ms)

  33. 0.08 0.07 0.06 0.05 0.04 0.03 0.02 0 5 10 15 20 25 30 35 40 Unambiguous Stimuli Only Clear effects of VOT B: p=.016* P: p<.001*** Linear Trend B: p=.012* P: p=.002** Experiment 1: Fixation Analysis Response= Response= Looks to Competitor Fixations Gradient sensitivity to acoustic detail Looks to Category Boundary VOT (ms)

  34. 0.11 0.1 0.09 Early (300-1100ms) 0.08 0.07 Late (1100-1900ms) 0.06 0.05 0.04 0.03 0.02 0.01 0 5 10 15 20 25 30 35 40 Main effect of VOT /b/: p=.015* /p/: p=.001*** Linear Trend for VOT /b/: p=.022* /p/: p=.009** No Interaction p>.1 Experiment 1: Effect of time Response= Response= Competitor Fixations Looks to Gradient sensitivity to acoustic detail Looks to Category Boundary VOT (ms)

  35. 0.11 0.1 0.09 Early (300-1100ms) 0.08 0.07 Late (1100-1900ms) 0.06 0.05 0.04 0.03 0.02 0.01 0 5 10 15 20 25 30 35 40 Unambiguous Stimuli Only Main effect of VOT /b/: p=.015* /p/: p=.001*** Linear Trend for VOT /b/: p=.022* /p/: p=.009** No Interaction p>.1 Experiment 1: Effect of time Response= Response= Competitor Fixations Looks to Gradient sensitivity to acoustic detail Looks to Category Boundary VOT (ms)

  36. Experiment 1: Summary Subphonemic acoustic differences in VOT have gradient effect on lexical activation. • Gradient effect of VOT on looks to the competitor. • Effect holds even for unambiguous stimuli. Gradient sensitivity to acoustic detail • Effect seems to be long-lasting. Conservative Test • Filter out “incorrect” responses. • Use unambiguous stimuli only. • VOT: “Categorical” phonetic dimension.

  37. Experiment 1: Objections You’ve probably already thought of these… 1080 is a lot of trials. Is this really tapping lexical activation or is it approaching an off-line “metalinguistic” task? Gradient sensitivity to acoustic detail Synthetic Speech—not very natural. Would effect hold up with natural tokens? VOT is not the only acoustic/phonetic dimension. Would effect hold up with other cues?

  38. P B Sh L Bear Experiment 2: A metalinguistic task Same Stimuli Same # of trials. 16 subjects Different task: Explicit phoneme decision on initial consonant Gradient sensitivity to acoustic detail

  39. 0.1 0.09 0.08 0.07 0.06 0.05 0.04 0.03 0.02 0.01 0 0 5 10 15 20 25 30 35 40 Experiment 2: Fixation Results Response=B Looks to B Response=P Looks to B Competitor Fixations Gradient sensitivity to acoustic detail Category Boundary VOT (ms) Gradient effects using the whole range of stimuli /b/: p<.001*** ptrend<.001*** /p/: p<.001*** ptrend=.021*

  40. 0.1 0.09 0.08 0.07 0.06 0.05 0.04 0.03 0.02 0.01 0 0 5 10 15 20 25 30 35 40 Experiment 2: Fixation Results Response=B Looks to B Response=P Looks to B Competitor Fixations Gradient sensitivity to acoustic detail Category Boundary VOT (ms) Smaller effects using “unambiguous” stimuli only. /b/: p=.013* ptrend=.026* /p/: p=.297 ptrend=.075

  41. 1 0.9 0.8 0.7 0.6 0.5 0.4 An even more explicit task (Ba/Pa continuum) yielded less sensitivity in ID curve. 0.3 0.2 0.1 Exp 1: Words proportion /p/ 0 Exp 2b: BP 0 5 10 15 20 25 30 35 40 B VOT (ms) P Experiment 2: Conclusions Reduced gradient sensitivity with more explicit task. Experiment 1 results not due to subjects adopting an explicit/metalinguistic strategy. Gradient sensitivity to acoustic detail

  42. Experiment 2: Conclusions Earlier I said: When subjects respond phonologically, performance looks categorical… This should be revised to: When subjects respond metalinguistically, performance looks categorical… Previous work overestimated categorical perception—CP is imposed on performance by the requirements of the task. Gradient sensitivity to acoustic detail

  43. Experiment 3: Natural Stimuli Are gradient effects of VOT an artifact of synthetic speech. Same Items, Taskas Experiment 1 17 Subjects Stimuli constructed from natural tokens with progressive cross-splicing. Gradient sensitivity to acoustic detail

  44. Experiment 3: Natural Stimuli Palm Bomb Gradient sensitivity to acoustic detail

  45. Experiment 3: Identification Results 1 0.9 Bale 0.8 Beach 0.7 Bear 0.6 Bomb Gradient sensitivity to acoustic detail 0.5 % /p/ Response Bump 0.4 Butter 0.3 0.2 0.1 0 0 5 10 15 20 25 30 35 40 VOT Normal looking identification functions. Wider variance in category boundaries between items.

  46. Experiment 3: Analysis of Fixations Category boundary computed from mouse-clicks for each subject, for each item. Tokens near these category boundaries excluded. Exact 5ms VOT steps could not be made with this method: VOT rounded to nearest 5ms bin. Gradient sensitivity to acoustic detail

  47. Driven by “bale” Experiment 3: Fixation Results Response= Response= 0.09 0.08 0.07 0.06 0.05 Competitor Fixations 0.04 Gradient sensitivity to acoustic detail Looks to 0.03 Looks to 0.02 0.01 0 0 10 20 30 40 VOT (ms) Gradient effects using the whole range of stimuli /b/: p=.043* ptrend=.055 /p/: p=.001** ptrend=.007**

  48. Experiment 3: Fixation Results Response= Response= 0.09 0.08 0.07 0.06 0.05 Competitor Fixations 0.04 Gradient sensitivity to acoustic detail Looks to 0.03 Looks to 0.02 0.01 0 0 10 20 30 40 VOT (ms) Unambiguous stimuli /b/: p<.001*** ptrend<.001*** /p/: p=.030 * ptrend=.018*

  49. Experiment 3: Conclusions Effect not due to synthetic speech. Gradient sensitivity to VOT replicates. Gradient sensitivity to acoustic detail

  50. Experiment 4: Other distinctions • Three R/L continua • Rake/Lake Race/Lace Ray/Lei • Six Fillers • Bees Beach Beak • Peas Peach Peak • (OK…actually another experiment) • F3 onset manipulated • Vowel held constant to reduce variability. • Same task. • Preliminary data (14 subjects). Gradient sensitivity to acoustic detail

More Related