1 / 18

Assessing the pragmatics of experiments with crowdsourcing: The case of scalar implicature

Assessing the pragmatics of experiments with crowdsourcing: The case of scalar implicature. Pranav Anand , Caroline Andrews, Matthew Wagers. University of California, Santa Cruz. Experiments & Pragmatic Processing. Case Study: (Embedded) Implicatures.

miloa
Download Presentation

Assessing the pragmatics of experiments with crowdsourcing: The case of scalar implicature

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Assessing the pragmatics of experiments with crowdsourcing: The case of scalar implicature PranavAnand, Caroline Andrews, Matthew Wagers University of California, Santa Cruz

  2. Experiments & Pragmatic Processing Case Study: (Embedded) Implicatures Each of the critics reviewed some of the movies. but not all ? Depending on the study: - no evidence of EI’s – evidence for EI’s, with different response choices Worry: Are we adequately testing the influence of methodologies on our data? Worry: How much do methodologies themselves influence judgements? Previous Limitation: Lack of Subjects and Money Crowd-sourcing addresses both problems

  3. Pragmatics of Experimental Situations Evaluation Apprehension – subjects know they are being judged Teleological Curiosity - Subjects hypothesizing “expected” behavior, matching an ideal Worry: How much do methodologies themselves influence judgements? The experiment itself is part of the pragmatic context See Rosenthal & Rosnow. (1975) The Volunteer Subject.

  4. Elements of Experimental Context Protocol – Social Context / Task Specification Response Structure – Response choices available to the subject e.g. True / False, Yes / No, 1-7 scale Prompt – the Question directions for the Response Structure Immediate Linguistic/Visual Context Our Goal: Explore variations of these elements in a systematic way Worry: How much do methodologies themselves influence judgements?

  5. Experimental Design Is this an accurate description? Some of the spices have red lids. Linguistic Contexts – All Relevant, All Irrelevant, No Context Protocol Experimental – normal experiment instructions Annotation – checking the work of unaffiliated annotators 4 Implicature Targets, 6 Some/All Controls, 20 Fillers

  6. Experiment 1:Social Context Focus on Protocol Annotation vs Experiment Accuracy Prompt - “Is this an accurate description?” Response Categories - Yes, No, Don’t Know Population: Undergraduates

  7. Experiment 1:Social Context Finding: Social context even when linguistic context does not. Linguistic Context: No Effect

  8. Experiment 1:Social Context Finding: Social context even when linguistic context does not. Lower SI rate for Annotation (p<0.05)

  9. Experiment 2Prompt Type Accuracy Prompt - “Is this an accurate description?” Response Categories - Yes, No, Don’t Know Informativity Prompt - “How Informative is this sentence?” Response Categories - Not Informative Enough Informative Enough Too Much Information False Population: Mechanical Turk Workers Systematic Debriefing Survey

  10. Experiment 2Prompt Type Effect for Prompt

  11. Experiment 2Prompt Type Effect for Prompt (p<0.001) Effect for Context (p<0.001)

  12. Experiment 2Prompt Type Effect for Prompt (p<0.001) Effect for Context (p<0.001) Weak Interaction: Prompt x Context (p<0.06)

  13. Experiment 2Prompt Type No Effect for Protocol

  14. Experiment 2Prompt Type Low SI rates overall But the debriefing survey indicates that (roughly) 70% of participants were aware of some/all contrast

  15. Populations Turkers – More sensitive to Linguistic Context Less sensitive to changes in changes in social context/ evaluation apprehension Undergraduates – More sensitive to Protocol

  16. Take Home Points • Methodological variables should be explored alongside conventional linguistic variables • Ideal: models of these processes (cf. Schutze 1996) • Crowdsourcing allows for cheap/fast exploration of parameter spaces • New Normal: Don’t guess, test. • Controls, norming, confounding … all testable online

  17. A potential check on exuberance • Undergraduates may be WEIRD*, but crowdsourcing engenders its own weirdness • High evaluation apprehension • Uncontrolled backgrounds, skillsets, focus levels • Unknown motivations • Ignorance does not necessarily mean diversity • This requires study if we rely on such participants more * Heinrich et al. (2010) The Weirdest People in the World? BBS

  18. Acknowledgments Thanks Jaye Padgett and to the attendees of two Semantics Lab presentations and the XPRAG conference for their comments, to the HUGRA committee for their generous award and support, and thanks to Rosie Wilson-Briggs for stimuli construction.

More Related