210 likes | 281 Views
This paper delves into the concept of the interviewer fallacy, the reluctance to create subsets of judgments that deviate too much from the expected distribution. Analyzing data from 10 years of MBA interviews, the study explores the impact of subsetting on ratings and examines alternative explanations like contrast effects and non-random sequencing. With a focus on average interview scores, the research challenges traditional beliefs and proposes new insights into the dynamics of interviewer judgments.
E N D
The Interviewer Fallacy:Evidence from 10 years of MBA interviews Photo not necessary Francesca Gino HBS Uri Simonsohn
Motivation • How is a journal editor like a venture capitalist? • Continuous flow of judgments “random” “daily” subsets. • Research question: Impact of subsetting? Narrow bracketing +Belief in law of small numbers interviewer fallacy Definition. Reluctance to create subsets of judgments that differ too much from expected distribution.
Paper in one slide • Data: 1-5 Rating of MBA interviewees • Handful per day. • corr[avg(so far), this interview]<0 • Ruled out alternatives: • Contrast effects • Non-random sequence
Data Description • A business school gave us data • 10 years: N=9,323, k=31 ***INTERRUPT THIS TALK TO COMMENT ON ANOTHER PROJECT*** False-Positive (PsychScience2011): “list all your variables” Naysayers: “love to, have too many” Authors of False-Positive: “really?” Uri: “watch me.”
Note: The .pdf weighs 13Kb. The Wharton logo from slide 1: 11kb A hardliner may say: Only reason to choose not to post is to hide information from readers.
Back to this talkData Description • A business school gave us data • 10 years: N=9,323, k=31* • Interviews per day M=4.5, SD=1.9 • Cluster SE [repeated measures] • Info on: • Applicant (e.g, GMAT scores, experience, race, gender) • Interviewer identity • Interview: time, date • Ratings (1-5 likert) • 5 subscores: communication, leader, etc. • Overall score (M=2.9, SD=0.9)
Would like to analyze like gambler fallacy • HHHHpr(T)↑ • Problem • Non-binary data • Covariates • Different interviewers
Instead: Scorek,i = OLS(average score so fari, covariates) k: Interviewee, 1 to N that day. i : Interviewer Prediction: <0
Effect Size • Average interview 1 point higher, • Equivalent to losing: • 40 GMAT points, or • 30 months of experience.
Alternative Explanations • Contrast effects • Non-random sequencing of interviews
Contrast vs. Interviewer Fallacy Two divergent predictions: • Same effect on the interview subscores? Explanation Prediction Contrast: yes, and stronger Int.Fallacy: no, or at least weaker. Data: • Every one of five subscores:n.s. • Average a-la Robyn Dawes:n.s. • Biggest point estimate, ¼ as big • one is >0
Contrast vs. Interviewer Fallacy Two divergent predictions: 2) Effect as end of day approaches. Explanation Prediction Contrast: weaker (arguably) Int.Fallacy: stronger (absolutely) Data: Estimate same regressions for: • last interview of day • 1 interview left • 2 interviews left
Alternative Explanations • Contrast effects • Non-random sequencing of interviews
If better candidates follow bad ones or vice-versa spurious finding. • Can we predict objective quality with average-interview-score-so-far? • Test: GMAT=OLS(avg.score) Job Experience = OLS(avg.score)
Possible Mechanisms • Gambler fallacy + confirmation bias • Mental Accounting • Accountability
A note on the internal validity of non-lab data • In the lab: hard to study interviewer fallacy • Participants could be learning about • Scale use • Distribution of underlying stimuli quality • Some psychological questions are better studied outside the lab. • This seems likes one of them.