1 / 31

Beliefs & Biases in Web Search

Beliefs & Biases in Web Search. Ryen White Microsoft Research ryenw@microsoft.com. Bias in IR and elsewhere. In IR, e.g., Domain bias – People prefer particular Web domains Rank bias – People favor high-ranked results Caption bias – People prefer captions with certain terms

wilbur
Download Presentation

Beliefs & Biases in Web Search

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Beliefs & Biases in Web Search Ryen White Microsoft Research ryenw@microsoft.com

  2. Bias in IR and elsewhere In IR, e.g., • Domain bias – People prefer particular Web domains • Rank bias – People favor high-ranked results • Caption bias – People prefer captions with certain terms In psychology, e.g., • Anchoring-and-adjustment, confirmation, availability, etc. • All impact user behavior • Opportunity to intersect psychology and IR

  3. Our Interest in Biases • Bias can be observed in IR in situations where searchersseek or are presented with information that significantly deviates from the truth

  4. Our Interest in Biases • Bias can be observed in IR in situations where searchersseek or are presented with information that significantly deviates from the truth User behavior Search engine behavior

  5. Example: Question : Can tea tree oil treat canker sores? The answer is yes, a health seeker may favor a particular out-come in light of their beliefs about the value of the oil, and seek or unconsciously prefer disaffirming information. • More on the “truth” later…

  6. Outline for Remainder of Talk • Initial Exploratory Questionnaire • Log Analysis • Labeling Content and Truth • Findings • Conclusions

  7. Initial Exploratory Questionnaire • Gain early insight into possible biases in search • Focus on Yes-No questions (answered with “Yes” or “No”) • Simplicity: Answers along single dimension (Yes  No) • Microsoft employees; recall recent Yes-No query (in last 2 weeks) • Asked about belief beforehand and afterwards • Multi-point scale: Yes / Lean Yes / Equal / Lean No / No • 200 respondents. Recalled questions such as: • “Does chocolate contain caffeine?”“Are shingles contagious?”

  8. Survey Results 48% believed lean yes,equal, or lean no compared to more than three quarters of respondents (77%) before searching Figure 1a provides evidence of a positive skew in respondents’ beliefs before they search. Specifically, 58% of respondents leaned toward yes and only 21% leaned toward no (the remaining 21% reported an equal belief in both outcomes). The fraction of searchers who believed yes after the search is more than double that of any other outcome, suggesting that respondents mostly shifted their prior beliefs from lean yes to yes.

  9. Survey Results • Two main findings: • 1. Respondents kept strongly-held beliefs (Yes-Yes and No-No) • 2. If Before = Equal, then 2x as likely to believe Yes after search Motivated us to:Further explore possible impact of biases on behavior and outcomes Post-search belief given Pre-search belief

  10. Log-Based Study of Yes-No Queries • Queries, clicks, and results from Bing logs (2 weeks) • Mined yes-no questions: start with “can”, “is”, “does”, etc. • Focused on health since it’s important and we could get truth • Randomly selected set of 1000 yes-no health questions • Each issued by at least 10 users, same top 10, same captions • Examples include: • “Is congestive heart failure a heart attack?”(answer = No)“Do food allergies make you tired?” (answer = Yes) we wanted to obtain answer labels from medical professionals, who were time constrained.

  11. Other Data Collected • Physician answers for the Yes-No questions Answer on four scale:yes,no,50/50,don’t know 50/50: there really was an equal split between yes and no more information be needed to provide an answer. don’t know: did not know the answer the query was not medical or was not a yes-no question • Yes-No Answer labels for captions and content of results

  12. Physician Answers • Two physicians reviewed the 1000 questions and gave answers The agreement matrix: • The Cohen’s free-marginal kappa (κ) inter-rater agreement • Cohen's kappa coefficient is a statistical measure of inter-rater agreement for qualitative items.

  13. Physician Answers (38.8%+31.5%)/(38.8%+8.2%+5.7%+31.5%)=83.4% (38.8%+8.2%)/(38.8%+8.2%+5.7%+31.5%)=55.82% (38.8%+5.7%)/ (38.8%+8.2%+5.7%+31.5%)=52.85% 55.82%*52.85%+(1-55.82%)*(1-52.85%)=50.33% K=(83.4%50.33%)/(150.33%)=0.668

  14. there is a high amount of uncertainty for the 50/50 and the don’t know categories, and they occur infrequently we focus on the cases where both judges were sufficiently confident to assign a rating of yes or no. 960*(38.8%+31.5%)=674 Physician Answers Distribution: 55% Yes and 45% No (used as TRUTH in our study)

  15. Answer Labeling Example Caption Labels Yes only • Captions and result content • Crowdsourced (Clickworker.com) • 3-5 judges/caption (consensus) • Task was to assign label of: • - Yes only- No only- Both (Yes and No)- Neither (not Yes and not No) • Agreement on 96% of captions • Performed similar labeling for each top 10 search results- Crowdsourced judges, agreement on 92% of pages No only Both Neither

  16. Answer Labeling Example Caption Labels Yes only • Captions and result content • Crowdsourced (Clickworker.com) • 3-5 judges/caption (consensus) • Task was to assign label of: • - Yes only- No only- Both (Yes and No)- Neither (not Yes and not No) • Agreement on 96% of captions • Performed similar labeling for each top 10 search results- Crowdsourced judges, agreement on 92% of pages No only Both Neither

  17. Answer Labeling Example Caption Labels Yes only • Captions and result content • Crowdsourced (Clickworker.com) • 3-5 judges/caption (consensus) • Task was to assign label of: • - Yes only- No only- Both (Yes and No)- Neither (not Yes and not No) • Agreement on 96% of captions • Performed similar labeling for each top 10 search results- Crowdsourced judges, agreement on 92% of pages No only Both Neither

  18. Answer Labeling Example Caption Labels Yes only • Captions and result content • Crowdsourced (Clickworker.com) • 3-5 judges/caption (consensus) • Task was to assign label of: • - Yes only- No only- Both (Yes and No)- Neither (not Yes and not No) • Agreement on 96% of captions • Performed similar labeling for each top 10 search results- Crowdsourced judges, agreement on 92% of pages No only Both Neither

  19. Using Physician Answers as Truth Used consensus physician answers as truth in three ways: • How closely does distribution of results match the truth? • How closely does interaction behavior match the truth? • How closely do answers that people reach match the truth? Bias = Distributions significantly differ from 55-45 Yes-No base rates

  20. Taking Stock of Our Data • We have: • 680 Yes-No health questions from search logs • Ground truth for each q via physicians’ consensus judgments • For each question we have: • HTML content of top 10 search results, plus: • Caption labels for Yes/No/Both/Neither • Result labels for Yes/No/Both/Neither • Clickthrough behavior from logs

  21. Analysis • Fourdirections for analysis: • Study ranking of results with Yes-No content • Study user behavior w.r.t. Yes-No content • Study answer accuracy for Yes-No questions • Study answer transitions for Yes-No questions

  22. Result Ranking • Volume of Yes-No content in the results • Percentage of captions or results with answer  More Yes content in top-10 than No content • Relative ranking of top Yes-No content when both in top 10 • Percentage of SERPs where top yes caption or result appears above (nearer the top of the ranking than) the top no  Yes content ranked above No more often (when both shown)

  23. User Behavior (Clickthrough rate) • Studied clickthrough rateson captions containing answers • Controlled for rank by just considering top result (r=1) SERP click likelihoods for different captions given variations in answer presence in SERPs/captions, and rank 3-4x as likely to click on captions with Yes content, even though TRUTH = 55% Yes / 45% No Just considering top search result

  24. User Behavior (Result skipping) Beyond clickthrough behavior, it is also worth considering the nature of the results that users skipped over prior to clicking on a particular caption. To study this in our context, we targeted SERPs with both yes-only and no-only answers in captions, and identified clicks on a caption with a yes or no answer where the user had skipped over another caption prior to clicking

  25. User Behavior (Result skipping) • Studied result skipping behavior • Frequency with which people skipped caption w/answer to click other caption • Distribution of clicks and skips by answer • Users more likely (5x) to skip No to click Yes than vice versa No Caption 1 No Caption 2 No Caption 3 Yes Caption 4

  26. Answer Accuracy • Examined accuracy of the top search result, as well as first click and last click in session • Findings show: 1. Top result accurate only 45% of time, less when truth is No 2. Users improve accuracy, but only slightly (limited by top 10)

  27. Answer Transitions • people reported being much more likely to search for confirmatory information than for information that challenged their hypothesis • Findings show: 1. no one transitioned from yes to no 2. confirmation was the primary motivation for pursuing information after the initial answer.

  28. Summary of Main Findings We observed: • Engines more likely to rank Yes above No, and return more Yes • People much more likely to click on Yes than No • Engine had wrong answer @ top rank for half of questions** Given that answer present at top position (~80% of queries) • Caveats: • Findings for our particular set of Yes-No health questions • More work needed to validate with other question sets, domains beyond health, etc.

  29. Discussion • Possible causes for observed bias include: • Search engines use behavior (hurt by common misconceptions) • Ranking algorithms consider query matche.g., for query: [can acid reflux cause back pain?]: Yes docs w/ “Acid reflux can cause back pain” better match (6 of 6 terms) than No docs w/ “Acid reflux cannot cause back pain” (5 of 6 terms) missing from query

  30. Conclusions • Studied potential bias in user behavior and outcomes • Showed effects on both from search engines • 2% of queries are Yes-No questions; Searchers want answers!! • To get users to accurate answers, engines should consider truth • Future directions: • Study availability of Yes-No content online; Move beyond Yes-No • Consider how truth should be determined and used in ranking • Follow-up user studies

  31. Thank you 2014/5/23

More Related