500 likes | 612 Views
This study explores how individuals construct coherent narratives from evidence during legal deliberations, focusing on the roles of reinforcement, reinterpretation, and discounting of evidence. It employs an Agent-Based Model (ABM) to simulate decision-making processes in a courtroom setting, analyzing how the interplay of various evidence agents leads to the formation of a "sufficiently strong" verdict. The findings highlight the importance of narrative consistency and coherence in enhancing confidence in judicial outcomes, demonstrating that different interpretations of the same evidence can lead to diverse verdicts. ###
E N D
Seeking Consistent Stories By Reinterpreting or Discounting Evidence: An Agent-Based Model Decision Consortium May Conference 13-Mar-2009
Agenda The Phenomenon Agent-Based Model (ABM) Primer The Model Sample Run Experiments
Real-Life Scenario: Bench Trial Prosecution (p) Defense Lawyer (d) “Guilty!” “Innocent!”
Sequential Evidence – What’s Normative Official Deliberation Story/ Verdict Evidence 1 Evidence 2 Evidence 3 Evidence N …
Empirical Literature on JDM • Form “coherent” stories that support option/verdict (Pennington & Hastie, 1988) • Confidence in option/verdict increases with “coherence” (Glockner et al., under review) • Decision threshold: “sufficiently strong” (supported by many consistent evidence) or “sufficiently stronger” than other stories (review by Hastie, 1993 • Narrative coherence: • consistency • causality • completeness • “Consistency” aspect of “good” stories • consistency between evidences in a story • consistency of evidence with favored story
Example Case (Pennington & Hastie, 1988) Scenario: Defendant Frank Johnson stabbed and killed Alan Caldwell • Evidence := facts or arguments given in support of a story/verdict • Facts— “Johnson took knife with him” ; “Johnson pulled out his knife” • Arguments—”Johnson pulled out knife because he wanted revenge” vs. “Johson pulled out knife because he was afraid” • Story := set of evidence supporting a given verdict • Same evidence can be framed to support multiple verdicts/stories!
Sequential Evidence – More Descriptive Official Deliberation Story/ Verdict Evidence 1 Evidence 2 Evidence 3 Evidence N Premature Story/ Verdict @ Evid n < N Compare, Deliberate, Interpret … (Brownstein, 2003; Russo et al., 2000)
People don’t just take evid @ face value, but are selective! Possible Reactions to New Evid in Light of Old: “reinforce” each other “reinterpret” less plausible one (Russo et al., etc.); e.g., misremember the info “discount” less plausible one (Winston) actively “seek” more evidence (not modeled here) Existing evidence A: “Johnson was not carrying a knife.” New evidence B1: “Johnson is nonviolent.” Inconsistent new evid B2: “Johnson pulled a knife.” Reinterpret B2: “Johnson grabbed a knife from Caldwell.” (i.e., explain it was Caldwell’s knife, not Johnson’s) Discount B2: “Witness must be mistaken.” Judge asks layers follow-up questions How People Deal with Incoming Evidence
Motivations • Study emergence of consistent stories via reinforcement, reinterpretation, and discounting mechanisms • Process, functions, consequences • Adaptive? • aid consistency? • speed-accuracy tradeoff? avoid indecisiveness • increases convergence rates? • order effects hurt accuracy?
Agenda The Phenomenon Agent-Based Model (ABM) Primer [1:43] The Model Sample Run Experiments
What is Agent-Based Modeling? A A A A A A • agents + interactions^ • start simple; build up^ • Key terminology • agents • system • dynamics • agent births and deaths • interactions/competitions • parameters A System of Agents
Symbiotic relationship: Behavioral Experiments ABM (Contributions ABMs Can Make) Input Test description: informs base assumptions understanding: study processes in detail parsimony: demo emergence of seemingly complex phenomen from small set of simple rules predictions: new observations/predictions
(Contrast with Bayesian & Algebraic Models) conglomerate of all previous evidence new evidence • Algebraic (additive), Bayesian (multiplicative): • “single meter” of overall plausibility • ABM allows: • revisiting and reconsidering previously-processed evidence • interaction/competition between individual evidences (not just conglomerate)
Contrast with Story Model & ECHO • Explanatory Coherence Model := Thagard’s Theory of Explanatory Coherence (TEC) + Story Model • Only implemented discounting, not reinterpretation • Local evidence-agent-level consistency, as opposed to global story-level consistency • Unlike previous ABMs, model agents within individual as system
Agenda The Phenomenon Agent-Based Model (ABM) Primer The Model [1:46] Sample Run Experiments
Goal • Model consistency-seeking process in story formation • Present evidence-agents to judge-system • Judge-system compares evidence-agents => keep, reinterpret, or discount evidence (agents “interact” & “compete”) • Until sufficiently strong/stronger story emerges
Agents Evid 1 Evid 1 Evid 2 Verdict (“G”,”N”,…) G N N Abstract features (binary) x x x x x y Plausibility index (0%-100%) 34% 14% 89% • Evidence-agents, composed of “features” • Operationalize consistency /b/ agents: • “similarity” in abstract features (Axelrod’s culture ABM, 1997) • “inverse Hamming distance” := % feature matches …
System & Agent Births • Judge-system represents judge’s mind • Evidence presentation = “agent birth”: • Initialization of N0 Agents: • Set up randomly-generated agents, OR… • uniform distributions—even for plausibility index due to prior beliefs (Kunda, 1990) or knowledge (Klein, 1993) • User-specified
results\sysout_081118_0455.log Printing 8 agents: 01 02 03 04 05 06 07 08 N G I G N I N N y y y y y y y y y y y y y y y y 100% 100% 100% 100% 100% 100% 100% 80% Dead/Rejected Evidence--Printing 9 agents: 01 02 03 04 05 06 07 08 09 I G G I G I N N N y y x x y x x x y y x x x y x x y y 00% 00% 00% 00% 00% 00% 00% 00% 00% To keep track of evidence-agents and their order of presentation: lists of agents in order of birth/presentation to the system. latest-born agent always appears at the end of a list. no geographical “topology” (Topology)
Stories Evid 1 Evid 2 Evid 1 Evid 3 N N G N y x x x y y x x 14% 34% 51% 89% Story promoting “Innocent” verdict Story promoting “Guilty” verdict strength = 99% / 3 = 33% strength = 89% := {evidence-agents supporting a verdict} • Consistency = inverse Hamming distance amongst evidence (:= SevidencePairs # feature matches / # evidencePairs / F ) • Plausibility Index (“Strength”) = average of plausibility indices
Births & Interactions • If time period t = k * I, where k is some constant, birth new agent. • Random selection of agent to compare with newest-born. • Interaction. Compare agents and compute consistency. Depending upon consistency [next slide]: • “reward consistency” • “increase consistency” • or “punish consistency” Agents with plausibility = 0% => death & removal from system • Gather stories in system. Check strengths. “Winning story found” if 1 ! story with strength >= S and/or |strength-strength | >= Sd for all competing stories; stop run early.
Possible Reactions to New Evid : Completely consistent (e.g., 100% features match) => => both are “winners” => “reinforce” both; “reward consistency” Inconsistent… but salvageable (50% features match) => “reinterpret” less plausible “loser” (by plausibility); “increaseinconsistency” not salvageable (0% features match) => “discount” “loser”; “punish inconsistency” => plaus(Evid1) > plaus(Evid2) => Evid2 “loser” Operationalizing Consistency--Examples Evid 2 Evid 1 Evid 2 Evid 1 Evid 2 Evid 1 Evid 1 Evid 2 Evid 1 Evid 2 Evid 1 Evid 2 x x x x x x y x y x x x y x y y y y y y y y x y 44% 51% 34% 24% 51% 61% 51% 34% 51% 51% 34% 34%
Agenda The Phenomenon Agent-Based Model (ABM) Primer The Model Sample Run [1:54] Experiments
Sample Output^ • Live Evidence--Printing 8 agents: • 01 02 03 04 05 06 07 08 • N G I G N I N N • y y y y y y y y • y y y y y y y y • 100% 100% 100% 100% 100% 100% 100% 80% • Dead/Rejected Evidence--Printing 9 agents: • 01 02 03 04 05 06 07 08 09 • I G G I G I N N N • y y x x y x x x y • y x x x y x x y y • 00% 00% 00% 00% 00% 00% 00% 00% 00% • 3 stories found: • -Verdict N supported by 4 evidence, with 1.00 consistency => 0.48 strength • -Verdict G supported by 2 evidence, with 1.00 consistency => 0.25 strength • -Verdict I supported by 2 evidence, with 1.00 consistency => 0.25 strength • *** Found winning story! Verdict N supported by 4 evidence, with 1.00 consistency => 0.48 strength
Judge-System can get Stuck… results\sysout_STUCK.log Live Evidence--Printing 17 agents: 01 02 03 04 05 06 07 08 09 10 11 12 13 14 15 16 17 N G G I G I I I I N I G G N N N G x x x x x x x x x x x x x x x x x y y y x y y y y y y y y y y y y y 100% 100% 100% 90% 100% 100% 100% 100% 100% 100% 84% 100% 79% 100% 100% 98% 100% Dead/Rejected Evidence--Printing 10 agents: 01 02 03 04 05 06 07 08 09 10 I I I I N I I G N G y x y y y y y y y y y x x x x x x y x x 00% 00% 00% 00% 00% 00% 00% 00% 00% 00% 3 stories found: -Verdict N supported by 5 evidence, with 1.00 consistency => 0.29 strength -Verdict G supported by 6 evidence, with 1.00 consistency => 0.34 strength -Verdict I supported by 6 evidence, with 1.00 consistency => 0.34 strength No winning story found.
Agenda The Phenomenon Agent-Based Model (ABM) Primer The Model Sample Run Experiments [1:56]
“control group” vary IV => various “treatment groups” Compare and contrast control group and treatment groups “base case” vary parameter => various other cases Compare and contrast base case and other cases Controlled Experiments in Simulation Adv: Don’t have to worry about assumptions b/c study inter-group differences rather than absolute outputs
5 Experiments • Experiment 1: Emergence of consistency Decision Speed • Experiment 2: Speedup • Experiment 3: Accuracy tradeoff Decision Accuracy—Order Effects • Experiment 4: Emergence of order effects • Experiment 5: Extending deliberation
Obtaining Consistent Stories – Q1 Q1: Evidence-level consistency sufficient? Which of the 3 mechanisms? Implementation: No stopping rules DV: Consistency of stories
Obtaining Consistent Stories – Q1 Results Reinterpret > Discount > Reinforcement
Speed-Accuracy Tradeoff – Q2 Q2: Reinterpretation & Discounting increase speed? Prediction: Reinterpretation & Discounting allow faster convergence DV: Time to Converge, Max nEvid (Max Consistency) Implementation: Both rules; all cases have reinforce
Speed-Accuracy Tradeoff – Q2 Results Results of 10 Runs—Time to First Convergence, Maximum Number of Evidence, Maximum Story Consistency
Speed-Accuracy Tradeoff – Q2 Results Medians of 10 Runs—Time to First Convergence, Maximum Number of Evidence, Maximum Story Consistency Reinterpret > Discount > Reinforcement only
Speed-Accuracy Tradeoff – Q3 • Q3: What would happen if allow process to continue even after having found winner? Any point to "holding off judgment" until all evidence presented? • DV: Which story wins? (Strength) • Prediction: Leader will only be strengthened; competing stories never get a foothold. • Implementation: Allow system to continue running even if found winner
Speed-Accuracy Tradeoff – Q3 Results^ Figure 2. Example run where winner switches • Over 20 runs, # runs same winner:# runs different winner = 15:1 • => Good heuristic to stop deliberation, for less time & effort
Order Effects – Q4 p d d d p d p d p d Heuristic may be ok only if randomized evidence…what if biased evidence? Q4: Is there an Order Effect? Took {20 randomly-generated evidence} and then “doctored” it; % D win = “accuracy” IV: Presentation order--P goes first, followed by D vs. interwoven evidence Prediction: earlier, weaker side (e.g., P) beats out later, stronger side (e.g., D); "Accuracy" of D…P… > PDPDPD… > P…D… DV: Time to Converge, (Projected) Winner
Order Effects – Q4 Results Figure 3. Sample random presentation order run.
Order Effects – Q4 Results Figure 4. Sample D…P… biased presentation order run.
Order Effects – Q4 Results Figure 5. Sample P…D… biased presentation order run.
Order Effects – Q4 Results Table 4. Results of 10 Runs Varying Presentation Order of Evidence • Strong primacy effect • Exper3 conclusion no longer holds; longer deliberation DOES help!
(Increasing Deliberation – Q5) Q5: Can deliberating more often between births reduce order effects (i.e., increase "accuracy")? Implementation: Use P…D… model from Exper4 IV: Varied I (I = 0 => wait till end to deliberate) DV: % runs that P wins
(Increasing Deliberation – Q5 Results) Too much lag time during trials can be detrimental!
Summary of Key Findings • Why reinterpret and discount evidence? • Maximizes consistency (Experiment 1) • Hastens convergence on decision (Experiment 2) Reinterpret > Discount > Reinforcement • Speed-accuracy tradeoff? Yes… • Accuracy ok if evidence balanced (Experiment 3) • Not ok/primacy effect if biased (Experiment 4) • => Important to interweave evidence, like in real trials! • Can reduce primacy effect by decreasing premature deliberation (Experiment 5) • All achieved by modeling consistency @ evidence level, not story level! more parsimonious & realistic(?)
Other Simplifying Assumptions • Features • Any combination of abstract features can be framed (e.g., by the lawyers) to support a verdict. All evidence are automatically categorizable into a verdict. • Interactions • Interactions only take place between the newest agent and another agent. • Computing Consistency • Verdict feature neither considered nor included in interactions • Comparing Stories • judge forms 1 story / verdict • If only 1 story is found at time t, use the "sufficiently strong" rule as opposed to the "sufficiently stronger" rule. (i.e., no “holding out” by judge) • Which layer drives consistency-seeking? • consistency at the individual agent interaction level, AND/OR • consistency at the story level
(Future Expansions) • Q6: What happens when introduce bias toward certain verdicts? • Prediction: Verdict-driven process takes less time to converge • DV: Time to Converge (Consistency of Stories) • Implementation: Add favoredVerdict • Q7: In general, what conditions lead to indecision?