1 / 30

CS 416 Artificial Intelligence

CS 416 Artificial Intelligence. Lecture 6 Informed Searches. Chess Match. Kasparov 1, Deep Junior 1, Draws 2. Compare two heuristics. Compare these two heuristics. h 2 is always better than h 1 for any node, n, h 2 (n) >= h 1 (n) h 2 dominates h 1

sumana
Download Presentation

CS 416 Artificial Intelligence

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. CS 416Artificial Intelligence Lecture 6 Informed Searches

  2. Chess Match • Kasparov 1, Deep Junior 1, Draws 2

  3. Compare two heuristics

  4. Compare these two heuristics • h2 is always better than h1 • for any node, n, h2(n) >= h1(n) • h2dominates h1 • Recall all nodes with f(n) < C* will be expanded? • This means all nodes, h(n) < C* - g(n), will be expanded • All nodes h2 expands will also be expanded by h1 and because h1 is smaller, others will be expanded as well

  5. Inventing admissible heuristic funcs • How can you create h(n)? • Simplify problem by reducing restrictions on actions • Allow 8-puzzle pieces to sit atop on another • Call this a relaxed problem • The cost of optimal solution to relaxed problem is admissible heuristic for original problem • It is at least as expensive for the original problem

  6. Examples of relaxed problems • A tile can move from square A to square B if • A is horizontally or vertically adjacent to B • and B is blank • A tile can move from A to B if A is adjacent to B (overlap) • A tile can move from A to B if B is blank (teleport) • A tile can move from A to B (teleport and overlap) • Solutions to these relaxed problems can be computed without search and therefore heuristic is easy to compute

  7. Multiple Heuristics • If multiple heuristics available: • h(n) = max {h1(n), h2(n), …, hm(n)}

  8. Use solution to subproblem as heuristic • What is optimal cost of solving some portion of original problem? • subproblem solution is heuristic of original problem

  9. Pattern Databases • Store optimal solutions to subproblems in database • We use an exhaustive search to solve every permutation of the 1,2,3,4 piece subproblem of the 8-puzzle • During solution of 8-puzzle, look up optimal cost to solve the 1,2,3,4 piece subproblem and use as heuristic

  10. Learning • Could also build pattern database while solving cases of the 8-puzzle • Must keep track of intermediate states and true final cost of solution • Inductive learning builds mapping of state -> cost • Because too many permutations of actual states • Construct important features to reduce size of space

  11. Local Search Algorithms andOptimization Problems

  12. Characterize Techniques • Uninformed Search • Looking for a solution where solution is a path from start to goal • At each intermediate point along a path, we have no prediction of value of path • Informed Search • Again, looking for a path from start to goal • This time we have insight regarding the value of intermediate solutions

  13. Now change things a bit • What if the path isn’t important, just the goal? • So the goal is unknown • The path to the goal need not be solved • Examples • What quantities of quarters, nickels, and dimes add up to $17.45 and minimizes the total number of coins • Is the price of Microsoft stock going up tomorrow?

  14. Local Search • Local search does not keep track of previous solutions • Instead it keeps track of current solution (current state) • Uses a method of generating alternative solution candidates • Advantages • Use a small amount of memory (usually constant amount) • They can find reasonable (note we aren’t saying optimal) solutions in infinite search spaces

  15. Optimization Problems • Objective Function • A function with vector inputs and scalar output • goal is to search through candidate input vectors in order to minimize or maximize objective function • Example • f (q, d, n) = 1,000,000 if q*0.25 + d*0.1 + n*0.05 != 17.45 = q + n + d otherwise • minimize f

  16. Search Space • The realm of feasible input vectors • Also called state space landscape • Usually described by • number of dimensions (3 for our change example) • domain of dimensions (q is discrete from 0 to 69…) • nature of relationship between input vector and objective function output • no relationship • smoothly varying • discontinuities

  17. Search Space • Looking for global maximum (or minimum)

  18. Hill Climbing • Also called Greedy Search • Select a starting point and set current • evaluate (current) • loop do • neighbor = highest value successor of current • if evaluate (neighbor) <= evaluate (current) • return current • else current = neighbor

  19. Hill climbing gets stuck • Hiking metaphor (you are wearing glasses that limit your vision to 10 feet) • Local maxima • Ridges • Plateau • why is this a problem?

  20. Hill Climbing Gadgets • Variants on hill climbing play special roles • stochastic hill climbing • don’t always choose the best successor • first-choice hill climbing • pick the first good successor you find • useful if number of successors is large • random restart • follow steepest ascent from multiple starting states • probability of finding global max increases with number of starts

  21. Hill Climbing Usefulness • It Depends • Shape of state space greatly influences hill climbing • local maxima are the Achilles heel • what is cost of evaluation? • what is cost of finding a random starting location?

  22. Simulated Annealing • A term borrowed from metalworking • We want metal molecules to find a stable location relative to neighbors • heating causes metal molecules to move around and to take on undesirable locations • during cooling, molecules reduce their movement and settle into a more stable position • annealing is process of heating metal and letting it cool slowly to lock in the stable locations of the molecules

  23. Simulated Annealing • Be the Ball • You have a wrinkled sheet of metal • Place a BB on the sheet and what happens? • BB rolls downhill • BB stops at bottom of hill (local or global min?) • BB momentum carries it out of hill into another (local or global) • By shaking metal sheet, your are adding energy (heat) • How hard do you shake?

  24. Our Simulated Annealing Algorithm • You’re not being the ball, Danny (Caddy Shack) • Gravity is great because it tells the ball which way is downhill at all times • We don’t have gravity, so how do we decide a successor state? • Randomness • aka Monte Carlo • aka Stochastic

  25. Algorithm Outline • Select some initial guess of evaluation function parameters: x0 • Evaluate evaluation function, E(x0)=v • Compute a random displacement, x’0 • The Monte Carlo event • Evaluate E(x’0) = v’ • If v’ < v; set new state, x1 = x’0 • Else set x1 = x’0 with Prob(E,T) • This is the Metropolis step • Repeat with updated state and temp

  26. Metropolis Step • We approximate nature’s alignment of molecules by allowing uphill transitions with probability • Prob (in energy state E) ~ exp (-E/kT) • Boltzmann Probabilty Distribution • Even when T is small, there is still a chance in high energy state • Prob (transferring from E1 to E2) = exp (-(E2-E1) / kT) • Metropolis Step • if E2 < E1, prob () is greater than 1 • if E2 > E1, we may transfer to higher energy state • The rate at which T is decreased and the amount it is decreased is prescribed by an annealing schedule

  27. What have we got? • Always move downhill if possible • Sometimes go uphill • More likely at start when T is high • Optimality guaranteed with slow annealing schedule • No need for smooth search space • We do not need to know what nearby successor is • Can be discrete search space • Traveling salesman problem

  28. Local Beam Search • Keep more previous states in memory • Simulated Annealing just kept one previous state in memory • This search keeps k states in memory Generate k initial states if any state is a goal, terminate else, generate all successors and select best k repeat

  29. Isn’t this steepest ascent in parallel? • Information is shared between k search points • Each k state generates successors • Best k successors are selected • Some search points may contribute none to best successors • One search point may contribute all k successors • If executed in parallel, no search points would be terminated like this • Draw it

  30. Beam Search • Premature termination of search paths? • Stochastic beam search • Choose k successors from total list at random

More Related