Heuristics slides 321

26
Heuristics Some further elaborations of the art of heuristics and examples.

Transcript of Heuristics slides 321

Page 1: Heuristics slides 321

Heuristics

Some further elaborations of the art of heuristics and examples.

Page 2: Heuristics slides 321

Goodness of heuristics

If a heuristic is perfect, search work is proportional to solution length

S = O(b*d) b is average branching, d depth of solution

If h1 and h2 are two heuristics, and h1 < h2 everywhere,

then A*(h2) will expand no more nodes than A*(h1)

If an heuristic never overestimates more than N % of the least cost, then the found solution is not more than N% over the optimal solution.

h()=0 is an admissible trivial heuristic, worst of all

In theory, we can always make a perfect heuristics, if we perform a

full breadth first search from each node, but that would be pointless.

Page 3: Heuristics slides 321

Example of good heuristics for the 8 puzzle

Transparency shows the heuristic

f(n)=g(n) + h(n)

h(n)=# misplaced tiles

On an A* search for the 8 - puzzle

Page 4: Heuristics slides 321
Page 5: Heuristics slides 321

Graph Search

%% Original version

function GRAPH-SEARCH(problem,fringe) returns a solution, or failure

closed <- an empty set

fringe <- INSERT(MAKE-NODE(INITIAL-STATE[problem]),fringe)

loop do

if EMPTY?(fringe) then return failure

node <- REMOVE-FIRST(fringe)

if GOAL-TEST[problem](STATE[node]) then return SOLUTION(node)

if STATE[node] is not in closed then

add STATE[node] to closed

fringe <- INSERT-ALL(EXPAND(node,problem),fringe)

Page 6: Heuristics slides 321

Heuristic Best First Search %% f [problem] is the heuristic selection function of the problem

function BEST-FIRST-SEARCH([problem])returns a solution,or failureOPEN <- an empty set // P1CLOSED <- an empty set // P2OPEN <- INSERT(MAKE-NODE(INITIAL-STATE[problem]),OPEN) // P3repeat if EMPTY?(OPEN) then return failure //P4 best <- the lowest f-valued node on OPEN // P5 remove best from OPEN // P6 if GOAL-TEST[problem](STATE[best]) then return SOLUTION(best) // P7 for all successors M of best if STATE[M] is not in CLOSED then // P8 OPEN <-INSERT(successor,OPEN) // P9 add STATE[best] to CLOSED // P10

Page 7: Heuristics slides 321

Heuristic Best First Search(A*)Java Pseudocode

// Instantiating OPEN, CLOSED OPEN = new Vector<Node>(); // P1 CLOSED = new Vector<Node>(); // P2

// Placing initial node on OPEN OPEN.add(0, initialnode); // P3

// After initial phase,we enter the main loop // of the A* algorithm

while (true) { // Check if OPEN is empty if (OPEN.size() == 0) { // P4

System.out.println("Failure :"); return;

// Locate next node on OPEN with heuristic lowIndex = 0; // P5 low = OPEN.elementAt(0).f; for (int i = 0; i < OPEN.size(); i++) { number = OPEN.elementAt(i).f;

if (number < low) { lowIndex = i; low = number;

// Move selected node from OPEN to n // P6 n = OPEN.elementAt(lowIndex); OPEN.removeElement(n);// Successful exit if n is goal node // P8 if (n.equals(goalnode)) return; // Retrieve all possible successors of n M = n.successors();//Compute f-,g- and h-value for each successor for (int i = 0; i < M.size(); i++) {

Node s = M.elementAt(i); s.g = n.g + s.cost; s.h = s.estimate(goalnode); s.f = s.g + s.h;

// Augmenting OPEN with suitable nodes from M for (int i = 0; i < M.size(); i++) // Insert node into OPEN if not on CLOSED// P9

if (!(on CLOSED)) OPEN.add(0, M.elementAt(i));

// Insert n into CLOSED CLOSED.add(0, n); // P10

Page 8: Heuristics slides 321

AStarJava Code

See exercise 7

http://www.idi.ntnu.no/emner/tdt4136/PRO/Astar.java

Page 9: Heuristics slides 321

ExampleMouse King Problem

(1,5) (5,5) ___________ | | | | | | | | |X| | | | | |X| | | | | |X| | | |M| |X| |C| ----------- (1,1) (5,1)

There is a 5X5 board.At (1,1) there is a mouse M which can move as a king on a chess board.The target is a cheese C at (5,1).

There is however a barrier XXXX at (3,1-4) which the mouse cannot gothrough, but the mouse heuristics is to ignore this.

Page 10: Heuristics slides 321

Heuristics for Mouse Kingpublic class MouseKingState extends State { public int[] value; public MouseKingState(int[] v) {value = v;} public boolean equals(State state) {…} public String toString() {…} public Vector<State> successors() {…}

public int estimate(State goal) { MouseKingState goalstate = (MouseKingState)goal; int[] goalarray = goalstate.value; int dx = Math.abs(goalarray[0] - value[0]);

int dy = Math.abs(goalarray[1] - value[1]); return Math.max(dx, dy);

}

} // End class MouseKingState

Page 11: Heuristics slides 321

Behaviour of Mouse King

The mouse will expand the following nodes: 1,1 2,1 2,2 2,3 1,2 1,3 2,4 1,4 2,5 3,5 4,4 4,3 4,2 5,1

Solution Path:start,f,g,h(1,1),4,0,4(2,2),4,1,3(2,3),5,2,3(2,4),6,3,3(3,5),8,4,4(4,4),8,5,3(4,3),8,6,2(4,2),8,7,1(5,1),8,8,0

(1,5) (5,5) _______________ | | 9|10| | | | 8| 7| |11| | | 6| 4| |12| | | 5| 3| |13| | | 1| 2| | |14| ---------------- (1,1) (5,1)

Order of expansion

(1,5) (5,5) _______________ | | |5 | | | | | 4| |6 | | | | 3| |7 | | | | 2| |8 | | | 1| | | |9 | ---------------- (1,1) (5,1)

Solution Path

Page 12: Heuristics slides 321

Perfect Heuristics Behaviour

If the heuristics had been perfect, the expansion of the nodes would have been equal to the solution path.

1,1 2,2 2,3 2,4 3,5 4,4 4,2 4,2 5,1 This means that a perfect heuristics ”encodes” all the relevant knowledge of the problem space.

Solution Path:start,f,g,h(1,1),8,0,8(2,2),8,1,7(2,3),8,2,6(2,4),8,3,5(3,5),8,4,4(4,4),8,5,3(4,3),8,6,2(4,2),8,7,1(5,1),8,8,0

(1,5) (5,5) _______________ | | |5 | | | | | 4| |6 | | | | 3| |7 | | | | 2| |8 | | | 1| | | |9 | ---------------- (1,1) (5,1)

Order of expansion

(1,5) (5,5) _______________ | | |5 | | | | | 4| |6 | | | | 3| |7 | | | | 2| |8 | | | 1| | | |9 | ---------------- (1,1) (5,1)

Solution Path

Page 13: Heuristics slides 321

Monotone (consistent) heuristicsA heuristic is monotone if the f-value is non-decreasing along any path from start to goal.

This is fulfilled if for every pair of nodes

n n’

f(n) <= f(n’) = g(n’) + h(n’)

= g(n) + cost(n,n’) + h(n’)

f(n) = g(n) + h(n)

Which gives the triangle inequality

h(n) <= cost(n.n’) + h(n’)

goalh(n)

h(n’)cost(n,n’)

Page 14: Heuristics slides 321

Properties of monotone heuristics

1) All monotone heuristics are admissible

2) Monotone heuristic is admissible at all nodes (if h(G)=0)

3) If a node is expanded using a monotone heuristic, A* has found the optimal route to that node

4) Therefore, there is no need to consider a node if it is already found.

5) If this is assumed, and the heuristic is monotone, the algorithm is still admissible

6) If the monotone assumption is not true, we risk a non optimal solution

7) However, most heuristics are monotone

Page 15: Heuristics slides 321

Some more notes on heuristics

If h1(n) and h2(n) are admissible heuristics, then the following are also admissible heuristics

• max(h1(n),h2(n))• α*h1(n) + β*h2(n) where α+β=1

Page 16: Heuristics slides 321

Monotone heuristic repair

Suppose a heuristic h is admissible but not consistent, i.e.

h(n) > c(n,n’)+h(n’), which means

f(n) > f(n’) (f is not monotone)

In that case, f’(n’) can be set to f(n) as a better heuristic,(higher but still underestimate), i.e. use

h’(n’) = max(h(n’),h(n)-c(n,n’))

Page 17: Heuristics slides 321

An example of a heuristicsConsider a knight (”horse”) on a chess board.It can move 2 squares in one direction and 1 to either side.The task is to get from one square to another in fewest possible steps (e.g. A1 to H8).

A proposed heuristics could be ManHattanDistance/2 Is it admissible ? Is it monotone ?

(Actually, it is not straightforward to find an heuristics that is both admissible and not monotone)

8 | | | | | | | | * | 7 | | | | | | | | | 6 | | | | | | | | | 5 | | | | | | | | | 4 | | | | | | | | | 3 | | | | | | | | | 2 | | | | | | | | | 1 | * | | | | | | | | A B C D E F G H

Page 18: Heuristics slides 321

Relaxed problemsMany heuristics can be found by using a relaxed (easier, simpler) model of the problem.

By definition, heuristics derived from relaxed models are underestimates of the cost of the original problem

For example, straight line distance presumes that we can move in straight lines.

For the 8-puzzle, the heuristics

W(n) = # misplaced tiles

would be exact if we could move tiles freely

The less relaxed, (and therefore better) heuristic

P(n) = distance from home ( Manhattan distance)

would allow tiles to be moved to an adjacent square even though there may already be a tile there.

Page 19: Heuristics slides 321

Generalized A*

f(n) = *g(n) + *h(n)

a==1 A*b= 0 Uniform cost

a = 0 Greedy searcha< 0, = 0 Depth first

a > Conservative (still admissible)a < Radical (not admissible)a , depending on g,h Dynamic

Page 20: Heuristics slides 321

Learning heuristics from experience

Where do heuristics come from ?

Heuristics can be learned as a computed (linear?) combination of features of the state. Example: 8-puzzle

Features:

x1(n) : number of misplaced tiles x2(n) : number of adjacent tiles that are also adjacent in the goal state

Procedure: make a run of searches from 100 random start states. Let h(ni) be the found minimal cost.

n1 h(n1) x1(n1) x2(n1) … …… …….. ………. n100 h(n100) x1(n100) x2(n100)

From these, use regression to estimate

h(n) = c1*x1(n) + c2*x2(n)

Page 21: Heuristics slides 321

Learning heuristics from experience (II)

Suppose the problem is harder than the heuristic (h1) indicates but that the hardness is assumed to be uniform over the state space. Then , it is an estimate to let an improved heuristics h2(x) = *h1(x).

Problem: move a piece from S to G using ChessKing heuristics h1(x) = # horisontal/vertical/diagonal moves.

h1(S)=7 h1(n) = 4

Assume problem is actual harder (in effect Manhattan disance h2(x), but we dont’ know that). It means g(n) = 6 We then estimate = g(n)/(h1(s)-h1(n))

h2(n) = g(n)/(h1(s)-h1(n)) * h1(n)

= 6/(7 – 4) *4 = 8 (correct)

| S | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | n | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | G|

Page 22: Heuristics slides 321

Learning heuristics from experience (III)

Suppose the problem is easier than the heuristic (h1) indicates but that the easiness is assumed to be uniform over the state space. Then , it is an estimate to let an improved heuristics h2(x) = *h1(x).

Problem: move a piece from S to G using Manhattan heuristics h1(x) = # horisontal/vertical moves

h1(S)=14 h1(n) = 8

Assume problem is actual easier (in effect Chess King disance h2(x), but we dont’ know that). It means g(n) = 3 We then estimate = g(n)/(h1(s)-h1(n))

h2(n) = g(n)/(h1(s)-h1(n)) * h1(n)

= 3/(14 – 8) *8 = 4 (correct)

| S | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | n | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | G|

Page 23: Heuristics slides 321

Practical example(Bus world scenario)

Find an optimal route from one place to

another by public transport

Nodes: Bus passing events

A bus route passes a station at a time

Actions: - enter bus

- leave bus

- wait

Page 24: Heuristics slides 321

Bus scenario search spaceSearch space space(2) X time(1)

Time

Space

Space

Bus 3

Bus 5

wait

Page 25: Heuristics slides 321

Heuristics for bus route plannerwhich route is best ?

K3

K1

K2Equivalent transfer

T2

T1

T3 = Z

T0 = A

N # bus transfers

A wait time 1. departure

Z wait time before arrival

T sum transfer waiting time

K sum driving time

Page 26: Heuristics slides 321

Planner discussion1. (T1+T2) = T critical if rain

2. If Z is to be minimised, must search backwards

3. Many equivalent transfers (same T and K)

4. In practice, A* is problematic

5. Waiting time A maybe unimportant

Solution: Relaxation

a) Find trasees independent of time

b) Eliminate equivalent transfers

c) For each trasee, find best route plan

d) Keep the best of these solutions