Informed Search
A* Algorithm
The key idea of A* algorithm comes from Uniform Cost Search. A* is same as UCS except that it has extra knowledge that can guide the search, which is known as a heuristic h(u). h(u) is an estimate of the cost of the shortest path from node u to a goal node w, and we have to compute h(u) also. For a time being, let us assume that we know h(u) for all the nodes, later we will discuss in detail about how to compute such heuristics. A* search evaluates nodes using an estimated cost of the optimal solution, therefore
ˆf(u)=ˆg(u)+h(u)Remember from previous lesson, we used the following notations:
Notation | Description |
---|---|
g(v) | minimum path cost from start to v. |
ˆg(v) | estimated minimum path cost from start to v. |
ˆgpop(v) | estimated minimum path cost from start to v when v is popped from F. |
Similar to the function g, we’ll use the function f, denoting the sum of g and h.
Algorithm 6 A* Algorithm: FindPathToGoal(u)
- F(Frontier) ← PriorityQueue(u) // This should be implemented with ˆf minimum
- E(Explored) ← {u}
- Initialize ˆg: ˆg[u]←0ˆg[v]=∞.∀v≠u
- while F is not empty do
- v←F.pop()
- if GoalTest (v) then
- return path (v)
- E.add (u)
- for all successors v of u do
- if v not in E then
- if v in F then
- ˆg[v]=min(ˆg[v],ˆg[u]+c(u,v))
- ˆf[v]=h[v]+ˆg[v]
- else
- F.push(v)
- ˆg[v]=ˆg[u]+c(u,v)
- ˆf[v]=h[v]+ˆg[v]
- return Failure
While in UCS, the frontier priority queue is implemented with ˆg ; in A* search, frontier priority queue should be implemented with ˆf , and ˆg . It is still used to keep track of the minimum path cost of reaching n discovered so far. A* is a mix of lowest-cost-first and best-first search, Algorithm 6 describes the A* algorithm.
Let us discuss the A* algorithm in detail.
As we have already discussed that UCS is optimal. The main reason for UCS being optimal was along the every possible optimal path S0,S1,…,Sk,Su , we know that before Si is popped, S1,…,Si−1 are already explored. To ensure this, UCS maintains ˆgpop(S0)≤ˆgpop(S1)≤…ˆgpop(Sk)≤ˆgpop(Su) , by ensuring ˆgpop(Si)=g(Si) . We want to use the same proof to show that A* is optimal.
Therefore in order to make A* optimal, we would like to look at the following property carefully:
Property:
ˆfpop(S0)≤ˆfpop(S1),≤,…,ˆfpop(Sk)≤ˆfpop(Su)Notice that If f(S0)≤f(S1)≤…≤f(Sk)≤f(u) is True, and ˆfpop(Si)=f(Si) holds, then this property would hold.
We know ∀i∈1,…,kf(Si)=g(Si)+h(Si), and let us assume f(S0)≤f(S1)≤…≤f(Sk)≤f(u).
∀i∈1,…,k, we have:
f(Si)≤f(Si+1)
g(Si)+h(Si)≤g(Si+1)+h(Si+1)
h(Si)≤c(Si,Si+1)+h(Si+1)
The property h(Si)≤c(SiSi+1)+h(Si+1) is known as consistency.
h(Si)≤c(Si,Si+1)+h(Si+1)
h(Si)≤c(Si,Si+1)+c(Si+1,Si+2)+h(Si+2)
..
..
h(Si)≤c(Si,Si+1)+c(Si+1,Si+2)+…+c(Sk,Su)+h(Su)
And, as we know, h(Su)=0,
h(Si)≤c(Si+1,Si)+c(Si+1,Si+2)+…+c(Sk,Su)
h(Si)≤Optimal(Si)
Where Optimal(Si) is optimal path cost from Si to goal u.
The property h(Si)≤Optimal(Si) is known as admissibility.
Therefore, consistency implies admissibility under the assumption that h(Su)=0 where Su is goal.
Claim: If h is consistent, then A* with graph search is optimal.
Proof: Let Frontier priority queue be implemented with ˆf(n), and ˆg(n), it is used to keep track of the minimum path cost of reaching n discovered so far. f(n),g(n) be the optimal cost, and fpop(n),gpop(n) represents the value of f and g for node n, when n is popped.
Given h is consistent, therefore:
h(Si)≤c(Si,Si+1)+h(Si+1) (5)
and, as discussed above, consistency implies admissibility, hence:
h(Si)≤Optimal(Si) (6)
Where Optimal is optimal value of cost of Si to goal. Also as per algorithm, for any path to goal u, we have:
∀i∈{1,…,u}f(Si)=g(Si)+h(Si) (7)
To prove A* is optimal, i.e ˆfpop(Su)=f(Su), where ˆfpop(Su) is estimated cost when Su is popped, and f(Su) is minimum cost to reach goal u.
By definition, we know:
ˆfpop(Sk+1)≥f(Sk+1) (8)
We will prove ˆfpop(Su)=f(Su) by induction. Let us assume we have a path Π to reach goal u.
Π=S1,S2,…,Sk,Sk+1,…,SuBase Case ˆfpop(S0)=f(S0)=h(S0)
Induction Hypothesis: ∀i∈1,…,kˆfpop(Si)=f(Si)
Observe that ˆfpop(Si)=f(Si) implies ˆgpop(Si)=g(Si)
Induction Step: When Sk is explored, the value of ˆf(Sk+1) is updated.
ˆfpop(Sk+1)≤min(ˆf(Sk+1),ˆgpop(Sk)+c(Sk,Sk+1)+h(Sk+1))
ˆfpop(Sk+1)≤ˆgpop(Sk)+c(Sk,Sk+1)+h(Sk+1)
ˆfpop(Sk+1)≤g(Sk)+c(Sk,Sk+1)+h(Sk+1) from Induction hypothesis
ˆfpop(Sk+1)≤g(Sk+1)+h(Sk+1)
ˆfpop(Sk+1)≤f(Sk+1) from equation 5
Hence, from above equation and equation 8
ˆfpop(Sk+1)=f(Sk+1)Completeness: A* is complete, and it will always find a solution if there is one, because the frontier always contains the initial part of a path to a goal, before that goal is selected. Also, whenever a node Si is popped from the Frontier, we have ˆgpop(Si)=g(Si). Therefore the maximum number of nodes that can be popped from the frontier before goal is popped is b⌊C∗ε⌋+1, which is finite.
A* for Tree Search Observe that admissibility is a weaker property than consistency, and with tree search, it is possible to design an optimal algorithm that can only uses admissibility property. In the case of graph search, a node is popped from fronter only once, so when it is explored, the optimal path must have been found, but with tree search that is not the case. A node can be popped multiple times, and it allows us to use relaxed conditions.
3.2 Design of Heuristic Function
We have discussed about informed search A*, and saw how an heuristic can help us to achieve optimality. We want an admissible and consistent heuristic function. There is no such standard strategy to design a consistent heuristic function but we have an approach for designing an admissible heuristic function.
There can be multiple good heuristic functions, but the function that we had used in the A* algorithm depends on the Euclidean distance. Euclidean distance is the straight line distance between 2 nodes. Let the Euclidean distance between node Si and Sj be d(Si,Sj), then h(Si)=d(Si,u) where u is the goal node.
In the above defined heuristic function, h is admissible, i.e h(si)≤Optimal(si). Also, h is consistent, and we can prove this claim by using the triangle inequality.
d(Si,u)≤d(Si,Si+1)+d(Si+1,u) d(Si,u)≤c(Si,Si+1)+d(Si+1,u) h(Si)≤c(Si,Si+1)+h(Si+1)Hence, h is consistent.
Heuristic function design is an art, and unfortunately there are no algorithm to automatically generate heuristic functions. Nevertheless, the purpose of a heuristic function is to reduce our search space, there are a few desired qualities in a good heuristic function.
Computable The heuristic function needs to be reasonably computable. For example, a possible heuristic function for a search problem could be the actual optimal cost to the goal. Although this is desired, this is not computable since finding the actual optimal cost, one would have to find the optimal path of a node to the goal which defeats the purpose of the heuristic function.
Informative Secondly, it needs to be informative. For example, a possible heuristic function is to set h(si)=0. Clearly, this fulfills both the criteria of consistency and admissibility, but this would be a bad heuristic, as it does not add any new information that narrows the search.
Let us try to find a good heuristic function for the example described in Figure 7 .Suppose we have a a grid of cells as shown in Figure 7 and we have an agent A that wants to reach the cell marked as G. Our agent can only travel in the 4 cardinal directions and cannot pass through cells marked with an X.
![]() |
---|
Figure 7: Example to design admissible heuristic function; Agent A wants to reach to agent G. |
There are 2 possible heuristic functions we can use, the Manhattan distance or the Euclidean distance. Notice that in this problem, both are equally easy to compute, but Manhattan distance provides more information because it takes into account the layout of our world while the Euclidean distance function does not. Hence, the better heuristic to use would be Manhattan distance.
3.3 Greedy-Best-First-Search
We have discussed UCS which orders the priority queue based on the estimated optimal cost ˆg of the node from the start node, and A* which orders the priority queue based on the combined cost f where f(si)=ˆg(si)+h(si).
We can also order the priority queue based solely on heuristic function h to first explore the nodes that are estimated to be closer to the goal. This approach is known as the Greedy-Best-First-Search (GBFS). It is easy to see that GBFS is not optimal and that can sometimes give bad results. Consider the graph shown in Figure 8.
![]() |
---|
Figure 8: Example to discuss Greedy-Best-First-Search; Roadmap from Home to School. |
As per example Figure 8, suppose the start node is Home and goal node is School. There are two paths to reach School from Home. Path1: Home, Station, School, and Path2: Home, BusStop1, BusStop2, School.
Consider the following values for h:
- h (Station) = 5
- h (Bus stop 1) = 20
- h (Bus stop 2) = 10
GBFS would first explore Path1, since h(Station)=5 is the minimum cost in the graph. However, we see that this is not the optimal path. The cost of Path1 is 105, but the cost of Path2 is 30.