The heuristic function
h(n) tells A* an estimate of the minimum cost from any vertex
n to the goal. It’s important to choose a good heuristic function.
The heuristic can be used to control A*’s behavior.
- At one extreme, if
h(n)is 0, then only
g(n)plays a role, and A* turns into Dijkstra’s algorithm, which is guaranteed to find a shortest path.
h(n)is always lower than (or equal to) the cost of moving from
nto the goal, then A* is guaranteed to find a shortest path. The lower
h(n)is, the more node A* expands, making it slower.
h(n)is exactly equal to the cost of moving from
nto the goal, then A* will only follow the best path and never expand anything else, making it very fast. Although you can’t make this happen in all cases, you can make it exact in some special cases. It’s nice to know that given perfect information, A* will behave perfectly.
h(n)is sometimes greater than the cost of moving from
nto the goal, then A* is not guaranteed to find a shortest path, but it can run faster.
- At the other extreme, if
h(n)is very high relative to
g(n), then only
h(n)plays a role, and A* turns into Greedy Best-First-Search.
Technically, the A* algorithm should be called simply A if the heuristic is an underestimate of the actual cost. However, I will continue to call it A* because the implementation is the same and the game programming community does not distinguish A from A*.
So we have an interesting situation in that we can decide what we want to get out of A*. At exactly the right point, we’ll get shortest paths really quickly. If we’re too low, then we’ll continue to get shortest paths, but it’ll slow down. If we’re too high, then we give up shortest paths, but A* will run faster.
In a game, this property of A* can be very useful. For example, you may find that in some situations, you would rather have a “good” path than a “perfect” path. To shift the balance between
h(n), you can modify either one.
A*’s ability to vary its behavior based on the heuristic and cost functions can be very useful in a game. The tradeoff between speed and accuracy can be exploited to make your game faster. For most games, you don’t really need the best path between two points. You just need something that’s close. What you need may depend on what’s going on in the game, or how fast the computer is.
Suppose your game has two types of terrain, Flat and Mountain, and the movement costs are 1 for flat land and 3 for mountains, A* is going to search three times as far along flat land as it does along mountainous land. This is because it’s possible that there is a path along flat terrain that goes around the mountains. You can speed up A*’s search by using 1.5 as the heuristic distance between two map spaces. A* will then compare 3 to 1.5, and it won’t look as bad as comparing 3 to 1. It is not as dissatisfied with mountainous terrain, so it won’t spend as much time trying to find a way around it. Alternatively, you can speed up up A*’s search by decreasing the amount it searches for paths around mountains―just tell A* that the movement cost on mountains is 2 instead of 3. Now it will search only twice as far along the flat terrain as along mountainous terrain. Either approach gives up ideal paths to get something quicker.
The choice between speed and accuracy does not have to be static. You can choose dynamically based on the CPU speed, the fraction of time going into pathfinding, the number of units on the map, the importance of the unit, the size of the group, the difficulty level, or any other factor. One way to make the tradeoff dynamic is to build a heuristic function that assumes the minimum cost to travel one grid space is 1 and then build a cost function that scales:
g'(n) = 1 + alpha * (g(n) - 1)
alpha is 0, then the modified cost function will always be 1. At this setting, terrain costs are completely ignored, and A* works at the level of simple passable/unpassable grid spaces. If
alpha is 1, then the original cost function will be used, and you get the full benefit of A*. You can set
alpha anywhere in between.
You should also consider switching from the heuristic returning the absolute minimum cost to returning the expected minimum cost. For example, if most of your map is grasslands with a movement cost of 2 but some spaces on the map are roads with a movement cost of 1, then you might consider having the heuristic assume no roads, and return
2 * distance.
The choice between speed and accuracy does not have to be global. You can choose some things dynamically based on the importance of having accuracy in some region of the map. For example, it may be more important to choose a good path near the current location, on the assumption that we might end up recalculating the path or changing direction at some point, so why bother being accurate about the faraway part of the path? Or perhaps it’s not so important to have the shortest path in a safe area of the map, but when sneaking past an enemy village, safety and quickness are essential.
f(n) = g(n) + h(n). To add two values, those two values need to be at the same scale. If
g(n) is measured in hours and
h(n) is measured in meters, then A* is going to consider
h too much or too little, and you either won’t get as good paths or you A* will run slower than it could.
If your heuristic is exactly equal to the distance along the optimal path, you’ll see A* expand very few nodes, as in the diagram shown in the next section. What’s happening inside A* is that it is computing
f(n) = g(n) + h(n) at every node. When
h(n) exactly matches
g(n), the value of
f(n) doesn’t change along the path. All nodes not on the right path will have a higher value of
f than nodes that are on the right path. Since A* doesn’t consider higher-valued
f nodes until it has considered lower-valued
f nodes, it never strays off the shortest path.
Precomputed exact heuristic
One way to construct an exact heuristic is to precompute the length of the shortest path between every pair of points. This is not feasible for most game maps. However, there are ways to approximate this heuristic:
- Fit a coarse grid on top of the fine grid. Precompute the shortest path between any pair of coarse grid locations.
- Precompute the shortest path between any pair of waypoints. This is a generalization of the coarse grid approach.
Then add in a heuristic
h' that estimates the cost of going from any location to nearby waypoints. (The latter too can be precomputed if desired.) The final heuristic will be:
h(n) = h'(n, w1) + distance(w1, w2) + h'(w2, goal)
or if you want a better but more expensive heuristic, evaluate the above with all pairs
w2 that are close to the node and the goal, respectively.
Linear exact heuristic
In a special circumstance, you can make the heuristic exact without precomputing anything. If you have a map with no obstacles and no slow terrain, then the shortest path from the starting point to the goal should be a straight line.
If you’re using a simple heuristic (one which does not know about the obstacles on the map), it should match the exact heuristic. If it doesn’t, then you may have a problem with scale or the type of heuristic you chose.
On a grid, there are well-known heuristic functions to use.
Use the distance heuristic that matches the allowed movement:
- On a square grid that allows 4 directions of movement, use Manhattan distance (L1).
- On a square grid that allows 8 directions of movement, use Diagonal distance (L∞).
- On a square grid that allows any direction of movement, you might or might not want Euclidean distance (L2). If A* is finding paths on the grid but you are allowing movement not on the grid, you may want to consider other representations of the map.
- On a hexagon grid that allows 6 directions of movement, use Manhattan distance adapted to hexagonal grids.
The standard heuristic for a square grid is the Manhattan distance. Look at your cost function and find the minimum cost
D for moving from one space to an adjacent space. In the simple case, you can set
D to be 1. The heuristic on a square grid where you can move in 4 directions should be
D times the Manhattan distance:
function heuristic(node) = dx = abs(node.x - goal.x) dy = abs(node.y - goal.y) return D * (dx + dy)
How do you pick D? Use a scale that matches your cost function. For the best paths, and an “admissible” heuristic, set D to the lowest cost between adjacent squares. In the absence of obstacles, and on terrain that has the minimum movement cost D, moving one step closer to the goal should increase
g by D and decrease
h by D. When you add the two,
f (which is set to
h) will stay the same; that’s a sign that the heuristic and cost function scales match. You can also give up optimal paths to make A* run faster by increasing D, or by decreasing the ratio between the lowest and highest edge costs.
(Note: the above image has a tie-breaker added to the heuristic.}
If your map allows diagonal movement you need a different heuristic (sometimes called the Chebyshev distance). The Manhattan distance for (4 east, 4 north) will be 8*D. However, you could simply move (4 northeast) instead, so the heuristic should be 4*D. This function handles diagonals, assuming that both straight and diagonal movement costs D:
function heuristic(node) = dx = abs(node.x - goal.x) dy = abs(node.y - goal.y) return D * max(dx, dy)
If the cost of diagonal movement is not D, but something like D2 = sqrt(2)*D, the above heuristic won’t be right for you. You will instead want something more sophisticated:
function heuristic(node) = dx = abs(node.x - goal.x) dy = abs(node.y - goal.y) return D * (dx + dy) + (D2 - 2 * D) * min(dx, dy)
Here we compute the number of steps you take if you can’t take a diagonal, then subtract the steps you save by using the diagonal. There are
min(dx, dy) diagonal steps, and each one costs
D2 but saves you
2*D non-diagonal steps.
Patrick Lester has a different way of writing this heuristic, using an explicit cases for
dx > dy vs
dx < dy. The above code has the same test but it’s hidden inside the call to
If your units can move at any angle (instead of grid directions), then you should probably use a straight line distance:
function heuristic(node) = dx = abs(node.x - goal.x) dy = abs(node.y - goal.y) return D * sqrt(dx * dx + dy * dy)
However, if this is the case, then you may have trouble with using A* directly because the cost function
g will not match the heuristic function
h. Since Euclidean distance is shorter than Manhattan or diagonal distance, you will still get shortest paths, but A* will take longer to run:
Euclidean distance, squared
I’ve seen several A* web pages recommend that you avoid the expensive square root in the Euclidean distance by just using distance-squared:
function heuristic(node) = dx = abs(node.x - goal.x) dy = abs(node.y - goal.y) return D * (dx * dx + dy * dy)
Do not do this! This definitely runs into the scale problem. The scale of
h need to match, because you’re adding them together to form
f. When A* computes
f(n) = g(n) + h(n), the square of distance will be much higher than the cost
g and you will end up with an overestimating heuristic. For longer distances, this will approach the extreme of
g(n) not contributing to
f(n), and A* will degrade into Greedy Best-First-Search:
To attempt to fix this you can scale the heuristic down. However, then you run into the opposite problem: for shorter distances, the heuristic will be too small compared to
g(n) and A* will degrade into Dijkstra’s algorithm.
If, after profiling, you find the cost of the square root is significant, either use a fast square root approximation with Euclidean distance or use the diagonal distance as an approximation to Euclidean.
If you want to search for any of several goals, construct a heuristic
h'(x) that is the minimum of
h3(x), ... where
h1, h2, h3 are heuristics to each of the nearby spots.
If you want to search for spot near a single goal, ask A* search to find a path to the center of the goal area. While processing nodes from the OPEN set, exit when you pull a node that is near enough.
In some grid maps there are many paths with the same length. For example, in flat areas without variation in terrain, using a grid will lead to many equal-length paths. A* might explore all the paths with the same
f value, instead of just one.
To solve this problem, we need to either adjust the
h values. The tie breaker needs to be deterministic with respect to the vertex (i.e., it shouldn’t just be a random number), and it needs to make the
f values differ. Since A* sorts by
f value, making them different means only one of the “equivalent”
f values will be explored.
One way to break ties is to nudge the scale of
h slightly. If we scale it downwards, then
f will increase as we move towards the goal. Unfortunately, this means that A* will prefer to expand vertices close to the starting point instead of vertices close to the goal. We can instead scale
h upwards slightly (even by 0.1%). A* will prefer to expand vertices close to the goal.
heuristic *= (1.0 + p)
p should be chosen so that
p <(minimum cost of taking one step)
/(expected maximum path length). Assuming that you don’t expect the paths to be more than 1000 steps long, you can choose p = 1/1000. (Note that this slightly breaks “admissibility” of the heuristic but in games it almost never matters.) The result of this tie-breaking nudge is that A* explores far less of the map than previously:
Tie-breaking scaling added to heuristic.
When there are obstacles of course it still has to explore to find a way around them, but note that after the obstacle is passed, A* explores very little:
Tie-breaking scaling added to heuristic, works nicely with obstacles.
Steven van Dijk suggests that a more straightforward way to do this would to pass
h to the comparison function. When the
f values are equal, the comparison function would break the tie by looking at
Another way to break ties is to add a deterministic random number to the heuristic or edge costs. (One way to choose a deterministic random number is to compute a hash of the coordinates.) This breaks more ties than adjusting
h as above. Thanks to Cris Fuhrman for suggesting this.
A different way to break ties is to prefer paths that are along the straight line from the starting point to the goal:
dx1 = current.x - goal.x dy1 = current.y - goal.y dx2 = start.x - goal.x dy2 = start.y - goal.y cross = abs(dx1*dy2 - dx2*dy1) heuristic += cross*0.001
This code computes the vector cross-product between the start to goal vector and the current point to goal vector. When these vectors don’t line up, the cross product will be larger. The result is that this code will give some slight preference to a path that lies along the straight line path from the start to the goal. When there are no obstacles, A* not only explores less of the map, the path looks very nice as well:
Tie-breaking cross-product added to heuristic, produces pretty paths.
However, because this tie-breaker prefers paths along the straight line from the starting point to the goal, weird things happen when going around obstacles (note that the path is still optimal; it just looks strange):
Tie-breaking cross-product added to heuristic, less pretty with obstacles.
To interactively explore the improvement from this tie breaker, see James Macgill’s A* applet [or try this mirror or this mirror]. Use “Clear” to clear the map, and choose two points on opposite corners of the map. When you use the “Classic A*” method, you will see the effect of ties. When you use the “Fudge” method, you will see the effect of the above cross product added to the heuristic.
Yet another way to break ties is to carefully construct your A* priority queue so that new insertions with a specific
f value are always ranked better (lower) than old insertions with the same
And yet another way to break ties on grids is to minimize turns. The change in x,y from the parent to the current node tells you what direction you were moving in. For all edges being considered from current to neighbor, if the change in x,y is different than the one from parent to current, then add a small penalty to the movement cost.
If ties are common, the above modifications to the heuristic may merely be a “band aid” fix to an underlying inefficiency. Ties occur when there are lots of paths that are equally good, leading to a large number of nodes to explore. Consider ways to “work smarter, not harder”:
- Alternate map representations can solve the problem by reducing the number of nodes in the graph. Collapsing multiple nodes into one, or by remove all but the important nodes. Rectangular Symmetry Reduction is a way to do this on square grids; also look at “framed quad trees”. Hierarchical pathfinding uses a high level graph with few nodes to find most of the path, then a low level graph with more nodes to refine the path.
- Some approaches leave the number of nodes alone but reduce the number of nodes visited. Jump Point Search skips over large areas of nodes that would contain lots of ties; it’s designed for square grids. Skip links add “shortcut” edges that skip over areas of the map. The AlphA* algorithm adds some depth-first searching to the usual breadth-first behavior of A*, so that it can explore a single path instead of processing all of them simultaneously.
- Fringe Search (PDF) solves the problem instead by making node processing fast. Instead of keeping the OPEN set sorted and visiting nodes one at a time, it processes nodes in batches, expanding only the nodes that have low f-values. This is related to the HOT queues approach.