The heuristic function `h(n)`

tells A* an *estimate* of the minimum cost from any vertex `n`

to the goal. It’s important to choose a good heuristic function.

## A*’s Use of the Heuristic#

The heuristic can be used to control A*’s behavior.

- At one extreme, if
`h(n)`

is 0, then only`g(n)`

plays a role, and A* turns into Dijkstra’s Algorithm, which is guaranteed to find a shortest path. - If
`h(n)`

is always lower than (or equal to) the cost of moving from`n`

to the goal, then A* is guaranteed to find a shortest path. The lower`h(n)`

is, the more node A* expands, making it slower. - If
`h(n)`

is exactly equal to the cost of moving from`n`

to the goal, then A* will only follow the best path and never expand anything else, making it very fast. Although you can’t make this happen in all cases, you can make it exact in some special cases. It’s nice to know that given perfect information, A* will behave perfectly. - If
`h(n)`

is sometimes greater than the cost of moving from`n`

to the goal, then A* is not guaranteed to find a shortest path, but it can run faster. - At the other extreme, if
`h(n)`

is very high relative to`g(n)`

, then only`h(n)`

plays a role, and A* turns into Greedy Best-First-Search.

So we have an interesting situation in that we can decide what we want to get out of A*. With 100% accurate estimates, we’ll get shortest paths really quickly. If we’re too low, then we’ll continue to get shortest paths, but it’ll slow down. If we’re too high, then we give up shortest paths, but A* will run faster.

In a game, this property of A* can be very useful. For example, you may find that in some situations, you would rather have a “good” path than a “perfect” path. To shift the balance between `g(n)`

and `h(n)`

, you can modify either one.

## Speed or accuracy?#

A*’s ability to vary its behavior based on the heuristic and cost functions can be very useful in a game. The tradeoff between speed and accuracy can be exploited to make your game faster. For most games, you don’t *really* need the **best** path between two points. You need something that’s close^{[1]}. What you need may depend on what’s going on in the game, or how fast the computer is. Using a function that *guarantees* it never overestimates the cost means that it will sometimes underestimate the cost by quite a bit.

Suppose your game has two types of terrain, Flat and Mountain, and the movement costs are 1 for flat land and 3 for mountains, A* is going to search three times as far along flat land as it does along mountainous land. This is because it’s *possible* that there is a path along flat terrain that goes around the mountains. You can speed up A*’s search by using 1.5 as the heuristic distance between two map spaces. A* will then compare 3 to 1.5, and it won’t look as bad as comparing 3 to 1. It is not as dissatisfied with mountainous terrain, so it won’t spend as much time trying to find a way around it. Alternatively, you can speed up up A*’s search by decreasing the amount it searches for paths around mountains—tell A* that the movement cost on mountains is 2 instead of 3. Now it will search only twice as far along the flat terrain as along mountainous terrain. Either approach gives up ideal paths to get something quicker.

The choice between speed and accuracy does not have to be static. You can choose dynamically based on the CPU speed, the fraction of time going into pathfinding, the number of units on the map, the importance of the unit, the size of the group, the difficulty level, or any other factor. One way to make the tradeoff dynamic is to build a heuristic function that assumes the minimum cost to travel one grid space is 1 and then build a cost function that scales:

g'(n) = 1 + alpha * (g(n) - 1)

If `alpha`

is 0, then the modified cost function will always be 1. At this setting, terrain costs are completely ignored, and A* works at the level of simple passable/unpassable grid spaces. If `alpha`

is 1, then the original cost function will be used, and you get the full benefit of A*. You can set `alpha`

anywhere in between.

You should also consider switching from the heuristic returning the *absolute* minimum cost to returning the *expected* minimum cost. For example, if most of your map is grasslands with a movement cost of 2 but some spaces on the map are roads with a movement cost of 1, then you might consider having the heuristic assume no roads, and return `2 * distance`

.

The choice between speed and accuracy does not have to be global. You can choose some things dynamically based on the importance of having accuracy in some region of the map. For example, it may be more important to choose a good path near the current location, on the assumption that we might end up recalculating the path or changing direction at some point, so why bother being accurate about the faraway part of the path? Or perhaps it’s not so important to have the shortest path in a safe area of the map, but when sneaking past an enemy village, safety and quickness are essential.

## Scale#

A* computes `f(n) = g(n) + h(n)`

. To add two values, those two values need to be at the same scale. If `g(n)`

is measured in hours and `h(n)`

is measured in meters, then A* is going to consider `g`

or `h`

too much or too little, and you either won’t get as good paths or you A* will run slower than it could.

## Exact heuristics#

If your heuristic is exactly equal to the distance along the optimal path, you’ll see A* expand very few nodes, as in the diagram shown in the next section. What’s happening inside A* is that it is computing `f(n) = g(n) + h(n)`

at every node. When `h(n)`

exactly matches `g(n)`

, the value of `f(n)`

doesn’t change along the path. All nodes not on the right path will have a higher value of `f`

than nodes that are on the right path. Since A* doesn’t consider higher-valued `f`

nodes until it has considered lower-valued `f`

nodes, it never strays off the shortest path.

### Precomputed exact heuristic#

One way to construct an exact heuristic is to precompute the length of the shortest path between every pair of points. This is not feasible for most game maps. However, there are ways to approximate this heuristic:

- Fit a coarse grid on top of the fine grid. Precompute the shortest path between any pair of coarse grid locations.
- Precompute the shortest path between any pair of waypoints. This is a generalization of the coarse grid approach.

Then add in a heuristic `h'`

that estimates the cost of going from any location to nearby waypoints. (The latter too can be precomputed if desired.) The final heuristic will be:

h(n) = h'(n, w1) + distance(w1, w2) + h'(w2, goal)

or if you want a better but more expensive heuristic, evaluate the above with all pairs `w1`

, `w2`

that are close to the node and the goal, respectively.

### Linear exact heuristic#

In a special circumstance, you can make the heuristic exact without precomputing anything. If you have a map with no obstacles and no slow terrain, then the shortest path from the starting point to the goal should be a straight line.

If you’re using a simple heuristic (one which does not know about the obstacles on the map), it should match the exact heuristic. If it doesn’t, then you may have a problem with scale or the type of heuristic you chose.

## Heuristics for grid maps#

On a grid, there are well-known heuristic functions to use.

**Use the distance heuristic that matches the allowed movement:**

- On a square grid that allows
**4 directions**of movement, use Manhattan distance (L_{1}). - On a square grid that allows
**8 directions**of movement, use Diagonal distance (L_{∞}). - On a square grid that allows
**any direction**of movement, you might or might not want Euclidean distance (L_{2}). If A* is finding paths on the grid but you are allowing movement not on the grid, you may want to consider other representations of the map^{[2]}. - On a hexagon grid that allows
**6 directions**of movement, use Manhattan distance adapted to hexagonal grids^{[3]}.

Multiply the distance in steps by the minimum cost for a step. For example, if you’re measuring in meters, the distance is 3 squares, and each square is 15 meters, then the heuristic would return 3 ⨉ 15 = 45 meters. If you’re measuring in time, the distance is 3 squares, and each square takes at least 4 minutes to cross, then the heuristic would return 3 ⨉ 4 = 12 minutes. The units (meters, minutes, etc.) returned by the heuristic should match the units used by the cost function.

### Manhattan distance#

The standard heuristic for a square grid is the Manhattan distance^{[4]}. Look at your cost function and find the minimum cost `D`

for moving from one space to an adjacent space. *In the simple case, you can set D to be 1.* The heuristic on a square grid where you can move in 4 directions should be

`D`

times the Manhattan distance:function heuristic(node) = dx = abs(node.x - goal.x) dy = abs(node.y - goal.y) return D * (dx + dy)

How do you pick D? Use a scale that matches your cost function. For the best paths, and an “admissible” heuristic, set D to the lowest cost between adjacent squares. In the absence of obstacles, and on terrain that has the minimum movement cost D, moving one step closer to the goal should *increase g* by D and

*decrease*by D. When you add the two,

`h`

`f`

(which is set to ```
g +
h
```

) will stay the same; that’s a sign that the heuristic and cost function scales match. You can also give up optimal paths to make A* run faster by increasing D, or by decreasing the ratio between the lowest and highest edge costs.(Note: the above image has a tie-breaker added to the heuristic.)

### Diagonal distance#

If your map allows diagonal movement you need a different heuristic. The Manhattan distance for (4 east, 4 north) will be 8⨉D. However, you could simply move (4 northeast) instead, so the heuristic should be 4⨉D2, where D2 is the cost of moving diagonally.

function heuristic(node) = dx = abs(node.x - goal.x) dy = abs(node.y - goal.y) return D * (dx + dy) + (D2 - 2 * D) * min(dx, dy)

Here we compute the number of steps you take if you can’t take a diagonal, then subtract the steps you save by using the diagonal. There are `min(dx, dy)`

diagonal steps, and each one costs `D2`

but saves you `2⨉D`

non-diagonal steps.

When D = 1 and D2 = 1, this is called the Chebyshev distance^{[5]}. When D = 1 and D2 = sqrt(2), this is called the *octile distance*.

Another way to write this is `D * max(dx, dy) + (D2-D) * min(dx, dy)`

. Patrick Lester writes it yet a different way^{[6]}, with `if (dx > dy) (D * (dx-dy) + D2 * dy) else (D * (dy-dx) + D2 * dx)`

. These are all equivalent.

### Euclidean distance#

If your units can move at any angle (instead of grid directions), then you should probably use a straight line distance:

function heuristic(node) = dx = abs(node.x - goal.x) dy = abs(node.y - goal.y) return D * sqrt(dx * dx + dy * dy)

However, if this is the case, then you may have trouble with using A* directly because the cost function `g`

will not match the heuristic function `h`

. Since Euclidean distance is shorter than Manhattan or diagonal distance, you will still get shortest paths, but A* will take longer to run:

There are several variants of A* that can help with this.

### Euclidean distance, squared#

I’ve seen several A* web pages recommend that you avoid the expensive square root in the Euclidean distance by using distance-squared:

function heuristic(node) = dx = abs(node.x - goal.x) dy = abs(node.y - goal.y) return D * (dx * dx + dy * dy)

**Do not do this!** This definitely runs into the scale problem. The scale of `g`

and `h`

need to match, because you’re adding them together to form `f`

. When A* computes `f(n) = g(n) + h(n)`

, the square of distance will be much higher than the cost `g`

and you will end up with an overestimating heuristic. For longer distances, this will approach the extreme of `g(n)`

not contributing to `f(n)`

, and A* will degrade into Greedy Best-First-Search:

To attempt to fix this you can scale the heuristic down. However, then you run into the opposite problem: for shorter distances, the heuristic will be too small compared to `g(n)`

and A* will degrade into Dijkstra’s Algorithm.

If, after profiling, you find the cost of the square root is significant, either use a fast square root approximation with Euclidean distance or use the diagonal distance as an approximation to Euclidean.

### Multiple goals#

If you want to search for *any* of several goals, construct a heuristic `h'(x)`

that is the minimum of ```
h1(x), h2(x),
h3(x), ...
```

where `h1, h2, h3`

are heuristics to each of the nearby spots.

One way to think about this is that we can add a new zero-cost edge from each of the goals to a new graph node. A path to that new node will necessarily go through one of the goal nodes.

If you want to search for paths to *all* of several goals, your best option may be Dijkstra’s Algorithm with early exit when you find all the goals. There may be a variant of A* that can calculate these paths; I don’t know.

If you want to search for spot *near* a single goal, ask A* search to find a path to the center of the goal area. While processing nodes from the OPEN set, exit when you pull a node that is near enough.

### Breaking ties#

In some grid maps there are many paths with the same length. For example, in flat areas without variation in terrain, using a grid will lead to many equal-length paths. A* might explore all the paths with the same `f`

value, instead of only one.

The quick hack to work around this problem is to either adjust the `g`

or `h`

values. The tie breaker needs to be deterministic with respect to the vertex (*i.e.,* it shouldn’t be a random number), and it needs to make the `f`

values differ. Since A* sorts by `f`

value, making them different means only one of the “equivalent” `f`

values will be explored.

One way to break ties is to nudge the scale of `h`

slightly. If we scale it downwards, then `f`

will increase as we move towards the goal. Unfortunately, this means that A* will prefer to expand vertices close to the starting point instead of vertices close to the goal. We can instead scale `h`

upwards slightly (even by 0.1%). A* will prefer to expand vertices close to the goal.

heuristic *= (1.0 + p)

The factor `p`

should be chosen so that `p <`

*(minimum cost of taking one step)*`/`

*(expected maximum path length)*. Assuming that you don’t expect the paths to be more than 1000 steps long, you can choose p = 1/1000. (Note that this slightly breaks “admissibility” of the heuristic but in games it almost never matters.) The result of this tie-breaking nudge is that A* explores far less of the map than previously:

When there are obstacles of course it still has to explore to find a way around them, but note that after the obstacle is passed, A* explores very little:

Steven van Dijk suggests that a more straightforward way to do this would to pass `h`

to the comparison function. When the `f`

values are equal, the comparison function would break the tie by looking at `h`

.

Another way to break ties is to add a deterministic random number to the heuristic or edge costs. (One way to choose a deterministic random number is to compute a hash of the coordinates.) This breaks more ties than adjusting `h`

as above. Thanks to Cris Fuhrman for suggesting this.

A different way to break ties is to prefer paths that are along the straight line from the starting point to the goal:

dx1 = current.x - goal.x dy1 = current.y - goal.y dx2 = start.x - goal.x dy2 = start.y - goal.y cross = abs(dx1*dy2 - dx2*dy1) heuristic += cross*0.001

This code computes the vector cross-product between the start to goal vector and the current point to goal vector. When these vectors don’t line up, the cross product will be larger. The result is that this code will give some slight preference to a path that lies along the straight line path from the start to the goal. When there are no obstacles, A* not only explores less of the map, the path looks very nice as well:

However, because this tie-breaker prefers paths along the straight line from the starting point to the goal, weird things happen when going around obstacles (note that the path is still optimal; it will look strange):

To interactively explore the improvement from this tie breaker, see James Macgill’s A* applet^{[7]} [or try this mirror^{[8]} or this mirror^{[9]}]. Use “Clear” to clear the map, and choose two points on opposite corners of the map. When you use the “Classic A*” method, you will see the effect of ties. When you use the “Fudge” method, you will see the effect of the above cross product added to the heuristic. [Most browsers will not run applets anymore as of 2021.]

Yet another way to break ties is to carefully construct your A* priority queue so that *new* insertions with a specific `f`

value are always ranked better (lower) than *old* insertions with the same `f`

value.

And yet another way to break ties on grids is to minimize turns. The change in x,y from the *parent* to the *current* node tells you what direction you were moving in. For all edges being considered from *current* to *neighbor*, if the change in x,y is different than the one from parent to current, then add a small penalty to the movement cost.

In my own projects I use a “checkerboard” approach. In Breadth First Search, I vary the *insertion order* so that paths alternate between horizontal and vertical steps. In Dijkstra’s Algorithm and A* I vary the *movement cost* so that paths alternate between horizontal and vertical steps. I have sample code and screenshots of both of these hacks^{[10]} in my A* implementation guide.

**The above modifications to the heuristic or costs are a “band aid” fix to an underlying inefficiency.** Ties occur when there are lots of paths that are equally good, leading to a large number of nodes to explore. Consider ways to “work smarter, not harder”:

- Walmsley, Goldstein, et al’s paper Path Counting for Grid-Based Navigation
^{[11]}uses Pascal’s Triangle to find the most aesthetically pleasing path in a grid. - Alternate map representations can solve the problem by
**reducing the number of nodes in the graph**. Collapsing multiple nodes into one, or by remove all but the important nodes. Rectangular Symmetry Reduction^{[12]}is a way to do this on square grids; also look at “framed quad trees”. Hierarchical pathfinding uses a high level graph with few nodes to find most of the path, then a low level graph with more nodes to refine the path. - Some approaches leave the number of nodes alone but
**reduce the number of nodes visited**. Jump Point Search^{[13]}skips over large areas of nodes that would contain lots of ties; it’s designed for square grids. Skip links add “shortcut” edges that skip over areas of the map. The AlphA* algorithm (PDF)^{[14]}adds some depth-first searching to the usual breadth-first behavior of A*, so that it can explore a single path instead of processing all of them simultaneously. - Fringe Search (PDF)
^{[15]}solves the problem instead by**making node processing fast**. Instead of keeping the OPEN set sorted and visiting nodes one at a time, it processes nodes in batches, expanding only the nodes that have low f-values. This is related to the HOT queues approach.

## Approximate heuristics#

A heuristic that has the exact distance is ideal for making A* fast but it’s usually impractical. We can often preprocess the graph to construct an approximate distance, and use that approximation in the A* heuristic.

ALT A*^{[16]} uses “landmarks” and the triangle inequality to preprocess the pathfinding graph in order to make pathfinding much faster. ALT also does a few other things, but the heuristic improvement is the part that got my attention. It’s surprisingly simple to implement, sometimes under 15 lines of code, and produces impressive speedups.

The name “landmark” is a little misleading. These points need to be placed on the outer edges of the map. Some authors call it “differential heuristics”.

The landmark approach stores lots of data that could be compressed. The Compressed Differential Heuristic^{[17]} shows the results of compressing the landmark data. You can store a lot more landmarks in the same space, so you get improved heuristic values.

Landmarks may be a special case of a more general approach. This paper^{[18]} explores transforming a map into a map where a regular distance metric works.

Distance oracles^{[19]} seem to be related but I haven’t looked into them yet.