# Minkowski–Bouligand dimension

(Redirected from Box-counting dimension)
Estimating the box-counting dimension of the coast of Great Britain

In fractal geometry, the Minkowski–Bouligand dimension, also known as Minkowski dimension or box-counting dimension, is a way of determining the fractal dimension of a set ${\displaystyle S}$ in a Euclidean space ${\displaystyle \mathbb {R} ^{n}}$, or more generally in a metric space ${\displaystyle (X,d)}$. It is named after the Polish mathematician Hermann Minkowski and the French mathematician Georges Bouligand.

To calculate this dimension for a fractal ${\displaystyle S}$, imagine this fractal lying on an evenly spaced grid and count how many boxes are required to cover the set. The box-counting dimension is calculated by seeing how this number changes as we make the grid finer by applying a box-counting algorithm.

Suppose that ${\displaystyle N(\varepsilon )}$ is the number of boxes of side length ${\displaystyle \varepsilon }$ required to cover the set. Then the box-counting dimension is defined as

${\displaystyle \dim _{\text{box}}(S):=\lim _{\varepsilon \to 0}{\frac {\log N(\varepsilon )}{\log(1/\varepsilon )}}.}$

Roughly speaking, this means that the dimension is the exponent ${\displaystyle d}$ such that ${\displaystyle N(1/n)\approx Cn^{d}}$, which is what one would expect in the trivial case where ${\displaystyle S}$ is a smooth space (a manifold) of integer dimension ${\displaystyle d}$.

If the above limit does not exist, one may still take the limit superior and limit inferior, which respectively define the upper box dimension and lower box dimension. The upper box dimension is sometimes called the entropy dimension, Kolmogorov dimension, Kolmogorov capacity, limit capacity or upper Minkowski dimension, while the lower box dimension is also called the lower Minkowski dimension.

The upper and lower box dimensions are strongly related to the more popular Hausdorff dimension. Only in very special applications is it important to distinguish between the three (see below). Yet another measure of fractal dimension is the correlation dimension.

## Alternative definitions

Examples of ball packing, ball covering, and box covering

It is possible to define the box dimensions using balls, with either the covering number or the packing number. The covering number ${\displaystyle N_{\text{covering}}(\varepsilon )}$ is the minimal number of open balls of radius ε required to cover the fractal, or in other words, such that their union contains the fractal. We can also consider the intrinsic covering number ${\displaystyle N'_{\text{covering}}(\varepsilon )}$, which is defined the same way but with the additional requirement that the centers of the open balls lie inside the set S. The packing number ${\displaystyle N_{\text{packing}}(\varepsilon )}$ is the maximal number of disjoint open balls of radius ε one can situate such that their centers would be inside the fractal. While N, Ncovering, N'covering and Npacking are not exactly identical, they are closely related and give rise to identical definitions of the upper and lower box dimensions. This is easy to prove once the following inequalities are proven:

${\displaystyle N_{\text{packing}}(\varepsilon )\leq N'_{\text{covering}}(\varepsilon )\leq N_{\text{covering}}(\varepsilon /2).}$

These, in turn, follow with a little effort from the triangle inequality.

The advantage of using balls rather than squares is that this definition generalizes to any metric space. In other words, the box definition is extrinsic — one assumes the fractal space S is contained in a Euclidean space, and defines boxes according to the external geometry of the containing space. However, the dimension of S should be intrinsic, independent of the environment into which S is placed, and the ball definition can be formulated intrinsically. One defines an internal ball as all points of S within a certain distance of a chosen center, and one counts such balls to get the dimension. (More precisely, the Ncovering definition is extrinsic, but the other two are intrinsic.)

The advantage of using boxes is that in many cases N(ε) may be easily calculated explicitly, and that for boxes the covering and packing numbers (defined in an equivalent way) are equal.

The logarithm of the packing and covering numbers are sometimes referred to as entropy numbers and are somewhat analogous to the concepts of thermodynamic entropy and information-theoretic entropy, in that they measure the amount of "disorder" in the metric space or fractal at scale ε and also measure how many bits or digits one would need to specify a point of the space to accuracy ε.

Another equivalent (extrinsic) definition for the box-counting dimension is given by the formula

${\displaystyle \dim _{\text{box}}(S)=n-\lim _{r\to 0}{\frac {\log {\text{vol}}(S_{r})}{\log r}},}$

where for each r > 0, the set ${\displaystyle S_{r}}$ is defined to be the r-neighborhood of S, i.e. the set of all points in ${\displaystyle R^{n}}$ that are at distance less than r from S (or equivalently, ${\displaystyle S_{r}}$ is the union of all the open balls of radius r centered at a point in S).

## Properties

Both box dimensions are finitely additive, i.e. if {A1, ..., An} is a finite collection of sets, then

${\displaystyle \dim(A_{1}\cup \dotsb \cup A_{n})=\max\{\dim A_{1},\dots ,\dim A_{n}\}.}$

However, they are not countably additive, i.e. this equality does not hold for an infinite sequence of sets. For example, the box dimension of a single point is 0, but the box dimension of the collection of rational numbers in the interval [0, 1] has dimension 1. The Hausdorff measure by comparison, is countably additive.

An interesting property of the upper box dimension not shared with either the lower box dimension or the Hausdorff dimension is the connection to set addition. If A and B are two sets in a Euclidean space, then A + B is formed by taking all the pairs of points ab where a is from A and b is from B and adding a + b. One has

${\displaystyle \dim _{\text{upper box}}(A+B)\leq \dim _{\text{upper box}}(A)+\dim _{\text{upper box}}(B).}$

## Relations to the Hausdorff dimension

The box-counting dimension is one of a number of definitions for dimension that can be applied to fractals. For many well behaved fractals all these dimensions are equal; in particular, these dimensions coincide whenever the fractal satisfies the open set condition (OSC).[1] For example, the Hausdorff dimension, lower box dimension, and upper box dimension of the Cantor set are all equal to log(2)/log(3). However, the definitions are not equivalent.

The box dimensions and the Hausdorff dimension are related by the inequality

${\displaystyle \dim _{\text{Haus}}\leq \dim _{\text{lower box}}\leq \dim _{\text{upper box}}.}$

In general, both inequalities may be strict. The upper box dimension may be bigger than the lower box dimension if the fractal has different behaviour in different scales. For example, examine the set of numbers in the interval [0, 1] satisfying the condition

for any n, all the digits between the 22n-th digit and the (22n+1 − 1)-th digit are zero.

The digits in the "odd place-intervals", i.e. between digits 22n+1 and 22n+2 − 1 are not restricted and may take any value. This fractal has upper box dimension 2/3 and lower box dimension 1/3, a fact which may be easily verified by calculating N(ε) for ${\displaystyle \varepsilon =10^{-2^{n}}}$ and noting that their values behave differently for n even and odd.

Another example: the set of rational numbers ${\displaystyle \mathbb {Q} }$, a countable set with ${\displaystyle \dim _{\text{Haus}}=0}$, has ${\displaystyle \dim _{\text{box}}=1}$ because its closure, ${\displaystyle \mathbb {R} }$, has dimension 1. In fact,

${\displaystyle \dim _{\text{box}}\left\{0,1,{\frac {1}{2}},{\frac {1}{3}},{\frac {1}{4}},\ldots \right\}={\frac {1}{2}}.}$

These examples show that adding a countable set can change box dimension, demonstrating a kind of instability of this dimension.

## References

1. ^ Wagon, Stan (2010). Mathematica in Action: Problem Solving Through Visualization and Computation. Springer-Verlag. p. 214. ISBN 0-387-75477-6.