# Free independence

In the mathematical theory of free probability, the notion of free independence was introduced by Dan Voiculescu.[1] The definition of free independence is parallel to the classical definition of independence, except that the role of Cartesian products of measure spaces (corresponding to tensor products of their function algebras) is played by the notion of a free product of (non-commutative) probability spaces.

In the context of Voiculescu's free probability theory, many classical-probability theorems or phenomena have free probability analogs: the same theorem or phenomenon holds (perhaps with slight modifications) if the classical notion of independence is replaced by free independence. Examples of this include: the free central limit theorem; notions of free convolution; existence of free stochastic calculus and so on.

Let ${\displaystyle (A,\phi )}$ be a non-commutative probability space, i.e. a unital algebra ${\displaystyle A}$ over ${\displaystyle \mathbb {C} }$ equipped with a unital linear functional ${\displaystyle \phi :A\to \mathbb {C} }$. As an example, one could take, for a probability measure ${\displaystyle \mu }$,

${\displaystyle A=L^{\infty }(\mathbb {R} ,\mu ),\phi (f)=\int f(t)\,d\mu (t).}$

Another example may be ${\displaystyle A=M_{N}}$, the algebra of ${\displaystyle N\times N}$ matrices with the functional given by the normalized trace ${\displaystyle \phi ={\frac {1}{N}}Tr}$. Even more generally, ${\displaystyle A}$ could be a von Neumann algebra and ${\displaystyle \phi }$ a state on ${\displaystyle A}$. A final example is the group algebra ${\displaystyle A=\mathbb {C} \Gamma }$ of a (discrete) group ${\displaystyle \Gamma }$ with the functional ${\displaystyle \phi }$ given by the group trace ${\displaystyle \phi (g)=\delta _{g=e},g\in \Gamma }$.

Let ${\displaystyle \{A_{i}:i\in I\}}$ be a family of unital subalgebras of ${\displaystyle A}$.

Definition. The family ${\displaystyle \{A_{i}:i\in I\}}$ is called freely independent if ${\displaystyle \phi (x_{1}x_{2}\cdots x_{n})=0}$ whenever ${\displaystyle \phi (x_{j})=0}$, ${\displaystyle x_{j}\in A_{i(j)}}$ and ${\displaystyle i(1)\neq i(2),i(2)\neq i(3),\dots }$.

If ${\displaystyle X_{i}\in A}$, ${\displaystyle i\in I}$ is a family of elements of ${\displaystyle A}$ (these can be thought of as random variables in ${\displaystyle A}$), they are called

freely independent if the algebras ${\displaystyle A_{i}}$ generated by ${\displaystyle 1}$ and ${\displaystyle X_{i}}$ are freely independent.

## Examples of free independence

• Let ${\displaystyle \Gamma }$ be the free product of groups ${\displaystyle \Gamma _{i},i\in I}$, let ${\displaystyle A=\mathbb {C} \Gamma }$ be the group algebra, ${\displaystyle \phi (g)=\delta _{g=e}}$ be the group trace, and set ${\displaystyle A_{i}=\mathbb {C} \Gamma _{i}\subset A}$. Then ${\displaystyle A_{i}:i\in I}$ are freely independent.
• Let ${\displaystyle U_{i}(N),i=1,2}$ be ${\displaystyle N\times N}$ unitary random matrices, taken independently at random from the ${\displaystyle N\times N}$ unitary group (with respect to the Haar measure). Then ${\displaystyle U_{1}(N),U_{2}(N)}$ become asymptotically freely independent as ${\displaystyle N\to \infty }$. (Asymptotic freeness means that the definition of freeness holds in the limit as ${\displaystyle N\to \infty }$).
• More generally, independent random matrices tend to be asymptotically freely independent, under certain conditions.

## References

1. ^ D. Voiculescu, K. Dykema, A. Nica, "Free Random Variables", CIRM Monograph Series, AMS, Providence, RI, 1992