# Random graph

In mathematics, random graph is the general term to refer to probability distributions over graphs. Random graphs may be described simply by a probability distribution, or by a random process which generates them.[1] The theory of random graphs lies at the intersection between graph theory and probability theory. From a mathematical perspective, random graphs are used to answer questions about the properties of typical graphs. Its practical applications are found all areas in which complex networks need to be modeled – a large number of random graph models are thus known, mirroring the diverse types of complex networks encountered in different areas. In a mathematical context, random graph refers almost exclusively to the Erdős–Rényi random graph model. In other contexts, any graph model may be referred to as a random graph.

## Random graph models

A random graph is obtained by starting with a set of n isolated vertices and adding successive edges between them at random. The aim of the study in this field is to determine at what stage a particular property of the graph is likely to arise.[2] Different random graph models produce different probability distributions on graphs. Most commonly studied is the one proposed by Edgar Gilbert, denoted G(n,p), in which every possible edge occurs independently with probability 0 < p < 1. The probability of a random graph with m edges is $p^m (1-p)^{\tbinom{n}{2}-m}$.[3] A closely related model, the Erdős–Rényi model denoted G(n,M), assigns equal probability to all graphs with exactly M edges. With the notation $N = \tbinom{n}{2}$ with 0 ≤ MN, G(n,p) has $\tbinom{N}{M}$ elements and every element occurs with probability $1/\tbinom{N}{M}$.[2] The latter model can be viewed as a snapshot at a particular time (M) of the random graph process $\tilde{G}_n$, which is a stochastic process that starts with n vertices and no edges, and at each step adds one new edge chosen uniformly from the set of missing edges.

If instead we start with an infinite set of vertices, and again let every possible edge occur independently with probability 0 < p < 1, then we get an object G called an infinite random graph. Except in the trivial cases when p is 0 or 1, such a G almost surely has the following property:

Given any n + m elements $a_1,\ldots, a_n,b_1,\ldots, b_m \in V$, there is a vertex c in V that is adjacent to each of $a_1,\ldots, a_n$ and is not adjacent to any of $b_1,\ldots, b_m$.

It turns out that if the vertex set is countable then there is, up to isomorphism, only a single graph with this property, namely the Rado graph. Thus any countably infinite random graph is almost surely the Rado graph, which for this reason is sometimes called simply the random graph. However, the analogous result is not true for uncountable graphs, of which there are many (nonisomorphic) graphs satisfying the above property.

Another model, which generalizes Gilbert's random graph model, is the random dot-product model. A random dot-product graph associates with each vertex a real vector. The probability of an edge uv between any vertices u and v is some function of the dot product uv of their respective vectors.

The network probability matrix models random graphs through edge probabilities, which represent the probability $p_{i,j}$ that a given edge $e_{i,j}$ exists for a specified time period. This model is extensible to directed and undirected; weighted and unweighted; and static or dynamic graphs.

For MpN, where N is the maximal number of edges possible, the two most widely used models, G(n,M) and G(n,p), are almost interchangeable.[4]

Random regular graphs form a special case, with properties that may differ from random graphs in general.

Once we have a model of random graphs, every function on graphs, becomes a random variable. The study of this model is to determine, or at least estimate the probability a property may occur.[3]

## Terminology

The term 'almost every' in the context of random graphs refers to a sequence of spaces and probabilities, such that the error probabilities tend to zero.[3]

## Properties of random graphs

The theory of random graphs studies typical properties of random graphs, those that hold with high probability for graphs drawn from a particular distribution. For example, we might ask for a given value of n and p what the probability is that G(n,p) is connected. In studying such questions, researchers often concentrate on the asymptotic behavior of random graphs—the values that various probabilities converge to as n grows very large. Percolation theory characterizes the connectedness of random graphs, especially infinitely large ones.

Percolation is related to the robustness of the graph (called also network). Given a random graph of n nodes and an average degree $\langle k\rangle$. Next we remove randomly a fraction 1−p of nodes and leave only a fraction p. There exists a critical percolation threshold $p_c=\tfrac{1}{\langle k\rangle}$ below which the network becomes fragmented while above pc a giant connected component exists.[1][5][4][6] [7][8]

(threshold functions, evolution of G~)

Random graphs are widely used in the probabilistic method, where one tries to prove the existence of graphs with certain properties. The existence of a property on a random graph can often imply, via the Szemerédi regularity lemma, the existence of that property on almost all graphs.

In random regular graphs, G(n,r-reg) are the set of r-regular graphs with r = r(n) such that n and m are the natural numbers, 3 ≤ r < n, and rn = 2m is even.[2]

The degree sequence of a graph G in Gn depends only on the number of edges in the sets[2]

$V_n^{(2)} = \left \{ij \ : \ 1 \leq j \leq n, i \neq j \right \} \subset V^{(2)}, \qquad i=1, \cdots, n.$

If edges, M in a random graph, GM is large enough to ensure that almost every GM has minimum degree at least 1, then almost every GM is connected and, if n is even, almost every GM has a perfect matching. In particular, the moment the last isolated vertex vanishes in almost every random graph, the graph becomes connected.[2]

Almost every graph process on an even number of vertices with the edge raising the minimum degree to 1 or a random graph with slightly more than (n/4)log(n) edges and with probability close to 1 ensures that the graph has a complete matching, with exception of at most one vertex.

For some constant c, almost every labelled graph with n vertices and at least cnlog(n) edges is Hamiltonian. With the probability tending to 1, the particular edge that increases the minimum degree to 2 makes the graph Hamiltonian.

## Coloring of Random Graphs

Given a random graph G of order n with the vertex V(G) = {1, ..., n}, by the greedy algorithm on the number of colors, the vertices can be colored with colors 1, 2, ... (vertex 1 is colored 1, vertex 2 is colored 1 if it is not adjacent to vertex 1, otherwise it is colored 2, etc).[2]

## Random trees

A random tree is a tree or arborescence that is formed by a stochastic process. In a large range of random graphs of order n and size M(n) the distribution of the number of tree components of order k is asymptotically Poisson. Types of random trees include uniform spanning tree, random minimal spanning tree, random binary tree, treap, rapidly exploring random tree, Brownian tree, and random forest.

## History

Random graphs were first defined by Paul Erdős and Alfréd Rényi in their 1959 paper "On Random Graphs"[8] and independently by Gilbert in his paper "Random graphs".[5]