Rejection sampling

From Wikipedia, the free encyclopedia
Jump to: navigation, search

In mathematics, rejection sampling is a basic technique used to generate observations from a distribution. It is also commonly called the acceptance-rejection method or "accept-reject algorithm" and is a type of Monte Carlo method. The method works for any distribution in \mathbb{R}^m with a density.

Rejection sampling is based on the observation that to sample a random variable one can sample uniformly from the region under the graph of its density function.[1][2][3][4]


To visualize the motivation behind rejection sampling, imagine graphing the density function of a random variable onto a large rectangular board and throwing darts at it. Assume that the darts are uniformly distributed around the board. Now take off (reject) all of the darts that are outside the area under the curve. The remaining darts will be distributed uniformly within the area under the curve, and the x-positions of these darts will be distributed according to the random variable's density. This is because there is the most room for the darts to land where the curve is highest and thus the probability density is greatest.

The visualization as just described is equivalent to a particular form of rejection sampling where the proposal distribution is uniform (hence its graph is a rectangle). The general form of rejection sampling assumes that the board is not necessarily rectangular but is shaped according to some distribution that we know how to sample from (e.g. using inversion sampling), and which is at least as high at every point as the distribution we want to sample from, so that the former completely encloses the latter. (Otherwise, there will be parts of the curved area we want to sample from that can never be reached.) Rejection sampling works as follows:

  1. Sample a point (an x-position) from the proposal distribution.
  2. Draw a vertical line at this x-position, up to the curve of the proposal distribution.
  3. Sample uniformly along this line (i.e. uniformly from 0 to the value of the proposal distribution (maximum of the probability density function)). If the sampled value is greater than the value of the desired distribution at this vertical line, return to step 1.

Note also that this algorithm can be used to sample from the area under any curve, regardless of whether the function integrates to 1. In fact, scaling a function by a constant has no effect on the sampled x-positions. This means that the algorithm can be used to sample from a distribution whose probability density function is only known up to a constant (i.e. whose normalizing constant is unknown), which is common in computational statistics.


Circle sampling.png

As a simple geometric example, suppose it is desired to generate a random point within the unit circle. Generate a candidate point (x,y) where x and y are independent uniformly distributed between −1 and 1. If it so happens that x^2+y^2 \leq 1 then the point is within the unit circle and should be accepted. If not then this point should be rejected and another candidate should be generated.

The ziggurat algorithm, a more advanced example, is used to efficiently generate normally-distributed pseudorandom numbers.


The rejection sampling method generates sampling values from an arbitrary probability distribution function f(x) by using an instrumental distribution g(x), under the only restriction that f(x)< M g(x) where M>1 is an appropriate bound on f(x)/g(x).

Rejection sampling is usually used in cases where the form of f(x) makes sampling difficult. Instead of sampling directly from the distribution f(x), we use an envelope distribution Mg(x) where sampling is easier. These samples from Mg(x) are probabilistically accepted or rejected.

This method relates to the general field of Monte Carlo techniques, including Markov chain Monte Carlo algorithms that also use a proxy distribution to achieve simulation from the target distribution f(x). It forms the basis for algorithms such as the Metropolis algorithm.

The unconditional acceptance probability is the proportion of proposed samples which are accepted, which is Pr(u<f(x)/Mg(x)) = E[f(x)/Mg(x)] =  \int(f(x)/Mg(x))g(x)dx = (1/M)\int f(x)dx = 1/M. If M is low, fewer samples are rejected, and the required number of samples for the target distribution is obtained more quickly. Because M must be no less than the maximum of f(x)/g(x), the unconditional acceptance probability is higher the less that ratio varies, however to obtain acceptance probability 1, f(x)=g(x), which defeats the purpose of sampling.


The algorithm (used by John von Neumann and dating back to Buffon and his needle) is as follows:

  • Sample x from g(x) and u from U(0,1) (the uniform distribution over the unit interval)
  • Check whether or not u<f(x)/Mg(x).
    • If this holds, accept x as a realization of f(x);
    • if not, reject the value of x and repeat the sampling step.

The validation of this method is the envelope principle: when simulating the pair (x,v=u\cdot Mg(x)), one produces a uniform simulation over the subgraph of Mg(x). Accepting only pairs such that u<f(x)/Mg(x) then produces pairs (x,v) uniformly distributed over the subgraph of f(x) and thus, marginally, a simulation from f(x).

This means that, with enough replicates, the algorithm generates a sample from the desired distribution f(x). There are a number of extensions to this algorithm, such as the Metropolis algorithm and the combination with ratio-of-uniforms approach.[5]


Rejection sampling can lead to a lot of unwanted samples being taken if the function being sampled is highly concentrated in a certain region, for example a function that has a spike at some location. For many distributions, this problem can be solved using an adaptive extension (see adaptive rejection sampling). In addition, as the dimensions of the problem get larger, the ratio of the embedded volume to the "corners" of the embedding volume tends towards zero, thus a lot of rejections can take place before a useful sample is generated, thus making the algorithm inefficient and impractical. See curse of dimensionality. In high dimensions, it is necessary to use a different approach, typically a Markov chain Monte Carlo method such as Metropolis sampling or Gibbs sampling. (However, Gibbs sampling, which breaks down a multi-dimensional sampling problem into a series of low-dimensional samples, may use rejection sampling as one of its steps.)

Adaptive rejection sampling[edit]

For many distributions, finding a proposal distribution that includes the given distribution without a lot of wasted space is difficult. An extension of rejection sampling that can be used to overcome this difficulty and efficiently sample from a wide variety of distributions (provided that they have log-concave density functions, which is in fact the case for most of the common distributions—even those whose density functions are not concave themselves!) is known as adaptive rejection sampling (ARS).

There are three basic ideas to this technique as ultimately introduced by Gilks in 1992:[6]

  1. If it helps, define your envelope distribution in log space (e.g. log-probability or log-density) instead. That is, work with  h\left(x\right) = \mathrm{ log }\; g\left(x\right) instead of  g\left(x\right) directly.
    • Often, distributions that have algebraically messy density functions have reasonably simpler log density functions (i.e. when f\left( x \right) is messy,  \mathrm{log}\; f\left(x\right) may be easier to work with or, at least, closer to piecewise linear).
  2. Instead of a single uniform envelope density function, use a piecewise linear density function as your envelope instead.
    • Each time you have to reject a sample, you can use the value of  f \left(x \right) that you evaluated, to improve the piecewise approximation  h \left(x \right) . This therefore reduces the chance that your next attempt will be rejected. Asymptotically, the probability of needing to reject your sample should converge to zero, and in practice, often very rapidly.
    • As proposed, any time we choose a point that is rejected, we tighten the envelope with another line segment that is tangent to the curve at the point with the same x-coordinate as the chosen point.
    • A piecewise linear model of the proposal log distribution results in a set of piecewise exponential distributions (i.e. segments of one or more exponential distributions, attached end to end). Exponential distributions are well behaved and well understood. The logarithm of an exponential distribution is a straight line, and hence this method essentially involves enclosing the logarithm of the density in a series of line segments. This is the source of the log-concave restriction: if a distribution is log-concave, then its logarithm is concave (shaped like an upside-down U), meaning that a line segment tangent to the curve will always pass over the curve.
    • If not working in log space, a piecewise linear density function can also be sampled via triangle distributions [7]
  3. We can take even further advantage of the (log) concavity requirement, to potentially avoid the cost of evaluating f \left( x \right) when your sample is accepted.
    • Just like we can construct a piecewise linear upper bound (the "envelope" function) using the values of  h \left(x \right) that we had to evaluate in the current chain of rejections, we can also construct a piecewise linear lower bound (the "squeezing" function) using these values as well.
    • Before evaluating (the potentially expensive) f \left( x \right) to see if your sample will be accepted, we may already know if it will be accepted by comparing against the (ideally cheaper)  g_l \left( x \right) (or h_l \left( x \right) in this case) squeezing function that have available.
    • This squeezing step is optional, even when suggested by Gilks. At best it saves you from only one extra evaluation of your (messy and/or expensive) target density. However, presumably for particularly expensive density functions (and assuming the rapid convergence of the rejection rate toward zero) this can make a sizable difference in ultimate runtime.

The method essentially involves successively determining an envelope of straight-line segments that approximates the logarithm better and better while still remaining above the curve, starting with a fixed number of segments (possibly just a single tangent line). Sampling from a truncated exponential random variable is straightfoward. Just take the log of a uniform random variable (with appropriate interva and corresponding truncation).

Unfortunately, ARS can only be applied from sampling from log-concave target densities. For this reason, several extensions of ARS have been proposed in literature for tackling non-log-concave target distributions.[8][9][10][11] Furthermore, different combinations of ARS and the Metropolis-Hastings method have been designed in order to obtain a universal sampler that builds a self-tuning proposal densities (i.e., a proposal automatically constructed and adapted to the target). This class of methods are often called as Adaptive Rejection Metropolis Sampling (ARMS) algorithms [12][13][14] (see some suitable Matlab code). The resulting adaptive techniques can be always applied but the generated samples are correlated in this case (although the correlation vanishes quickly to zero as the number of iterations grows).

See also[edit]


  1. ^ Casella, George; Robert, Christian P.; Wells, Martin T. Generalized Accept-Reject sampling schemes. pp. 342–347. doi:10.1214/lnms/1196285403. 
  2. ^ Martino, Luca; Míguez, Joaquín (2010-11-01). "Generalized rejection sampling schemes and applications in signal processing". Signal Processing 90 (11): 2981–2995. doi:10.1016/j.sigpro.2010.04.025. 
  3. ^ Neal, Radford M. (2003). "Slice Sampling". Annals of Statistics 31 (3): 705–767. doi:10.1214/aos/1056562461. MR 1994729. Zbl 1051.65007. 
  4. ^ Bishop, Christopher (2006). "11.4: Slice sampling". Pattern Recognition and Machine Learning. Springer. ISBN 0-387-31073-8. 
  5. ^ Luengo, D.; Martino, L. "Efficient random variable generation: ratio of uniforms and polar rejection sampling". Electronics Letters 48 (6). doi:10.1049/el.2012.0206. 
  6. ^ Adaptive Rejection Sampling for Gibbs Sampling.
  7. ^ D.B. Thomas and W. Luk , Non-uniform random number generation through piecewise linear approximations, 2006.
  8. ^ Hörmann, Wolfgang (1995-06-01). "A Rejection Technique for Sampling from T-concave Distributions". ACM Trans. Math. Softw. 21 (2): 182–193. doi:10.1145/203082.203089. ISSN 0098-3500. 
  9. ^ Evans, M.; Swartz, T. (1998-12-01). "Random Variable Generation Using Concavity Properties of Transformed Densities". Journal of Computational and Graphical Statistics 7 (4): 514–528. doi:10.2307/1390680. 
  10. ^ Martino, Luca; Míguez, Joaquín (2010-08-25). "A generalization of the adaptive rejection sampling algorithm". Statistics and Computing 21 (4): 633–647. doi:10.1007/s11222-010-9197-9. ISSN 0960-3174. 
  11. ^ Görür, Dilan; Teh, Yee Whye (2011-01-01). "Concave-Convex Adaptive Rejection Sampling". Journal of Computational and Graphical Statistics 20 (3): 670–691. doi:10.1198/jcgs.2011.09058. ISSN 1061-8600. 
  12. ^ Gilks, W. R.; Best, N. G.; Tan, K. K. C. (1995-01-01). "Adaptive Rejection Metropolis Sampling within Gibbs Sampling". Journal of the Royal Statistical Society. Series C (Applied Statistics) 44 (4): 455–472. doi:10.2307/2986138. 
  13. ^ Meyer, Renate; Cai, Bo; Perron, François (2008-03-15). "Adaptive rejection Metropolis sampling using Lagrange interpolation polynomials of degree 2". Computational Statistics & Data Analysis 52 (7): 3408–3423. doi:10.1016/j.csda.2008.01.005. 
  14. ^ Martino, L.; Read, J.; Luengo, D. (2015-06-01). "Independent Doubly Adaptive Rejection Metropolis Sampling Within Gibbs Sampling". IEEE Transactions on Signal Processing 63 (12): 3123–3138. doi:10.1109/TSP.2015.2420537. ISSN 1053-587X. 
  • Robert, C.P. and Casella, G. "Monte Carlo Statistical Methods" (second edition). New York: Springer-Verlag, 2004.
  • J. von Neumann, "Various techniques used in connection with random digits. Monte Carlo methods", Nat. Bureau Standards, 12 (1951), pp. 36–38.