In computer graphics, the rendering equation is an integral equation in which the equilibrium radiance leaving a point is given as the sum of emitted plus reflected radiance under a geometric optics approximation. It was simultaneously introduced into computer graphics by David Immel et al. and James Kajiya in 1986. The various realistic rendering techniques in computer graphics attempt to solve this equation.
The physical basis for the rendering equation is the law of conservation of energy. Assuming that L denotes radiance, we have that at each particular position and direction, the outgoing light (Lo) is the sum of the emitted light (Le) and the reflected light. The reflected light itself is the sum from all directions of the incoming light (Li) multiplied by the surface reflection and cosine of the incident angle.
The rendering equation may be written in the form
- is a particular wavelength of light
- is time
- is the location in space
- is the direction of the outgoing light
- is the negative direction of the incoming light
- is the total spectral radiance of wavelength directed outward along direction at time , from a particular position
- is emitted spectral radiance
- is the unit hemisphere containing all possible values for
- is an integral over
- is the bidirectional reflectance distribution function, the proportion of light reflected from to at position , time , and at wavelength
- is spectral radiance of wavelength coming inward toward from direction at time
- is the weakening factor of inward irradiance due to incident angle, as the light flux is smeared across a surface whose area is larger than the projected area perpendicular to the ray
Two noteworthy features are: its linearity—it is composed only of multiplications and additions, and its spatial homogeneity—it is the same in all positions and orientations. These mean a wide range of factorings and rearrangements of the equation are possible.
Note this equation's spectral and time dependence — may be sampled at or integrated over sections of the visible spectrum to obtain, for example, a trichromatic color sample. A pixel value for a single frame in an animation may be obtained by fixing motion blur can be produced by averaging over some given time interval (by integrating over the time interval and dividing by the length of the interval).
Solving the rendering equation for any given scene is the primary challenge in realistic rendering. One approach to solving the equation is based on finite element methods, leading to the radiosity algorithm. Another approach using Monte Carlo methods has led to many different algorithms including path tracing, photon mapping, and Metropolis light transport, among others.
Although the equation is very general, it does not capture every aspect of light reflection. Some missing aspects include the following:
- Transmission, which occurs when light is transmitted through the surface, like for example when it hits a glass object or a water surface,
- Subsurface scattering, where the spatial locations for incoming and departing light are different. Surfaces rendered without accounting for subsurface scattering may appear unnaturally opaque — however, it is not necessary to account for this if transmission is included in the equation, since that will effectively include also light scattered under the surface,
- Polarization, where different light polarizations will sometimes have different reflection distributions, for example when light bounces at a water surface,
- Phosphorescence, which occurs when light or other electromagnetic radiation is absorbed at one moment in time and emitted at a later moment in time, usually with a longer wavelength (unless the absorbed electromagnetic radiation is very intense),
- Interference, where the wave properties of light are exhibited,
- Fluorescence, where the absorbed and emitted light have different wavelengths,
- Non-linear effects, where very intense light can increase the energy level of an electron with more energy than that of a single photon (this can occur if the electron is hit by two photons at the same time), and emission of light with higher frequency than the frequency of the light that hit the surface suddenly becomes possible, and
- Relativistic Doppler effect, where light that bounces on an object that is moving in a very high speed will get its wavelength changed; if the light bounces at an object that is moving towards it, the impact will compress the photons, so the wavelength will become shorter and the light will be blueshifted and the photons will be packed more closely so the photon flux will be increased; if it bounces at an object that is moving away from it, it will be redshifted and the photons will be packed more sparsely so the photon flux will be decreased.
For scenes that are either not composed of simple surfaces in a vacuum or for which the travel time for light is an important factor, researchers have generalized the rendering equation to produce a volume rendering equation suitable for volume rendering and a transient rendering equation for use with data from a time-of-flight camera.
- Immel, David S.; Cohen, Michael F.; Greenberg, Donald P. (1986), "A radiosity method for non-diffuse environments", Siggraph 1986: 133, doi:10.1145/15922.15901, ISBN 0-89791-196-2
- Kajiya, James T. (1986), "The rendering equation", Siggraph 1986: 143, doi:10.1145/15922.15902, ISBN 0-89791-196-2
- Owen, Scott (September 5, 1999). "Reflection: Theory and Mathematical Formulation". Retrieved 2008-06-22.
- Kajiya, James T.; Von Herzen, Brian P. (1984), "Ray tracing volume densities", Siggraph 1984 18 (3): 165, doi:10.1145/964965.808594
- Smith, Adam M.; Skorupski, James, Davis, James (2008). Transient Rendering (Technical report). UC Santa Cruz. UCSC-SOE-08-26.
- Lecture notes from Stanford University course CS 348B, Computer Graphics: Image Synthesis Techniques