Signal-to-noise ratio (imaging)

From Wikipedia, the free encyclopedia
Jump to navigation Jump to search

Signal-to-noise ratio (SNR) is used in imaging to characterize image quality. The sensitivity of a (digital or film) imaging system is typically described in the terms of the signal level that yields a threshold level of SNR.

Industry standards define sensitivity in terms of the ISO film speed equivalent, using SNR thresholds (at average scene luminance) of 40:1 for "excellent" image quality and 10:1 for "acceptable" image quality.[1]

SNR is sometimes quantified in decibels (dB) of signal power relative to noise power, though in the imaging field the concept of "power" is sometimes taken to be the power of a voltage signal proportional to optical power; so a 20 dB SNR may mean either 10:1 or 100:1 optical power, depending on which definition is in use.

Definition of SNR[edit]

Traditionally, SNR is defined to be the ratio of the average signal value to the standard deviation of the signal [citation needed]:

when the signal is an optical intensity, or as the square of this value if the signal and noise are viewed as amplitudes (field quantities).

See also[edit]

References[edit]

  1. ^ ISO 12232: 1997 Photography – Electronic Still Picture Cameras – Determining ISO Speed

Further reading[edit]