# Magnitude (astronomy)

(Redirected from Second magnitude)
• Top: Light sources of different magnitudes. A very bright satellite flare can be seen in the night sky.
• Bottom: The Hubble Ultra-Deep Field detected objects as faint as 30th magnitude (left). Comet Borrelly, the colors show its brightness over the range of three orders of magnitude (right).

In astronomy, magnitude is a logarithmic measure of the brightness of an object, measured in a specific wavelength or passband, usually in the visible or near-infrared spectrum. An imprecise but systematic determination of the magnitude of objects was introduced in ancient times by Hipparchus.

Astronomers use two different definitions of magnitude: apparent magnitude and absolute magnitude. The apparent magnitude (m, or vmag for the visible spectrum) is the brightness of an object as it appears in the night sky from Earth, while the absolute magnitude (Mv, V and H) describes the intrinsic brightness of an object as it would appear if it were placed at a certain distance from Earth. This distance is 10 parsecs for stars and 1 astronomical unit for planets and small Solar System bodies. A minor planet's size is typically estimated based on its absolute magnitude in combination with its presumed albedo.[1]

The brighter an object appears, the lower the value of its magnitude, with the brightest objects reaching negative values. The Sun has an apparent magnitude of −27, the full moon −13, the brightest planet Venus measures −5, and Sirius, the brightest visible star in the night sky, is at −1.5. An apparent magnitude can also be assigned to man-made objects in Earth orbit. The brightest satellite flares are ranked at −9, and the International Space Station, ISS, appears at a magnitude of −6. The scale is logarithmic, and defined such that each step of one magnitude changes the brightness by a factor of the fifth root of 100, or approximately 2.512. For example, a magnitude 1 star is exactly a hundred times brighter than a magnitude 6 star, as the difference of five magnitude steps corresponds to 2.5125 or 100.[2]

## History

The magnitude system dates back roughly 2000 years to the Greek astronomer Hipparchus (or the Alexandrian astronomer Ptolemy—references vary) who classified stars by their apparent brightness, which they saw as size (magnitude means "bigness, size"[3]). To the unaided eye, a more prominent star such as Sirius or Arcturus appears larger than a less prominent star such as Mizar, which in turn appears larger than a truly faint star such as Alcor. In 1736, the mathematician John Keill described the ancient naked-eye magnitude system in this way:

The fixed Stars appear to be of different Bignesses, not because they really are so, but because they are not all equally distant from us.[note 1] Those that are nearest will excel in Lustre and Bigness; the more remote Stars will give a fainter Light, and appear smaller to the Eye. Hence arise the Distribution of Stars, according to their Order and Dignity, into Classes; the first Class containing those which are nearest to us, are called Stars of the first Magnitude; those that are next to them, are Stars of the second Magnitude ... and so forth, 'till we come to the Stars of the sixth Magnitude, which comprehend the smallest Stars that can be discerned with the bare Eye. For all the other Stars, which are only seen by the Help of a Telescope, and which are called Telescopical, are not reckoned among these six Orders. Altho' the Distinction of Stars into six Degrees of Magnitude is commonly received by Astronomers; yet we are not to judge, that every particular Star is exactly to be ranked according to a certain Bigness, which is one of the Six; but rather in reality there are almost as many Orders of Stars, as there are Stars, few of them being exactly of the same Bigness and Lustre. And even among those Stars which are reckoned of the brightest Class, there appears a Variety of Magnitude; for Sirius or Arcturus are each of them brighter than Aldebaran or the Bull's Eye, or even than the Star in Spica; and yet all these Stars are reckoned among the Stars of the first Order: And there are some Stars of such an intermedial Order, that the Astronomers have differed in classing of them; some putting the same Stars in one Class, others in another. For Example: The little Dog was by Tycho placed among the Stars of the second Magnitude, which Ptolemy reckoned among the Stars of the first Class: And therefore it is not truly either of the first or second Order, but ought to be ranked in a Place between both.[4]

Note that the brighter the star, the smaller the magnitude: Bright "first magnitude" stars are "1st-class" stars, while stars barely visible to the naked eye are "sixth magnitude" or "6th-class". The system was a simple delineation of stellar brightness into six distinct groups but made no allowance for the variations in brightness within a group.

Tycho Brahe attempted to directly measure the “bigness” of the stars in terms of angular size, which in theory meant that a star's magnitude could be determined by more than just the subjective judgment described in the above quote. He concluded that first magnitude stars measured 2 arc minutes (2′) in apparent diameter (130 of a degree, or 115 the diameter of the full moon), with second through sixth magnitude stars measuring 1 12′, 1 112′, 34′, 12′, and 13′, respectively.[5] The development of the telescope showed that these large sizes were illusory—stars appeared much smaller through the telescope. However, early telescopes produced a spurious disk-like image of a star that was larger for brighter stars and smaller for fainter ones. Astronomers from Galileo to Jaques Cassini mistook these spurious disks for the physical bodies of stars, and thus into the eighteenth century continued to think of magnitude in terms of the physical size of a star.[6] Johannes Hevelius produced a very precise table of star sizes measured telescopically, but now the measured diameters ranged from just over six seconds of arc for first magnitude down to just under 2 seconds for sixth magnitude.[6][7] By the time of William Herschel astronomers recognized that the telescopic disks of stars were spurious and a function of the telescope as well as the brightness of the stars, but still spoke in terms of a star's size more than its brightness.[6] Even well into the nineteenth century the magnitude system continued to be described in terms of six classes determined by apparent size, in which

There is no other rule for classing the stars but the estimation of the observer; and hence it is that some astronomers reckon those stars of the first magnitude which others esteem to be of the second.[8]

However, by the mid-nineteenth century astronomers had measured the distances to stars via stellar parallax, and so understood that stars are so far away as to essentially appear as point sources of light. Following advances in understanding the diffraction of light and astronomical seeing, astronomers fully understood both that the apparent sizes of stars were spurious and how those sizes depended on the intensity of light coming from a star (this is the star's apparent brightness, which can be measured in units such as watts/cm2) so that brighter stars appeared larger.

### Modern definition

Early photometric measurements (made, for example, by using a light to project an artificial “star” into a telescope’s field of view and adjusting it to match real stars in brightness) demonstrated that first magnitude stars are about 100 times brighter than sixth magnitude stars.

Thus in 1856 Norman Pogson of Oxford proposed that a logarithmic scale of ‹The template Sqrt is being considered for merging.› 5100 ≈ 2.512 be adopted between magnitudes, so five magnitude steps corresponded precisely to a factor of 100 in brightness.[9][10] Every interval of one magnitude equates to a variation in brightness of ‹The template Sqrt is being considered for merging.› 5100 or roughly 2.512 times. Consequently, a first magnitude star is about 2.5 times brighter than a second magnitude star, 2.52 brighter than a third magnitude star, 2.53 brighter than a fourth magnitude star, and so on.

This is the modern magnitude system, which measures the brightness, not the apparent size, of stars. Using this logarithmic scale, it is possible for a star to be brighter than “first class”, so Arcturus or Vega are magnitude 0, and Sirius is magnitude −1.46.

## Scale

As mentioned above, the scale appears to work 'in reverse', with objects with a negative magnitude being brighter than those with a positive magnitude. The 'larger' the negative value, the brighter.

Objects appearing farther to the left on this line are brighter, while objects appearing farther to the right are dimmer. Thus zero appears in the middle, with the brightest objects on the far left, and the dimmest objects on the far right.

## Apparent and absolute magnitude

Two of the main types of magnitudes distinguished by astronomers are:

The difference between these concepts can be seen by comparing two stars. Betelgeuse (apparent magnitude 0.5, absolute magnitude -5.8) appears slightly dimmer in the sky than Alpha Centauri (apparent magnitude 0.0, absolute magnitude 4.4) even though it emits thousands of times more light, because Betelgeuse is much farther away. Usually only apparent magnitude is mentioned since it can be measured directly. Absolute magnitude can be calculated from apparent magnitude and distance from:

${\displaystyle m-M=5\left(\log _{10}d-1\right).}$

This is known as the distance modulus, where d is the distance to the star measured in parsecs, m is the apparent magnitude, and M is the absolute magnitude.

If the line of sight between the object and observer is affected by extinction due to absorption of light by interstellar dust particles, then the object's apparent magnitude will be correspondingly fainter. For A magnitudes of extinction, the relationship between apparent and absolute magnitudes becomes

${\displaystyle m-M=5\left(\log _{10}d-1\right)+A.}$

Other magnitude scales exist, such as bolometric magnitude.

### Apparent magnitude

Under the modern logarithmic magnitude scale, two objects, one of which is used as a reference or baseline, whose intensities (brightnesses) measured from Earth in units of power per unit area (such as watts per square metre, W m−2) are I1 and Iref, will have magnitudes m1 and mref related by

${\displaystyle m_{1}-m_{\rm {ref}}=-2.5\log _{10}\left({\frac {I_{1}}{I_{\rm {ref}}}}\right).}$

Using this formula, the magnitude scale can be extended beyond the ancient magnitude 1–6 range, and it becomes a precise measure of brightness rather than simply a classification system. Astronomers can now measure differences as small as one-hundredth of a magnitude. Stars that have magnitudes between 1.5 and 2.5 are called second-magnitude; there are some 20 stars brighter than 1.5, which are first-magnitude stars (see the list of brightest stars). For example, Sirius is magnitude −1.46, Arcturus is −0.04, Aldebaran is 0.85, Spica is 1.04, and Procyon (the little Dog) is 0.34. Under the ancient magnitude system, all of these stars might have been classified as "stars of the first magnitude".

Magnitudes can also be calculated for objects far brighter than stars (such as the Sun and Moon), and for objects too faint for the human eye to see (such as Pluto).

### Examples

The following is a table giving apparent magnitudes for celestial objects and artificial satellites ranging from the Sun to the faintest object visible with the Hubble Space Telescope (HST):

Apparent
magnitude
Brightness
relative to
magnitude 0
Example Apparent
magnitude
Brightness
relative to
magnitude 0
Example Apparent
magnitude
Brightness
relative to
magnitude 0
Example
−27 6.31×1010 Sun −7 631 SN 1006 supernova 13 6.31×10−6 3C 273 quasar
limit of 4.5–6 in (11–15 cm) telescopes
−26 2.51×1010 −6 251 ISS (max.) 14 2.51×10−6 Pluto (max.)
limit of 8–10 in (20–25 cm) telescopes
−25 1×1010 −5 100 Venus (max.) 15 1×10−6
−24 3.98×109 −4 39.8 Faintest objects visible during the day with the naked eye when the sun is high[11] 16 3.98×10−7 Charon (max.)
−23 1.58×109 −3 15.8 Jupiter (max.), Mars (max.) 17 1.58×10−7
−22 6.31×108 −2 6.31 Mercury (max.) 18 6.31×10−8
−21 2.51×108 −1 2.51 Sirius 19 2.51×10−8
−20 1×108 0 1 Vega, Saturn (max.) 20 1×10−8
−19 3.98×107 1 0.398 Antares 21 3.98×10−9 Callirrhoe (satellite of Jupiter)
−18 1.58×107 2 0.158 Polaris 22 1.58×10−9
−17 6.31×106 3 0.0631 Cor Caroli 23 6.31×10−10
−16 2.51×106 4 0.0251 Acubens 24 2.51×10−10
−15 1×106 5 0.01 Vesta (max.), Uranus (max.) 25 1×10−10 Fenrir (satellite of Saturn)
−14 3.98×105 6 3.98×10−3 typical limit of naked eye[note 2] 26 3.98×10−11
−13 1.58×105 full moon 7 1.58×10−3 Ceres (max.) 27 1.58×10−11 visible light limit of 8m telescopes
−12 6.31×104 8 6.31×10−4 Neptune (max.) 28 6.31×10−12
−11 2.51×104 9 2.51×10−4 29 2.51×10−12
−10 1×104 10 1×10−4 typical limit of 7×50 binoculars 30 1×10−12
−9 3.98×103 Iridium flare (max.) 11 3.98×10−5 Proxima Centauri 31 3.98×10−13
−8 1.58×103 12 1.58×10−5 32 1.58×10−13 visible light limit of HST

### Other scales

Under the Vega system for measuring the brightness of astronomical brightness, the star Vega is defined to have an apparent magnitude of zero as measured through all filters, although this is only an approximation e.g. its actual brightness has been measured to be 0.03 in the V (visual) band. The brightest star, Sirius, has a Vega magnitude of −1.46. or −1.5. However, Vega has been found to vary in brightness, and other standards are in common use.[12] One such system is the AB magnitude system, in which the reference is a source with a constant flux density per unit frequency. Another is the STMAG system, in which the reference source is instead defined to have constant flux density per unit wavelength.

## Problems

The human eye is easily fooled, and Hipparchus's scale has had problems. For example, the human eye is more sensitive to yellow and red light than to blue, and photographic film more to blue than to yellow/red, giving different values of visual magnitude and photographic magnitude. Apparent magnitude can also be affected by factors such as dust in the atmosphere or light cloud cover absorbing some of the light.

Furthermore, many people find it counter-intuitive that a high magnitude star is dimmer than a low magnitude star.

## Notes

1. ^ Today astronomers know that the brightness of stars is a function of both their distance and their own luminosity.
2. ^ Under very dark skies, such as are found in remote rural areas

## References

1. ^ "Glossary—Absolute magnitude (H)". NASA. 21 August 2015.
2. ^ "Apparent & absolute magnitude". ESA—educational support. 14 May 2013.
3. ^ Heifetz, M.; Tirion, W. (2004). A Walk Through the Heavens: A Guide to Stars and Constellations and Their Legends. Cambridge: Cambridge University Press. p. 6.
4. ^ Keill, J. (1739). An introduction to the true astronomy (3rd ed.). London. pp. 47–48.
5. ^ Thoren, V. E. (1990). The Lord of Uraniborg. Cambridge: Cambridge University Press. p. 306.
6. ^ a b c Graney, C. M.; Grayson, T. P. (2011). "On the Telescopic Disks of Stars: A Review and Analysis of Stellar Observations from the Early 17th through the Middle 19th Centuries". Annals of Science. 68 (3): 351–373. doi:10.1080/00033790.2010.507472.
7. ^ Graney, C. M. (2009). "17th Century Photometric Data in the Form of Telescopic Measurements of the Apparent Diameters of Stars by Johannes Hevelius". Baltic Astronomy. 18 (3–4): 253–263. Bibcode:2009BaltA..18..253G. arXiv:.
8. ^ Ewing, A.; Gemmere, J. (1812). Practical Astronomy. Burlington, NJ: Allison. p. 41.
9. ^ Hoskin, M. (1999). The Cambridge Concise History of Astronomy. Cambridge: Cambridge University Press. p. 258.
10. ^ Tassoul, J. L.; Tassoul, M. (2004). A Concise History of Solar and Stellar Physics. Princeton, NJ: Princeton University Press. p. 47.
11. ^ http://sky.velp.info/daystars.php
12. ^ Milone, E. F. (2011). Astronomical Photometry: Past, Present and Future. New York: Springer. pp. 182–184. ISBN 978-1-4419-8049-6.