Atomic clock

From Wikipedia, the free encyclopedia
Jump to: navigation, search
For a clock updated by radio signals which is sometimes called an "atomic clock", see Radio clock. For the clock as a measure for risk of catastrophic destruction, see Doomsday Clock. For the album by Zion I, see Atomic Clock (album).
Atomic clock
FOCS-1.jpg
FOCS 1, a continuous cold caesium fountain atomic clock in Switzerland, started operating in 2004 at an uncertainty of one second in 30 million years.
Classification Clock
Industry Telecommunications, Science
Application GPS
Fuel source Electricity
Powered Yes
Inventor US National Bureau of Standards
Invented 1949
The master atomic clock ensemble at the U.S. Naval Observatory in Washington D.C., which provides the time standard for the U.S. Department of Defense.[1] The rack mounted units in the background are Symmetricom (formerly HP) 5071A caesium beam clocks. The black units in the foreground are Symmetricom (formerly Sigma-Tau) MHM-2010 hydrogen maser standards.

An atomic clock is a clock device that uses an electronic transition frequency in the microwave, optical, or ultraviolet region[2] of the electromagnetic spectrum of atoms as a frequency standard for its timekeeping element. Atomic clocks are the most accurate time and frequency standards known, and are used as primary standards for international time distribution services, to control the wave frequency of television broadcasts, and in global navigation satellite systems such as GPS.

The principle of operation of an atomic clock is not based on nuclear physics, but rather on atomic physics and using the microwave signal that electrons in atoms emit when they change energy levels. Early atomic clocks were based on masers at room temperature. Currently, the most accurate atomic clocks first cool the atoms to near absolute zero temperature by slowing them with lasers and probing them in atomic fountains in a microwave-filled cavity. An example of this is the NIST-F1 atomic clock, one of the U.S.'s national primary time and frequency standards.

The accuracy of an atomic clock depends on two factors. The first factor is temperature of the sample atoms—colder atoms move much more slowly, allowing longer probe times. The second factor is the frequency and intrinsic width of the electronic transition. Higher frequencies and narrow lines increase the precision.

National standards agencies in most industrialised and semi-industrialised countries maintain an accuracy of 10−9 seconds per day (approximately 1 part in 1014), and a precision set by the radio transmitter pumping the maser. These clocks collectively define a continuous and stable time scale, International Atomic Time (TAI). For civil time, another time scale is disseminated, Coordinated Universal Time (UTC). UTC is derived from TAI, but approximately synchronised, by using leap seconds, to UT1, which is based on actual rotations of the Earth with respect to the solar time.

History[edit]

The idea of using atomic transitions to measure time was first suggested by Lord Kelvin in 1879.[3] Magnetic resonance, developed in the 1930s by Isidor Rabi, became the practical method for doing this.[4] In 1945, Rabi first publicly suggested that atomic beam magnetic resonance might be used as the basis of a clock.[5] The first atomic clock was an ammonia maser device built in 1949 at the U.S. National Bureau of Standards (NBS, now NIST). It was less accurate than existing quartz clocks, but served to demonstrate the concept.[6] The first accurate atomic clock, a caesium standard based on a certain transition of the caesium-133 atom, was built by Louis Essen in 1955 at the National Physical Laboratory in the UK.[7] Calibration of the caesium standard atomic clock was carried out by the use of the astronomical time scale ephemeris time (ET).[8] This led to the internationally agreed definition of the latest SI second being based on atomic time. Equality of the ET second with the (atomic clock) SI second has been verified to within 1 part in 1010.[9] The SI second thus inherits the effect of decisions by the original designers of the ephemeris time scale, determining the length of the ET second.

May 2009- JILA's strontium optical atomic clock is based on neutral atoms. Shining a blue laser onto ultracold strontium atoms in an optical trap tests how efficiently a previous burst of light from a red laser has boosted the atoms to an excited state. Only those atoms that remain in the lower energy state respond to the blue laser, causing the fluorescence seen here.[10]

Since the beginning of development in the 1950s, atomic clocks have been based on the hyperfine transitions in hydrogen-1, cesium-133, and rubidium-87. The first commercial atomic clock was the Atomichron, manufactured by the National Company. More than 50 were sold between 1956 and 1960. This bulky and expensive instrument was subsequently replaced by much smaller rack-mountable devices, such as the Hewlett-Packard model 5060 caesium frequency standard, released in 1964.[4]

In the late 1990s four factors contributed to major advances in clocks:[11]

Chip-scale atomic clocks, such as this one unveiled in 2004, are expected to greatly improve GPS location.

In August 2004, NIST scientists demonstrated a chip-scale atomic clock.[12] According to the researchers, the clock was believed to be one-hundredth the size of any other. It requires no more than 125 mW,[13] making it suitable for battery-driven applications. This technology became available commercially in 2011.[13] Ion trap experimental optical clocks are more precise than the current cesium standard.

Mechanism[edit]

Since 1967, the International System of Units (SI) has defined the second as the duration of 9192631770cycles of radiation corresponding to the transition between two energy levels of the caesium-133 atom.[14]

This definition makes the caesium oscillator the primary standard for time and frequency measurements, called the caesium standard. Other physical quantities, e.g., the volt and the metre, rely on the definition of the second in their own definitions.[15]

The actual time-reference of an atomic clock consists of an electronic oscillator operating at microwave frequency. The oscillator is arranged so that its frequency-determining components include an element that can be controlled by a feedback signal. The feedback signal keeps the oscillator tuned in resonance with the frequency of the electronic transition of caesium or rubidium.

The core of the atomic clock is a tunable microwave cavity containing the gas. In a hydrogen maser clock the gas emits microwaves (the gas mases) on a hyperfine transition, the field in the cavity oscillates, and the cavity is tuned for maximum microwave amplitude. Alternatively, in a caesium or rubidium clock, the beam or gas absorbs microwaves and the cavity contains an electronic amplifier to make it oscillate. For both types the atoms in the gas are prepared in one electronic state prior to filling them into the cavity. For the second type the number of atoms which change electronic state is detected and the cavity is tuned for a maximum of detected state changes.

Most of the complexity of the clock lies in this adjustment process. The adjustment tries to correct for unwanted side-effects, such as frequencies from other electron transitions, temperature changes, and the spreading in frequencies caused by ensemble effects.[clarification needed] One way of doing this is to sweep the microwave oscillator's frequency across a narrow range to generate a modulated signal at the detector. The detector's signal can then be demodulated to apply feedback to control long-term drift in the radio frequency. In this way, the quantum-mechanical properties of the atomic transition frequency of the caesium can be used to tune the microwave oscillator to the same frequency, except for a small amount of experimental error. When a clock is first turned on, it takes a while for the oscillator to stabilize. In practice, the feedback and monitoring mechanism is much more complex than described above.

Historical accuracy of atomic clocks from NIST

A number of other atomic clock schemes are in use for other purposes. Rubidium standard clocks are prized for their low cost, small size (commercial standards are as small as 17 cm3)[13] and short-term stability. They are used in many commercial, portable and aerospace applications. Hydrogen masers (often manufactured in Russia) have superior short-term stability compared to other standards, but lower long-term accuracy.

Often, one standard is used to fix another. For example, some commercial applications use a rubidium standard periodically corrected by a global positioning system receiver. This achieves excellent short-term accuracy, with long-term accuracy equal to (and traceable to) the U.S. national time standards.

The lifetime of a standard is an important practical issue. Modern rubidium standard tubes last more than ten years, and can cost as little as US$50.[citation needed] Caesium reference tubes suitable for national standards currently last about seven years and cost about US$35,000. The long-term stability of hydrogen maser standards decreases because of changes in the cavity's properties over time.

Modern clocks use magneto-optical traps to cool the atoms for improved precision.

Physics package realisations[edit]

A number of methods exist for utilizing hyperfine atomic transitions. These methods, with their respective benefits and drawbacks, have influenced the development of commercial devices and laboratory standards. By tradition, the hardware that is used to probe the atoms is called the physics package.

One of the actual Hewlett Packard HP5061A cesium beam atomic clock units used in the Hafele–Keating "flying clock" experiment in 1971 to test relativity theory.

Atomic beam standard[edit]

The atomic beam standard is a direct extension of the Stern-Gerlach atomic splitting experiment. The atoms of choice are heated in an oven to create gas, which is collimated into a beam. This beam passes through a state-selector magnet A, where atoms of the wrong state are separated out from the beam. The beam is exposed to an RF field at or near the transition. The beam then passes through a space before it is again exposed to the RF field. The RF field and a static homogeneous magnetic field from the C-field coil will change the state of the atoms. After the second RF field exposure the atomic beam passes through a second state selector magnet B, where the atom state being selected out of the beam at the A magnet is being selected. This way, the detected amount of atoms will relate to the ability to match the atomic transition. After the second state-selector a mass-spectrometer using an ionizer will detect the rate of atoms being received.

Modern variants of this beam mechanism use optical pumping to transition all atoms to the same state rather than dumping half the atoms. Optical detection using scintillation can also be used.

The most common isotope for beam devices is caesium (133Cs), but rubidium (87Rb) and thallium (205Tl) are examples of others used in early research.

The frequency errors can be made very small for a beam device, or predicted (such as the magnetic field pull of the C-coil) in such a way that a high degree of repeatability and stability can be achieved. This is why an atomic beam can be used as a primary standard.

Atomic gas cell standard[edit]

The atomic gas cell standard builds on a confined reference isotope (often an alkali metal such as Rubidium (87Rb)) inside an RF cavity. The atoms are excited to a common state using optical pumping; when the applied RF field is swept over the hyperfine spectrum, the gas will absorb the pumping light, and a photodetector provides the response. The absorption peak steers the fly-wheel oscillator.

A typical rubidium gas-cell uses a rubidium (87Rb) lamp heated to 108-110 degrees Celsius, and an RF field to excite it to produce light, where the D1 and D2 lines are the significant wavelengths. An 85Rb cell filters out the D1 line so that only the D2 line pumps the 87Rb gas cell in the RF cavity.

Among the significant frequency pulling mechanisms inherent to the gas cell are wall-shift, buffer-gas shift, cavity-shift and light-shift. The wall-shift occurs as the gas bumps into the wall of the glass container. Wall-shift can be reduced by wall coating and compensation by buffer gas. The buffer gas shift comes from the reference atoms which bounce into buffer gas atoms such as neon and argon; these shifts can be both positive and negative. The cavity shift comes from the RF cavity, which can deform the resonance amplitude response; this depends upon cavity centre frequency and resonator Q-value. Light-shift is an effect where frequency is pulled differently depending on the light intensity, which often is modulated by the temperature shift of the rubidium lamp and filter cell.

There are thus many factors in which temperature and ageing can shift frequency over time, and this is why a gas cell standard is unfit for a primary standard, but can become a very inexpensive, low-power and small-size solution for a secondary standard or where better stability compared to crystal oscillators is needed, but not the full performance of a caesium beam standard. The rubidium gas standards have seen use in telecommunications systems and portable instruments.

Active maser standard[edit]

The active maser standard is a development from the atomic beam standard in which the observation time was incremented by using a bounce-box. By controlling the beam intensity spontaneous emission will provide sufficient energy to provide a continuous oscillation, which is being tapped and used as a reference for a fly-wheel oscillator.

The active maser is sensitive to wall-shift and cavity pulling. The wall-shift is mitigated by using PTFE coating (or other suitable coating) to reduce the effect. The cavity pulling effect can be reduced by automatic cavity tuning. In addition the magnetic field pulls the frequency.

While not being long-term stable as caesium beams, it remains one of the most stable sources available. The inherent pulling effects makes repeatability troublesome and does prohibit its use as being primary standard, but it makes an excellent secondary standard. It is used as low-noise fly-wheel standard for caesium beam standards.

Fountain standard[edit]

The fountain standard is a development from the beam standard where the beam has been folded back to itself such that the first and second RF field becomes the same RF cavity. A ball of atoms is laser cooled, which reduces black body temperature shifts. Phase errors between RF cavities are essentially removed. The length of the beam is longer than many beams, but the speed is also much slower such that the observation time becomes significantly longer and hence a higher Q value is achieved in the Ramsey fringes.

Cesium fountains have been implemented in many laboratories, but rubidium has even greater ability to provide stability in the fountain configuration.

Ion trap standard[edit]

The ion trap standard is a set of different approaches, but their common property is that a cooled ion is confined in an electrostatic trap. The hyperfine region of the available electron is then being tracked similar to that of a gas cell standard.

Ion traps have been used for numerous ions. 199Hg+ was an early candidate. Quantum logic spectroscopy of a single Al ion became the most precise[16] in 2008. In 2010 an improved setup using a Mg+ logic ion instead of Be was demonstrated[17]

Power consumption[edit]

The power consumption of atomic clocks varies with their size. Atomic clocks on the scale of one chip require less than 30 mW;[18] The United States Time Standard atomic clocks, NIST-F1 and NIST-F2, use far greater quantities of power.[12][19]

Research[edit]

A caesium atomic clock from 1975 (upper unit) and battery backup (lower unit).[20]

Most research focuses on the often conflicting goals of making the clocks smaller, cheaper, more accurate, and more reliable.

Optical clocks[edit]

New technologies, such as femtosecond frequency combs, optical lattices, and quantum information, have enabled prototypes of next-generation atomic clocks. These clocks are based on optical rather than microwave transitions. A major obstacle to developing an optical clock is the difficulty of directly measuring optical frequencies. This problem has been solved with the development of self-referenced mode-locked lasers, commonly referred to as femtosecond frequency combs. Before the demonstration of the frequency comb in 2000, terahertz techniques were needed to bridge the gap between radio and optical frequencies, and the systems for doing so were cumbersome and complicated. With the refinement of the frequency comb, these measurements have become much more accessible and numerous optical clock systems are now being developed around the world.

As in the radio range, absorption spectroscopy is used to stabilize an oscillator—in this case a laser. When the optical frequency is divided down into a countable radio frequency using a femtosecond comb, the bandwidth of the phase noise is also divided by that factor. Although the bandwidth of laser phase noise is generally greater than stable microwave sources, after division it is less.

The two primary systems under consideration for use in optical frequency standards are single ions isolated in an ion trap and neutral atoms trapped in an optical lattice.[21] These two techniques allow the atoms or ions to be highly isolated from external perturbations, thus producing an extremely stable frequency reference.

Optical clocks have already achieved better stability and lower systematic uncertainty than the best microwave clocks.[21] This puts them in a position to replace the current standard for time, the cesium fountain clock.[22]

Atomic systems under consideration include Al+, Hg+/2+,[21] Hg, Sr, Sr+/2+, In+/3+, Mg, Ca, Ca+, Yb+/2+/3+ and Yb.

The rare-earth element ytterbium (Yb) is valued not so much for its mechanical properties but for its complement of internal energy levels. "A particular transition in Yb atoms, at a wavelength of 578 nm, currently provides one of the world's most accurate optical atomic frequency standards," said Marianna Safronova.[23] The estimated amount of uncertainty achieved corresponds to a Yb clock uncertainty of about one second over the lifetime of the universe so far, 15 billion years, according to scientists at the Joint Quantum Institute (JQI) and the University of Delaware in December 2012.

In 2013 optical lattice clocks (OLCs) were shown to be as good as or better than cesium fountain clocks. Two optical lattice clocks containing about 10 000 atoms of strontium-87 were able to stay in synchrony with each other at a precision of at least 1.5 × 10-16, which is as accurate as the experiment could measure. These clocks have been shown to keep pace with all three of the caesium clocks at the Paris Observatory. There are two reasons for the possibly better precision. Firstly, the frequency is measured using light, which has a much higher frequency than microwaves, and secondly, by using many atoms, any errors are averaged.[24] Using ytterbium atoms, a new record for stability with a precision of 1.6 × 10-18 was published on 22 August 2013. At this stability, the two optical lattice clocks the NIST research team used would differ less than a tenth of a second over the age of the universe; this was 10 times better than previous experiments. The clock relies on 10 000 ytterbium atoms cooled to 10 microkelvin and trapped in an optical lattice. A laser at 578 nm excites the atoms between two of their energy levels.[25] Having established the stability of the clocks, the researchers are studying external influences and evaluating the remaining systematic uncertainties, in the hope that they can bring the clock's accuracy down to the level of its stability.[26] An improved OLC was described in a 2014 Nature paper.[27]

Quantum clocks[edit]

Further information: Quantum clock

In March 2008, physicists at NIST described a quantum logic clock based on individual ions of beryllium and aluminium. This clock was compared to NIST's mercury ion clock. These were the most accurate clocks that had been constructed, with neither clock gaining nor losing time at a rate that would exceed a second in over a billion years.[28] In February 2010, NIST physicists described a second, enhanced version of the quantum logic clock based on individual ions of magnesium and aluminium. Considered the world's most precise clock, it offers more than twice the precision of the original.[29] [30]

Evaluated accuracy[edit]

In 2011, the NPL-CsF2 caesium fountain clock operated by the National Physical Laboratory (NPL), which serves as the United Kingdom primary time and frequency standard, was improved regarding the two largest sources of measurement uncertainties — distributed cavity phase and microwave lensing frequency shifts. As of 2011 this resulted in an evaluated frequency uncertainty reduction from 4.1 x 10−16 to 2.3 x 10−16 — the lowest value for any primary national standard at the time.[31] At this frequency uncertainty, the NPL-CsF2 is expected to neither gain nor lose a second in more than 138 million years.[32][33][34]

The NIST-F2 caesium fountain clock operated by the National Institute of Standards and Technology (NIST), was launched in April 2014, to serve as a new U.S. civilian time and frequency standard, along with the NIST-F1 standard. "NIST-F2 would neither gain nor lose one second in about 300 million years, making it about three times as accurate as NIST-F1, which has served as the standard since 1999."[35] NIST-F2 was designed using lessons learned from NIST-F1. The key advance is that the vertical flight tube is now chilled inside a container of liquid nitrogen, at −193 °C (−315.4 °F). This cycled cooling dramatically lowers the background radiation and thus reduces some of the very small measurement errors that must be corrected in NIST-F1.[36][37]

Applications[edit]

The development of atomic clocks has led to many scientific and technological advances such as a worldwide system of precise position measurement (the Global Positioning System), and applications in the Internet, which depend critically on frequency and time standards. Atomic clocks are installed at sites of time signal radio transmitters. They are used at some long wave and medium wave broadcasting stations to deliver a very precise carrier frequency.[citation needed] Atomic clocks are used in many scientific disciplines, such as for long-baseline interferometry in radioastronomy.[38]

Global Positioning System[edit]

The Global Positioning System (GPS) provides very accurate timing and frequency signals. A GPS receiver works by measuring the relative time delay of signals from a minimum of four, but usually more, GPS satellites, each of which has at least two onboard caesium and as many as two rubidium atomic clocks. The relative times are mathematically transformed into three absolute spatial coordinates and one absolute time coordinate. The time is accurate to within about 50 nanoseconds. However, inexpensive GPS receivers may not assign a high priority to updating the display, so the displayed time may differ perceptibly from the internal time. Precision time references that use GPS are marketed for use in computer networks, laboratories, and cellular communications networks, and do maintain accuracy to within about 50ns.

Time signal radio transmitters[edit]

A radio clock is a clock that automatically synchronizes itself by means of government radio time signals received by a radio receiver. Many retailers market radio clocks inaccurately as atomic clocks;[39] although the radio signals they receive originate from atomic clocks, they are not atomic clocks themselves. They are inexpensive time-keeping devices with an accuracy of about a second. Instrument grade time receivers provide higher accuracy. Such devices incur a transit delay of approximately 1 ms for every 300 kilometres (186 mi) of distance from the radio transmitter. Many governments operate transmitters for time-keeping purposes.

See also[edit]

References[edit]

  1. ^ USNO Master Clock
  2. ^ McCarthy, Dennis; Seidelmann, P. Kenneth (2009). TIME from Earth Rotation to Atomic Physics. Weinheim: Wiley-VCH. ch. 10 & 11. 
  3. ^ Sir William Thomson (Lord Kelvin) and Peter Guthrie Tait, Treatise on Natural Philosophy, 2nd ed. (Cambridge, England: Cambridge University Press, 1879), vol. 1, part 1, page 227.
  4. ^ a b M.A. Lombardi, T.P. Heavner, S.R. Jefferts (2007). "NIST Primary Frequency Standards and the Realization of the SI Second". Journal of Measurement Science 2 (4): 74. 
  5. ^ See:
    • Isidor I. Rabi, "Radiofrequency spectroscopy" (Richtmyer Memorial Lecture, delivered at Columbia University in New York, New York, on 20 January 1945).
    • "Meeting at New York, January 19 and 20, 1945" Physical Review, vol. 67, pages 199-204 (1945).
    • William L. Laurence, "'Cosmic pendulum' for clock planned," New York Times, 21 January 1945, page 34. Reprinted on page 77 of: Lombardi, Michael A.; Heavner, Thomas P.; and Jefferts, Steven R. (December 2007) "NIST primary frequency standards and the realization of the SI second," NCSLI Measure, vol. 2, no. 4, pages 74-89.
  6. ^ D.B. Sullivan (2001). "Time and frequency measurement at NIST: The first 100 years". "2001 IEEE International Frequency Control Symposium". NIST. pp. 4–17. 
  7. ^ Essen, L.; Parry, J. V. L. (1955). "An Atomic Standard of Frequency and Time Interval: A Cæsium Resonator". Nature 176 (4476): 280. Bibcode:1955Natur.176..280E. doi:10.1038/176280a0.  edit
  8. ^ W. Markowitz, R.G. Hall, L. Essen, J.V.L. Parry (1958). "Frequency of cesium in terms of ephemeris time". Physical Review Letters 1: 105–107. Bibcode:1958PhRvL...1..105M. doi:10.1103/PhysRevLett.1.105. 
  9. ^ W. Markowitz (1988). "Comparisons of ET(Solar), ET(Lunar), UT and TDT'". In A.K. Babcock, G.A. Wilkins. "The Earth's Rotation and Reference Frames for Geodesy and Geophysics, International Astronomical Union Symposia #128". pp. 413–418. . Pages 413–414, gives the information that the SI second was made equal to the second of ephemeris time as determined from lunar observations, and was later verified in this relation, to 1 part in 1010.
  10. ^ D. Lindley (20 May 2009). "Coping With Unusual Atomic Collisions Makes an Atomic Clock More Accurate". National Science Foundation. Retrieved 10 July 2009. 
  11. ^ J. Ye, H. Schnatz, L.W. Hollberg (2003). "Optical frequency combs: From frequency metrology to optical phase control". IEEE Journal of Selected Topics in Quantum Electronics 9 (4): 1041. [dead link]
  12. ^ a b "Chip-Scale Atomic Devices at NIST". NIST. 2007. Archived from the original on 7 January 2008. Retrieved 17 January 2008.  Available on-line at: NIST.gov
  13. ^ a b c "SA.45s CSAC Chip Scale Atomic Clock". 2011. Retrieved 12 June 2013. 
  14. ^ "International System of Units (SI)" (8th ed.). Bureau International des Poids et Mesures. 2006. 
  15. ^ "FAQs". Franklin Instrument Company. 2007. Archived from the original on 16 February 2008. Retrieved 17 January 2008. 
  16. ^ P. O. Schmidt, T. Rosenband, C. Langer, W. M. Itano, J. C. Bergquist, D. J. Wineland (2005). "Spectroscopy using quantum logic". Science 309, 749. 
  17. ^ "Frequency Comparison of Two High-Accuracy Al+ Optical Clocks". Phys Rev Lett. 10 Feb 2010. doi:10.1103/PhysRevLett.104.070802. 
  18. ^ Lutwak, Robert (26–29 November 2007). "The Chip-Scale Atomic Clock — Prototype Evaluation". 36th Annual Precise Time and Time Interval (PTTI) Systems and Applications Meeting. 
  19. ^ http://www.nist.gov/pml/div688/nist-f2-atomic-clock-040314.cfm
  20. ^ "President Piñera Receives ESO's First Atomic Clock". ESO Announcement. 15 November 2013. Retrieved 20 November 2013. 
  21. ^ a b c W.H. Oskay et al. (2006). "Single-atom optical clock with high accuracy". Physical Review Letters 97 (2): 020801. Bibcode:2006PhRvL..97b0801O. doi:10.1103/PhysRevLett.97.020801. PMID 16907426. 
  22. ^ "Time for a change? Super precise optical lattice clock could 'redefine the second' and be used as a standard for the world's time". Daily Mail. 10 July 2013. Retrieved 10 July 2013. 
  23. ^ "Blackbody Radiation Shift: Quantum Thermodynamics Will Redefine Clocks". Retrieved 5 December 2012. 
  24. ^ "Precise atomic clock may redefine time". 9 July 2013. Retrieved 24 August 2013. 
  25. ^ "NIST Ytterbium Atomic Clocks Set Record for Stability". 22 August 2013. Retrieved 24 August 2013. 
  26. ^ "New atomic clock sets the record for stability". 27 August 2013. Retrieved 19 January 2014. 
  27. ^ "An optical lattice clock with accuracy and stability at the 10−18 level". Nature. 22 January 2014. doi:10.1038/nature12941. 
  28. ^ "NIST 'Quantum Logic Clock' Rivals Mercury Ion as World's Most Accurate Clock". PhysOrg.com. 6 March 2008. Retrieved 24 October 2009. 
  29. ^ NIST's Second 'Quantum Logic Clock' Based on Aluminum Ion is Now World's Most Precise Clock, NIST, 4 February 2010
  30. ^ C.W Chou, D. Hume, J.C.J. Koelemeij, D.J. Wineland, and T. Rosenband (17 February 2010). "Frequency Comparison of Two High-Accuracy Al+ Optical Clocks". NIST. Retrieved 9 February 2011. 
  31. ^ Evaluation of the frequency of the H-maser 1401708 by the primary frequency standard NPL-CsF2, National Physical Laboratory, February 2010
  32. ^ NPL’s atomic clock revealed to be the world’s most accurate
  33. ^ NPL-CsF2: now the atomic clock with the world's best long-term accuracy
  34. ^ Improved accuracy of the NPL-CsF2 primary frequency standard: evaluation of distributed cavity phase and microwave lensing frequency shifts, Ruoxin Lia, Kurt Gibblea, and Krzysztof Szymaniecb, August 2011
  35. ^ "NIST launches a new US time standard: NIST-F2 atomic clock". phys.org. 3 April 2014. Retrieved 3 April 2014. 
  36. ^ "NIST launches a new US time standard: NIST-F2 atomic clock". nist.gov. 3 April 2014. Retrieved 3 April 2014. 
  37. ^ "Background: How NIST-F2 Works". nist.gov. 2 April 2014. Retrieved 4 April 2014. 
  38. ^ McCarthy, D. D.; Seidelmann, P. K. (2009). TIME—From Earth Rotation to Atomic Physics. Weinheim: Wiley-VCH Verlag GmbH & Co. KGaA. p. 266. ISBN 978-3-527-40780-4. 
  39. ^ Michael A. Lombardi, "How Accurate is a Radio Controlled Clock?, National Institute of Standards and Technology, 2010.

External links[edit]