Pleiades (supercomputer)

From Wikipedia, the free encyclopedia
Jump to: navigation, search
Pleiades (supercomputer)
Pleiades supercomputer.jpg
Active 2008 – Present
Sponsors National Aeronautics and Space Administration (NASA),  USA
Operators NAS, SGI
Location NASA Advanced Supercomputing Division at NASA Ames Research Center, Moffett Field, California
Architecture 163 SGI Altix ICE 8400/X racks (11,176 nodes), 184,800 Intel Xeon processors, InfiniBand QDR/FDR interconnect
Operating system Linux[1]
Memory 502 terabytes
Storage 15 petabytes (RAID)
Speed 1.54 petaflops (sustained), 3.59 petaflops (peak)
Ranking TOP500: 16, November 2013
Legacy Ranked Third in TOP500 LINPACK at 487 teraflops, November 2008
Web site NAS Pleiades Homepage

Pleiades (pronounced /pliː ə diːz/ or /pʰ le diːz/) is a petascale supercomputer housed at the NASA Advanced Supercomputing (NAS) facility at NASA Ames Research Center located at Moffett Field near Mountain View, California. It is maintained by NASA and partners Silicon Graphics (SGI) and Intel.

As of November 2013 it is ranked the world's sixteenth fastest computer on the TOP500 list with a LINPACK rating of 1.54 petaflops (1.54 quadrillion floating point operations per second)[2] and as of April 2014, it has a peak performance of 3.59 petaflops following a large hardware upgrade.[3] The ultimate goal of the partnership between NASA and SGI is to attain a peak supercomputing performance of 10 petaflops, or 10 quadrillion floating point operations per second, in order to "increase the computational capabilities for research, [and] modeling and simulation work at the NASA Advanced Supercomputing (NAS) facility."[4]

History[edit]

Anatomy of a Pleiades node, shown on display at the NASA Ames Exploration Center, in Mountain View, California

Built in 2008 and named for the Pleiades open star cluster, the supercomputer debuted as the third most powerful supercomputer in the world at 487 teraflops.[5] It originally contained 100 SGI Altix ICE 8200EX racks with 12,800 Intel Xeon quad-core E5472 Harpertown processors connected with more than 20 miles of InfiniBand double data rate (DDR) cabling.[6]

With the addition of ten more racks of quad-core X5570 Nehalem processors in 2009, Pleiades ranked sixth on the November 2009 TOP500 with 14,080 processors running at 544 teraflops.[7] In January 2010, the scientists and engineers at NAS successfully completed a “live integration” of another ICE 8200 rack by connecting the new rack’s InfiniBand dual port fabric via 44 fibre cables while the supercomputer was still running a full workload, saving 2 million hours in productivity that would previously have been lost.[8]

Another expansion in 2010 added 32 new SGI Altix ICE 8400 racks with Intel Xeon six-core X5670 Westmere processors, bringing up to 18,432 processors (81,920 cores in 144 racks) at a theoretical peak of 973 teraflops and a LINPACK rating of 773 teraflops.[9] NASA also put an emphasis on keeping Pleiades energy efficient, increasing the power efficiency with each expansion so that in 2010 it was three times more power-efficient than the original 2008 components, which were the most power-efficient at the time. The integration of the six-core Westmere nodes also required new quad data rate (QDR) and hybrid DDR/QDR InfiniBand cabling, making the world's largest InfiniBand interconnect network with more than 45 miles of cable.[9]

After another 14 ICE 8400 racks containing Westmere processors were added in 2011, Pleiades ranked seventh on the TOP500 list in June of that year at a LINPACK rating of 1.09 petaflops, or 1.09 quadrillion floating point operations per second.[10]

InfiniBand DDR and QDR fiber cables are used to connect the all of nodes to each other, as well as to the mass storage systems at NAS and the hyperwall visualization system, creating a network made up of more than 65 miles of InfiniBand fabric, the largest of its kind in the world. Pleiades is built in a partial 11-D hypercube technology, where each node has eleven connections to eleven other nodes, with some making up to twelve connections to form a 12-D hypercube.[11]

In 2012, NASA and partners SGI and Intel began working on the integration of 24 new Altix ICE X racks with Intel Xeon eight-core E5-2760 Sandy Bridge processors to replace 27 of the original Alitx 8200 racks containing quad-core Harpertown processors. With a total of 126,720 processor cores and over 233 terabytes of RAM across 182 racks, the expansion increased Pleiades' available computing capacity 40 percent.[12] Each new Sandy Bridge node has four networking links using fourteen data rate (FDR) InfiniBand cable for a total transfer bandwidth of 56 gigabits (about 7 gigabytes) per second.[13]

In early 2013, work began on a larger hardware refresh for Pleiades, ultimately removing all of the original 4-core Harpertown processors and adding 46 SGI ICE X racks with 10-core Intel Xeon E5-2680V2 (Ivy Bridge) processors. When installation was complete in August 2013, the system's overall peak performance increased 62% from 1.78 petaflops to 2.87 petaflops. [14] The system was slowly upgraded again between January and April of 2014, adding another 29 racks of Ivy Bridge nodes and increasing the system's theoretical computational capability to 3.59 petaflops.[3] To make room for the expansion, all of the system's remaining Nehalem nodes and 12 Westmere nodes were removed.

Role at NASA[edit]

Pleiades is part of NASA’s High-End Computing Capability (HECC) Project and represents NASA’s state-of-the-art technology for meeting the agency’s supercomputing requirements, enabling NASA scientists and engineers to conduct high-fidelity modeling and simulation for NASA missions in Earth studies, space science, aeronautics research, as well as human and robotic space exploration.[15]

Some of the scientific and engineering projects run on Pleiades include:

  • The Kepler Mission, a space observatory launched in March 2009 to locate Earth-like planets, monitors a section of space containing more than 200,000 stars and takes high-resolution images every 30 minutes. After the operations center gathers this data, it is pipelined to Pleiades in order to calculate the size, orbit, and location of the planets surrounding these stars.[16] As of February 2012, the Kepler mission has discovered 1,235 planets, 5 of which are approximately Earth-sized and orbit within the "habitable zone" where water can exist in all three forms (solid, liquid, gas).[17] After setbacks following the failure of two of Kepler's four reaction wheels, responsible for keeping the spacecraft pointed in the correct direction, in 2013, the Kepler team moved the entire data pipeline to Pleiades, which continues to run light curve analyses from the existing Kepler data.[18]
  • Research and development of next generation space launch vehicles is done on Pleiades using cutting-edge analysis tools and computational fluid dynamics (CFD) modeling and simulation in order to create more efficient and affordable space launch system and vehicle designs. Research has also been done on reducing noise created by the landing gear of aircraft using CDF code application to detect where the sources of noise are within the structures.[19]
  • Astrophysics research into the formation of galaxies is run on Pleiades to create simulations of how our own Milky Way Galaxy was formed and what forces might have caused it to form in its signature disk-shape. [20] Pleiades has also been the supercomputing resource for dark matter research and simulation, helping to discover gravitationally bound “clumps” of dark matter within galaxies in one of the largest simulations ever done, in terms of particle numbers.[21]
  • Visualization of the Earth's ocean currents using a NASA-built data synthesis model for the Estimating the Circulation and Climate of the Ocean (ECCO) Project between MIT and the NASA Jet Propulsion Laboratory in Pasadena, California. According to NASA, the "ECCO model-data syntheses are being used to quantify the ocean’s role in the global carbon cycle, to understand the recent evolution of the polar oceans, to monitor time-evolving heat, water, and chemical exchanges within and between different components of the Earth system, and for many other science applications."[22]

Gallery[edit]

References[edit]

  1. ^ http://top500.org/system/177259
  2. ^ "TOP500 November 2013 List". TOP500. 
  3. ^ a b "Pleiades Upgrade Completed". NASA's High-End Computing Capability Project. 
  4. ^ "NASA Selects Next-Generation SGI ICE HPC Platform for Technology Upgrade". Silicon Graphics November 2011. 
  5. ^ "TOP500 Pleiades 8200 System Ranking". TOP500. 
  6. ^ "NASA Supercomputer Ranks Among World’s Fastest". NASA November 2008. 
  7. ^ "NASA Supercomputer Ranks Among World’s Fastest". NASA November 2008. 
  8. ^ "’Live’ Integration of Pleiades Rack Saves 2 Million Hours". NAS February 2010. 
  9. ^ a b "NASA Supercomputer Doubles Capacity, Increases Efficiency". NASA June 2010. 
  10. ^ "NASA's Pleiades Supercomputer Ranks Among World's Fastest". NASA June 2011. Archived from the original on 2011-10-21. 
  11. ^ "NAS/HECC User Knowledge Base Pleiades Interconnect". NAS. 
  12. ^ "Pleiades Supercomputer Gets a Little More Oomph". NAS June 2012. 
  13. ^ "HECC Project Hardware Overview: Sandy Bridge Processors". NAS. 
  14. ^ "NASA's Pleiades Supercomputer Upgraded, Harpertown Nodes Repurposed". NAS August 2013. 
  15. ^ "NASA Advanced Supercomputing Division Mission Goals". NAS. 
  16. ^ Law, Gillian (2012). "Simulating the Stars". Scientific Computing World. February/March (22): 41–43. 
  17. ^ "NASA Finds Earth-Size Planet Candidates In Habitable Zone, Six Planet System". NASA February 2012. Archived from the original on 2011-10-19. 
  18. ^ "Kepler Lives On: It's All in the Data". NASA@SC13 November 2013. 
  19. ^ Khorrami, Mehdi R. (2011). "High-Fidelity Simulation of Landing Gear Noise". HPC Source (Scientific Computing). Spring 2011: 18–22. 
  20. ^ "First Glimpse into the Birth of the Milky Way". University of Zurich August 2011. 
  21. ^ "A Dark Matter for Astrophysics Research". HPCwire May 2011. 
  22. ^ "NASA Views Our Perpetual Ocean". NASA April 2012. Archived from the original on 2012-05-05. 

External links[edit]