Dark rubik cube

Major supercomputer drive for US in 2015

Pro
(Source: Stockfresh)

14 March 2014

Supercomputing speed is typically boosted by adding more processors, but two new systems funded by the US National Science Foundation (NSF) due to go live next January will take an unconventional approach to speed up calculations and data analysis.

Arrays of memory and flash storage, with potentially petabytes of storage, will be loaded on the Wrangler supercomputer at Texas Advanced Computing Centre (TACC) at the University of Texas at Austin and the Comet supercomputer at the San Diego Supercomputer Centre (SDSC) at the University of California. The supercomputers, which are currently under construction, have a new design with high levels of storage relative to the number of processors in the system.

The supercomputers will provide better throughput, in-memory and caching features, which could be a faster and more efficient way to solve complex problems, said NSF in a budget request published as part of President Barack Obama’s 2015 $3.9 trillion (€2.8 billion) budget proposal sent to Congress.

The new batch of supercomputers will support research in disciplines such as economics, geosciences, medicine, earthquake engineering and climate and weather modelling.

NSF is requesting $7 billion (€5 billion) to fund scientific research, of which $894 million (€643 million) is dedicated to research in areas such as software, chip manufacturing, semiconductors, cybersecurity and cognitive computing systems. NSF also funds the construction of supercomputers so scientists have access to computing resources for simulation and other tasks. The supercomputers are being built as part of NSF’s Extreme Digital (XD) program, in which scientists share computing resources to advance research.

Compared to what NSF has funded in the past, including IBM’s Blue Waters, the new servers have a different design, said Dan Olds, principal analyst at Gabriel Consulting Group.

Processors and other computing resources already deliver high levels of performance, but the real bottleneck has been throughput. NSF wants more sophisticated supercomputing designs so bits and bytes move between processing elements faster, Olds said.

“It has to do with the changing nature of high-performance computing,” Olds said. “They want to control massive data streams instead of handling batch [jobs].”

The Comet supercomputer is more “suitable for both high throughput and data-intensive computing,” NSF said. “Its heterogeneous configuration will support not only complex simulations, but also advanced analytics and visualisation of output.”

Servers are increasingly packing large arrays of DRAM for in-memory computing, which is considered beneficial for databases and other data-intensive applications. Solid-state drives are being used as a cache layer on which data is temporarily stored before being processed. SSDs are also becoming primary storage at the expense of hard drives, which are slower and more power hungry.

Comet will be built by Dell, have 1,024 processor cores, a massive 7PB array of high-performance storage and 6PB of “durable storage for data reliability,” according to specifications published by SDSC. The supercomputer will use Intel Xeon chips and Nvidia graphics processors. Each node will have 128GB of memory and 320GB of flash, though it is unclear how many nodes the supercomputer will have. There will also be special nodes with 1.5TB of memory. It will have 100 Gigabit Ethernet and the InfiniBand interconnect for throughput. The system is built on the Lustre file system, which is designed to overcome bottlenecks on distributed computing systems.

“The Comet project…is designed to efficiently deliver significant computing capacity (two petaflops) for the 98% of research that requires fewer than 1,000 simultaneous and tightly coupled cores to be conducted,” NSF said.

SDSC is not saying much more about Comet as it goes through validation and deployment, said Jan Zverina, director of communications and media relations at the Centre, in an email. More details are likely to be shared later this year, Zverina said.

TACC’s Wrangler will combine 120 servers with Intel-based Xeon server chips code-named Haswell. It was touted by NSF as the “most powerful data analysis system allocated in XD, with 10 petabytes (PB) of replicated, secure, high performance data storage.” It will have 3,000 processing cores dedicated to data analysis, and flash storage layers for analytics. The supercomputer’s bandwidth will be 1 terabytes per second (TBps) and 275 million input/output operations per second (IOPS).

NSF’s research priorities are relevant to the problems faced in computing today, Olds said, adding that the government agency is heading in the right direction on supercomputer development.

 

Agam Shah, IDG News Service

Read More:


Back to Top ↑

TechCentral.ie