Visit additional Tabor Communication Publications
September 01, 2009
Large-memory resource first of its kind among major HPC systems
Sept. 1 -- Leveraging lightning-fast technology already familiar to many from the micro storage world of digital cameras, thumb drives and laptop computers, the San Diego Supercomputer Center (SDSC) at the University of California, San Diego today unveiled a "super-sized" version -- a "flash" memory-based supercomputer that accelerates investigation of a wide range of data-intensive science problems.
The new High-Performance Computing (HPC) system, dubbed "Dash," is an element of the Triton Resource, an integrated, data-intensive resource primarily designed to support UC San Diego and UC researchers that went online earlier this summer. As envisioned, this "system within a system" will help researchers looking for solutions to particularly data-intensive problems that arise in astrophysics, genomics and many other domains of science.
While Dash, which already has begun trial runs, is a medium-sized system as supercomputers go with a peak speed of 5.2 teraflops (TF), it has several unique properties, including the first use of flash memory technology in an HPC system, using Intel High-Performance SATA Solid-State Drives. Four of its nodes are specially configured as I/O nodes each serving up 1 terabyte (TB) of flash memory to any other node, courtesy of new I/O controllers also developed by Intel Corporation and integrated by Appro International, Inc. (One terabyte equals one trillion bytes of storage capacity).
The system features 68 Appro GreenBlade servers with dual-socket quad-core Intel Xeon processor 5500 series (formerly codenamed Nehalem) nodes linked to an InfiniBand interconnect. In its current configuration, Dash has 48 gigabytes (GB) of DRAM memory on each node, and employs vSMP Foundation software from ScaleMP, Inc. that provides virtual symmetric multiprocessing capabilities and aggregates memory across 16 nodes into shared memory "supernodes," giving users access to as much as 768 GB of shared DRAM memory in addition to 1 TB of flash memory per "supernode".
"Dash's use of flash memory for fast file-access and swap space -- as opposed to spinning discs that have much slower latency or I/O times -- along with vSMP capabilities for large shared memory will facilitate scientific research," said Michael Norman, interim director of SDSC. "Today's high-performance instruments, simulations and sensor networks are creating a deluge of data that presents formidable challenges to store and analyze, challenges that Dash helps to overcome."
For example, Dash will have the capability to search sky survey data for near-earth asteroids and brown dwarfs that may help researchers better understand periodic extinctions on Earth, and it will speed up investigations to establish relationships among species based on their genes. Such research not only could yield new information regarding evolution, but help biomedical researchers mine these complex data sets for clues to develop new drugs or cures for a variety of diseases.
"Dash can do random data accesses one order-of-magnitude faster than other machines," said Allan Snavely, associate director at SDSC. "This means it can solve data-mining problems that are looking for the proverbial 'needle in the haystack' more than 10 times faster than could be done on even much larger supercomputers that still rely on older 'spinning disk' technology."
Dash is currently being tested but soon will be made available to users of the TeraGrid, the nation's largest open-access scientific discovery infrastructure, for evaluation and development of application codes that can take advantage of flash memory and virtual "supernodes" technology. For additional information about access and allocations, see www.teragrid.org.
As an organized research unit of UC San Diego, SDSC is a national leader in creating and providing cyberinfrastructure for data-intensive research. Cyberinfrastructure refers to an accessible and integrated network of computer-based resources and expertise, focused on accelerating scientific inquiry and discovery. SDSC recently doubled its size to 160,000 square feet with a new, energy-efficient building and datacenter extension, and is a founding member of TeraGrid, the nation's largest open-access scientific discovery infrastructure.
San Diego Supercomputer Center (SDSC): www.sdsc.edu
UC San Diego: www.ucsd.edu
In quieter times, sounding the bell of funding big science with big systems tends to resonate further than when ears are already burning with sour economic and national security news. For exascale's future, however, the time could be ripe to instill some sense of urgency....
In a recent solicitation, the NSF laid out needs for furthering its scientific and engineering infrastructure with new tools to go beyond top performance, Having already delivered systems like Stampede and Blue Waters, they're turning an eye to solving data-intensive challenges. We spoke with the agency's Irene Qualters and Barry Schneider about..
Large-scale, worldwide scientific initiatives rely on some cloud-based system to both coordinate efforts and manage computational efforts at peak times that cannot be contained within the combined in-house HPC resources. Last week at Google I/O, Brookhaven National Lab’s Sergey Panitkin discussed the role of the Google Compute Engine in providing computational support to ATLAS, a detector of high-energy particles at the Large Hadron Collider (LHC).
May 23, 2013 |
The study of climate change is one of those scientific problems where it is almost essential to model the entire Earth to attain accurate results and make worthwhile predictions. In an attempt to make climate science more accessible to smaller research facilities, NASA introduced what they call ‘Climate in a Box,’ a system they note acts as a desktop supercomputer.
May 22, 2013 |
At some point in the not-too-distant future, building powerful, miniature computing systems will be considered a hobby for high schoolers, just as robotics or even Lego-building are today. That could be made possible through recent advancements made with the Raspberry Pi computers.
May 16, 2013 |
When it comes to cloud, long distances mean unacceptably high latencies. Researchers from the University of Bonn in Germany examined those latency issues of doing CFD modeling in the cloud by utilizing a common CFD and its utilization in HPC instance types including both CPU and GPU cores of Amazon EC2.
May 15, 2013 |
Supercomputers at the Department of Energy’s National Energy Research Scientific Computing Center (NERSC) have worked on important computational problems such as collapse of the atomic state, the optimization of chemical catalysts, and now modeling popping bubbles.
05/10/2013 | Cleversafe, Cray, DDN, NetApp, & Panasas | From Wall Street to Hollywood, drug discovery to homeland security, companies and organizations of all sizes and stripes are coming face to face with the challenges – and opportunities – afforded by Big Data. Before anyone can utilize these extraordinary data repositories, however, they must first harness and manage their data stores, and do so utilizing technologies that underscore affordability, security, and scalability.
04/15/2013 | Bull | “50% of HPC users say their largest jobs scale to 120 cores or less.” How about yours? Are your codes ready to take advantage of today’s and tomorrow’s ultra-parallel HPC systems? Download this White Paper by Analysts Intersect360 Research to see what Bull and Intel’s Center for Excellence in Parallel Programming can do for your codes.
In this demonstration of SGI DMF ZeroWatt disk solution, Dr. Eng Lim Goh, SGI CTO, discusses a function of SGI DMF software to reduce costs and power consumption in an exascale (Big Data) storage datacenter.
The Cray CS300-AC cluster supercomputer offers energy efficient, air-cooled design based on modular, industry-standard platforms featuring the latest processor and network technologies and a wide range of datacenter cooling requirements.