The Week in HPC Research

By Nicole Hemsoth

February 21, 2013

The top research stories of the week have been hand-selected from prominent journals and leading conference proceedings. Here’s another diverse set of items, including one on GPU programming, distributed file systems, exhaustive search with parallel agents, the benefits of invasive computing, and an HPC cloud proof-of-concept.

Extending OpenMP for GPU Programming

The International Journal of Computational Science and Engineering (Volume 8, Number 1/2013) includes an interesting research item from Seyong Lee (Computer Science and Mathematics Division, Oak Ridge National Laboratory) and Rudolf Eigenmann (School of Electrical and Computer Engineering, Purdue University). The duo have developed a directive-based OpenMP extension to address programmability and tunability issues relevant to the GPGPU developer community.

GPGPU computing provides an inexpensive parallel computing platform for compute-intensive applications, yet programming complexity can challenge developers hindering more widespread adoption, the authors note. “Even though the compute unified device architecture (CUDA) programming model offers better abstraction, developing efficient GPGPU code is still complex and error–prone,” they argue.

Thus the authors propose a new programming interface, called OpenMPC, comprised of standard OpenMP and a new set of compiler directives and environment variables that have been extended for CUDA. They argue that OpenMPC, a directive–based, high–level programming model, offers better programmability and tunability for GPGPU code.

“We have developed a fully automatic compilation and user–assisted tuning system supporting OpenMPC. In addition to a range of compiler transformations and optimisations, the system includes tuning capabilities for generating, pruning, and navigating the search space of compilation variants. Evaluation using 14 applications shows that our system achieves 75% of the performance of the hand–coded CUDA programmes (92% if excluding one exceptional case),” they write.

Next >>

Six Distributed File Systems

A trio of French scientists provide a thorough analysis of six distributed file systems in this recent 39-page research paper, appearing in the HAL/INRIA open archive. The authors, one from SysFera and two from Laboratoire MIS at the Universite de Picardie Jules Verne, start with the observation that a large number of HPC applications rely on distributed computing environments to process and analyze large amounts of data. (Examples provided include probabilistic analysis, weather forecasting and aerodynamic research.) They further note the emergence of new infrastructures designed to handle the increased computational demand. Most of these new architectures, the authors assert, involve some manner of distributed computing, such that the computing process is spread among the nodes of a large distributed computing platform.

Furthermore the team believes that the growing body of scientific data will likewise necessitate innovations in distributed storage. “Easy to use and reliable storage solutions” are essential for scientific computing, they argue, and the community already has a “well-tried solution to this issue,” in the form of Distributed File Systems (DFSs).

The paper offers a comparison of six modern DFSs as to three fundamental issues: scalability, transparency and fault tolerance. For their study, the authors selected popular, widely-used and frequently updated DFSs: HDFS, MooseFS, iRODS, Ceph, GlusterFS, and Lustre.

Next >>

Exhaustive Search with Parallel Agents

In a recent paper, Macedonia researcher Toni Draganov Stojanovski from University for Information Science And Technology in the Republic of Macedonia sets out to examine the performance of exhaustive search when it is conducted with many search agents working in parallel.

Stojanovski and his research team observe that the advance of manycore processors and more sophisticated distributed processing offers more opportunities for exhaustive search via the use of multiple search agents. While there are a selection of elegant algorithms available for solving complex problems, exhaustive search remains as the best or only solution for real-life problems with no regular structure.

The paper reviews the performance that is achieved using the exhaustive search approach in conjunction with several different search agents with special attention to the following parameters:

• Differences in speeds of search agents.

• Length of allocated search subregions.

• Type of communication between central server and agents.

The findings reveal that the performance of the search improves with the increase in the level of mutual assistance between agents. Furthermore, nearly identical performance outcomes can be achieved with homogeneous and heterogeneous search agents as long as “the lengths of subregions allocated to individual search regions follow the differences in the speeds of heterogeneous search agents.” The research team also demonstrate how to achieve the optimum search performance by means of increasing the dimension of the search region.

The work appears in the January issue of the Turkish Journal of Electrical Engineering & Computer Sciences.

Next >>

The Benefits of Invasive Computing

In their paper, titled Invasive Computing on High Performance Shared Memory Systems, three researchers from the Department of Informatics, at Garching, Germany, offer new approaches for improving the throughput of runtime-adaptive applications on cutting-edge HPC systems. Their work was published as a chapter in Facing the Multicore Challenge III.

According to the team, there are multiple issues at play:

A first issue is the, in general, missing information about the actual impact of unforeseeable workload by adaptivity and of the unknown number of time steps or iterations on the runtime of adaptive applications. Another issue is that resource scheduling on HPC systems is currently done before an application is started and remains unchanged afterwards, even in case of varying requirements. Furthermore, an application cannot be started after another running application allocated all resources.

The authors propose a solution that involves the design of algorithms that adapt their use of resources during runtime, e.g., by relinquishing or adding compute cores. In the event that concurrent applications are competing for resources, they recommend that an appropriate resource management solution be adopted.

To improve the throughput of runtime-adaptive applications, the computer scientists employed invasive paradigms that start applications and schedule resources during runtime. Scheduling work can be achieved through the use of a global resource manager, and scalability graphs help improve load balancing of multiple applications. In the case of adaptive simulations, several scalability graphs are employed.

The paper includes a proof-of-concept that demonstrates runtime/throughput results for a fully adaptive shallow-water simulation.

Next >>

Easy to Use Cloud Service

Among the many HPC cloud research pieces that were published this week was an Australian endeavor that seeks to transform complicated HPC applications into easy-to-use SaaS cloud services. Researchers Adam K.L. Wonga and Andrzej M. Goscinskia from the School of Information Technology at Deakin University in Australia set out to develop and test a unified framework for HPC applications as services in clouds.

The duo acknowledge the benefits of HPC cloud. Scalable, affordable and accessible on demand, the use of HPC resources in a cloud environment have been a natural fit for many scientific disciplines, including biology, medicine, chemistry, they write. Still they have observed a steep learning curve when it comes to preparing for and deploying HPC applications in the cloud. This they say has stood in the way of many innovative HPC-backed discoveries.

To remedy this situation and improve ease of use and access to HPC resources, the researchers are looking to the world of Web-based tools, but as they write “high-performance computational research are both unique and complex, which make the development of web-based tools for this research difficult.”

The paper describes their approach to developing a unified cloud framework – one that makes it easier for various domain users to deploy HPC applications in public clouds as services. Their proof-of-concept integrates three components:

(i) Amazon EC2 public cloud for providing HPC infrastructure.

(ii) a HPC service software library for accessing HPC resources.

(iii) the Galaxy web-based platform for exposing and accessing HPC application services.

The authors conclude that “this new approach can reduce the time and money needed to deploy, expose and access discipline HPC applications in clouds.”

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

Insights from Optimized Codes on Cineca’s Marconi

February 15, 2019

What can you do with 381,392 CPU cores? For Cineca, it means enabling computational scientists to expand a large part of the world’s body of knowledge from the nanoscale to the astronomic, from calculating quantum effe Read more…

By Ken Strandberg

What Will IBM’s AI Debater Learn from Its Loss?

February 14, 2019

The utility of IBM’s latest man-versus-machine gambit is debatable. At the very least its Project Debater got us thinking about the potential uses of artificial intelligence as a way of helping humans sift through al Read more…

By George Leopold

ClusterVision in Bankruptcy, Fate Uncertain

February 13, 2019

ClusterVision, European HPC specialists that have built and installed over 20 Top500-ranked systems in their nearly 17-year history, appear to be in the midst of bankruptcy proceedings. According to Dutch news site Drimb Read more…

By Tiffany Trader

HPE Extreme Performance Solutions

HPE Systems With Intel Omni-Path: Architected for Value and Accessible High-Performance Computing

Today’s high-performance computing (HPC) and artificial intelligence (AI) users value high performing clusters. And the higher the performance that their system can deliver, the better. Read more…

IBM Accelerated Insights

Medical Research Powered by Data

“We’re all the same, but we’re unique as well. In that uniqueness lies all of the answers….”

  • Mark Tykocinski, MD, Provost, Executive Vice President for Academic Affairs, Thomas Jefferson University

Getting the answers to what causes some people to develop diseases and not others is driving the groundbreaking medical research being conducted by the Computational Medicine Center at Thomas Jefferson University in Philadelphia. Read more…

South African Weather Service Doubles Compute and Triples Storage Capacity of Cray System

February 13, 2019

South Africa has made headlines in recent years for its commitment to HPC leadership in Africa – and now, Cray has announced another major South African HPC expansion. Cray has been awarded contracts with Eclipse Holdings Ltd. to upgrade the supercomputing system operated by the South African Weather Service (SAWS). Read more…

By Oliver Peckham

Insights from Optimized Codes on Cineca’s Marconi

February 15, 2019

What can you do with 381,392 CPU cores? For Cineca, it means enabling computational scientists to expand a large part of the world’s body of knowledge from th Read more…

By Ken Strandberg

ClusterVision in Bankruptcy, Fate Uncertain

February 13, 2019

ClusterVision, European HPC specialists that have built and installed over 20 Top500-ranked systems in their nearly 17-year history, appear to be in the midst o Read more…

By Tiffany Trader

UC Berkeley Paper Heralds Rise of Serverless Computing in the Cloud – Do You Agree?

February 13, 2019

Almost exactly ten years to the day from publishing of their widely-read, seminal paper on cloud computing, UC Berkeley researchers have issued another ambitious examination of cloud computing - Cloud Programming Simplified: A Berkeley View on Serverless Computing. The new work heralds the rise of ‘serverless computing’ as the next dominant phase of cloud computing. Read more…

By John Russell

Iowa ‘Grows Its Own’ to Fill the HPC Workforce Pipeline

February 13, 2019

The global workforce that supports advanced computing, scientific software and high-speed research networks is relatively small when you stop to consider the magnitude of the transformative discoveries it empowers. Technical conferences provide a forum where specialists convene to learn about the latest innovations and schedule face-time with colleagues from other institutions. Read more…

By Elizabeth Leake, STEM-Trek

Trump Signs Executive Order Launching U.S. AI Initiative

February 11, 2019

U.S. President Donald Trump issued an Executive Order (EO) today launching a U.S Artificial Intelligence Initiative. The new initiative - Maintaining American L Read more…

By John Russell

Celebrating Women in Science: Meet Four Women Leading the Way in HPC

February 11, 2019

One only needs to look around at virtually any CS/tech conference to realize that women are underrepresented, and that holds true of HPC. SC hosts over 13,000 H Read more…

By AJ Lauer

IBM Bets $2B Seeking 1000X AI Hardware Performance Boost

February 7, 2019

For now, AI systems are mostly machine learning-based and “narrow” – powerful as they are by today's standards, they're limited to performing a few, narro Read more…

By Doug Black

Assessing Government Shutdown’s Impact on HPC

February 6, 2019

After a 35-day federal government shutdown, the longest in U.S. history, government agencies are taking stock of the damage -- and girding for a potential secon Read more…

By Tiffany Trader

Quantum Computing Will Never Work

November 27, 2018

Amid the gush of money and enthusiastic predictions being thrown at quantum computing comes a proposed cold shower in the form of an essay by physicist Mikhail Read more…

By John Russell

Cray Unveils Shasta, Lands NERSC-9 Contract

October 30, 2018

Cray revealed today the details of its next-gen supercomputing architecture, Shasta, selected to be the next flagship system at NERSC. We've known of the code-name "Shasta" since the Argonne slice of the CORAL project was announced in 2015 and although the details of that plan have changed considerably, Cray didn't slow down its timeline for Shasta. Read more…

By Tiffany Trader

The Case Against ‘The Case Against Quantum Computing’

January 9, 2019

It’s not easy to be a physicist. Richard Feynman (basically the Jimi Hendrix of physicists) once said: “The first principle is that you must not fool yourse Read more…

By Ben Criger

AMD Sets Up for Epyc Epoch

November 16, 2018

It’s been a good two weeks, AMD’s Gary Silcott and Andy Parma told me on the last day of SC18 in Dallas at the restaurant where we met to discuss their show news and recent successes. Heck, it’s been a good year. Read more…

By Tiffany Trader

Intel Reportedly in $6B Bid for Mellanox

January 30, 2019

The latest rumors and reports around an acquisition of Mellanox focus on Intel, which has reportedly offered a $6 billion bid for the high performance interconn Read more…

By Doug Black

US Leads Supercomputing with #1, #2 Systems & Petascale Arm

November 12, 2018

The 31st Supercomputing Conference (SC) - commemorating 30 years since the first Supercomputing in 1988 - kicked off in Dallas yesterday, taking over the Kay Ba Read more…

By Tiffany Trader

Looking for Light Reading? NSF-backed ‘Comic Books’ Tackle Quantum Computing

January 28, 2019

Still baffled by quantum computing? How about turning to comic books (graphic novels for the well-read among you) for some clarity and a little humor on QC. The Read more…

By John Russell

Contract Signed for New Finnish Supercomputer

December 13, 2018

After the official contract signing yesterday, configuration details were made public for the new BullSequana system that the Finnish IT Center for Science (CSC Read more…

By Tiffany Trader

Leading Solution Providers

SC 18 Virtual Booth Video Tour

Advania @ SC18 AMD @ SC18
ASRock Rack @ SC18
DDN Storage @ SC18
HPE @ SC18
IBM @ SC18
Lenovo @ SC18 Mellanox Technologies @ SC18
NVIDIA @ SC18
One Stop Systems @ SC18
Oracle @ SC18 Panasas @ SC18
Supermicro @ SC18 SUSE @ SC18 TYAN @ SC18
Verne Global @ SC18

Deep500: ETH Researchers Introduce New Deep Learning Benchmark for HPC

February 5, 2019

ETH researchers have developed a new deep learning benchmarking environment – Deep500 – they say is “the first distributed and reproducible benchmarking s Read more…

By John Russell

ClusterVision in Bankruptcy, Fate Uncertain

February 13, 2019

ClusterVision, European HPC specialists that have built and installed over 20 Top500-ranked systems in their nearly 17-year history, appear to be in the midst o Read more…

By Tiffany Trader

IBM Quantum Update: Q System One Launch, New Collaborators, and QC Center Plans

January 10, 2019

IBM made three significant quantum computing announcements at CES this week. One was introduction of IBM Q System One; it’s really the integration of IBM’s Read more…

By John Russell

Nvidia’s Jensen Huang Delivers Vision for the New HPC

November 14, 2018

For nearly two hours on Monday at SC18, Jensen Huang, CEO of Nvidia, presented his expansive view of the future of HPC (and computing in general) as only he can do. Animated. Backstopped by a stream of data charts, product photos, and even a beautiful image of supernovae... Read more…

By John Russell

HPC Reflections and (Mostly Hopeful) Predictions

December 19, 2018

So much ‘spaghetti’ gets tossed on walls by the technology community (vendors and researchers) to see what sticks that it is often difficult to peer through Read more…

By John Russell

IBM Bets $2B Seeking 1000X AI Hardware Performance Boost

February 7, 2019

For now, AI systems are mostly machine learning-based and “narrow” – powerful as they are by today's standards, they're limited to performing a few, narro Read more…

By Doug Black

The Deep500 – Researchers Tackle an HPC Benchmark for Deep Learning

January 7, 2019

How do you know if an HPC system, particularly a larger-scale system, is well-suited for deep learning workloads? Today, that’s not an easy question to answer Read more…

By John Russell

Intel Confirms 48-Core Cascade Lake-AP for 2019

November 4, 2018

As part of the run-up to SC18, taking place in Dallas next week (Nov. 11-16), Intel is doling out info on its next-gen Cascade Lake family of Xeon processors, specifically the “Advanced Processor” version (Cascade Lake-AP), architected for high-performance computing, artificial intelligence and infrastructure-as-a-service workloads. Read more…

By Tiffany Trader

  • arrow
  • Click Here for More Headlines
  • arrow
Do NOT follow this link or you will be banned from the site!
Share This