Since 1986 - Covering the Fastest Computers in the World and the People Who Run Them

Language Flags

Topics » Developer Tools


Altair Crash Tests New Xeon E7s

Mar 12, 2014 |

RADIOSS, one of the top structural analysis solvers in the industry, is used by auto companies like Ford and PSA in France to run car crash and other simulations to test safety and viability of a broad array of products. Like similar CAE codes, it is highly parallelized for top performance on large, powerful clusters–and Read more…

The Future of Accelerator Programming

Jan 9, 2014 |

Many of the latest supercomputers are based on accelerators, including the two fastest systems according to the 11/2013 TOP500 list. Accelerators are also becoming widespread in PCs and are even starting to appear in handheld devices, which will further boost the interest in accelerator programming. This broad adoption is the result of high performance, good Read more…

Adding MUSCLE to Multiscale Simulations

Dec 11, 2013 |

Multiscale models help understand phenomena with a wider scope or an increased level of detail. These models allow us to take the best from multiple worlds, for example by combining models with a fine-grained time or space resolution with models that capture systems over a large baseline. Classical examples of multiscale modeling include coupling atomistic Read more…

Compilers and More: Accelerated Programming

Dec 3, 2013 |

Having just returned from SC13, one burning issue is the choice of a standard approach for programming the next generation HPC systems. While not guaranteed, these systems are likely to be large clusters of nodes with multicore CPUs and some sort of attached accelerators. A standard programming approach is necessary to convince developers, and particularly Read more…

OpenACC Broadens Appeal with GCC Compiler Support

Nov 14, 2013 |

As the non-profit standards group behind the push for wider adoption via easier use of accelerators, OpenACC has quite a big job ahead. Although analysts agree that accelerators sit along a comfortable adoption curve, usability, programmability and portability are key concerns, among others. Over the last couple of years, OpenACC has worked with user groups Read more…

Short Takes

The Portability Mandate

Jul 24, 2014 |

Argonne National Laboratory recently published several sessions from its Summer 2013 Extreme-Scale Computing program to YouTube. One of these is a lesson on combining performance and portability presented by Argonne Assistant Computational Scientist Jeff Hammond. For some reason the video image does not match the lecture, but you will find a link to Hammond’s slide deck here. Read more…

Parallel Computing Trends

Jul 22, 2014 |

One of the most pressing issues faced by the HPC community is how to go about attracting and training the next generation of HPC users. The staff at Argonne National Laboratory is tackling this challenge head on by holding an intensive summer school in extreme-scale computing. One of the highlights of the 2013 summer program was a Read more…

The Case for a Parallel Programming Alternative

Jul 2, 2014 |

Cray engineers have been working on a new parallel computing language, called Chapel. Aimed at large-scale parallel computing environments, Chapel was designed with a focus on productivity and accessibility. The project originated from the DARPA High Productivity Computing Systems (HPCS) program, which challenged HPC vendors to improve the productivity of high-end computing systems. To explain Read more…

Programmability Matters

Jun 30, 2014 |

While discussions of HPC architectures have long centered on performance gains, that is not the only measure of success, according to Petteri Laakso of Vector Fabrics. Spurred by ever-proliferating core counts, programmability is taking on new prominence. Vector Fabrics is a Netherlands-based company that specializes in multicore software parallelization tools, so programmability is high on Read more…

An Easier, Faster Programming Language?

Jun 18, 2014 |

The HPC community has turned out supercomputers surpassing tens of petaflops of computing power by stringing together thousands of multicore processors, often in tandem with accelerators like NVIDIA GPUs and Intel Phi coprocessors. Of course, these multi-million dollar systems are only as useful as the programs that run on them, and developing applications that can Read more…

Off the Wire