November 7, 2017
Enlisting computational technologies in the war on cancer isn’t new but it has taken on an increasingly decisive role. At SC17, Eric Stahlberg, director of th Read more…
January 4, 2017
Twenty years ago high performance computing was nearly absent from life sciences. Today it’s used throughout life sciences and biomedical research. Genomics and the data deluge from modern lab instruments are the main drivers, but so is the longer-term desire to perform predictive simulation in support of Precision Medicine (PM). There’s even a specialized life sciences supercomputer, ‘Anton’ from D.E. Shaw Research, and the Pittsburgh Supercomputing Center is standing up its second Anton 2 and actively soliciting project proposals. There’s a lot going on. Read more…
December 7, 2016
Sometime in Q2 2017 the first ‘results’ of the Joint Design of Advanced Computing Solutions for Cancer (JDACS4C) will become publicly available according to Rick Stevens. He leads one of three JDACS4C pilot projects pressing deep learning (DL) into service in the War on Cancer. Read more…
July 31, 2013
Officials at Australian National University (ANU) in Canberra today took the wraps off Raijin, a new 1.2 petaflop supercomputer that is fastest ever in the country. The new cluster is housed at the new National Computational Infrastructure (NCI) facility on the ANU campus, and will be used for climate modeling, among other uses. Read more…
Whether an organization chooses a cloud for general business needs or a highly tailored workload, the spectrum of offerings and configurations can be overwhelming. To help you navigate the various cloud options available today, we're breaking down your options, exploring pros and cons, and sharing ways to keep your options open and your business agile as you execute your cloud strategy.
Researchers in academic labs and commercial R&D groups continue to need more compute capacity, which means leveraging the latest innovations in HPC technologies as well as an assortment of resources to meet the unique needs of different workloads. Increasingly, systems based on Arm processors are stepping into that role, offering low power consumption and strategic advantages for HPC workloads.
Whether it's for fraud detection, personalized medicine, manufacturing, smart cities, autonomous vehicles and many other areas, advanced-scale computing has exploded beyond the realm of academia and government and into the private sector. And with data-intensive workloads on the rise, commercial users are turning to HPC-based infrastructure to run the AI, ML and cognitive computing applications that their organizations depend on.
© HPCwire. All Rights Reserved. A Tabor Communications Publication
Reproduction in whole or in part in any form or medium without express written permission of Tabor Communications, Inc. is prohibited.