HPC Innovator’s Work Spans Two Continents

By Nicole Hemsoth

May 28, 2010

Over the past two decades, Dr. Ashwini Nanda has been at the center of some of the most cutting-edge HPC projects and initiatives in the world.  At IBM’s T.J. Watson Research Center, in New York, Dr. Nanda led the development of the Cell processor-based  systems (QS20, QS21, and QS22 blades) and software technologies for high performance computing. That work culminated in the construction of the Roadrunner supercomputer for Los Alamos National Laboratory. Later, as the head of Computational Research Laboratories (CRL) in Pune, India, he directed the development of the “Eka” HPC cluster that, in 2007, held the title of Asia’s fastest supercomputer. He also established the shared memory systems group at IBM Research, worked on the Amazon superscalar architecture at Texas Instruments, and developed parallel computers for India’s missile defense systems at Wipro, Bangalore.

Recently he moved back to India, where he founded HPC Links, a company that offers parallel programming software tools and services for the global high performance computing community. We got the opportunity to ask Dr. Nanda about how he sees the HPC industry today, how it’s changing, and what led him to start up his company.

HPCwire: Maybe we can start with some of your thoughts about the development of high performance computing in India. What are the principal challenges in building the country’s HPC capability? And how do you see the way forward for HPC and supercomputing in India — for both vendors and users?
 
Ashwini Nanda: India has less than 1 percent of the compute power in the global TOP500 supercomputing list today. This tiny share of the pie is not indicative of the financial and high-tech might of India. The gap indicates there is significant room for growth in HPC infrastructure in the country. The gap one sees today is perhaps due to the lack of broad awareness of the benefits of HPC to scientific/technical research and to business enterprises. This is despite the commendable progress made by organizations such as CDAC, CRL, SERC/IISc and HiPC in spreading HPC awareness in the country.

India’s per capita spending in high performance computing is negligible compared to the US, Japan and Germany, and for that matter compared even to China, Spain and Russia. The Unique ID (Adhaar) project of the government will likely fuel growth in HPC related to data mining, consummer security and national security. The government has procured quite a few HPC clusters for weather and climate prediction which would foster research and application development in these areas. India has very successful government enterprises in nuclear energy, space and defence, which would benefit tremendously from the use of HPC. The US counterparts in these sectors are the prime movers of the HPC industry there. The Indian government has the financial strength and must spend heavily in HPC in order to make these sectors globally competitive.
 
India has vibrant pharmaceutical, financial, entertainment and manufacturing industries, all of whom would gain significant productivity and competitiveness by using HPC. We built the Eka machine at CRL soley with private investment from Tata. The other industrial powerhouses in India have the financial means, and they will likely follow suit once they see the benefits of HPC to their business. Talking to leaders from government, academia and industry across the country, one can sense a growing awareness of the potential of HPC, which I believe will translate into a faster pace of growth in HPC infrastructure and services during the next five years or so in India. When that happens, it will benefit the multinational vendors as well as the local vendors.

HPCwire: What do you think will be the role of cloud computing for HPC users in India?

Nanda: Culturally, and economically, reuse and sharing of resources are well accepted in the Indian society. Once the users discover the utility of HPC, I think sharing resources through cloud platforms will come naturally to them, especially the small and medium scale users of HPC. But more importantly, India could also become a key global host of cloud computing infrastructure and services due to attractive low-cost operations and highly skilled technical manpower.

HPCwire: You had a rather prominent role in the development of Cell processor-based systems and software at IBM. The Cell helped usher in the petaflop era, but overall it looks like the impact of this technology in supercomputing is going to be relatively limited. What do you think the Cell brought to the HPC space and what lessons were learned?

Nanda: The Cell based systems from IBM made two important contributions to HPC technology, besides helping cross the petaflop barrier. First they set the new trend of using hybrid multicore clusters — with CPU-GPU combinations — to build the most cost-effective, power-efficient and best-performing supercomputers. Secondly, creating a software ecosystem to harness the compute power of a revolutionary processor like Cell seemed prohibitive in the beginning. But the Cell systems effort at IBM Research showed that an effective collaboration of government, academia and industry can indeed build a respectable software ecosystem for a new architecture.

It’s true that the momentum around Cell has has been lost, but credit should go to IBM for laying the foundations of a new era that would see the proliferation of hybrid multicore CPU-GPU combination clusters in solving key HPC problems. We are already seeing glimpses of this through the momentum building around such offerings from AMD, Intel and NVIDIA.

HPCwire: In HPC, which hardware and software technologies do you see becoming increasing important over the next, say, five years?

Nanda: I believe in terms of hardware, hybrid multicore GPU-CPU combination technologies and commodity InfiniBand and Ethernet technologies will continue to lead the way and take us through the exaflop mark. In terms of software we will see more emphasis being put on tools that make it easier to write parallel applications and increase productivity.

HPCwire: Switching to your current role as the founder and CEO of HPC Links: What is the company about and what was your motivation to launch this as a business?

Nanda: While doing the systems work at IBM and at Tata over the last few years, the obvious realization came to me that now the commoditization of HPC platforms is complete. We have been building affordable parallel machines all around the world, and almost any one who needs access to a parallel machine can access one today. But are these machines utilized well enough? Not really. Are most people, or industries, who could benefit from HPC, taking advantage of these platforms today? The answer is clearly no. So what is the problem? The US government Council on Competitiveness identified two years back that there are three primary barriers to mass adoption of HPC. Namely, lack of parallel programming skills, lack of parallel applications, and high cost of adoption.

HPC Links was formed last year to help address these customer pain points. Our goal is to help businesses stay competitive by alleviating these pain points and achieving high efficiency, faster time to market and enhanced product quality through innovative use of HPC, cloud and multicore solutions. We are addressing all the three key barriers in HPC adoption through our parallel application services offerings, system integration offerings and the software tools and packages under development.

HPCwire: What is unique about the company?
 
Nanda: Our uniqueness today, if I can point to the most significant one, is our interdisciplinary skill pool. We have Ph.D.’s and Masters in multitude of scientific and engineering disciplines, all adept at various flavors of parallel programming, on all kinds of hardware and software platforms. Mix that with the breadth of domain knowledge the HPC Links team has across industry verticals, and you get a really unique, comprehensive parallel programming skill pool in the world.

HPCwire: What do you see as the principal challenges in developing parallel applications for science and technical computing codes today?

Nanda: The hard challenges in parallel application development as I see are lack of skills, and productive tools. Tackling the challenge of productive tools for parallel programming, I believe, will take longer than tackling the issue of lack of skills. There is the general lack of parallel programmers in the world today, which the universities have started to address pretty effectively. And then there is the problem that people who are good in parallel programming are not necessarily trained in various application domains, and people who are domain experts are often times not proficient in parallel programming. The key is to bring domain experts and parallel programmers together and cross train them on the job.

HPCwire: Who do think will be your main customers for these services?

Nanda: In the near term we see majority of our customers outside of India where the HPC market has higher momentum and awareness — especially in the US and Europe. The Indian HPC market is in its nascent stage, but has great potential in the longer term. Our services are targeted toward any one who can benefit from an experienced parallel programming skill pool in research and development, scaling, testing and optimization of parallel applications in a wide range of domains. For example, recently we signed up with Microsoft to provide this kind of parallel application programming services for their HPC Server platform. This makes our services available to the Windows HPC Server users and ISVs in all industry segments. We have deep Linux cluster programming expertise, and are working with universities and national labs, as well as industry HPC users.

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

Insights from Optimized Codes on Cineca’s Marconi

February 15, 2019

What can you do with 381,392 CPU cores? For Cineca, it means enabling computational scientists to expand a large part of the world’s body of knowledge from the nanoscale to the astronomic, from calculating quantum effe Read more…

By Ken Strandberg

What Will IBM’s AI Debater Learn from Its Loss?

February 14, 2019

The utility of IBM’s latest man-versus-machine gambit is debatable. At the very least its Project Debater got us thinking about the potential uses of artificial intelligence as a way of helping humans sift through al Read more…

By George Leopold

ClusterVision in Bankruptcy, Fate Uncertain

February 13, 2019

ClusterVision, European HPC specialists that have built and installed over 20 Top500-ranked systems in their nearly 17-year history, appear to be in the midst of bankruptcy proceedings. According to Dutch news site Drimb Read more…

By Tiffany Trader

HPE Extreme Performance Solutions

HPE Systems With Intel Omni-Path: Architected for Value and Accessible High-Performance Computing

Today’s high-performance computing (HPC) and artificial intelligence (AI) users value high performing clusters. And the higher the performance that their system can deliver, the better. Read more…

IBM Accelerated Insights

Medical Research Powered by Data

“We’re all the same, but we’re unique as well. In that uniqueness lies all of the answers….”

  • Mark Tykocinski, MD, Provost, Executive Vice President for Academic Affairs, Thomas Jefferson University

Getting the answers to what causes some people to develop diseases and not others is driving the groundbreaking medical research being conducted by the Computational Medicine Center at Thomas Jefferson University in Philadelphia. Read more…

South African Weather Service Doubles Compute and Triples Storage Capacity of Cray System

February 13, 2019

South Africa has made headlines in recent years for its commitment to HPC leadership in Africa – and now, Cray has announced another major South African HPC expansion. Cray has been awarded contracts with Eclipse Holdings Ltd. to upgrade the supercomputing system operated by the South African Weather Service (SAWS). Read more…

By Oliver Peckham

Insights from Optimized Codes on Cineca’s Marconi

February 15, 2019

What can you do with 381,392 CPU cores? For Cineca, it means enabling computational scientists to expand a large part of the world’s body of knowledge from th Read more…

By Ken Strandberg

ClusterVision in Bankruptcy, Fate Uncertain

February 13, 2019

ClusterVision, European HPC specialists that have built and installed over 20 Top500-ranked systems in their nearly 17-year history, appear to be in the midst o Read more…

By Tiffany Trader

UC Berkeley Paper Heralds Rise of Serverless Computing in the Cloud – Do You Agree?

February 13, 2019

Almost exactly ten years to the day from publishing of their widely-read, seminal paper on cloud computing, UC Berkeley researchers have issued another ambitious examination of cloud computing - Cloud Programming Simplified: A Berkeley View on Serverless Computing. The new work heralds the rise of ‘serverless computing’ as the next dominant phase of cloud computing. Read more…

By John Russell

Iowa ‘Grows Its Own’ to Fill the HPC Workforce Pipeline

February 13, 2019

The global workforce that supports advanced computing, scientific software and high-speed research networks is relatively small when you stop to consider the magnitude of the transformative discoveries it empowers. Technical conferences provide a forum where specialists convene to learn about the latest innovations and schedule face-time with colleagues from other institutions. Read more…

By Elizabeth Leake, STEM-Trek

Trump Signs Executive Order Launching U.S. AI Initiative

February 11, 2019

U.S. President Donald Trump issued an Executive Order (EO) today launching a U.S Artificial Intelligence Initiative. The new initiative - Maintaining American L Read more…

By John Russell

Celebrating Women in Science: Meet Four Women Leading the Way in HPC

February 11, 2019

One only needs to look around at virtually any CS/tech conference to realize that women are underrepresented, and that holds true of HPC. SC hosts over 13,000 H Read more…

By AJ Lauer

IBM Bets $2B Seeking 1000X AI Hardware Performance Boost

February 7, 2019

For now, AI systems are mostly machine learning-based and “narrow” – powerful as they are by today's standards, they're limited to performing a few, narro Read more…

By Doug Black

Assessing Government Shutdown’s Impact on HPC

February 6, 2019

After a 35-day federal government shutdown, the longest in U.S. history, government agencies are taking stock of the damage -- and girding for a potential secon Read more…

By Tiffany Trader

Quantum Computing Will Never Work

November 27, 2018

Amid the gush of money and enthusiastic predictions being thrown at quantum computing comes a proposed cold shower in the form of an essay by physicist Mikhail Read more…

By John Russell

Cray Unveils Shasta, Lands NERSC-9 Contract

October 30, 2018

Cray revealed today the details of its next-gen supercomputing architecture, Shasta, selected to be the next flagship system at NERSC. We've known of the code-name "Shasta" since the Argonne slice of the CORAL project was announced in 2015 and although the details of that plan have changed considerably, Cray didn't slow down its timeline for Shasta. Read more…

By Tiffany Trader

The Case Against ‘The Case Against Quantum Computing’

January 9, 2019

It’s not easy to be a physicist. Richard Feynman (basically the Jimi Hendrix of physicists) once said: “The first principle is that you must not fool yourse Read more…

By Ben Criger

AMD Sets Up for Epyc Epoch

November 16, 2018

It’s been a good two weeks, AMD’s Gary Silcott and Andy Parma told me on the last day of SC18 in Dallas at the restaurant where we met to discuss their show news and recent successes. Heck, it’s been a good year. Read more…

By Tiffany Trader

Intel Reportedly in $6B Bid for Mellanox

January 30, 2019

The latest rumors and reports around an acquisition of Mellanox focus on Intel, which has reportedly offered a $6 billion bid for the high performance interconn Read more…

By Doug Black

US Leads Supercomputing with #1, #2 Systems & Petascale Arm

November 12, 2018

The 31st Supercomputing Conference (SC) - commemorating 30 years since the first Supercomputing in 1988 - kicked off in Dallas yesterday, taking over the Kay Ba Read more…

By Tiffany Trader

Looking for Light Reading? NSF-backed ‘Comic Books’ Tackle Quantum Computing

January 28, 2019

Still baffled by quantum computing? How about turning to comic books (graphic novels for the well-read among you) for some clarity and a little humor on QC. The Read more…

By John Russell

Contract Signed for New Finnish Supercomputer

December 13, 2018

After the official contract signing yesterday, configuration details were made public for the new BullSequana system that the Finnish IT Center for Science (CSC Read more…

By Tiffany Trader

Leading Solution Providers

SC 18 Virtual Booth Video Tour

Advania @ SC18 AMD @ SC18
ASRock Rack @ SC18
DDN Storage @ SC18
HPE @ SC18
IBM @ SC18
Lenovo @ SC18 Mellanox Technologies @ SC18
NVIDIA @ SC18
One Stop Systems @ SC18
Oracle @ SC18 Panasas @ SC18
Supermicro @ SC18 SUSE @ SC18 TYAN @ SC18
Verne Global @ SC18

Deep500: ETH Researchers Introduce New Deep Learning Benchmark for HPC

February 5, 2019

ETH researchers have developed a new deep learning benchmarking environment – Deep500 – they say is “the first distributed and reproducible benchmarking s Read more…

By John Russell

ClusterVision in Bankruptcy, Fate Uncertain

February 13, 2019

ClusterVision, European HPC specialists that have built and installed over 20 Top500-ranked systems in their nearly 17-year history, appear to be in the midst o Read more…

By Tiffany Trader

IBM Quantum Update: Q System One Launch, New Collaborators, and QC Center Plans

January 10, 2019

IBM made three significant quantum computing announcements at CES this week. One was introduction of IBM Q System One; it’s really the integration of IBM’s Read more…

By John Russell

Nvidia’s Jensen Huang Delivers Vision for the New HPC

November 14, 2018

For nearly two hours on Monday at SC18, Jensen Huang, CEO of Nvidia, presented his expansive view of the future of HPC (and computing in general) as only he can do. Animated. Backstopped by a stream of data charts, product photos, and even a beautiful image of supernovae... Read more…

By John Russell

HPC Reflections and (Mostly Hopeful) Predictions

December 19, 2018

So much ‘spaghetti’ gets tossed on walls by the technology community (vendors and researchers) to see what sticks that it is often difficult to peer through Read more…

By John Russell

IBM Bets $2B Seeking 1000X AI Hardware Performance Boost

February 7, 2019

For now, AI systems are mostly machine learning-based and “narrow” – powerful as they are by today's standards, they're limited to performing a few, narro Read more…

By Doug Black

The Deep500 – Researchers Tackle an HPC Benchmark for Deep Learning

January 7, 2019

How do you know if an HPC system, particularly a larger-scale system, is well-suited for deep learning workloads? Today, that’s not an easy question to answer Read more…

By John Russell

Intel Confirms 48-Core Cascade Lake-AP for 2019

November 4, 2018

As part of the run-up to SC18, taking place in Dallas next week (Nov. 11-16), Intel is doling out info on its next-gen Cascade Lake family of Xeon processors, specifically the “Advanced Processor” version (Cascade Lake-AP), architected for high-performance computing, artificial intelligence and infrastructure-as-a-service workloads. Read more…

By Tiffany Trader

  • arrow
  • Click Here for More Headlines
  • arrow
Do NOT follow this link or you will be banned from the site!
Share This