Quantum Bits: D-Wave and VW; Google Quantum Lab; IBM Expands Access

By John Russell

March 21, 2017

For a technology that’s usually characterized as far off and in a distant galaxy, quantum computing has been steadily picking up steam. Just how close real-world applications are depends on whom you talk to and for what kinds of applications. Los Alamos National Lab, for example, has an active application development effort for its D-Wave system and LANL researcher Susan Mniszewski and colleagues have made progress on using the D-Wave machine for aspects of quantum molecular dynamics (QMD) simulations.

At CeBIT this week D-Wave and Volkswagen will discuss their pilot project to monitor and control taxi traffic in Beijing using a hybrid HPC-quantum system – this is on the heels of recent customer upgrade news from D-Wave (more below). Last week IBM announced expanded access to its five-qubit cloud-based quantum developer platform. In early March, researchers from the Google Quantum AI Lab published an excellent commentary in Nature examining real-world opportunities, challenges and timeframes for quantum computing more broadly. Google is also considering making its homegrown quantum capability available through the cloud.

As an overview, the Google commentary provides a great snapshot, noting soberly that challenges such as the lack of solid error correction and the small size (number of qubits) in today’s machines – whether “universal” digital machines like IBM’s or “analog” adiabatic annealing machines like D-Wave’s – have prompted many observers to declare useful quantum computing is still a decade way. Not so fast, says Google.

“This conservative view of quantum computing gives the impression that investors will benefit only in the long term. We contend that short-term returns are possible with the small devices that will emerge within the next five years, even though these will lack full error correction…Heuristic ‘hybrid’ methods that blend quantum and classical approaches could be the foundation for powerful future applications. The recent success of neural networks in machine learning is a good example,” write Masoud Mohseni, Peter Read, and John Martinis (a 2017 HPCwire Person to Watch) and colleagues (Nature, March 8, “Commercialize early quantum technologies”)

The D-Wave/VW project is a good example of a hybrid approach (details to follow) but first here’s a brief summary of recent quantum computing news:

  • IBM released a new API and upgraded simulator for modeling circuits up to 20 qubits on its 5-qubit platform. It also announced plans for a software developer kit by mid-year for building “simple” quantum applications. So far, says IBM, its quantum cloud has attracted about 40,000 users, including, for example, the Massachusetts Institute of Technology, which used the cloud service for its online quantum information science course. IBM also noted heavy use of the service by Chinese researchers. (See HPCwire coverage, IBM Touts Hybrid Approach to Quantum Computing)
  • D-Wave has been actively extending its development ecosystem (qbsolv (D-wave) and qmasm (LANL, et al.) and says researchers have recently been able to simulate a 20,000 qubit system on 1,000-qubit machine using qbsolv (more below). After announcing a 2,000-quibit machine in the fall, the company has begun deploying them. The first will be for a new customer, Temporal Defense System, and another is planned for the Google/NASA/USRA partnership which has a 1,000-qubit machine now. D-wave also just announced Virginia Tech and the Hume Center will begin using D-Wave systems for work on defense and intelligence applications.
  • Google’s commentary declares: “We anticipate that, within a few years, well-controlled quantum systems may be able to perform certain tasks much faster than conventional computers based on CMOS (complementary metal oxide–semiconductor) technology. Here we highlight three commercially viable uses for early quantum-computing devices: quantum simulation, quantum-assisted optimization and quantum sampling. Faster computing speeds in these areas would be commercially advantageous in sectors from artificial intelligence to finance and health care.”
D-Wave 2000Q System

Clearly there is a lot going on even at this stage of quantum computing’s development. There’s also been a good deal of wrangling over just what is a quantum computer and the differences between IBM’s “universal” digital approach – essentially a machine able to do anything computers do now – and D-Wave’s adiabatic annealing approach, which is currently intended to solve specific classes of optimization problems.

“They are different kinds of machines. No one has a universal quantum computer now, so you have to look at each case individually for its particular strengths and weaknesses,” explained Martinis to HPCwire. “The D-wave has minimal quantum coherence (it loses the information exchanged between qubits quite quickly), but makes up for it by having many qubits.”

“The IBM machine is small, but the qubits have quantum coherence enough to do some standard quantum algorithms. Right now it is not powerful, as you can run quantum simulations on classical computers quite easily. But by adding qubits the power will scale up quickly. It has the architecture of a universal machine and has enough quantum coherence to behave like one for very small problems,” Martinis said.

Noteworthy, Google has developed 9-qubit devices that have 3-5x more coherence than IBM, according to Martinis, but they are not on the cloud yet. “We are ready to scale up now, and plan to have this year a ‘quantum supremacy’ device that has to be checked with a supercomputer. We are thinking of offering cloud also, but are more or less waiting until we have a hardware device that gives you more power than a classical simulation.”

Quantum supremacy as described in the Google commentary is a term coined by theoretical physicist John Preskill to describe “the ability of a quantum processor to perform, in a short time, a well-defined mathematical task that even the largest classical supercomputers (such as China’s Sunway TaihuLight) would be unable to complete within any reasonable time frame. We predict that, in a few years, an experiment achieving quantum supremacy will be performed.”

Bo Ewald

For the moment, D-Wave is the only vendor offering near-production machines versus research machines, said Bo Ewald, the company’s ever-cheerful evangelist. He quickly agrees though that at least for now there aren’t any production-ready applications. Developing a quantum tool/software ecosystem is a driving focus at D-wave. The LANL app dev work, though impressive, still represents proto-application development. Nevertheless the ecosystem of tools is growing quickly.

“We have defined a software architecture that has several layers starting at the quantum machine instruction layer where if you want to program in machine language you are certainly welcome to do that; that is kind of the way people had to do it in the early days,” said Ewald.

“The next layer up is if you want to be able to create quantum machine instructions from C or C++ or Python. We have now libraries that run on host machines, regular HPC machines, so you can use those languages to generate programs that run on the D-Wave machine but the challenge that we have faced, that customers have faced, is that our machines had 500 qubits or 1,000 qubits and now 2,000; we know there are problems that are going to consume many more qubits than that,” he said.

For D-Wave systems, qbsolv helps address this problem. It allows a meta-description of the machine and the problem you want to solve as quadratic unconstrained binary optimization or QUBO. It’s an intermediate representation. D-Wave then extended this capability to what it calls virtual QUBOs likening it to virtual memory.

“You can create QUBOs or representations of problems which are much larger than the machine itself and then using combined classical computer and quantum computer techniques we could partition the problem and solve them in chunks and then kind of glue them back together after we solved the D-Wave part. We’ve done that now with the 1,000-qubit machine and run problems that have the equivalent of 20,000 qubits,” said Ewald, adding the new 2,000-qubit machines will handle problems of even greater size using this capability.

At LANL, researcher Scott Pakin has developed another tool – a quantum macro assembler for D-Wave systems (QMASM). Ewald said part of the goal of Pakin’s work was to determine, “if you could map gates onto the machine even though we are not a universal or a gate model. You can in fact model gates on our machine and he has started to [create] a library of gates (or gates, and gates, nand gates) and you can assemble those to become macros.”

Pakin said,My personal research interest has been in making the D-Wave easier to program. I’ve recently built something really nifty on top of QMASM: edif2qmasm, which is my answer to the question: Can one write classical-style code and run it on the D-Wave?

“For many difficult computational problems, solution verification is simple and fast. The idea behind edif2qmasm is that one can write an ordinary(-ish) program that reports if a proposed solution to a problem is in fact valid. This gets compiled for the D-Wave then run _backwards_, giving it ‘true’ for the proposed solution being valid and getting back a solution to the difficult computational problem.”

Pakin noted there are many examples on github to provide a feel for the power of this tool.

“For example, mult.v is a simple, one-line multiplier. Run it backwards, and it factors a number, which underlies modern data decryption. In a dozen or so lines of code, circsat.v evaluates a Boolean circuit. Run it backwards, and it tells you what inputs lead to an output of “true”, which used in areas of artificial intelligence, circuit design, and automatic theorem proving. map-color.v reports if a map is correctly colored with four colors such that no two adjacent regions have the same color. Run it backwards, and it _finds_ such a coloring.

“Although current-generation D-Wave systems are too limited to apply this approach to substantial problems, the trends in system scale and engineering precision indicate that some day we should be able to perform real work on this sort of system. And with the help of tools like edif2qmasm, programmers won’t need an advanced degree to figure out how to write code for it,” he explained.

The D-Wave/VW collaboration, just a year or so old, is one of the more interesting quantum computing proof-of-concept efforts because it tackles an optimization problem of the kind that is widespread in everyday life. As described by Ewald, VW CIO Martin Hoffman was making his yearly swing through Silicon Valley and stopped in at D-Wave and talk turned to the many optimization challenges big automakers face, such as supply logistics, vehicle delivery, and various machine learning tasks and doing a D-Wave project around one of them. Instead, said Ewald, VW eventually settled on a more driver-facing problem.

It turns out there are about 10,000 taxis in Beijing, said Ewald. Each has a GPS device and their positions are recorded every five seconds. Traffic congestion, of course, is a huge problem in Beijing. The idea was to explore if it was possible to create an application running on both traditional computer resources and D-Wave to help monitor and guide taxi movement more quickly and effectively.

“Ten thousand taxis on all of the streets in Beijing is way too big for our machine at this point, but they came to this same idea we talked about with qbsolve where you partition problems,” said Ewald. “On the traditional machines VW created a map and grid and subdivided the grid into quadrants and would find the quadrant that was the most red.” That’s red as in long cab waits.

The problem quadrant was then sent to D-Wave to be solved. “We would optimize the flow, basically minimize the wait time for all of the taxis within the quadrant, send that [solution] back to the traditional machine which would then send us the next most red, and we would try to turn it green,” said Ewald.

According to Ewald, VW was able to relatively create the “hybrid” solutions quickly and “get what they say are pretty good results.” They have talked about then being able to extend this project to predict where traffic jams are going to be and give people perhaps 45 minute warnings that there’s the potential for a traffic jam at such and such intersection. The two companies have a press conference planned this week at CeBIT to showcase the project.

It’s good to emphasize that the VW/D-wave exercise is developmental – what Ewald labels as a proto application: “But just the fact that they were able to get it running is a great step forward in many ways in that we believe our machine will be used side by side with existing machines, much like GPUs were used in the early days on graphics. In this case VW has demonstrated quite clearly how our machine, our QPU if you will, can be used in helping accelerate the work being done on a traditional HPC machines.”

Image art, chip diagram: D-Wave

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

With New Owner and New Roadmap, an Independent Omni-Path Is Staging a Comeback

July 23, 2021

Put on a shelf by Intel in 2019, Omni-Path faced a uncertain future, but under new custodian Cornelis Networks, OmniPath is looking to make a comeback as an independent high-performance interconnect solution. A "significant refresh" – called Omni-Path Express – is coming later this year according to the company. Cornelis Networks formed last September as a spinout of Intel's Omni-Path division. Read more…

PEARC21 Panel Reviews Eight New NSF-Funded HPC Systems Debuting in 2021

July 23, 2021

Over the past few years, the NSF has funded a number of HPC systems to further supply the open research community with computational resources to meet that community’s changing and expanding needs. A review of these systems at the PEARC21 conference (July 19-22) highlighted... Read more…

Chameleon’s HPC Testbed Sharpens Its Edge, Presses ‘Replay’

July 22, 2021

“One way of saying what I do for a living is to say that I develop scientific instruments,” said Kate Keahey, a senior fellow at the University of Chicago and a computer scientist at Argonne National Laboratory, as s Read more…

PEARC21 Plenary Session: AI for Innovative Social Work

July 21, 2021

AI analysis of social media poses a double-edged sword for social work and addressing the needs of at-risk youths, said Desmond Upton Patton, senior associate dean, Innovation and Academic Affairs, Columbia University. S Read more…

Summer Reading: “High-Performance Computing Is at an Inflection Point”

July 21, 2021

At last month’s 11th International Symposium on Highly Efficient Accelerators and Reconfigurable Technologies (HEART), a group of researchers led by Martin Schulz of the Leibniz Supercomputing Center (Munich) presented a “position paper” in which they argue HPC architectural landscape... Read more…

AWS Solution Channel

Accelerate innovation in healthcare and life sciences with AWS HPC

With Amazon Web Services, researchers can access purpose-built HPC tools and services along with scientific and technical expertise to accelerate the pace of discovery. Whether you are sequencing the human genome, using AI/ML for disease detection or running molecular dynamics simulations to develop lifesaving drugs, AWS has the infrastructure you need to run your HPC workloads. Read more…

PEARC21 Panel: Wafer-Scale-Engine Technology Accelerates Machine Learning, HPC

July 21, 2021

Early use of Cerebras’ CS-1 server and wafer-scale engine (WSE) has demonstrated promising acceleration of machine-learning algorithms, according to participants in the Scientific Research Enabled by CS-1 Systems panel Read more…

With New Owner and New Roadmap, an Independent Omni-Path Is Staging a Comeback

July 23, 2021

Put on a shelf by Intel in 2019, Omni-Path faced a uncertain future, but under new custodian Cornelis Networks, OmniPath is looking to make a comeback as an independent high-performance interconnect solution. A "significant refresh" – called Omni-Path Express – is coming later this year according to the company. Cornelis Networks formed last September as a spinout of Intel's Omni-Path division. Read more…

Chameleon’s HPC Testbed Sharpens Its Edge, Presses ‘Replay’

July 22, 2021

“One way of saying what I do for a living is to say that I develop scientific instruments,” said Kate Keahey, a senior fellow at the University of Chicago a Read more…

Summer Reading: “High-Performance Computing Is at an Inflection Point”

July 21, 2021

At last month’s 11th International Symposium on Highly Efficient Accelerators and Reconfigurable Technologies (HEART), a group of researchers led by Martin Schulz of the Leibniz Supercomputing Center (Munich) presented a “position paper” in which they argue HPC architectural landscape... Read more…

PEARC21 Panel: Wafer-Scale-Engine Technology Accelerates Machine Learning, HPC

July 21, 2021

Early use of Cerebras’ CS-1 server and wafer-scale engine (WSE) has demonstrated promising acceleration of machine-learning algorithms, according to participa Read more…

15 Years Later, the Green500 Continues Its Push for Energy Efficiency as a First-Order Concern in HPC

July 15, 2021

The Green500 list, which ranks the most energy-efficient supercomputers in the world, has virtually always faced an uphill battle. As Wu Feng – custodian of the Green500 list and an associate professor at Virginia Tech – tells it, “noone" cared about energy efficiency in the early 2000s, when the seeds... Read more…

Frontier to Meet 20MW Exascale Power Target Set by DARPA in 2008

July 14, 2021

After more than a decade of planning, the United States’ first exascale computer, Frontier, is set to arrive at Oak Ridge National Laboratory (ORNL) later this year. Crossing this “1,000x” horizon required overcoming four major challenges: power demand, reliability, extreme parallelism and data movement. Read more…

Quantum Roundup: IBM, Rigetti, Phasecraft, Oxford QC, China, and More

July 13, 2021

IBM yesterday announced a proof for a quantum ML algorithm. A week ago, it unveiled a new topology for its quantum processors. Last Friday, the Technical Univer Read more…

ExaWind Prepares for New Architectures, Bigger Simulations

July 10, 2021

The ExaWind project describes itself in terms of terms like wake formation, turbine-turbine interaction and blade-boundary-layer dynamics, but the pitch to the Read more…

AMD Chipmaker TSMC to Use AMD Chips for Chipmaking

May 8, 2021

TSMC has tapped AMD to support its major manufacturing and R&D workloads. AMD will provide its Epyc Rome 7702P CPUs – with 64 cores operating at a base cl Read more…

Intel Launches 10nm ‘Ice Lake’ Datacenter CPU with Up to 40 Cores

April 6, 2021

The wait is over. Today Intel officially launched its 10nm datacenter CPU, the third-generation Intel Xeon Scalable processor, codenamed Ice Lake. With up to 40 Read more…

Berkeley Lab Debuts Perlmutter, World’s Fastest AI Supercomputer

May 27, 2021

A ribbon-cutting ceremony held virtually at Berkeley Lab's National Energy Research Scientific Computing Center (NERSC) today marked the official launch of Perlmutter – aka NERSC-9 – the GPU-accelerated supercomputer built by HPE in partnership with Nvidia and AMD. Read more…

Ahead of ‘Dojo,’ Tesla Reveals Its Massive Precursor Supercomputer

June 22, 2021

In spring 2019, Tesla made cryptic reference to a project called Dojo, a “super-powerful training computer” for video data processing. Then, in summer 2020, Tesla CEO Elon Musk tweeted: “Tesla is developing a [neural network] training computer called Dojo to process truly vast amounts of video data. It’s a beast! … A truly useful exaflop at de facto FP32.” Read more…

Google Launches TPU v4 AI Chips

May 20, 2021

Google CEO Sundar Pichai spoke for only one minute and 42 seconds about the company’s latest TPU v4 Tensor Processing Units during his keynote at the Google I Read more…

CentOS Replacement Rocky Linux Is Now in GA and Under Independent Control

June 21, 2021

The Rocky Enterprise Software Foundation (RESF) is announcing the general availability of Rocky Linux, release 8.4, designed as a drop-in replacement for the soon-to-be discontinued CentOS. The GA release is launching six-and-a-half months after Red Hat deprecated its support for the widely popular, free CentOS server operating system. The Rocky Linux development effort... Read more…

CERN Is Betting Big on Exascale

April 1, 2021

The European Organization for Nuclear Research (CERN) involves 23 countries, 15,000 researchers, billions of dollars a year, and the biggest machine in the worl Read more…

Iran Gains HPC Capabilities with Launch of ‘Simorgh’ Supercomputer

May 18, 2021

Iran is said to be developing domestic supercomputing technology to advance the processing of scientific, economic, political and military data, and to strengthen the nation’s position in the age of AI and big data. On Sunday, Iran unveiled the Simorgh supercomputer, which will deliver.... Read more…

Leading Solution Providers

Contributors

HPE Launches Storage Line Loaded with IBM’s Spectrum Scale File System

April 6, 2021

HPE today launched a new family of storage solutions bundled with IBM’s Spectrum Scale Erasure Code Edition parallel file system (description below) and featu Read more…

Julia Update: Adoption Keeps Climbing; Is It a Python Challenger?

January 13, 2021

The rapid adoption of Julia, the open source, high level programing language with roots at MIT, shows no sign of slowing according to data from Julialang.org. I Read more…

10nm, 7nm, 5nm…. Should the Chip Nanometer Metric Be Replaced?

June 1, 2020

The biggest cool factor in server chips is the nanometer. AMD beating Intel to a CPU built on a 7nm process node* – with 5nm and 3nm on the way – has been i Read more…

GTC21: Nvidia Launches cuQuantum; Dips a Toe in Quantum Computing

April 13, 2021

Yesterday Nvidia officially dipped a toe into quantum computing with the launch of cuQuantum SDK, a development platform for simulating quantum circuits on GPU-accelerated systems. As Nvidia CEO Jensen Huang emphasized in his keynote, Nvidia doesn’t plan to build... Read more…

Microsoft to Provide World’s Most Powerful Weather & Climate Supercomputer for UK’s Met Office

April 22, 2021

More than 14 months ago, the UK government announced plans to invest £1.2 billion ($1.56 billion) into weather and climate supercomputing, including procuremen Read more…

Q&A with Jim Keller, CTO of Tenstorrent, and an HPCwire Person to Watch in 2021

April 22, 2021

As part of our HPCwire Person to Watch series, we are happy to present our interview with Jim Keller, president and chief technology officer of Tenstorrent. One of the top chip architects of our time, Keller has had an impactful career. Read more…

Quantum Roundup: IBM, Rigetti, Phasecraft, Oxford QC, China, and More

July 13, 2021

IBM yesterday announced a proof for a quantum ML algorithm. A week ago, it unveiled a new topology for its quantum processors. Last Friday, the Technical Univer Read more…

Senate Debate on Bill to Remake NSF – the Endless Frontier Act – Begins

May 18, 2021

The U.S. Senate today opened floor debate on the Endless Frontier Act which seeks to remake and expand the National Science Foundation by creating a technology Read more…

  • arrow
  • Click Here for More Headlines
  • arrow
HPCwire