New IBM Blog Details Path to Quantum Advantage in 2023

By John Russell

July 19, 2022

At IBM’s Quantum Summit last November, the company issued a detailed roadmap and declared it would deliver quantum advantage (better than classical computers) on selected applications in 2023. It was a bold claim that was greeted with enthusiasm in some quarters and skepticism in others. Today, IBM reiterated that claim with a lengthy blog on how error mitigation is the key to achieving quantum advantage (QA) while still using noisy intermediate scale quantum (NISQ) computers.

“There’s still a common refrain in the field if you want value from a quantum computer, it needs to be a large, fault tolerant quantum processor. Until then we’re stuck with noisy devices, inferior to classical counterparts,” an IBM spokesperson told HPCwire. “Our team has a different view. They have developed new techniques to draw ever more value from noisy qubits.”

Jay Gambetta, IBM

The blog, written by Jay Gambetta, IBM fellow and vice president of quantum and HPCwire Person to Watch in 2022, and research colleagues Kristan Temme, Ewout van den Berg, and Abhinav Kandala argues that probabilistic error cancellation is the secret sauce, which when combined with other advances, will enable IBM to make good on its promise to deliver selective quantum advantage sometime next year. It’s an important part of IBM’s long-term path to fault-tolerant quantum computing.

The IBM researchers write:

“The history of classical computing is one of advances in transistor and chip technology yielding corresponding gains in information processing performance. Although quantum computers have seen tremendous improvements in their scale, quality and speed in recent years, such a gradual evolution seems to be missing from the narrative. Indeed, it is widely accepted that one must first build a large fault-tolerant quantum processor before any of the quantum algorithms with proven super-polynomial speed-up can be implemented. Building such a processor therefore is the central goal for our development.

“However, recent advances in techniques we refer to broadly as quantum error mitigation allow us to lay out a smoother path towards this goal. Along this path, advances in qubit coherence, gate fidelities, and speed immediately translate to measurable advantage in computation, akin to the steady progress historically observed with classical computers.

“Of course, the ultimate litmus test for practical quantum computing is to provide an advantage over classical approaches for a useful problem. Such an advantage can take many forms, the most prominent one being a substantial improvement of a quantum algorithm’s runtime over the best classical approaches. For this to be possible, the algorithm should have an efficient representation as quantum circuits, and there should be no classical algorithm that can simulate these quantum circuits efficiently. This can only be achieved for specific kinds of problems.”

That’s the IBM plan.

Many in the quantum community agree. Others demure. Photonic-based quantum computing start-up PsiQuantum has consistently said that pursuit of full fault-tolerance is the only realistic way to produce practical, broadly useful quantum computers. PsiQuantum has embarked just such a strategy. (See HPCwire article, PsiQuantum’s Path to 1 Million Qubits.)

Analyst Bob Sorensen, chief quantum analyst for Hyperion Research, noted that IBM’s portrayal of the quantum industry, or at least big portions of it, as clinging to the idea that achieving fault-tolerant quantum computing is necessary before achieving usefulness may have been overstated. “I think there’s a lot more optimism for error-tolerant NISQ systems than is being suggested,” noted Sorensen.

Nevertheless, he was impressed with IBM’s activities. “What IBM is doing, as a full stack company, is addressing error correction at every level, down to the qubit structure and all the way up to software. They can cover everything in between. That’s a key advantage for a full stack vendor. So I think IBM is trying to aggressively take advantage of its full stack perspective. [Moreover,] they’re not putting everything under the hood. They’re saying, here’s our error correction. Here’s how you guys can play in that game as well,” said Sorensen.

Sorensen said, “I was more struck by IBM’s emphasis on runtime improvement and how all of its efforts, not just error correction per se, are aimed at reducing runtime. They’re moving away from a somewhat more abstract, classical versus quantum argument, to an emphasis on how their runtime is improving on these applications. That’s something that every classical guy can resonate with. That emphasis on runtime was an interesting perspective, because it embodies all of the things that they’re looking at – circuit depth, error rates, and the algorithms themselves. And it’s a much more accessible version of in my mind of the quantum volume measure (benchmark measure).”

Sorensen thought the blog, ostensibly about error mitigation and achieving quantum advantage, was really a broader positioning statement. The lengthy IBM blog did encompass a mini-summary of its steady advances to date, along with a deeper description of its error mitigation efforts.

Looking at error mitigation, IBM has developed two general purpose methods: one called zero noise extrapolation (ZNE); and one called probabilistic error cancellation (PEC). “The ZNE method cancels subsequent orders of the noise affecting the expectation value of a noisy quantum circuit by extrapolating measurement outcomes at different noise strength,” write the authors. PEC, says IBM, can already enable noise-free estimators of quantum circuits on noisy quantum computers.

IBM issued a PEC paper in January, which noted, “Noise in pre-fault-tolerant quantum computers can result in biased estimates of physical observables. Accurate bias-free estimates can be obtained using probabilistic error cancellation (PEC), which is an error-mitigation technique that effectively inverts well-characterized noise channels. Learning correlated noise channels in large quantum circuits, however, has been a major challenge and has severely hampered experimental realizations. Our work presents a practical protocol for learning and inverting a sparse noise model that is able to capture correlated noise and scales to large quantum devices.”

IBM was able to demonstrate PEC on a superconducting quantum processor with crosstalk errors. Here’s a somewhat lengthy excerpt (to avoid garbling) from today’s blog digging into PEC:

“In PEC, we learn and effectively invert the noise of the circuit of interest for the computation of expectation values by averaging randomly sampled instances of noisy circuits. The noise inversion, however, leads to pre-factors in the measured expectation values that translate into a circuit sampling overhead. This overhead is exponential in the number of qubits n and circuit depth d. The base of this exponential, represented as γ̄, is a property of the experimentally learned noise model and its inversion. We can therefore conveniently express the circuit sampling overhead as a quantum runtime, J:

     J = γ̄ nd β d

“Here, γ̄ is a powerful quality metric of quantum processors — in technical terms, γ̄ is a measure of the negativity of the quasi-probability distribution used to represent the inverse of the noise channel. Improvements in qubit coherence, gate fidelity and crosstalk will reflect in lower γ̄ values and consequently reduce the PEC runtime dramatically. Meanwhile, β is a measure of the time per circuit layer operation (see CLOPS), an important speed metric.

“The above expression therefore highlights why the path to quantum advantage will be one driven by improvements in the quality and speed of quantum systems as their scale grows to tackle increasingly complex circuits. In recent years, we have pushed the needle on all three fronts.”

Here’s IBM’s overall checklist of notable milestones:

  • “We unveiled our 127-qubit Eagle processor, pushing the scale of these processors well beyond the boundaries of exact classical stimulability.
  • “We also introduced a metric to quantify the speed of quantum systems — CLOPs — and demonstrated a 120x reduction in the runtime of a molecular simulation.
  • “The coherence times of our transmon qubits exceeded 1 ms, an incredible milestone for superconducting qubit technology.
  • “Since then, these improvements have extended to our largest processors, and our 65-qubit Hummingbird processors have seen a 2-3x improvement in coherence, which further enables higher fidelity gates.
  • “In our latest Falcon r10 processor, IBM Prague, two-qubit gate errors dipped under 0.1%, yet another first for superconducting quantum technology, allowing this processor to demonstrate two steps in Quantum Volume of 256 and 512.”

The blog, while not at the depth of a technical paper, has a fuller discussion of the error mitigation schemes, and IBM presents a few examples of error-corrected circuits run on its processors versus classical.

IBM points out that while the improvements in γ̄ might seem modest, the run-time implications (exponential) of these γ̄ improvements are huge — for instance, for a 100-qubit circuit of depth 100, the runtime overhead going from the quality of Hummingbird r2 (version of processor) to Falcon r10 is dramatically reduced by 110 orders of magnitude.

IBM presents a second example of how reductions in two-qubit gate errors will lead to dramatic improvements in quantum runtime. “Consider the example of 100 qubit Trotter-type circuits of varying depth representing the evolution of an Ising spin chain. This is at a size well beyond exact classical simulation. In the figure below we estimate the PEC circuit overhead as a function of two-qubit gate error.

“As discussed above, two-qubit gate errors on our processors have recently dipped below 1e-3. This suggests that if we could further reduce our gate error down to ~ 2-3e-4, we could have access to noise-free observables of 100 qubit, depth 400 circuits in less than a day of runtime. Furthermore, this runtime will be reduced even further as we simultaneously improve the speed of our quantum systems.”

The IBM team write, “We can therefore think of our proposed path to practical quantum computing as a game of exponentials. Despite the exponential cost of PEC, this framework, and the mathematical guarantees on the accuracy of error mitigated expectation values, allow us to lay down a concrete path and a metric (γ̄) to quantify when even noisy processors can achieve computational advantage over classical machines. The metrics γ̄ and β provide numerical criteria for when the runtime of quantum circuits will outperform classical methods that produce accurate expectation values, even before the advent of fault tolerant quantum computation.

“To understand this better, let us consider the example of a specific circuit family: hardware-efficient circuits with layers of Clifford entangling gates and arbitrary single-qubit gates. These circuits are extremely relevant for a host of applications ranging from chemistry to machine learning. The table below compares the runtime scaling of PEC with the runtime scaling of the best classical circuit simulation methods, and estimates the γ̄ required for error mitigated quantum computation to be competitive with best-known classical techniques.”

The blog is best read directly and provides a better view of what IBM sees as its path to quantum advantage using error mitigation techniques. So far, IBM has generally hit its milestones (hardware and software). The latest roadmap includes delivery of previously announced single-chip 433-qubit Osprey processor this year and 1,121-qubit Condor processor next year, but introduces newly architected processors intended for use in multi-chip modules, starting with a 133-qubit (Heron) processor in 2023, a 408-qubit processor (Crossbill) and 462-qubit processor (Flamingo) in 2024, and leading to the 1386-qubit processor (multiple chips) Kookaburra in 2025.

The blog’s concluding statement was quite broad:

“At IBM Quantum, we plan to continue developing our hardware and software with this path in mind. As we improve the scale, quality, and speed of our systems, we expect to see decreases in γ̄ and β resulting in improvements in quantum runtime for circuits of interest.

“At the same time, together with our partners and the growing quantum community, we will continue expanding the list of problems that we can map to quantum circuits and develop better ways of comparing quantum circuit approaches to traditional classical methods to determine if a problem can demonstrate quantum advantage. We fully expect that this continuous path that we have outlined will bring us practical quantum computing.”

Link to IBM blog, https://research.ibm.com/blog/gammabar-for-quantum-advantage

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industry updates delivered to you every week!

Google’s DataGemma Tackles AI Hallucination

September 18, 2024

The rapid evolution of large language models (LLMs) has fueled significant advancement in AI, enabling these systems to analyze text, generate summaries, suggest ideas, and even draft code. However, despite these impress Read more…

Quantum and AI: Navigating the Resource Challenge

September 18, 2024

Rapid advancements in quantum computing are bringing a new era of technological possibilities. However, as quantum technology progresses, there are growing concerns about the availability of resources—a challenge remin Read more…

Intel’s Falcon Shores Future Looks Bleak as It Concedes AI Training to GPU Rivals

September 17, 2024

Intel's Falcon Shores future looks bleak as it concedes AI training to GPU rivals On Monday, Intel sent a letter to employees detailing its comeback plan after an abysmal second-quarter earnings report with critics calli Read more…

AI Helps Researchers Discover Catalyst for Green Hydrogen Production

September 16, 2024

Researchers from the University of Toronto have used AI to generate a “recipe” for an exciting new catalyst needed to produce green hydrogen fuel. As the effects of climate change begin to become more apparent in our Read more…

The Three Laws of Robotics and the Future

September 14, 2024

Isaac Asimov's Three Laws of Robotics have captivated imaginations for decades, providing a blueprint for ethical AI long before it became a reality. First introduced in his 1942 short story "Runaround" from the "I, R Read more…

Microsoft, Quantinuum Use Hybrid Workflow to Simulate Catalyst

September 13, 2024

Microsoft and Quantinuum reported the ability to create 12 logical qubits on Quantinuum's H2 trapped ion system this week and also reported using two logical qubits on an H1 system to simulate an iron catalyst's low ener Read more…

Google’s DataGemma Tackles AI Hallucination

September 18, 2024

The rapid evolution of large language models (LLMs) has fueled significant advancement in AI, enabling these systems to analyze text, generate summaries, sugges Read more…

Quantum and AI: Navigating the Resource Challenge

September 18, 2024

Rapid advancements in quantum computing are bringing a new era of technological possibilities. However, as quantum technology progresses, there are growing conc Read more…

Shutterstock_2176157037

Intel’s Falcon Shores Future Looks Bleak as It Concedes AI Training to GPU Rivals

September 17, 2024

Intel's Falcon Shores future looks bleak as it concedes AI training to GPU rivals On Monday, Intel sent a letter to employees detailing its comeback plan after Read more…

The Three Laws of Robotics and the Future

September 14, 2024

Isaac Asimov's Three Laws of Robotics have captivated imaginations for decades, providing a blueprint for ethical AI long before it became a reality. First i Read more…

GenAI: It’s Not the GPUs, It’s the Storage

September 12, 2024

A recent news release from Data storage company WEKA and S&P Global Market Intelligence unveiled the findings of their second annual Global Trends in AI rep Read more…

Shutterstock 793611091

Argonne’s HPC/AI User Forum Wrap Up

September 11, 2024

As fans of this publication will already know, AI is everywhere. We hear about it in the news, at work, and in our daily lives. It’s such a revolutionary tech Read more…

Quantum Software Specialist Q-CTRL Inks Deals with IBM, Rigetti, Oxford, and Diraq

September 10, 2024

Q-CTRL, the Australia-based start-up focusing on quantum infrastructure software, today announced that its performance-management software, Fire Opal, will be n Read more…

AWS’s High-performance Computing Unit Has a New Boss

September 10, 2024

Amazon Web Services (AWS) has a new leader to run its high-performance computing GTM operations. Thierry Pellegrino, who is well-known in the HPC community, has Read more…

Everyone Except Nvidia Forms Ultra Accelerator Link (UALink) Consortium

May 30, 2024

Consider the GPU. An island of SIMD greatness that makes light work of matrix math. Originally designed to rapidly paint dots on a computer monitor, it was then Read more…

AMD Clears Up Messy GPU Roadmap, Upgrades Chips Annually

June 3, 2024

In the world of AI, there's a desperate search for an alternative to Nvidia's GPUs, and AMD is stepping up to the plate. AMD detailed its updated GPU roadmap, w Read more…

Nvidia Shipped 3.76 Million Data-center GPUs in 2023, According to Study

June 10, 2024

Nvidia had an explosive 2023 in data-center GPU shipments, which totaled roughly 3.76 million units, according to a study conducted by semiconductor analyst fir Read more…

Shutterstock_1687123447

Nvidia Economics: Make $5-$7 for Every $1 Spent on GPUs

June 30, 2024

Nvidia is saying that companies could make $5 to $7 for every $1 invested in GPUs over a four-year period. Customers are investing billions in new Nvidia hardwa Read more…

Comparing NVIDIA A100 and NVIDIA L40S: Which GPU is Ideal for AI and Graphics-Intensive Workloads?

October 30, 2023

With long lead times for the NVIDIA H100 and A100 GPUs, many organizations are looking at the new NVIDIA L40S GPU, which it’s a new GPU optimized for AI and g Read more…

Shutterstock 1024337068

Researchers Benchmark Nvidia’s GH200 Supercomputing Chips

September 4, 2024

Nvidia is putting its GH200 chips in European supercomputers, and researchers are getting their hands on those systems and releasing research papers with perfor Read more…

IonQ Plots Path to Commercial (Quantum) Advantage

July 2, 2024

IonQ, the trapped ion quantum computing specialist, delivered a progress report last week firming up 2024/25 product goals and reviewing its technology roadmap. Read more…

Google Announces Sixth-generation AI Chip, a TPU Called Trillium

May 17, 2024

On Tuesday May 14th, Google announced its sixth-generation TPU (tensor processing unit) called Trillium.  The chip, essentially a TPU v6, is the company's l Read more…

Leading Solution Providers

Contributors

Intel’s Next-gen Falcon Shores Coming Out in Late 2025 

April 30, 2024

It's a long wait for customers hanging on for Intel's next-generation GPU, Falcon Shores, which will be released in late 2025.  "Then we have a rich, a very Read more…

Atos Outlines Plans to Get Acquired, and a Path Forward

May 21, 2024

Atos – via its subsidiary Eviden – is the second major supercomputer maker outside of HPE, while others have largely dropped out. The lack of integrators and Atos' financial turmoil have the HPC market worried. If Atos goes under, HPE will be the only major option for building large-scale systems. Read more…

xAI Colossus: The Elon Project

September 5, 2024

Elon Musk's xAI cluster, named Colossus (possibly after the 1970 movie about a massive computer that does not end well), has been brought online. Musk recently Read more…

Department of Justice Begins Antitrust Probe into Nvidia

August 9, 2024

After months of skyrocketing stock prices and unhinged optimism, Nvidia has run into a few snags – a  design flaw in one of its new chips and an antitrust pr Read more…

MLPerf Training 4.0 – Nvidia Still King; Power and LLM Fine Tuning Added

June 12, 2024

There are really two stories packaged in the most recent MLPerf  Training 4.0 results, released today. The first, of course, is the results. Nvidia (currently Read more…

Spelunking the HPC and AI GPU Software Stacks

June 21, 2024

As AI continues to reach into every domain of life, the question remains as to what kind of software these tools will run on. The choice in software stacks – Read more…

Nvidia H100: Are 550,000 GPUs Enough for This Year?

August 17, 2023

The GPU Squeeze continues to place a premium on Nvidia H100 GPUs. In a recent Financial Times article, Nvidia reports that it expects to ship 550,000 of its lat Read more…

Shutterstock 1886124835

Researchers Say Memory Bandwidth and NVLink Speeds in Hopper Not So Simple

July 15, 2024

Researchers measured the real-world bandwidth of Nvidia's Grace Hopper superchip, with the chip-to-chip interconnect results falling well short of theoretical c Read more…

  • arrow
  • Click Here for More Headlines
  • arrow
HPCwire