Visit additional Tabor Communication Publications
July 04, 2012
CAMBRIDGE, Mass., July 3 -- It’s a challenge that’s long been one of the holy grails of quantum computing: how to create the key building blocks known as quantum bits, or qubits, that exist in a solid-state system at room temperature.
Most current systems, by comparison, rely on complex and expensive equipment designed to trap a single atom or electron in a vacuum and then cool the entire system to close to absolute zero.
A group of Harvard scientists, led by Professor of Physics Mikhail Lukin and including graduate students Georg Kucsko and Peter Maurer and postdoctoral researcher Christian Latta, say they’ve cracked the problem, and they did it by turning to one of the purest materials on Earth: diamonds.
Using a pair of impurities in ultra-pure, laboratory-grown diamonds, the researchers were able to create quantum bits and store information in them for nearly two seconds, an increase of nearly six orders of magnitude over the life span of earlier systems. The work, described in the June 8 issue of Science, is a critical first step in the eventual construction of a functional quantum computer, and has a host of other potential applications.
“What we’ve been able to achieve in terms of control is quite unprecedented,” Lukin said. “We have a qubit, at room temperature, that we can measure with very high efficiency and fidelity. We can encode data in it, and we can store it for a relatively long time. We believe this work is limited only by technical issues, so it looks feasible to increase the life span into the range of hours. At that point, a host of real-world applications become possible.”
In addition to a practical quantum computer, Lukin envisions the system being used in applications that include “quantum cash” (a payment system for bank transactions and credit cards that relies on the coding of quantum bits to thwart counterfeiters) and quantum networks (a highly secure communications method that uses quantum bits to transmit data).
“This research is an important step forward in research toward one day building a practical quantum computer,” said Kucsko, who works in Lukin’s lab and is one of two first authors of the paper. “For the first time, we have a system that has a reasonable timescale for memory and simplicity, so this is now something we can pursue.”
The groundwork for Lukin’s breakthrough was laid several years ago, when researchers discovered that nitrogen-vacancy (NV) centers, atomic-scale impurities in lab-grown diamonds, behave in the same way as single atoms. Like individual atoms, each center possesses a spin, which can be polarized, similar to on a bar magnet. Using lasers, researchers are able not only to control the spin, but to detect its orientation as it changes over time.
But the idea of using the NV centers to form the backbone of a quantum computer simply wasn’t practical, largely because they can only hold data for about one-millionth of a second before their quantum properties — and any data they may have held — are lost.
The culprit, Lukin said, was another impurity in the diamond crystal.
In initial experiments, the team used diamonds that contained 99 percent carbon-12 atoms, which have no spin. The remainder, however, was made up of carbon-13 atoms, a tricky isotope that contains a spin in the atom’s nucleus. Though weak, the interaction with those spins was causing the NV centers’ short life spans.
With this latest research, however, Lukin and his team turned what was once a challenge — the interaction between the NV center and carbon-13 atoms — to their advantage.
“The nuclear spin of the carbon-13 makes an ideal quantum bit, because they are very isolated,” Lukin said. “Because they interact with so few outside forces, they have relatively long coherence times. Of course, the same properties that make them ideal qubits also make them difficult to measure and manipulate.”
The solution Lukin and his team came up with was surprisingly elegant. Rather than trying to find a way to measure the spin of the carbon atoms, they used the NV center to do it for them.
Working with researchers at Element Six, a British-based company that specializes in manufacturing artificial diamonds, they developed a new technique to create crystals that were even more pure: 99.99 percent carbon-12. Researchers then bombard the crystal with nitrogen to create the NV center, which interacts with a nearby carbon-13 atom.
The result of that interaction is that the NV center mirrors the state of the carbon atom, meaning researchers can encode a bit of information into the spin of the atom, then “read” that data by monitoring the NV center.
“The system we’ve developed uses this very local probe, the NV center, to allow us to monitor that spin,” Lukin said. “As a result, for the first time, we can encode a bit of information into that spin, and use this system to read it out.”
However, encoding information into the spin of the carbon-13 atom and reading it out using the NV center is only a step on the road to a quantum computer. To truly be useful, researchers had to determine how to take advantage of the atom’s quantum properties — that is, its ability to occupy two states simultaneously.
That ability to be in two states at the same time is a key principle of quantum computers. As opposed to traditional computers, which encode bits of information as either zero or one, quantum computers rely on atomic-scale quantum mechanics to give quantum bits both values at once. That property, in theory, allows quantum computers to perform multiple computations in parallel, making them vastly more powerful than traditional computers, which perform operations in sequence.
The solution, Lukin explained, was a two-step process.
The first step is to cut the connection between the NV center and the carbon atom. Using massive amounts of laser light, researcher are able to effectively keep the NV center occupied and prevent it from interacting with the carbon atom. In step two, the diamond crystal is bombarded with a specific set of radio frequency pulses, suppressing the interaction between the carbon-13 atom and any nearby atoms.
“By limiting interactions with the carbon-13 atom, we can extend the life of the qubit and hold the data for longer,” Lukin said. “The end result is that we’re able to push the coherence time from a millisecond to nearly two seconds.”
Researchers at the California Institute of Technology and the Max-Planck-Institut für Quantenoptik also participated in the research. Funding was provided by the National Science Foundation, the Center for Ultracold Atoms, the Defense Advanced Research Projects Agency, Element 6, the Packard Foundation, the European Union, the Swiss National Science Foundation, and the Sherman Fairchild Foundation.
Source: Harvard University
Large-scale, worldwide scientific initiatives rely on some cloud-based system to both coordinate efforts and manage computational efforts at peak times that cannot be contained within the combined in-house HPC resources. Last week at Google I/O, Brookhaven National Lab’s Sergey Panitkin discussed the role of the Google Compute Engine in providing computational support to ATLAS, a detector of high-energy particles at the Large Hadron Collider (LHC).
The Xeon Phi coprocessor might be the new kid on the high performance block, but out of all first-rate kickers of the Intel tires, the Texas Advanced Computing Center (TACC) got the first real jab with its new top ten Stampede system.We talk with the center's Karl Schultz about the challenges of programming for Phi--but more specifically, the optimization...
Although Horst Simon was named Deputy Director of Lawrence Berkeley National Laboratory, he maintains his strong ties to the scientific computing community as an editor of the TOP500 list and as an invited speaker at conferences.
May 16, 2013 |
When it comes to cloud, long distances mean unacceptably high latencies. Researchers from the University of Bonn in Germany examined those latency issues of doing CFD modeling in the cloud by utilizing a common CFD and its utilization in HPC instance types including both CPU and GPU cores of Amazon EC2.
May 15, 2013 |
Supercomputers at the Department of Energy’s National Energy Research Scientific Computing Center (NERSC) have worked on important computational problems such as collapse of the atomic state, the optimization of chemical catalysts, and now modeling popping bubbles.
May 10, 2013 |
Program provides cash awards up to $10,000 for the best open-source end-user applications deployed on 100G network.
May 09, 2013 |
The Japanese government has revealed its plans to best its previous K Computer efforts with what they hope will be the first exascale system...
05/10/2013 | Cleversafe, Cray, DDN, NetApp, & Panasas | From Wall Street to Hollywood, drug discovery to homeland security, companies and organizations of all sizes and stripes are coming face to face with the challenges – and opportunities – afforded by Big Data. Before anyone can utilize these extraordinary data repositories, however, they must first harness and manage their data stores, and do so utilizing technologies that underscore affordability, security, and scalability.
04/15/2013 | Bull | “50% of HPC users say their largest jobs scale to 120 cores or less.” How about yours? Are your codes ready to take advantage of today’s and tomorrow’s ultra-parallel HPC systems? Download this White Paper by Analysts Intersect360 Research to see what Bull and Intel’s Center for Excellence in Parallel Programming can do for your codes.
In this demonstration of SGI DMF ZeroWatt disk solution, Dr. Eng Lim Goh, SGI CTO, discusses a function of SGI DMF software to reduce costs and power consumption in an exascale (Big Data) storage datacenter.
The Cray CS300-AC cluster supercomputer offers energy efficient, air-cooled design based on modular, industry-standard platforms featuring the latest processor and network technologies and a wide range of datacenter cooling requirements.