Since 1987 - Covering the Fastest Computers in the World and the People Who Run Them

September 19, 2012

For Energy-Efficient HPC, Less is More

Ian Armas Foster

Near-Threshold Voltage, or NTV, has the potential to significantly cut energy requirements for high performance computing. This is becoming especially important for the largest supercomputers, which are already well into the multi-megawatt realm and are expected to hit tens of megawatts in the exascale era.

Intel recently demonstrated their NTV capabilities at ISSCC 2012, operating an x86 microprocessor on only 2 milliWatts of power. The company published three papers on their results, which were analyzed and discussed in an article at Real World Technologies by David Kanter.

The threshold voltage is the voltage required to generate a minimum of current across a transistor. Intel has discovered that the most efficient use of energy would be to operate a circuit near that of its threshold voltage, that is, the amount required to turn a transistor on.

There are a couple of intrinsically tricky things about operating at such a low voltage. The first is limiting dI/dt, the fancy mathematical way of expressing change in current over time. Rapid spikes or drops in current, especially those that would occur as a result of a particular transistor accidentally dropping beneath the threshold, can create computational errors.

Ideally, all transistors would be created equally. Statistically, however, since there sometimes lie billions of transistors on a given chip, some will perform worse that others.

Another challenge to overcome is the resulting power loss. The power available is proportional to the square of the voltage, such that a 10 percent reduction in voltage leads to a 19 percent reduction in power. While this reduced voltage would be a great way to increase efficiency, it would also be a great way to ensure your CPU does not have the juice required to run what it needs to.

Further, NTV significantly decreases frequency. “The 32nm Pentium core,” Kanter said about a core that ran using NTV “increased efficiency by about 5×, by running at slightly under 100MHz. The maximum frequency was 915MHz, so the absolute performance decreased by about an order of magnitude.”

As he notes, NTV would be impractical for general-purpose CPUs, as they are generally used for applications that expect reasonable single-threaded performance. Thus they require the higher voltages needed to drive faster clocks. On the other hand, HPC and its massively parallel computing environment could benefit greatly from NTV.

“Based on our analysis of these papers,” Kanter wrote, “Near-Threshold Voltage computing techniques are most applicable to highly parallel workloads. Generally, NTV is an ideal fit for HPC workloads and works very well for graphics, but not general purpose CPUs.”

Since HPC is highly parallelized and requires backups and fail-safe mechanisms throughout a computation, it can withstand the consequences of a single transistor giving out. HPC computations are also not expected to happen anywhere near real time, making the frequency decrease less of a problem. This is especially true of “throughput” accelerators like GPGPUs and Intel’s Xeon Phi, which are naturally frequency-constrained because of their high core counts.

There is a sense that this technology is being developed to specifically benefit HPC rather than it accidentally doing so. This is not only hinted at by the Intel papers themselves, but is also indirectly supported by the people funding the papers, specifically the US Government. “Perhaps most telling,” Kanter wrote, “US government grants typically focus on areas of national interest. Graphics simply is not vital to the country, whereas HPC is a critical tool for the Departments of Defense, Energy, and any number of intelligence agencies.”

Full story at Real World Technologies

Share This