Tag: data analytics
For all the accolades one hears about German engineering, there are few IT vendors native to that country. Recently though, we got the opportunity to talk with one such company, ParStream, a Cologne-based startup that has developed a bleeding-edge CPU/GPU-based analytics platform that marries high performance computing to big data.
The Weekly Top Five features the five biggest HPC stories of the week, condensed for your reading pleasure. This week, we cover the NC State effort to overcome the memory limitations of multicore chips; the sale of the first-ever commercial quantum computing system; Cray’s first GPU-accelerated machine; speedier machine learning algorithms; and the connection between shrinking budgets and increased reliance on modeling and simulation.
Culling together massive data has provided some profound opportunities for a wide array of analytics projects but has created a number of complications for those who want to gain actionable intelligence from it. While the “big data” movement is still unfolding, a number of companies have emerged to help simplify access and use, especially of unstructured information. HPC stalwart Platform Computing entered the race to refine handling of vast datasets — not to mention the management behind such operations to stake their claim in this emerging space.
Advanced data analytics applications are replacing flesh-and-blood analysts.
When announced in 2006, the Cray XMT supercomputer attracted little attention. The machine was originally targeted for high-end data mining and analysis for a particular set of government clients in the intelligence community. While the feds have given the XMT support over the past five years, Cray is now looking to move these machines into the commercial sphere. And with the next generation XMT-2 on the horizon, the company is gearing up to accelerate that strategy in 2011.
Yes, there is life beyond Xeons, Opterons and GPGPUs.
Complex event processing may be a technology whose time has come.