Visit additional Tabor Communication Publications
July 13, 2010
The Universe might be expanding, but at least it's getting easier to see. On Monday, at the annual Microsoft Research Faculty Summit, the software maker unveiled the largest and clearest image of the night sky ever assembled. This so-called "TeraPixel" sky map was generated with the help of some of Microsoft's latest HPC and parallel software assets.
The TeraPixel project from the folks at Microsoft Research was essentially a recomputation of the image data collected by Digitized Sky Survey over the last 50 years. The input data was made up of 1,791 pairs of red-light and blue-light plates produced by two ground high-powered telescopes: the Palomar telescope in California (US) and the UK Schmidt telescope in New South Wales (Australia). Between them, the two installations covered the night sky of the Northern and Southern hemispheres.
As one might suspect of photographs collected over a long period of time with different equipment and under different conditions, the quality of the images varied considerably. Different color saturation, contrast, noise, and brightness, as well as the presence of vignetting (darkening toward the image corners) meant that the data would require a lot of post-processing to produce what the researchers were going for: a seamless photograph of the entire sky.
Compared to the old sky image, the TeraPixel version is much more refined. With all the artifacts, seams and inconsistencies processed away, it looks like a true unified image of the sky above. It's like going from Super Mario Brothers on 1985-era Nintendo consoles to Halo 2 on Xbox 360s.
According to Dan Fay, the director of Earth, Energy and Environment at Microsoft Research, to get this level of refinement, all the images had to go through a four-stage processing to correct for the irregularities. The first stage attacked the vignetting artifact to brighten up the dark corners. The next step was more complex. Since each plate had a red and blue version for each area, these had to be processed separately and then realigned into one image. They even had to account for multiple overlapping plates. In some cases, Fay says, they chose the best pixels on the various plates to come up with the highest quality image. The third step involved stitching the individual images together and smoothing out the seams. Lastly, the multi-resolution images were generated so that users could zoom in for greater detail. The final result was a spherical panorama of the night sky in 24-bit RGB format.
Much of the software relied on Microsoft software as well as Microsoft programmers. The project used the global image optimization program developed by Hugues Hoppe and Dinoj Surendran of Microsoft Research and Michael Kazhdan of Johns Hopkins. The DryadLINQ and the .NET parallel extensions framework was employed to construct and manage the applications. DryadLINQ is a programming environment for running parallel applications across a cluster, using LINQ (Language Integrated Query) as a query engine on top of the Dryad runtime. The latter takes the queries and distribute them across the nodes. Windows HPC Server was used to schedule the more tightly-coupled jobs and the Project Trident Workbench was employed to manage the entire workflow.
By HPC standards, the hardware platform was relatively modest. A 16-node Intel Xeon cluster was used to process the TeraPixel image, but the final runs were done on a 64-node system. The image was built iteratively since the algorithms were continuously tweaked to get better refinement. A full run on the 16-node machine took three days, while on the larger machine, it took just over half a day.
One of the costliest operations, time-wise, was shuffling the images around the cluster. "Some of the biggest issues were data movement," notes Fay. "When you start getting to that many nodes and parallel jobs, moving the data ends up taking a lot of the time." Just transferring the final 1,025 files (802 GB total) off the cluster took 2.5 hours using a 1 Gbps link.
The TeraPixel image can be viewed by researchers and the general public on Microsoft Research's WorldWide Telescope Web site. It also can be accessed via Bing Maps, via a plug-in, where you can see a street-wise view of the sky above. Because of the high resolution of the imagery, viewers are able to zoom into any area of the sky and see greater detail of specific star systems.
The sky image they've produced has been verified by astronomers, who made sure that nothing is rotated incorrectly or is otherwise erroneous. According to Fay, the feedback from the community has been gratifying. "No one has ever seen an image of the sky like this," he says.
May 23, 2013 |
The study of climate change is one of those scientific problems where it is almost essential to model the entire Earth to attain accurate results and make worthwhile predictions. In an attempt to make climate science more accessible to smaller research facilities, NASA introduced what they call ‘Climate in a Box,’ a system they note acts as a desktop supercomputer.
May 22, 2013 |
At some point in the not-too-distant future, building powerful, miniature computing systems will be considered a hobby for high schoolers, just as robotics or even Lego-building are today. That could be made possible through recent advancements made with the Raspberry Pi computers.
May 16, 2013 |
When it comes to cloud, long distances mean unacceptably high latencies. Researchers from the University of Bonn in Germany examined those latency issues of doing CFD modeling in the cloud by utilizing a common CFD and its utilization in HPC instance types including both CPU and GPU cores of Amazon EC2.
May 15, 2013 |
Supercomputers at the Department of Energy’s National Energy Research Scientific Computing Center (NERSC) have worked on important computational problems such as collapse of the atomic state, the optimization of chemical catalysts, and now modeling popping bubbles.
05/10/2013 | Cleversafe, Cray, DDN, NetApp, & Panasas | From Wall Street to Hollywood, drug discovery to homeland security, companies and organizations of all sizes and stripes are coming face to face with the challenges – and opportunities – afforded by Big Data. Before anyone can utilize these extraordinary data repositories, however, they must first harness and manage their data stores, and do so utilizing technologies that underscore affordability, security, and scalability.
04/15/2013 | Bull | “50% of HPC users say their largest jobs scale to 120 cores or less.” How about yours? Are your codes ready to take advantage of today’s and tomorrow’s ultra-parallel HPC systems? Download this White Paper by Analysts Intersect360 Research to see what Bull and Intel’s Center for Excellence in Parallel Programming can do for your codes.
In this demonstration of SGI DMF ZeroWatt disk solution, Dr. Eng Lim Goh, SGI CTO, discusses a function of SGI DMF software to reduce costs and power consumption in an exascale (Big Data) storage datacenter.
The Cray CS300-AC cluster supercomputer offers energy efficient, air-cooled design based on modular, industry-standard platforms featuring the latest processor and network technologies and a wide range of datacenter cooling requirements.