Visit additional Tabor Communication Publications
July 16, 2009
Ongoing collaborations with the Data Intensive Cyber Environments group aims to make NC a data leader
CHAPEL HILL, NC, July 16 -- Almost a year after the Data Intensive Cyber Environments (DICE) research group moved from the University of California at San Diego to the University of North Carolina at Chapel Hill, the internationally recognized research group has established deep ties to the Renaissance Computing Institute (RENCI).
Together, DICE and RENCI are working to link data repositories across the state and make them accessible and easy to use by researchers, businesses and government. Through a variety of collaborations, the two groups hope to establish North Carolina as an international leader in solving the pressing problems of managing and sharing today's deluge of digital data in ways that provide a boost to research, business and education.
DICE members, long of UCSD's, San Diego Supercomputer Center, made the cross county trek to UNC Chapel Hill early last fall, bringing with them a research portfolio in excess of $10 million. The researchers hold appointments in Carolina's nationally recognized School of Information and Library Science (SILS) and also serve as chief scientists at RENCI, with research space at RENCI's off campus Chapel Hill location.
For more than 10 years the DICE group's data grid technologies have been used in research projects worldwide to manage large, distributed data collections and support discovery, access, retrieval, replication, archiving, and analysis tasks. The researchers most recently released version 2.1 of iRODS, the open source Integrated Rule-Oriented Data System, which introduced user-settable rules that automate complex management policies, helping users handle today's mushrooming collections of digital data.
"Having RENCI in North Carolina was one of the main considerations in coming here," said Reagan Moore, head of the DICE group, director of the new DICE Center at UNC-Chapel Hill and chief scientist for data intensive cyber environments at RENCI. "RENCI works to provide the technology infrastructure that thriving universities and business communities need and a data infrastructure is part of that. Working together, RENCI and DICE can develop the kind of technology infrastructure that opens up all kinds of opportunities."
RENCI works closely with the DICE Center, which was launched in May with funding from the National Science Foundation, the National Archives and Records Administration and the National Historical Publications and Records Commission. The center is working to establish an interoperable data center that spans the 17-campus UNC system that allows campus researchers to easily access, search, share and mine vast stores of data. The bulk of the center's data cache will be located at RENCI, which currently provides about 100 terabytes of data storage to the DICE center. Eventually, said Moore, DICE hopes to have a data infrastructure, supported by RENCI, that provides more than 1 petabyte of data management and storage capacity.
"The indispensable role of digital data across society and the increasing size and complexity of data collections are reaching a critical point," said Richard Marciano, executive director of the DICE center and RENCI chief scientist for persistent archives and digital preservation. "With the growing need for practical digital data technologies, the new DICE Center is already collaborating with many important projects across UNC-Chapel Hill as well as national and international partners, helping them harness their digital data collections and working with them to efficiently create, share and preserve new knowledge."
Regional hubs through Data Grid
A premier DICE-RENCI project is the RENCI Data Grid, which links RENCI's engagement centers at UNC-Chapel Hill, Duke, NC State, UNC Charlotte, UNC Asheville, and East Carolina University as interconnected regional data hubs managed by iRODS. The project began in late 2008 by adding multi-terabyte storage capacity at the RENCI engagement centers and at RENCI's offices in Chapel Hill. Those data hubs will house RENCI visualization files, the data on which the visualizations are based, and the visual analytics software tools that enable many additional levels of insight into the data. But Data Grid's potential is much greater, according to Ray Idaszak, RENCI's director of visualization and collaborative environments.
"Data Grid is the start of an effort by DICE and RENCI to give the state and its universities a data resource that will be a tool used for decision making, for increasing productivity or for examining data in the context of other data," said Idaszak.
When completed, the Data Grid in action might work like this: Data on development patterns around the North Carolina would be stored at RENCI at UNC Charlotte, where researchers at the RENCI engagement center study urban growth patterns and their implications. An urban planner in eastern North Carolina would be able to access that data as well as the software tools that allow it to be viewed in a visual, intuitive format. Those same researchers also would be able to access coastal floodplain maps and storm surge visualizations stored at other data hubs and to use all of the information to plan sustainable coastal developments.
One of the largest datasets the Data Grid plans to offer will be orthophotographic images--an extremely high resolution photographic map of the state with data points at every six inches across a statewide grid. When overlaid with other datasets--for example, census data or information on the location of healthcare facilities--the orthophotographs become tools useful to emergency managers, city planners, researchers and area businesses.
All of the Data Grid hubs will be operational by fall 2009 and new datasets will be added to the collection as more storage capacity becomes available.
Other projects on which RENCI and DICE collaborate include:
Large-scale, worldwide scientific initiatives rely on some cloud-based system to both coordinate efforts and manage computational efforts at peak times that cannot be contained within the combined in-house HPC resources. Last week at Google I/O, Brookhaven National Lab’s Sergey Panitkin discussed the role of the Google Compute Engine in providing computational support to ATLAS, a detector of high-energy particles at the Large Hadron Collider (LHC).
The Xeon Phi coprocessor might be the new kid on the high performance block, but out of all first-rate kickers of the Intel tires, the Texas Advanced Computing Center (TACC) got the first real jab with its new top ten Stampede system.We talk with the center's Karl Schultz about the challenges of programming for Phi--but more specifically, the optimization...
Although Horst Simon was named Deputy Director of Lawrence Berkeley National Laboratory, he maintains his strong ties to the scientific computing community as an editor of the TOP500 list and as an invited speaker at conferences.
May 16, 2013 |
When it comes to cloud, long distances mean unacceptably high latencies. Researchers from the University of Bonn in Germany examined those latency issues of doing CFD modeling in the cloud by utilizing a common CFD and its utilization in HPC instance types including both CPU and GPU cores of Amazon EC2.
May 15, 2013 |
Supercomputers at the Department of Energy’s National Energy Research Scientific Computing Center (NERSC) have worked on important computational problems such as collapse of the atomic state, the optimization of chemical catalysts, and now modeling popping bubbles.
May 10, 2013 |
Program provides cash awards up to $10,000 for the best open-source end-user applications deployed on 100G network.
May 09, 2013 |
The Japanese government has revealed its plans to best its previous K Computer efforts with what they hope will be the first exascale system...
05/10/2013 | Cleversafe, Cray, DDN, NetApp, & Panasas | From Wall Street to Hollywood, drug discovery to homeland security, companies and organizations of all sizes and stripes are coming face to face with the challenges – and opportunities – afforded by Big Data. Before anyone can utilize these extraordinary data repositories, however, they must first harness and manage their data stores, and do so utilizing technologies that underscore affordability, security, and scalability.
04/15/2013 | Bull | “50% of HPC users say their largest jobs scale to 120 cores or less.” How about yours? Are your codes ready to take advantage of today’s and tomorrow’s ultra-parallel HPC systems? Download this White Paper by Analysts Intersect360 Research to see what Bull and Intel’s Center for Excellence in Parallel Programming can do for your codes.
In this demonstration of SGI DMF ZeroWatt disk solution, Dr. Eng Lim Goh, SGI CTO, discusses a function of SGI DMF software to reduce costs and power consumption in an exascale (Big Data) storage datacenter.
The Cray CS300-AC cluster supercomputer offers energy efficient, air-cooled design based on modular, industry-standard platforms featuring the latest processor and network technologies and a wide range of datacenter cooling requirements.