November 14, 2023
In 2021, Intel famously declared its goal to get to zettascale supercomputing by 2027, or scaling today's Exascale computers by 1,000 times. Moving forward t Read more…
April 24, 2023
February 14th to April 15th – it’s been a long run for the 2023 Winter Classic Student Cluster Competition. 63 students from HBCU and HSI schools learned ha Read more…
April 7, 2023
The close of the 2023 Winter Classic Invitational Student Cluster Competition is coming up fast, and I have to get some material out to you, our vast viewing au Read more…
January 20, 2023
UT-Battelle, which manages Oak Ridge National Laboratory (ORNL) for the U.S. Department of Energy, has appointed Jeff Smith as interim director for the lab as t Read more…
June 28, 2022
With the Linpack exaflops milestone achieved by the Frontier supercomputer at Oak Ridge National Laboratory, the United States is turning its attention to the next crop of exascale machines, some 5-10x more performant than Frontier. At least one such system is being planned for the 2025-2030 timeline, and the DOE is soliciting input from the vendor community... Read more…
May 30, 2022
In April 2018, the U.S. Department of Energy announced plans to procure a trio of exascale supercomputers at a total cost of up to $1.8 billion dollars. Over the ensuing four years, many announcements were made, many deadlines were missed, and a pandemic threw the world into disarray. Now, at long last, HPE and Oak Ridge National Laboratory (ORNL) have announced that the first of those... Read more…
May 16, 2022
The Department of Energy's Oak Ridge National Laboratory (ORNL) has selected Doug Kothe to be the next Associate Laboratory Director for its Computing and Computational Sciences Directorate (CCSD), HPCwire has learned. Kothe will fill the vacancy created by the retirement of Jeff Nichols, whose last day is July 1. Kothe will transition into the position on June 6. As director of the United States' Exascale Computing Project... Read more…
October 1, 2021
Last year, Summit was one of the first major supercomputing deployments in the fight against COVID-19, heralding a deluge that crescendoed with the cumulative efforts of virtually every research system in the world. While news around COVID-oriented supercomputing research... Read more…
Data centers are experiencing increasing power consumption, space constraints and cooling demands due to the unprecedented computing power required by today’s chips and servers. HVAC cooling systems consume approximately 40% of a data center’s electricity. These systems traditionally use air conditioning, air handling and fans to cool the data center facility and IT equipment, ultimately resulting in high energy consumption and high carbon emissions. Data centers are moving to direct liquid cooled (DLC) systems to improve cooling efficiency thus lowering their PUE, operating expenses (OPEX) and carbon footprint.
This paper describes how CoolIT Systems (CoolIT) meets the need for improved energy efficiency in data centers and includes case studies that show how CoolIT’s DLC solutions improve energy efficiency, increase rack density, lower OPEX, and enable sustainability programs. CoolIT is the global market and innovation leader in scalable DLC solutions for the world’s most demanding computing environments. CoolIT’s end-to-end solutions meet the rising demand in cooling and the rising demand for energy efficiency.
Divergent Technologies developed a digital production system that can revolutionize automotive and industrial scale manufacturing. Divergent uses new manufacturing solutions and their Divergent Adaptive Production System (DAPS™) software to make vehicle manufacturing more efficient, less costly and decrease manufacturing waste by replacing existing design and production processes.
Divergent initially used on-premises workstations to run HPC simulations but faced challenges because their workstations could not achieve fast enough simulation times. Divergent also needed to free staff from managing the HPC system, CAE integration and IT update tasks.
© 2024 HPCwire. All Rights Reserved. A Tabor Communications Publication
HPCwire is a registered trademark of Tabor Communications, Inc. Use of this site is governed by our Terms of Use and Privacy Policy.
Reproduction in whole or in part in any form or medium without express written permission of Tabor Communications, Inc. is prohibited.