November 13, 2023
The fall 2023 TOP500 list is out and Frontier retains its top spot and is still the only exascale machine. However, five new or upgraded systems have shaken up Read more…
July 31, 2023
Esperanto Technologies has ambitious plans for its next RISC-V processor: to undo the accelerator model and build a chip that has both CPU and GPU capabilities Read more…
December 2, 2022
The Frontier supercomputer – still fresh off its chart-topping 1.1 Linpack exaflops run and maintaining its number-one spot on the Top500 list – was still v Read more…
November 14, 2022
Nvidia’s H100 GPU, the flagship of its Hopper architecture, has debuted on the Top500 and Green500 lists at SC22. The new GPU appears in the relatively small Read more…
June 8, 2022
Back in 2008, the U.S. Defense Advanced Research Projects Agency (DARPA) set an ambitious target: an exascale supercomputer in a 20-megawatt envelope. That targ Read more…
December 2, 2021
“This is the 30th Green500,” said Wu Feng, custodian of the Green500 list, at the list’s SC21 birds-of-a-feather session. “You could say 15 years of Green500, which makes it, I guess, the crystal anniversary.” Indeed, HPCwire marked the 15th anniversary of the Green500 – which ranks supercomputers by flops-per-watt, rather than just by flops – earlier this year with... Read more…
July 15, 2021
The Green500 list, which ranks the most energy-efficient supercomputers in the world, has virtually always faced an uphill battle. As Wu Feng – custodian of the Green500 list and an associate professor at Virginia Tech – tells it, “noone" cared about energy efficiency in the early 2000s, when the seeds... Read more…
June 28, 2021
The 57th Top500, revealed today from the ISC 2021 digital event, showcases many of the same systems as the previous edition, with Fugaku holding its significant lead and only one new entrant in the top 10 cohort: the Perlmutter system at the DOE Lawrence Berkeley National Laboratory enters the list at number five with 65.69 Linpack petaflops. Perlmutter is the largest... Read more…
Data centers are experiencing increasing power consumption, space constraints and cooling demands due to the unprecedented computing power required by today’s chips and servers. HVAC cooling systems consume approximately 40% of a data center’s electricity. These systems traditionally use air conditioning, air handling and fans to cool the data center facility and IT equipment, ultimately resulting in high energy consumption and high carbon emissions. Data centers are moving to direct liquid cooled (DLC) systems to improve cooling efficiency thus lowering their PUE, operating expenses (OPEX) and carbon footprint.
This paper describes how CoolIT Systems (CoolIT) meets the need for improved energy efficiency in data centers and includes case studies that show how CoolIT’s DLC solutions improve energy efficiency, increase rack density, lower OPEX, and enable sustainability programs. CoolIT is the global market and innovation leader in scalable DLC solutions for the world’s most demanding computing environments. CoolIT’s end-to-end solutions meet the rising demand in cooling and the rising demand for energy efficiency.
Divergent Technologies developed a digital production system that can revolutionize automotive and industrial scale manufacturing. Divergent uses new manufacturing solutions and their Divergent Adaptive Production System (DAPS™) software to make vehicle manufacturing more efficient, less costly and decrease manufacturing waste by replacing existing design and production processes.
Divergent initially used on-premises workstations to run HPC simulations but faced challenges because their workstations could not achieve fast enough simulation times. Divergent also needed to free staff from managing the HPC system, CAE integration and IT update tasks.
© 2024 HPCwire. All Rights Reserved. A Tabor Communications Publication
HPCwire is a registered trademark of Tabor Communications, Inc. Use of this site is governed by our Terms of Use and Privacy Policy.
Reproduction in whole or in part in any form or medium without express written permission of Tabor Communications, Inc. is prohibited.