The current data center environment was not designed to address the high heat density presented by the new research computing environment.
(PRWEB) July 08, 2014
George Washington University is the largest institution of higher education in the District of Columbia with more than 20,000 students from all 50 states, the District of Columbia and more than 130 countries. It offers a rich range of disciplines: from forensic science and creative writing to international affairs and computer engineering, as well as medicine, public health, the law and public policy
To support new research computing requirements, the university installed a High Performance Computer (HPC) cluster with 24kW rack densities. Additional heat loads from the new cluster proved too much for the hot aisle containment / chimney strategy GWU had implemented in the data center. So as a remediation, chimney fans were installed. Even though the cold aisle was supplying 50° F air, they were barely able to keep up. The result was serious concerns that the hot aisle containment / chimney cooling strategy would not be able to keep up as additional racks of equipment were brought online.
GWU began a search for a data center cooling solution that would address current cooling demands and satisfy future growth requirements. Coolcentric’s High Density Rear Door Heat Exchangers (RDHx-HD) became the cooling solution of choice. The new RDHx-HD is a high capacity passive heat exchanger, closely coupled to the rear of the rack, and capable of removing 30kW of heat before it enters the data center.
Stephen Dobbins, Assistant Director, Data Center said,” The current data center environment was not designed to address the high heat density presented by the new research computing environment. Our goal was to find a solution which presented the HPC environment as a net zero heat load to the airflow capability of the data center. Coolcentric provided the best approach to creating a zero heat load addition to the data center while being able to support the airflow and cooling requirements of the HPC equipment."