- A new supercomputing cooling method on the rise: Refrigeration.
- Testing is underway at National Renewable Energy Lab, and Sandia Labs is in line for next year.
- Refrigeration can save a HPC center millions of gallons of water annually — hundreds of millions nationwide.
In different parts of the country, people discuss gray-water recycling and rainwater capture to minimize the millions of gallons of groundwater required to cool large data centers. But the simple answer in many climates, says Sandia National Laboratories researcher David J. Martinez, is to use liquid refrigerant.
Martinez — engineering project lead for Sandia’s infrastructure computing services — is helping design and monitor a cooling system at Sandia. If successful, it is expected to save 4 million to 5 million gallons annually in New Mexico, and hundreds of millions of gallons nationally if the method is widely adopted. It’s now being tested at the National Renewable Energy Laboratory in Colorado, which expects to save a million gallons annually.
The system, built by Johnson Controls and called the Thermosyphon Cooler Hybrid System, cools like a refrigerator without the expense and energy needs of a compressor.
Currently, many data centers use water to remove waste heat from servers. The warmed water is piped to cooling towers, where a separate stream of water is turned to mist and evaporates into the atmosphere. Like sweat evaporating from the body, the process removes heat from the piped water, which returns to chill the installation.
But large-scale replenishment of the evaporated water is needed to continue the process. Thus, an increasing amount of water will be needed worldwide to evaporate heat from the growing number of data centers, which themselves are increasing in size as more users put information into the cloud.
“My job is to eventually put cooling towers out of business,” Martinez says.
“Ten years ago, I gave a talk on the then-new approach of using water to directly cool supercomputers. There were 30 people at the start of my lecture and only 10 at the end. ‘Dave,’ they said, ‘no way water can cool a supercomputer. You need air.’”
“So now most data centers use water to cool themselves, but I’m always looking at the future and I see refrigerant cooling coming in for half the data centers in the US, north and west of Texas, where the climate will make it work.”
The prototype method uses a liquid refrigerant instead of water to carry away heat, and works like this:
Water heated by the computing center is pumped within a closed system into proximity with another system containing refrigerant. The refrigerant absorbs heat from the water so that the water, now cooled, can circulate to cool again. Meanwhile the heated refrigerant vaporizes and rises in its closed system to exchange heat with the atmosphere. As heat is removed from the refrigerant, it condenses and sinks to absorb more heat, and the cycle repeats.
An increasing amount of water will be needed worldwide to evaporate heat from the growing number of data centers, which themselves are increasing in size as more users put information into the cloud.
“There’s no water loss like there is in a cooling tower that relies on evaporation,” Martinez says. “We also don’t have to add chemicals such as biocides, another expense. This system does not utilize a compressor, which would incur more costs. The system utilizes phase-changing refrigerant and only requires outside air that’s cool enough to absorb the heat.”
In New Mexico, that would occur in spring, fall and winter, saving millions of gallons of water.
In summer, the state’s ambient temperature is high enough that a cooling tower or some method of evaporation could be used. But more efficient computer architectures can raise the acceptable temperature for servers to operate and make the occasional use of cooling towers even less frequent.
“If you don’t have to cool a data center to 45 degrees Fahrenheit but instead only to 65 to 80 degrees, then a warmer outside air temperature — just a little cooler than the necessary temperature in the data center — could do the job,” Martinez says.
For indirect air cooling in a facility, better design brings the correct amount of cooling to the right location, allowing operating temperatures to be raised and allowing the refrigerant cycle to be used more during the year.
“At Sandia, we used to have to run at 45 degrees Fahrenheit. Now we’re at 65 to 78. We arranged for air to flow more smoothly instead of ignoring whorls as it cycled in open spaces. We did that by working with supercomputer architects and manufacturers of cooling units so they designed more efficient air-flow arrangements. Also, we installed fans sensitive to room temperature, so they slow down as the room cools from decreased computer usage and go faster as computer demand increases. This results in a more efficient and economical way to circulate air in a data center.”
Big jobs that don’t have to be completed immediately can be scheduled at night when temperatures are cooler.
“Improving efficiencies inside a system raises efficiencies in the overall system,” Martinez says. “That saves still more water by allowing more use of the water-saving refrigerant system.”