20 Oct 2021
We are ignoring the true cost of water-guzzling data centres
Article by Masaō Ashtine, Oxford e-Research Centre, University of Oxford and David Mytton, Imperial College London.
This article is republished from The Conversation under a Creative Commons license. Read the original article.
The 1960s ushered in a new age of processing digital information, driven by the intelligence needs of the Cold War. Moore’s law meant microchips doubled in speed every two years, shrinking costs and miniaturising machines that once filled entire rooms. Today, the smartphone probably being used to read this article is millions of times more powerful than the computer that landed the Apollo missions on the moon.
While those huge supercomputers have disappeared, the proliferation of the cloud and the Internet of Things, with everything down to our socks being able to connect to the internet, means more and more computer processors that need to communicate with data centres around the world. Even something as simple as scrolling down on this article triggers communications that may eventually pass through a distant data centre.
Data centres can range in size from small cabinets through to vast “hyperscale” warehouses the size of stadiums. Inside, are computers called servers which support the software, apps and websites we use every day.
As of the end of 2020, 597 hyperscale data centres were in operation (39% in the US, 10% in China, 6% Japan), up by almost 50% since 2015. Amazon, Google and Microsoft account for more than half of these and a further 219 are in various stages of planning.
Energy eaters
Data centres accounted for around 1% or 2% of global electricity demand in 2020. All that processing power generates lots of heat, so data centres must keep cool to prevent damage. While some companies are using cool air on mountain sites and Microsoft has used the cold waters of Scotland to experiment with underwater data centres, up to 43% of data centre electricity in the US is used for cooling.
This energy goes into cooling water which is either sprayed into air flowing past the servers, or evaporated to transfer heat away from the servers. Not only is energy required to cool the water (unless the system is specifically designed as a closed loop) but that water is lost as it evaporates. In a relatively small 1 megawatt data centre (that uses enough electricity to power 1,000 houses), these traditional types of cooling would use 26 million litres of water per year.
Water directly used for cooling is what most data centre operators focus on, but the largest source of water usage is actually electricity generation. This comes from how water is heated to generate steam which turns a turbine and generates electricity. Fossil fuels and nuclear power all consume water in this way, and even hydroelectric power involves some water loss from reservoirs.
The transition to renewables is therefore important to reduce both water and carbon footprints. By 2030, wind and solar energy could reduce water withdrawals related to power generation by 50% in the UK, 25% in the US, Germany and Australia, and 10% in India.
More than the daily water recommendation?
Data centre water demand is more complicated than the carbon footprint. Reaching zero carbon is a reasonable goal, but zero-water is not necessarily the right choice. Consumption goals need more context.
Sundry Photography / shutterstock
Some data centres are in regions with abundant water, easily accessible without competing with other potential users. However, others may be built in areas of drought with degrading infrastructure.
Regional water stress must be considered for each data centre, not just in relation to the water used for on-site cooling, but also linked to the power plants that generate the electricity that power the centre.
For example, a recent US study showed that the western US has more water stress compared to the eastern US, and that electricity generated in the south west is more water intensive due to the use of more hydropower. Despite this, the west and south west have more data centres.
Md Abu Bakar Siddik et al 2021 Environ. Res. Lett. 16 064017
Opposition is starting to grow even as new projects are being approved. The US has seen local communities protest against new data centres, which may be why in the past Google has considered its usage of water to be a trade secret. Similar concerns led to a temporary ban on new data centres in The Netherlands, and France is in the process of passing new laws to require more transparency.
We do not give water enough value
Companies are not pricing water risk into their calculations when picking locations for data centres. A lower price does not necessarily mean lower risk. When Microsoft assessed its water footprint at a data centre in San Antonio, Texas, it found the true cost of water was 11 times more than it was paying.
This is similar to carbon footprints. We undervalue or ignore abatement costs associated with greenhouse emissions, and the impacts are hardly marginal. Carbon dioxide and water are inextricably linked, and climate change is already stressing drought-prone areas around the world.
The first step is transparency. Some companies like Microsoft and Facebook already publish aggregated water data, but others need to do the same. Every operator needs to publish their water efficiency plan and back it up with the relevant regional numbers.
Most data centre owners have received the message about reducing their carbon footprint and the transition to renewable energy. We regularly see new projects announced with net-zero carbon goals. They now need to do something similar for water.
Masaō Ashtine, Postdoctoral Researcher in Local Energy Systems, University of Oxford and David Mytton, Researcher, Centre for Environmental Policy, Imperial College London
This article is republished from The Conversation under a Creative Commons license. Read the original article.