How much of our computing now happens in the cloud? A lot. Providers of public cloud services alone take in more than a quarter of a trillion U.S. dollars a year. That’s why Amazon, Google, and Microsoft maintain massive data centers all around the world. Apple and Facebook, too, run similar facilities, all stuffed with high-core-count CPUs, sporting terabytes of RAM and petabytes of storage.
These machines do the heavy lifting to support what’s been called “surveillance capitalism”: the endless tracking, user profiling, and algorithmic targeting used to distribute advertising. All that computing rakes in a lot of dollars, of course, but it also consumes a lot of watts: Bloomberg recently estimated that about 1 percent of the world’s electricity goes to cloud computing.
That figure is poised to grow exponentially over the next decade. Bloomberg reckons that, globally, we might exit the 2020s needing as much as 8 percent of all electricity to power the future cloud. That might seem like a massive jump, but it’s probably a conservative estimate. After all, by 2030, with hundreds of millions of augmented-reality spectacles streaming real-time video into the cloud, and with the widespread adoption of smart digital currencies seamlessly blending money with code, the cloud will provide the foundation for nearly every financial transaction and user interaction with data.
How much energy can we dedicate to all this computing? In an earlier time, we could have relied on Moore’s Law to keep the power budget in check as we scaled up our computing resources. But now, as we wring out the last bits of efficiency from the final few process nodes before we reach atomic-scale devices, those improvements will hit physical limits. It won’t be long until computing and power consumption will once again be strongly coupled—as they were 60 years ago, before integrated CPUs changed the game.
We seem to be hurtling toward a brick wall, as the rising demand for computing collides with decreasing efficiencies. We can’t devote the whole of the planet’s electricity generation to support the cloud. Something will have to give.
The most immediate solutions will involve processing more data at the edge, before it goes into the cloud. But that only shifts the burden, buying time for rethinking how to manage our computing in the face of limited power resources.
Software and hardware engineering will no doubt reorient their design practices around power efficiency. More code will find its way into custom silicon. And that code will find more reasons to run infrequently, asynchronously, and as minimally as possible. All of that will help, but as software progressively eats more of the world—to borrow a now-famous metaphor—we will confront this challenge in ever-wider realms.
We can already spy one face of this future in the nearly demonic coupling of energy consumption and private profit that provides the proof-of-work mechanism for cryptocurrencies like Bitcoin. Companies like Square have announced investments in solar energy for Bitcoin mining, hoping to deflect some of the bad press associated with this activity. But more than public relations is at stake.
Bitcoin asks us right now to pit the profit motive against the health of the planet. More and more computing activities will do the same in the future. Let’s hope we never get to a point where the fate of the Earth hinges on the fate of the transistor.
This article appears in the August 2021 print issue as “Cloud Computing’s Dark Cloud.”