U.S. says , Cloud computing slows energy demand

Ten years ago, power usage at data centers was growing at an unsustainable rate, soaring 24% from 2005 to 2010. But a shift to virtualization, cloud computing and improved data center management is reducing energy demand.

According to a new study, data center energy use is expected to increase just 4% from 2014 to 2020, despite growing demand for computing resources.
Total data center electricity usage in the U.S., which includes powering servers, storage, networking and the infrastructure to support it, was at 70 billion kWh (kilowatt hours) in 2014, representing 1.8% of total U.S. electricity consumption.
Based on current trends, data centers are expected to consume approximately 73 billion kWh in 2020, becoming nearly flat over the next four years. "Growth in data center energy consumption has slowed drastically since the previous decade," according to a study by the U.S. Department of Energy's Lawrence Berkeley National Laboratory. "However, demand for computations and the amount of productivity performed by data centers continues to rise at substantial rates."
Improved efficiency is most evident in the growth rate of physical servers.
From 2000 to 2005, server shipments increased 15% each year, resulting in a near doubling of servers in data centers. From 2005 to 2010, the annual shipment increases fell to 5%, but some of this decline was due to the recession. Nonetheless, this server growth rate is now at 3%, a pace that is expected to continue through 2020.
The reduced server growth rate is a result of the increase in server efficiency, better utilization thanks to virtualization, and a shift to cloud computing. This includes concentration of workloads in so-called "hyperscale" data centers, defined as 400,000 square feet in size and above.

0 comments:

Post a Comment