Energy consumption vs % utilization?
Andre Oppermann
nanog-list at nrg4u.com
Wed Oct 27 09:13:37 UTC 2004
Steven M. Bellovin wrote:
> In message <Pine.WNT.4.61.0410261429110.3340 at vanadium.hq.nac.net>, Alex Rubenst
> ein writes:
>>
>>Hello,
>>
>>I've done quite a bit of studyin power usage and such in datacenters over
>>the last year or so.
>>
>>>I'm looking for information on energy consumption vs percent utilization. In
>>
>>>other words if your datacenter consumes 720 MWh per month, yet on average
>>>your servers are 98% underutilized, you are wasting a lot of energy (a hot
>>>topic these days). Does anyone here have any real data on this?
>>
>>I've never done a study on power used vs. CPU utilization, but my guess is
>>that the heat generated from a PC remains fairly constant -- in the grand
>>scheme of things -- no matter what your utilization is.
>
> I doubt that very much, or we wouldn't have variable speed fans. I've
> monitored CPU temperature when doing compilations; it goes up
> significantly. That suggests that the CPU is drawing more power at
> such times.
From running a Colo in a place with ridiculus high electricity engery
costs (Zurich/Switzerland) I can tell you that the energy consuption
of routers/telco (70%) and servers (30%) changes changes significantly
throughout the day. It pretty much follows the traffic graph. There
is a solid base load just because the stuff is powered up and from there
it goes up as much as 20-30% depending on the routing/computing load of
the boxes. To simplify things you can say that per packet you have that
many "mWh" (milli-Watt-hours) per packet switched/routed or http requests
answered over the base load. I haven't tried to calulate how much energy
routing a packet on a Cisco 12k or Juniper M40 cost though. Would be
very interesting if someone (student) could do that calculation.
--
Andre
More information about the NANOG
mailing list