cost to operate.
Brian Cluff
brian at SnapTek.com
Fri Jul 23 11:16:17 MST 2021
Power supply ratings are their maximum output they are capable of.
Computer power supplies are going to be oversized (if the computer was
built right) otherwise they wouldn't last for very long and would run
very hot. Computers, especially modern ones, power usage is going to
vary wildly from one second to the next based on it's load and what's
connected to it. If your system is just sitting there, on, doing
nothing, it will likely be under 100watts, especially if the monitor is
off, asleep or non-existent. Servers will tend to draw more, because
they have a lot more fans, hard drives, and power profiles that don't
allow for them throttle as much.
Even if you do have a system that only uses 50 watts normally, I still
recommend getting something low power like a raspberry pi to serve your
house because even if you have to buy the PI and the existing computer
is free, the PI will quickly pay for itself and after that it's almost
free to run it... and a lot more quiet and you also don't have to pay
for your air conditioner to cool off the room that your higher power
computer heated up which is also a very real cost that hasn't really
been mentioned yet.
I had to argue with an electrician about power supply sizes when I build
a computer lab with custom built computers with massively oversized
power supplies. He went around adding out all the wattage ratings of
the power supplies and decided that my 30 computer lab would require a
minimum of 15 circuits in order to not pop breakers. I never could
convince him that I was right, and that the breakers wouldn't pop and he
finally did want I asked him to do which was to add 4 circuits, which we
never has any problems with.
Brian Cluff
On 7/23/21 10:22 AM, Keith Smith via PLUG-discuss wrote:
>
> Based on what we have been discussing I assume my 400 watt power
> supply may be drawing much less power based on actual usage. Therefore
> maybe my computer might only be using 60 watts... making the cost lower.
>
> Your thoughts.
>
>
>
> On 2021-07-22 21:39, Mike Bushroe via PLUG-discuss wrote:
>> I usually use a mental rule of thumb that for every watt of 24/7/365
>> power consumption costs about $1 per year. Obviously this is failing
>> as electric rates keep going up. So to first order of magnitude a 100
>> watt server would cost around $100 a year, but if the server was using
>> the whole 400 watts it would cost more like $400 a year.
>>
>>>> If my home web server is using 100 watts an hour that mean 100
>> watts *>> 30 days * 24 hours or 72K watts.
>>>>
>>>> I'm thinking 72 * .1085 = $7.81 a month.
>>
>> KINDNESS
>>
>> is most VALUABLE when it is GIVEN AWAY for
>>
>> FREE
>> ---------------------------------------------------
>> PLUG-discuss mailing list - PLUG-discuss at lists.phxlinux.org
>> To subscribe, unsubscribe, or to change your mail settings:
>> https://lists.phxlinux.org/mailman/listinfo/plug-discuss
> ---------------------------------------------------
> PLUG-discuss mailing list - PLUG-discuss at lists.phxlinux.org
> To subscribe, unsubscribe, or to change your mail settings:
> https://lists.phxlinux.org/mailman/listinfo/plug-discuss
More information about the PLUG-discuss
mailing list