[lug] server spec

Bear Giles bgiles at coyotesong.com
Wed Jun 13 22:33:08 MDT 2007

Nate Duehr wrote:
> Also, your reply assumes the original poster was putting his 1RU 
> machines in a data center environment.  Maybe he wasn't planning on 
> that, but I may have missed it.  When you have the luxury of paying 
> someone for rack space in a nicely controlled environment, great.
> Or maybe I should say -- they shouldn't NEED to put a 1RU PeeCee in a 
> datacenter with "properly designed airflow".  If they need to, that 
> quality level of machine should NOT be called a "server".
> REAL well-engineered servers shouldn't keel over dead at the first 
> sign of an 85 degree room, or a little "hot spot" at their air intakes.

Well-engineered systems take economics into account.  How much would it 
cost, in terms of materials, physical space and power consumption, to 
make these servers robust enough to handle higher temperatures?  Now 
multiply that thousands (or even tens of thousands) of servers in a data 
center.  Suddenly the cost of redundant cooling systems starts to look 
pretty cheap.

BTW I've read that some next-generation servers take DC power instead of 
handling AC conversion themselves.  The idea is to take that heat source 
out of the boxes and put it someplace else.  I don't know how well it 
works in practice, but it's an interesting thought.  It makes you wonder 
how much else can be removed from the box.

More information about the LUG mailing list