The following is part of a series of posts I wrote called "Building a data center at home".
Living in the SF bay area, you can acquire 2nd hand data center equipment relatively cheap. The following are a series of posts detailing my deep dive in building a cluster with data-center equipment at home (often called a homelab) which consists of 48 CPU cores, 576Gb RAM, 33.6TB storage across 60 x 6Gb/s HDD's with a combined weight of just over 380lb/170kg within a budget of $3000.
I recently built out a Kubernetes cluster using a few old computers that I wasn’t using which you can read about here.
Living in the San Francisco bay area, it turns out that obtaining old data center equipment is actually pretty inexpensive due to the large number of tech companies and the low demand for old equipment. A single new server unit can cost many 10s or 100s of thousands, which sounds like a lot, but given the costs surrounding running a data center such as power, staff and cooling, this is not such a significant cost. In this case, having older less efficient equipment can actually end up costing more in the long run, hence the lower demand for such hardware.
I found a liquidation business around Hunters Point that sold hardware that they obtained
in bulk from companies that were doing wholesale changes to their data centers. I managed to pick
up 3 Dell
R710s and 1 Dell
R610 for $350. They even threw in two Force10 24 port ethernet
switches and an old monitor as well.
Having not done had any experience with proper rack mount servers before, booting these up for the first time was quite interesting. Admittedly, I was totally unaware that different hardware at this level can contain it’s own BIOS, so during the boot process you are able to enter the BIOS configuration for actual hardware not just the standard motherboard hardware.
The first thing I needed to do was to just into the Perc 6/i BIOS for each server and set up the
RAID configuration given that the HDD’s that were supplied where totally wiped. Each of the
have 6 3.5” bays while the
R610 has 6 2.5” bays.
Initially I set up each machine running with a RAID-1 across two 500GB enterprise level disks. Due to the fact that I’m going to be running a cluster of Kubernetes machines, I don’t care too much about local storage as this is likely going to be handled elsewhere on the network.
I install Ubuntu server 18.04 on each via a USB flashed with etcher and the entire process is pretty much standard as if I was installing on a laptop. I was actually expecting there to be more issues that I’d need to find solutions for but it was actually quite an easy process.
I was originally thinking that these severs were going to be LOUD, but they’re not too bad.
When the air conditioning is on in the house, they somewhat become inaudible. Funnily enough,
the Force10 actually sounds a little louder than the servers although I think that this is due to
the frequencies that it produces being more uneven. The
R710s sound to me like they
have a flatter sound profile, so more of a white noise, while the switch seems to have a more
uneven sound profile and is more noticeable.
Either way, might be interesting to test this theory in future by running each through a spectrum analyzer.
After setting everything up I add each to the existing kubernetes cluster via
everything just works as expected. This was all much easier than I was expecting. Happy days!
As you can see from the photos, storing the servers on a table is not the best option so I order
this rack from Amazon to put everything. With the
1 unit switch, 3
R710s at 2 units each, and 1 unit
R610, it leaves 4 units free for a storage
unit which I’ll look to add next.
The situation with power in the States is interesting coming from Australia.
This creates an interesting situation as I need to plug these servers into two outlets to get the required power without over loading an outlet where as for the power required, in Australia these servers would run on one outlet.