4

I am looking to buy HDDs for a single server which will be used as development server. The server will use ESXi as hypervisor and will have some but not too many virtual machines (web server, database server, file server)

We have already taken a Samsung 950 Pro 256 GB for the virtual machines but I struggle in finding an answer to the question:

In a RAID 10: Are lots of slow disks (5400/5900 rpm) faster than fewer fast disks (7200 rpm)? For example 8x 5400rpm vs 4x 7200rpm

Would the 8 slow be a better choice compared to the 4 fast?

My choice was between 1TB WD Red which run at 5400rpm vs 2TB WD Se 7200rpm.

13
  • 5
    Given the tremendous price differences and the fact your developers work for free, I understand your reluctance to go all SSD. Except your developres do not work for free and a 1tb SSD does not cost a lot. Your setup has low performance and high maintenance. Go with SSD.
    – TomTom
    Mar 31, 2016 at 9:01
  • 2
    @TomTom I completely agree, but would note that it's axiomatic that devs (however many there are) will always exhaust the server (however big it is). Especially if there are databases involved.
    – MadHatter
    Mar 31, 2016 at 9:37
  • 3
    Generally no. My production server is all SSD too - there is nothing cheaper than SSD for a database server because unless you love dead data you ARE limited by IOPS.
    – TomTom
    Mar 31, 2016 at 12:21
  • 2
    @Noctalin Are SSD drives as reliable as mechanical drives (2013)? (And yes, you have to keep backups. Unless of course you want those 12 people to stop working for several days and lose Who Knows How Much time getting back to where they were.)
    – user
    Mar 31, 2016 at 15:43
  • 3
    Since when do 7200 rpm spinners count as "fast" in a modern multi-user environment?
    – user
    Mar 31, 2016 at 15:44

4 Answers 4

10

This is hard to answer directly, because the answer is very much 'it depends on your caching'.

You see - hard disks are slow. Very slow. But actually, most of the time you don't notice, if you have a decent cache, and a reasonably predicable access pattern.

Read caching means recently accessed things will be coming off cache. Write caching means that recently written things can be deferred and coalesced into full stripe writes.

But when it comes to measuring speed - there's two elements to it. Write performance ... actually probably doesn't matter too much, because of the caching - you can always cache writes, and so as long as your sustained write IO doesn't exceed that of your combined RAID group, it's irrelevant.

But read IO it's much more of a problem - you cannot service a read until you've got your data, and to get your data you have to wait for the disk if it's not already in cache. This is where you see the performance cost - it shows up in read latency.

And that's why slower disks are a problem - because it doesn't actually matter how fast the combined RAID group is, when the data is at rest on one of two drives - you have to access that drive to service the read.

So - at risk of sounding a bit wishy washy - "it depends" a lot on your workload. You can usually get by with some really slow disks and some awesome caching if you want really large amounts of terabytes cheap.

However, I'd only bother doing so when we start talking at 100+TB or so, which is where the cost of rack space, cooling, floor footprint etc. start to become significant.

For your application - I'd say buy the fastest drives you can afford, because it's a lot easier to buy more drives later if you need them, than to realise you can't because you've filled your drive bays and need a new shelf.

But I'd strongly suggest also considering SSDs - their price-per-gig isn't amazing, but their price-per-IOP really is. And whilst no one seems to care about IOPs, that's the thing they really do care about when it comes to 'being a bit slow'.

5
  • SSDs would do the job but we are going to create a VM to backup our remote servers and also our local data on our computers. That is why I opted for HDD RAID.
    – Noctalin
    Mar 31, 2016 at 10:02
  • 1
    You're creating a VM to act as a backup server? Why? Why not just run a backup server on the physical tin? If performance isn't a consideration, then your whole question is moot - just buy the cheapest per TB, and don't even bother with RAID 10 in the first place.
    – Sobrique
    Mar 31, 2016 at 10:05
  • I will actually do it that way and get a separate machine for backups. A QNAP/Synology with Mirror should do the work? Use the 950 Pro for the VM and an additional big SSD as storage.
    – Noctalin
    Mar 31, 2016 at 10:11
  • +1 for hard disks are slow. Another mental +1 for the rest because I can't vote up twice.
    – user
    Mar 31, 2016 at 15:46
  • 1
    If you are looking at storage solutions in the multiple TB and up range, you also have to consider how much of that is your active data set. My guess is that many such workloads are more like write-and-archive than they are read-heavy across the whole data set. When you've got large bulk storage of data that is mostly at rest, and a smaller working set that is accessed a lot, you can look into things like having a large but slow set of spinners fronted by some SSDs. ZFS' L2ARC makes that easy to maintain (but needs time to prime the cache, so uptime), but there are other options as well.
    – user
    Mar 31, 2016 at 20:20
5

Based on calculations from RAID Performance Calculator, eight 5400rpm disks will yield about 304 IO/s. Whereas four 7200rpm disks will yield about 210 IO/s.

7
  • 4
    The irony is - that's about the same as a single 15K SAS drive, and no where even close to even a single SSD.
    – Sobrique
    Mar 31, 2016 at 9:28
  • 1
    @Sobrique, why is that ironic? Mar 31, 2016 at 12:19
  • 1
    @AshBurlaczenko: Because the poinf of using HDD's over SSD's is to save money. But here, the HDD solution has less IOPS for more $.
    – MSalters
    Mar 31, 2016 at 15:33
  • 1
    @MSalters, I can't see any pricing on this page so how can you say its more. Where can I buy 8TB of SSD storage for less than than 8TB of HDD? Mar 31, 2016 at 15:44
  • 5
    When it comes to enterprise storage: capacity is cheap. Gigs at rest cost negligible amounts. What's important - and expensive - is how fast you can move it around. But too many people argue about whether a bus or a truck is better, when they should be thinking about whether they actually need a motorbike.
    – Sobrique
    Mar 31, 2016 at 17:16
1

The main difference between your two choices is 3 year warranty vs 5 years and SOHO usage area vs. certification for datacenters. In terms of raw I/O performance, there is not much of a difference between 5400/5900 and 7200 rpm S-ATA disks nowadays (that was different a few years ago).

In the end therefore the array with more disks will very likely show a superior performance, as RAID10 scales with the number of disks.

What I would guess is more important in this setup, then the number of drives is, get a proper hardware RAID controller with BBU+Cache if you can afford it - or even better go with SSD only from the start on.

Also please be aware, that ESXi natively only supports hardware RAID! While there are workarounds to use software RAID with ESXi, I personally wouldn't advice using these, if you are looking for a decent performance.

Protip: If you are on a really tight budget to self-build this storage server and we are talking about a non mission-critical system, then go fetch yourself an OEM version of a proper RAID controller (IBM M5015, DELL PERC H710, etc.) on ebay.

12
  • Well we code our sites and mobile apps currently on remote server and would rather have the development happen here on a local server. It is not supposed to be accessible from outside the office. So a hardware RAID with HDD is what you recommend?
    – Noctalin
    Mar 31, 2016 at 9:59
  • 3
    The problem is, that if you want to run ESXi you need a hardware RAID controller anyway, as ESXi doesn't support software RAID (also no motherboard onboard "fake" RAID). Besides that a proper hardware RAID controller will influence the performance of the array far more then the choice between 5400 vs. 7200 rpm disks.
    – s1lv3r
    Mar 31, 2016 at 10:03
  • 1
    It certainly works, but of course you have a lot of overhead (iSCSI stack etc.) and a more complex system setup. Personally I wouldn't recommend it, if you are looking for performance.
    – s1lv3r
    Mar 31, 2016 at 10:11
  • 1
    Does it really matter whether ESXi supports SW RAID? It just needs to expose the disks to the guest VMs, and they can manage their software RAIDs. Sure, not as CPU-efficient as HW RAID, but the problem for HW RAID is that CPU performance is cheap.
    – MSalters
    Mar 31, 2016 at 15:56
  • 1
    @Noctalin Yeah, looks like those are your options. Maybe you also want to consider to switch to Hyper-V or Xen. They both offer similiar features while supporting soft RAID. ESXi always sucks for hardware compatility anyway (I think they want to enforce use of certified hardware only) ...
    – s1lv3r
    Apr 5, 2016 at 10:50
0

8x 5400 RPM disks will generally outperform 4x 7200 RPM disks, save for single-threaded sequential read/write.

As your use case (virtual machines) is rich of small random reads/writes, you should prefer the solution with more spindles (ie: the more the disks, the better).

Anyway, considered how cheap are 7200 RPM disks, I strongly suggest to use 8x such disks ;)

You must log in to answer this question.

Not the answer you're looking for? Browse other questions tagged .