Hardware Raid is Dead and is a Bad Idea in 2022
Video Statistics and Information
Channel: Level1Techs
Views: 484,050
Rating: undefined out of 5
Keywords: technology, science, design, ux, computers, hardware, software, programming, level1, l1, level one
Id: l55GfAwa8RI
Channel Id: undefined
Length: 22min 18sec (1338 seconds)
Published: Tue Apr 05 2022
Please note that this website is currently a work in progress! Lots of interesting data and statistics to come.
He mentions with BTRFS that you can set a RAID policy on a per folder basis, I don't believe this is true. From what I understand there was talk at one point about making RAID policies set per subvolume, but this never came to fruition. What you can do is set a different RAID policy for data and metadata, but that's not really the same.
The video has some great information in it regardless.
I am glad he did it on his main channel. I have been sending people this link to his 7 yr old video on his third channel. It is a pain for me to find in the "algorithm" even when I know what to search.
https://youtu.be/yAuEgepZG_8
Beyond raid 1/10 for hypervisor hardware raid has been dead for quite a few years already.
I'm a little split on this. On one hand ZFS is great, and the ability to detect and recover from bitrot is hard to get unless you're paying lots of money for a SAN or specific hardware RAID cards. So I can't understate how much I agree with him that the resiliency of file systems like ZFS is a large benefit.
On the other hand it's not quite as dire as Wendell points out (hardware RAID is definitely not dead!). HDDs and SSDs both have error-correcting codes as part of the sector (separate to the 520 byte sectors that RAID cards can store checksums in). So in most cases bitrot is going to be caught by the HDD and appear as a read error to the hardware controller, allowing it to recover the data from the parity. So the case of the hardware RAID controller getting bad data from the disk is quite rare, and will do a decent job at avoiding bitrot.
Also in my professional life I've never actually seen a problem that can be directly attributed to bitrot. Though this is likely just a function of how much data you deal with. Most large things that I've worked with are either SAN based, or use a replicated storage system with its own checksum and recovery. And to be honest I'd always prefer moving that check-summing and recovery as close to the application as possible.
0:11 - "Support for it has gone away at a hardware level a long time ago"
No idea what he's on about. Modern HW RAID cards can do SATA/SAS and even NVMe (called TriMode) and work fine on Ice Lake and Milan based servers.
They also now measure in the hundreds of thousand to millions of IOPs as well.
What does "gone away" even mean in his context.
1:45 - What he's talking about is NOT HW RAID. It's SW RAID with a HW Offload to an Nvidia GPU. Nowhere near the same tech as LSI/Broadcom or Microchip/SmartROC cards. (The 2 biggest vendors folks like HPE, Dell, Cisco and Lenovo use)
12:15 - Battery backed cache. Actually Batteries are still offered, along with a hybrid SuperCap/Battery module as well. But many controllers now include some NAND (SSD basically) on the Controller or Cache module and the Battery/SuperCap only needs to provide power long enough to dump the RAM cache to the NAND and do a CRC check, then the card powers down. At this point the server can remain un-powered for days or weeks. When the server powers back on, the NAND is checked and any data found is pulled back to cache, CRC checked again, and then flushed to the drives before the OS has even had a chance to boot.
20:00 - PLP - hahaha, no. It's got a DRAM based cache and the PLP is to protect the data in flight in the DRAM so it can be written to the NAND before the card loses power.
But Casper, how can you be sure? Wendel is SOOO much smarter than you.
https://www.samsung.com/us/business/computing/memory-storage/enterprise-solid-state-drives/983-dct-1-9tb-mz-1lb1t9ne/
"to provide enough time to transfer the cached data in the DRAM to the flash memory"
Gee, maybe because it's on the damn website for the SSD...
While I agree that old school HW RAID isn't a viable alternative for large Enterprise systems anymore, he glosses right over that this is not what most people USE HW RAID for anymore.
Smaller deployments, Edge or simple RAID 1 Boot drives for example, are FAR and away the majority in the Enterprise.
Large data pools are either Storage Arrays, Software Defined Storage, or giant PB scale Distributed storage systems like Ceph, Qumulo or Scality Ring.
And those Strorage Arrays like NetApp, Nimble, 3PAR and others, often ARE doing the RAID and storage management in a CPU in SW and some have a HW offload accelerator already as well.
Videos like this are why I personally can't stand L1Techs and LTT.
They come off so smug and don't leave room for any alternate viewpoint other than theirs.
Find me a big company like Coke or Disney or Bank of America who is using ZFS.
I'd bet <5% of them touch it.
Yet folks like L1 and LTT think it's the be all end all to data storage.