Why you DONβT want a 20TB Hard Drive
Video Statistics and Information
Channel: LMG Clips
Views: 876,444
Rating: 4.8817048 out of 5
Keywords: linus, tech, tips, wan, luke, podcast, clip, snippet, opinion, technology, gaming, pc, hardware
Id: 8IRjFZ9xEj8
Channel Id: undefined
Length: 20min 13sec (1213 seconds)
Published: Wed Jan 01 2020
Please note that this website is currently a work in progress! Lots of interesting data and statistics to come.
These WD SMR drives are host-managed SMR, not drive-managed SMR: Seagate's "archive" and now-endemic unlabeled drives, nor the 2.5" drive-managed SMRs.
Linux has some beta-level support for host-managed SMR file systems AFAIK, but I wouldn't trust a single copy of any data files to it yet.
Personally I pretty much agree with Linus' interpretations but these drives still could be useful for long-term storage arrays. You just have to use a topography without distributed parity and make sure any dedicated parity drives in the system are non-SMR. This will optimize performance as much as is possible with this hardware.
It's a mix of a click-bait title and some "Linus points" that are as usual and expected not what the "other normal" people would care about (and I'm vastly enlarging the usual "normal" definition to include most DHers too).
He starts by acknowledging the speed isn't an issue at all during normal use but then it's a big problem:
when doing data migrations. While raw bandwidth does scale up with the number of drives the "logical" migrations (as opposed to I don't know, doing dd on a bunch of disks in parallel) aren't really speed-scaling with the number of drives and more drives usually mean more headaches. Last time you bought one or a bunch of large disks, lets say 10TBs you think you could migrate faster if you bought 5 times more 2TB drives (even if they -assume- had about the same raw speed)? Only in very theoretical cases, in practice they would've been a nightmare.
when doing RAID rebuilds he's speaking about biting his nails for the whole day while doing a recovery on a 160TB array in RAID-5 (hypothetically on the 160TBs with the new 20TB drives but he's done it on smaller ones). Well, of course but he's Linus and this shit of RAID with no backup is his hallmark. It has nothing to do with speed or even with losing one disk at all. He had a PRODUCTION WITH NO BACKUP array messed up and it wasn't because of (lack of) speed, in fact they were very expensive SSDs and NONE actually failed!
As someone who only updates roughly 100GB/month to my server, these would be fine - though I can still see why everyone dislikes smr.
Looks like these drives are Host Managed, so I'd assume some optimizations could be done system side for random IO and/or TRIM.
tl;dr the video - Current drive write speeds are too slow, and HDDs still cost too much money. To make this size drive viable, write speeds need to increase significantly, and prices need to come down a lot.
Assuming 100 MB/sec transfer speed:
1 min = 6 GB
1 hour = 360 GB
24 hours = 8,640 GB
So on average, 1 TB (931 GB) every 3 hours.
I'm currently in the process of migrating and organizing data across 4x 6TB WD Reds (2 of which are SMR). I pretty much see a solid transfer speed of about 100 MB/sec when moving data. I only transfer about 100-200 GB at a time, and all the files are 3-5 GB mp4 video files. All the drives are connected via HBA, no raid.
If I needed to rebuild/clone a single drive in one sitting, it would take ~18 hours for 6 TB.
Seems like a lot of people in this thread are missing the point of this clip, and instead are getting hung up on the SMR bit. If you had a array/storage pool that is made up of 4x 20 TB HDDs, it would take ~ 60 hours to rebuild a single disk. That's a long time to hope nothing else breaks while transferring data. If the HDDs were cheaper though, it would mitigate the risk since you could literally just throw more HDDs at the problem to solve it. But given that 20TB HDDs are ~ $500, thats a very expensive solution.
I had this same fear when 1TB HDD appeared on the market - when I got up to about 9 drives, I made the switch to 6TB (@ $300 a drive) - bought 2 at first, then another 2, and just recently, another 2.
Linus needs to close his mouth more often.
would these be good for an unraid setup? data is only written then read