

I’ve seen ZFS in production use on pools with hundreds of TBs, clustered together into clusters of many PBs. The people running that don’t even think about BTRFS, and certainly won’t actively consider it for anything.
- BTRFS once had data corruption bugs. ZFS also had that, but only in very specific edge cases. That case was taken very seriously, but basically, ZFS has a reputation for not fucking up your bits even close to BTRFS
- ZFS was built and tested for use on large pools from the beginning, by Sun engineers from back when Sun was fucking amazing and not a pile of Oracle managed garbage. BTRFS still doesn’t have stable RAID5/6.
- ZFS send/recv is amazing for remote replication of large filesystems.
- DRAID is kind o the only sane thing to do with todays disk sizes, speeds and pool sizes.
But those are ancillary reasons. I’ll quote the big reason from the archwiki:
The RAID 5 and RAID 6 modes of Btrfs are fatally flawed, and should not be used for "anything but testing with throw-away data”.
For economic reasons, you need erasure coding for bigger pools, either classic RAID5/6 or DRAID. BTRFS will either melt your data in RAID5/6 or not support DRAID at all.
I need Emacs, a terminal emulator and a web browser to be productive, but basically nothing else. (Give me my tiling window manager, with a config I haven’t bothered to update the past few years for an extra 3% bump in efficiency.)
It’s weird, I know how all the components in a modern desktop environment work and fit together but I don’t want to care anymore. I want someone to hold my hand, manage my system and make all the thinking go away, right up until I ssh out from my desktop and out into a fleet of servers and start spewing out esoteric commands and orchestration.
My dream is to have someone manage my desktop for me, so I don’t even have to think about it.