45
submitted 1 month ago by poinck@lemmy.world to c/linux@lemmy.ml

I want to transfer 80 TB of data to another locatio . I already have the drives for it. The idea is to copy everything to it, fly it to the target and use or copy the data on/to the server.

What filesystem would you use and would you use a raid configuration? Currently I lean towards 8 single disk filesystems on the 10 TB drives with ext4, because it is simple. I considered ZFS because of the possiblity to scrub at the target destination and/or pool all drives. But ZFS may not be available at the target.

There is btrfs which should be available everywhere because it is in mainline linux and ZFS is not. But from my knowledge btrfs would require lvm to pool disks together like zfs can do natively.

Pooling the drives would also be a problem if one disk gets lost during transit. If I have everything on 8 single disks at least the remaining data can be used at the target and they only have to wait for the missing data.

I like to read about your opinions or practical experience with similar challanges.

you are viewing a single comment's thread
view the rest of the comments
[-] atzanteol@sh.itjust.works -2 points 1 month ago

Either way, traveling with that much spinning rust there is always a good chance for bit-flips or damage.

What? Lol no. They'll travel fine.

[-] danielquinn@lemmy.ca 8 points 1 month ago* (last edited 1 month ago)

Multiple disks with many moving parts, containing 80TB of data on magnetic platters flying at high altitude where they'll be subjected to far more physical impacts, radiation, and cosmic rays than at sea level.

Yeah, it's a risk.

Here's a really amazing Radiolab episode about bit flips!

[-] atzanteol@sh.itjust.works 8 points 1 month ago

You kids think HDDs just failed daily or something. I flew all over the place with a laptop with an HDD for years, as did many others. It'll be fine. Especially since it's unlikely they would be using the drives while traveling.

[-] mko@discuss.tchncs.de 4 points 1 month ago

From a position of handling corporate data on a daily basis, I am pretty confident that data integrity is top of mind.

[-] poinck@lemmy.world 2 points 4 weeks ago

I agree with both of you. Somehow I don't worry about the drive in my laptop but 80 TB of scientific data is another thing, and I want to make sure it is the same data when it arrives.

[-] frongt@lemmy.zip 7 points 1 month ago

Really, then why is there an explicit SMART conveyance test?

It's to test for damage that may have occurred during shipping.

[-] atzanteol@sh.itjust.works 1 points 1 month ago

And how often does it happen?

[-] mko@discuss.tchncs.de 1 points 1 month ago

How do you ensure that is doesn’t happen? If this is corporate data that can be key.

[-] poinck@lemmy.world 3 points 4 weeks ago

this is scientific data.

Funfact, I recently did a scrub on my offline backup drive of my work PC. It correct around 250 errors. I wouldn't have noticed any problems if I had used ext4 instead of btrfs.

[-] frongt@lemmy.zip 0 points 1 month ago

Often enough that there's a test designed to detect it specifically. If you want hard data you'll have to find it on your own, I don't have any handy.

this post was submitted on 25 Nov 2025
45 points (95.9% liked)

Linux

57274 readers
395 users here now

From Wikipedia, the free encyclopedia

Linux is a family of open source Unix-like operating systems based on the Linux kernel, an operating system kernel first released on September 17, 1991 by Linus Torvalds. Linux is typically packaged in a Linux distribution (or distro for short).

Distributions include the Linux kernel and supporting system software and libraries, many of which are provided by the GNU Project. Many Linux distributions use the word "Linux" in their name, but the Free Software Foundation uses the name GNU/Linux to emphasize the importance of GNU software, causing some controversy.

Rules

Related Communities

Community icon by Alpár-Etele Méder, licensed under CC BY 3.0

founded 6 years ago
MODERATORS