I’m writing a program that wraps around dd to try and warn you if you are doing anything stupid. I have thus been giving the man page a good read. While doing this, I noticed that dd supported all the way up to Quettabytes, a unit orders of magnitude larger than all the data on the entire internet.

This has caused me to wonder what the largest storage operation you guys have done. I’ve taken a couple images of hard drives that were a single terabyte large, but I was wondering if the sysadmins among you have had to do something with e.g a giant RAID 10 array.

  • Urist@lemmy.ml
    link
    fedilink
    English
    arrow-up
    32
    arrow-down
    1
    ·
    8 months ago

    I obviously downloaded a car after seeing that obnoxious anti-piracy ad.

  • fuckwit_mcbumcrumble@lemmy.dbzer0.com
    link
    fedilink
    English
    arrow-up
    20
    ·
    8 months ago

    Entire drive/array backups will probably be by far the largest file transfer anyone ever does. The biggest I’ve done was a measly 20TB over the internet which took forever.

    Outside of that the largest “file” I’ve copied was just over 1TB which was a SQL file backup for our main databases at work.

    • cbarrick@lemmy.world
      link
      fedilink
      English
      arrow-up
      3
      ·
      8 months ago

      +1

      From an order of magnitude perspective, the max is terabytes. No “normal” users are dealing with petabytes. And if you are dealing with petabytes, you’re not using some random poster’s program from reddit.

      For a concrete cap, I’d say 256 tebibytes…

  • Hugin@lemmy.world
    link
    fedilink
    arrow-up
    15
    ·
    8 months ago

    It was something around 40 TB X2 . We were doing a terrain analysis of the entire Earth. Every morning for 25 days I would install two fresh drives in the cluster doing the data crunching and migrate the filled drives to our file server rack.

    The drives were about 80% full and our primary server was mirrored to two other 50 drive servers. At the end of the month the two servers were then shipped to customer locations.

  • ramble81@lemm.ee
    link
    fedilink
    arrow-up
    14
    ·
    8 months ago

    I’ve done a 1PB sync between a pair of 8-node SAN clusters as one was being physically moved since it’d be faster to seed the data and start a delta sync rather than try to do it all over a 10Gb pipe. M

  • Trigger2_2000@sh.itjust.works
    link
    fedilink
    arrow-up
    11
    ·
    8 months ago

    I once abused an SMTP relay (my own) by emailing Novell a 400+ MB memory dump. Their FTP site kept timing out.

    After all that, and them swearing they had to have it, the OS team said “Nope, we’re not going to look at it”. Guess how I feel about Novell after that?

    This was in the mid-90’s.

  • brygphilomena@lemmy.world
    link
    fedilink
    arrow-up
    10
    ·
    8 months ago

    In the middle of something 200tb for my Plex server going from a 12 bay system to a 36 LFF system. But I’ve also literally driven servers across the desert because it was faster than trying to move data from one datacenter to another.

  • d00phy@lemmy.world
    link
    fedilink
    English
    arrow-up
    8
    ·
    8 months ago

    I’ve migrated petabytes from one GPFS file system to another. More than once, in fact. I’ve also migrated about 600TB of data from D3 tape format to 9940.

  • HarriPotero@lemmy.world
    link
    fedilink
    arrow-up
    5
    ·
    8 months ago

    I worked at a niche factory some 20 years ago. We had a tape robot with 8 tapes at some 200GB each. It’d do a full backup of everyone’s home directories and mailboxes every week, and incremental backups nightly.

    We’d keep the weekly backups on-site in a safe. Once a month I’d do a run to another plant one town over with a full backup.

    I guess at most we’d need five tapes. If they still use it, and with modern tapes, it should scale nicely. Today’s LTO-tapes are 18TB. Driving five tapes half an hour would give a nice bandwidth of 50GB/s. The bottleneck would be the write speed to tape at 400MB/s.

  • weker01@sh.itjust.works
    link
    fedilink
    arrow-up
    4
    ·
    8 months ago

    I once moved ~5TB of research data over the internet. It took days and unfortunately it also turned out that the data was junk :/

  • IphtashuFitz@lemmy.world
    link
    fedilink
    English
    arrow-up
    4
    ·
    8 months ago

    Back in the late 90’s I worked for an internet search company, long before Google was a thing. We would regularly physically drive a dozen SCSI drives from a RAID array between two datacenters about 20 miles apart.

  • southsamurai@sh.itjust.works
    link
    fedilink
    arrow-up
    4
    ·
    8 months ago

    I think 16 terabytes? Might have been twelve. I was consolidating a bunch of old drives and data into a nas for a friend. He just didn’t have the time, between working and school and brought me all the hardware and said “go” lol.

  • HappyTimeHarry@lemm.ee
    link
    fedilink
    English
    arrow-up
    3
    ·
    8 months ago

    I downloaded that 200gb leak from national public data the other day, maybe not the biggest total but certainly the largest single text file ive ever messed with

  • Joelk111@lemmy.world
    link
    fedilink
    arrow-up
    3
    ·
    edit-2
    8 months ago

    When I was moving from a Windows NAS (God, fuck windows and its permissions management) on an old laptop to a Linux NAS I had to copy about 10TB from some drives to some other drives so I could re-format the drives as a Linux friendly format, then copy the data back to the original drives.

    I was also doing all of this via terminal, so I had to learn how to copy in the background, then write a script to check and display the progress every few seconds. I’m shocked I didn’t loose any data to be completely honest. Doing shit like that makes me marvel at modern GUIs.

    Took about 3 days in copying files alone. When combined with all the other NAS setup stuff, ended up taking me about a week just in waiting for stuff to happen.

    I cannot reiterate enough how fucking difficult it was to set up the Windows NAS vs the Ubuntu Server NAS. I had constant issues with permissions on the Windows NAS. I’ve had about 1 issue in 4 months on the Linux NAS, and it was much more easily solved.

    The reason the laptop wasn’t a Linux NAS is due to my existing Plex server instance. It’s always been on Windows and I haven’t yet had a chance to try to migrate it to Linux. Some day I’ll get around to it, but if it ain’t broke… Now the laptop is just a dedicated Plex server and serves files from the NAS instead of local. It has much better hardware than my NAS, otherwise the NAS would be the Plex server.

    • calm.like.a.bomb@lemmy.dbzer0.com
      link
      fedilink
      English
      arrow-up
      2
      ·
      8 months ago

      so I had to learn how to copy in the background, then write a script to check and display the progress every few seconds

      I hope you learned about terminal multiplexers in the meantime… They make your life much easier in cases like this.