I’m writing a program that wraps around dd to try and warn you if you are doing anything stupid. I have thus been giving the man page a good read. While doing this, I noticed that dd supported all the way up to Quettabytes, a unit orders of magnitude larger than all the data on the entire internet.

This has caused me to wonder what the largest storage operation you guys have done. I’ve taken a couple images of hard drives that were a single terabyte large, but I was wondering if the sysadmins among you have had to do something with e.g a giant RAID 10 array.

  • Davel23@fedia.io
    link
    fedilink
    arrow-up
    82
    ·
    30 days ago

    Not that big by today’s standards, but I once downloaded the Windows 98 beta CD from a friend over dialup, 33.6k at best. Took about a week as I recall.

    • pete_the_cat@lemmy.world
      link
      fedilink
      English
      arrow-up
      31
      ·
      30 days ago

      I remember downloading the scene on American Pie where Shannon Elizabeth strips naked over our 33.6 link and it took like an hour, at an amazing resolution of like 240p for a two minute clip 😂

    • absGeekNZ@lemmy.nz
      link
      fedilink
      English
      arrow-up
      17
      ·
      30 days ago

      Yep, downloaded XP over 33.6k modem, but I’m in NZ so 33.6 was more advertising than reality, it took weeks.

    • 50MYT@aussie.zone
      link
      fedilink
      arrow-up
      1
      ·
      29 days ago

      In similar fashion, downloaded dude where’s my car, over dialup, using at the time the latest tech method - a file download system that would split the file into 2mb chunks and download them in order.

      It took like 4 days.

  • psmgx@lemmy.world
    link
    fedilink
    arrow-up
    7
    ·
    30 days ago

    Currently pushing about 3-5 TB of images to AI/ML scanning per day. Max we’ve seen through the system is about 8 TB.

    Individual file? Probably 660 GB of backups before a migration at a previous job.

  • d00phy@lemmy.world
    link
    fedilink
    English
    arrow-up
    18
    ·
    30 days ago

    I’ve migrated petabytes from one GPFS file system to another. More than once, in fact. I’ve also migrated about 600TB of data from D3 tape format to 9940.

  • krazylink@lemmy.world
    link
    fedilink
    arrow-up
    2
    ·
    29 days ago

    I recently copied ~1.6T from my old file server to my new one. I think that may be my largest non-work related transfer.

  • Urist@lemmy.ml
    link
    fedilink
    English
    arrow-up
    63
    arrow-down
    1
    ·
    30 days ago

    I obviously downloaded a car after seeing that obnoxious anti-piracy ad.

  • freijon@lemmings.world
    link
    fedilink
    arrow-up
    60
    ·
    29 days ago

    I’m currently backing up my /dev folder to my unlimited cloud storage. The backup of the file /dev/random is running since two weeks.

    • Mike1576218@lemmy.ml
      link
      fedilink
      arrow-up
      9
      ·
      29 days ago

      No wonder. That file is super slow to transfer for some reason. but wait till you get to /dev/urandom. That file hat TBs to transfer at whatever pipe you can throw at it…

      • PlexSheep@infosec.pub
        link
        fedilink
        arrow-up
        3
        ·
        28 days ago

        /dev/random and other “files” in /dev are not really files, they are interfaces which van be used to interact with virtual or hardware devices. /dev/random spits out cryptographically secure random data. Another example is /dev/zero, which spits out only zero bytes.

        Both are infinite.

        Not all “files” in /dev are infinite, for example hard drives can (depending on which technology they use) be accessed under /dev/sda /dev/sdb and so on.

  • Nibodhika@lemmy.world
    link
    fedilink
    arrow-up
    3
    ·
    29 days ago

    Why would dd have a limit on the amount of data it can copy, afaik dd doesn’t check not does anything fancy, if it can copy one bit it can copy infinite.

    Even if it did any sort of validation, if it can do anything larger than RAM it needs to be able to do it in chunks.

    • CrabAndBroom@lemmy.ml
      link
      fedilink
      English
      arrow-up
      1
      arrow-down
      1
      ·
      29 days ago

      Well they do nickname it disk destroyer, so if it was unlimited and someone messed it up, it could delete the entire simulation that we live in. So its for our own good really.

    • nik9000@programming.dev
      link
      fedilink
      arrow-up
      2
      ·
      29 days ago

      Not looking at the man page, but I expect you can limit it if you want and the parser for the parameter knows about these names. If it were me it’d be one parser for byte size values and it’d work for chunk size and limit and sync interval and whatever else dd does.

      Also probably limited by the size of the number tracking. I think dd reports the number of bytes copied at the end even in unlimited mode.

    • data1701d (He/Him)@startrek.websiteOP
      link
      fedilink
      English
      arrow-up
      1
      ·
      29 days ago

      It’s less about dd’s limits and more laughs the fact that it supports units that might take decades or more for us to read a unit that size.

    • FooBarrington@lemmy.world
      link
      fedilink
      arrow-up
      3
      ·
      29 days ago

      No, it can’t copy infinite bits, because it has to store the current address somewhere. If they implement unbounded integers for this, they are still limited by your RAM, as that number can’t infinitely grow without infinite memory.

  • brygphilomena@lemmy.world
    link
    fedilink
    arrow-up
    18
    ·
    29 days ago

    In the middle of something 200tb for my Plex server going from a 12 bay system to a 36 LFF system. But I’ve also literally driven servers across the desert because it was faster than trying to move data from one datacenter to another.

  • Yeahboiiii@lemm.ee
    link
    fedilink
    arrow-up
    7
    ·
    30 days ago

    Largest one I ever did was around 4.something TB. New off-site backup server at a friends place. Took me 4 months due to data limits and an upload speed that maxed out at 3MB/s.