I have backups on a backup hard drive and also synced to B2, but I am thinking about backing up to some format to put in the cupboard.

The issue I see is that if I don’t have a catastrophic failure and instead just accidentally delete some files one day while organising and don’t realise, at some point the oldest backup state is removed and the files are gone.

The other thing is if I get hit by a bus and no one can work out how to decrypt a backup or whatever.

So I’m thinking of a plain old unencrypted copy of photos etc that anyone could find and use. Bonus points if I can just do a new CD or whatever each year with additions.

I have about 700GB of photos and videos which is the main content I’m concerned about. Do people use DVDs for this or is there something bigger? I am adding 60GB or more each year, would be nice to do one annual addition or something like that.

  • schizo@forum.uncomfortable.business
    link
    fedilink
    English
    arrow-up
    7
    ·
    1 day ago

    So, 50 years isn’t a reasonable goal unless you have a pretty big budget for this. Essentially no media is likely to survive that long and be readable unless they’re stored in a vault, under perfect climate controlled conditions. And even if the media is fine, finding an ancient drive to read a format that no longer exists is not a guaranteed proposition.

    You frankly should be expecting to have to replace everything every couple of years, and maybe more often if your routine tests of the media show it’s started rotting.

    Long term archival storage really isn’t just a dump it to some media and lock it up and never look at ever again.

    Alternately, you could just make someone else pay for all of this, and shove all of this to something like Glacier and make the media Amazon’s problem. (Assuming Amazon is around that long and that nothing catches fire.)

    • Dave@lemmy.nzOP
      link
      fedilink
      English
      arrow-up
      2
      ·
      1 day ago

      Hmm damn. I don’t really think cloud is the right answer for what I’m trying to do.

      I disagree that formats like JPEG won’t be readable in 50 years. I feel like there would be big demand for being able to read the format even if it’s been superceded, on account of all the JPEGs that still living people have.

      Maybe I get a big drive. Each year I copy over files from the last year. Every X years I swap the hard drive for a new one, copy all data.

      How can I tell if individual files get corrupted? Like the hard drive failed in that section, then I copy the corrupted file to the new drive, and I’d never know. Can I test in bulk? 50k+ photos and videos so far.

      • sugar_in_your_tea@sh.itjust.works
        link
        fedilink
        English
        arrow-up
        2
        ·
        edit-2
        8 hours ago

        How can I tell if individual files get corrupted?

        Checksums. A good filesystem will do this for you, but you can do it yourself if you want.

        If you sync a drive with rsync or something periodically, it’ll replace files with different checksums, fixing any corruption as you go. Then smart tests should tell you if there’s any corruption the drive is aware of. I’m sure automated backup tools have options for this.

        • Dave@lemmy.nzOP
          link
          fedilink
          English
          arrow-up
          1
          ·
          4 hours ago

          I specifically don’t want to be touching previous files on the drive, it should be addition only. So I may need to write a script to do the checks, or compare against a mirror drive. I can do this with the right filesystem, but I’m worried that if I use a filesystem not readable by Windows then it may not be layman-proof enough.

          • sugar_in_your_tea@sh.itjust.works
            link
            fedilink
            English
            arrow-up
            2
            ·
            31 minutes ago

            Then I’d go with FAT on a USB, which should be plenty portable into the future. You’ll want to replace it every 5-10 years, and check on it every other year or so.

            That’s about as easy to use as I can think of. Decades down the road, physical media like DVDs and tapes may be difficult to find readers for, but USB is versatile enough that someone is bound to have access. Micro SD cards may also be a good option, as long as you keep a couple USB readers around.

            • Dave@lemmy.nzOP
              link
              fedilink
              English
              arrow-up
              1
              ·
              19 minutes ago

              I have a terrible track record with USB sticks, including completely losing a stack of photos because of a USB stick.

              I’m now thinking the benefits of a nice error-correcting file system probably outweigh the benefits of using a widely supported one. So I might use a pair of mirrored hard drives with SATA->USB cable, then include instructions along the lines of “plug into my linux laptop to access, or take to a computer repair show if you can’t work it out”.

          • Krik@lemmy.dbzer0.com
            link
            fedilink
            English
            arrow-up
            2
            ·
            1 hour ago

            Any file systems Windows can read out-of-the-box are no good file systems. What Windows read? FAT and NTFS. Former is so basic it has no mechanisms to detect errors and bitrot and the later one is a mess.
            You should stick to ext4, btrfs and zfs.

            If you want to make if fool-proof then add a sticker with ‘bring me to a computer shop to access my content’.

            • Dave@lemmy.nzOP
              link
              fedilink
              English
              arrow-up
              1
              ·
              1 hour ago

              I have considered that exact message. It does seem making it easily plug and play may be out of the question if I want the error correction capabilities.

              • Krik@lemmy.dbzer0.com
                link
                fedilink
                English
                arrow-up
                2
                ·
                58 minutes ago

                Btrfs and zfs are self-healing.

                You can make a script to check for errors and autocorrection yourself but that needs at least a second hdd. On both drives are the same data and a file or database with the checksums of the data. The script then compares the actual checksums of the two copies and the db checksum. If they match -> perfect. If they don’t match the file where there are two matching checksum is the good one and replaces the faulty one or corrects the db entry, whichever is defect. That’s it. It doesn’t have to be more complicated.

                • Dave@lemmy.nzOP
                  link
                  fedilink
                  English
                  arrow-up
                  2
                  ·
                  46 minutes ago

                  Yip I think this is the setup I will want (probably both - zfs + a custom script for validation, just to be sure). Two mirrored drives. I do need to read up some about zfs mirroring to understand it a bit more but I think I have a path to follow now.

      • schizo@forum.uncomfortable.business
        link
        fedilink
        English
        arrow-up
        7
        ·
        1 day ago

        The format is the tape in the drive, or the disk or whatever.

        Tape existed 50 years ago: nothing modern and in production can read those tapes.

        The problem is, given a big enough time window, the literal drives to read it will simply no longer exist, and you won’t be able to access even non-rotted media because of that.

        As for data integrity, there’s a lot of options: you can make a md5 sum of each file, and then do it again and see if anything is different.

        The only caveat here is you have to make sure whatever you’re using to make the checksums gets stored somewhere that’s not JUST on the drive because if the drive DOES corrupt itself, and your only record of the “good” hashes is on the drive, well, you can’t necessarily trust those hashes either.

        • Dave@lemmy.nzOP
          link
          fedilink
          English
          arrow-up
          1
          ·
          22 hours ago

          Ah good thinking. I am thinking a spare drive that I update once a year with new content and replace every few years with a new drive is a good idea.

          • schizo@forum.uncomfortable.business
            link
            fedilink
            English
            arrow-up
            2
            ·
            8 hours ago

            That could probably work.

            Were it me, I’d build a script that would re-hash and compare all the data to the previous hash as the first step of adding more files, and if the data comes out consistent, I’d copy the files over, hash everything again, save the hash results elsewhere and then repeat as needed.

            • Dave@lemmy.nzOP
              link
              fedilink
              English
              arrow-up
              1
              ·
              4 hours ago

              Yeah I think I should do something like this. I really want to make sure the files are not getting corrupted in storage without me knowing.