I am using duplicati and thinking of switching to Borg. What do you use and why?

  • GnomeComedy@beehaw.org
    link
    fedilink
    arrow-up
    2
    ·
    edit-2
    1 year ago

    I’m curious - how much data are you backing up with that method and how frequently are you doing your backups? Doesn’t sound like it would scale well, but I’m also wondering if maybe this is perfect and I’ve just been over thinking it.

    • flatbield@beehaw.org
      link
      fedilink
      arrow-up
      3
      ·
      1 year ago

      There is not a size limit. Lot of these other methods actually use GNU Tar behind the scenes anyway. More then that GNU tar has been used for decades for this purpose. Pull out any Unix book from 2 decades ago and you will see “tar”, “cpio”, and “dump/restore” as the way. The new tool out there is Pax and in fact GNU Tar supports the new “pax” format. Moreover GNU Tar with Pax format can backup almost full disk structure including hard links, ACLs, and extended attributes which a lot of tools do not do. It is still useful to archive some things at a lower level like your partition table, and boot blocks of course. You also have to decide what run-level (such as rescue) you want to archive in, and/or what services you should stop, or provide separate to file system dumps for depending on your system. Databases, and things like ecryptfs take some special thought (thought it does for any tool). It is also good to do test restores to verify your disaster plan.

      I use tar on many systems. My workstation is about 1TB of data. Backup is about 11 hours though I think it could be faster if I disabled compression (I currently use the standard gzip compression which is not optimal). I think the process is CPU bound by the compression at the moment. Going to uncompressed or using parallel gzip at level 2 is probably the fastest you can do and should really speed things up by 4X or more. I have played with this some for my wife and her raw backup is a lot faster now. My wife uses USB 3 external drives specifically plugged into USB 3 ports (the one with the SS symbol and the blue interior), and with a USB 3 related cable. I use 6TB naked SATA drives I insert into a hot mount enclosure and store in storage boxes. My backup system can theoretically do incrementals too, but it has some issues since I have moved to BTRFS so I do not use that at the moment. Did always use before. I have an idea how to fix, but need to debug and test incrementals now.

      How often: I backup monthly. When my incrementals were working I use to do it weekly or whenever I got nervous. Other option for the BTRFS file systems would be to use their native backup tools. Not sure though, I like to use generic stuff. Lot to be said for generic.

      Big downside of tar is the mind numbing man page. Getting the options correct takes some real thought. You also have to be comfortable with the shell and Bash scripting. Big upside you can customize exactly what you want.

        • flatbield@beehaw.org
          link
          fedilink
          arrow-up
          3
          ·
          1 year ago

          Yes, I actually did not know how far back, thanks. Wikipedia seems to say 1979. I know my system admin book dated 1992 talks about it and it was common then. I think my brother use to use it in the early 1980s for his job and maybe I did too a few times. Wikipedia says GNU Tar is newer and traces back to 1987. The formats have changed some and there are several. The PAX format is much newer which I think was standardized in 2001 but GNU Tar would have taken time to implement it. I do not know that date.

          People seem to forget that tar worked well back then and still does.

          • davefischer@beehaw.org
            link
            fedilink
            arrow-up
            2
            ·
            1 year ago

            I had the chance to play with late 70s Unix for a bit a few years ago. (Hardware on loan from a museum.) VERY minimal, but still recognizable. (Well, my Unix reflexes are old - I started in the mid 80s.)

            • flatbield@beehaw.org
              link
              fedilink
              arrow-up
              2
              ·
              1 year ago

              Interesting. About then I was using a VAX. Somehow I spend most of my time on other stuff until I switched to Linux around 2000.

              • davefischer@beehaw.org
                link
                fedilink
                arrow-up
                2
                ·
                1 year ago

                My first Unix was 4.3BSD on a VAX-11/750. (There was another 11/750 running VMS, but I didn’t like that nearly as much.)

                • flatbield@beehaw.org
                  link
                  fedilink
                  arrow-up
                  2
                  ·
                  1 year ago

                  Yes VMS. That was what I was using. Unix. I did use it for something a few times. The university had one of those mini-supper computers that were a thing for awhile.

                  • davefischer@beehaw.org
                    link
                    fedilink
                    arrow-up
                    2
                    ·
                    1 year ago

                    Oooh - what mini super? Something weird, or just a small vector machine? That was an interesting niche…

    • winety@dataterm.digital
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      I am using similarly “dumb” back-up system. I’ve two external USB HDDs, to which I copy my home folder every 4 to 6 months. The back-up folder currently has circa 250 GiB, but I don’t use any compression and I also probably do not have to back up my Steam library multiple times.

      Yes, it doesn’t scale very well, but at the same time, I do not need to hoard 5 year old data. Yes, I should have an off-site back-up, but if my house burns down, I have bigger problems than losing my old photos.