I use boring old zfs snapshot
+ zfs send -i
.
It's not pretty, but it's reliable.
Linux
From Wikipedia, the free encyclopedia
Linux is a family of open source Unix-like operating systems based on the Linux kernel, an operating system kernel first released on September 17, 1991 by Linus Torvalds. Linux is typically packaged in a Linux distribution (or distro for short).
Distributions include the Linux kernel and supporting system software and libraries, many of which are provided by the GNU Project. Many Linux distributions use the word "Linux" in their name, but the Free Software Foundation uses the name GNU/Linux to emphasize the importance of GNU software, causing some controversy.
Rules
- Posts must be relevant to operating systems running the Linux kernel. GNU/Linux or otherwise.
- No misinformation
- No NSFW content
- No hate speech, bigotry, etc
Related Communities
Community icon by Alpár-Etele Méder, licensed under CC BY 3.0
I've used a combination of
- Managing ZFS snapshots with pyznap
- Plain old rsync to copy important files that happen not to be on ZFS filesystems to ZFS.
If I were doing this over today, I'd probably consider https://zrepl.github.io/ instead of pyznap, as pyznap is no longer receiving real active development.
In the past I've used rdiff-backup, which is great but it's hard to beat copy-on-write snapshots for speed and being lightweight.
rsnapshot
If only restic deduplicated... But other than that it does okay.
restic does do deduplication.
At this moment I use too many tools.
For user data on my PC and on home server I mostly use Duplicacy. It is fast and efficient. All data backed up locally on NAS box over SFTP, and a subset of that data is backed up to S3 cloud storage.
I have a Mac, this one is using TimeMachine, storing data on NAS, then it's synced to S3 cloud storage one a day.
And on top of that VMs and containers from home server are backed up by Proxmox built in tool to NAS. These mostly exclude user data.
An external hard drive works 100%. And relying on .dotfiles to redownload the whole thing back.
...I mean, it takes like less than 3 minutes to redownload and 5 reconfiguring everything manually, so eh.
Restic in the homelab and Veeam at work. I’m pretty happy with both!
I use Raspberry Pi 4 with connected external HDD and installed Nextcloud
Restic (local repo) which I sync onto a Hetzner Storagebox using rclone.
i simply use freefilesync
@dustyData I have hundreds of thousands of files that need to be backed up locally and in the cloud. I use either Vorta or Pika. Both are interfaces for Borg. Easy to use and their deduplication feature manages to save a lot of diskspace. I tried so many backup solutions and none worked as reliably.
I use Pika and Timeshift.
I’ve recently started using proxmox -backup-client. Works well. Goes to my backup server along with my vm image backups. Works nicely with full deducing and such. Quite good savings if you are backing up multiple machines.
I the. Rsync this up to cloud once a day.
Deja Dup backs my local machines to my Synology NAS. That uses Hyper-backup to send everything to Dropbox.
Dejadup backup is neat if you need a GUI. But TBH, you really don't need a GUI, restic will work just fine as long as you target a few folders. It mostly boils down to file/folder hygiene.
I just map my entire documents, pictures and other important home folders to subfolders inside Dropbox. This propagates all of my files across all of my computers via the cloud and makes everything accessible from my phone as well.
I don't worry about backing up my operating system, though important configuration file locations are also mapped into Dropbox for easily setting things up again. Complete portable apps are also located in Dropbox.
I just use MegaSync, which backsup my config folder and documents folder.
On phone, I use syncthing to backup to home server (I never knew syncthing can backup over WAN), then synced to MegaSync. I also keep all the files on MegaSync on my server just in case megasync suddenly goes down one day.