I don't. I lose my data like all the cool (read: fool) kids.
I too rawdog linux like a chad
Too real
Timeshift is a great tool for creating incremental backups. Basically it's a frontend for rsync and it works great. If needed you can also use it in CLI
I use Borg backup with Vorta for a GUI. Hasn't let me down yet.
Borgmatic is also a great option, cli only.
I use PikaBackup which I think uses Borg. Super good looking Gnome app that has worked for me.
Is it just me or the backup topic is recurring each few days on !linux@lemmy.ml and !selfhosted@lemmy.world?
To be on topic as well - I use restic+autorestic combo. Pretty simple, I made repo with small script to generate config for different machines and that's it. Storing between machines and b2.
It is a critical one. Maybe needs to be part of an FAQ with link to discussion.
I have a bash script that backs all my stuff up to my Homeserver with Borg. My servers have cronjobs that run similar scripts.
Exactly like you think. Cronjob runs a periodic rsync of a handful of directories under /home. My OS is on a different drive that doesn't get backed up. My configs are in an ansible repository hosted on my home server and backed up the same way.
I like rsnapshot, run from a cron job at various useful intervals. backups are hardlinked and rotated so that eventually the disk usage reaches a very slowly growing steady state.
I also use it. Big benefit is also that you don‘t need a special software to access your backup.
Pika Backup (GUI for borgbackup
) is a great app for backups. It has all the features you might expect from backup software and "just works".
I use restic
(https://restic.net/) which can use rclone
to connect to a variety of backends (eg. onedrive, mega, dropbox etc.). Also, resticprofile
(https://restic.net/) makes it easier to run (hides flags in the config file). I use it manually but a cron job would be easy to implement (a tutorial is here: https://forum.yunohost.org/t/daily-automated-backups-using-restic/16812).
Restic does not need rclone and can use many remote storage services directly. I do restic backups directly to Backblaze.
I use Back In Time to backup my important data on an external drive. And for snapshots I use timeshift.
I do periodic backups of my system from live usb via Borg Backup to a samba share.
Duplicity (cli) with deja-dup (gui) has saved my sorry ass many times.
rsync + backblaze B2. Bafkblaze is stupid cheap.
Cost is about $10 per year.
Kopia or Restic. Both do incremental, deduplicated backups and support many storage services.
Kopia provides UI for end user and has integrated scheduling. Restic is a powerfull cli tool thatlyou build your backup system on, but usually one does not need more than a cron job for that. I use a set of custom systems jobs and generators for my restic backups.
Keep in mind, than backups on local, constantly connected storage is hardly a backup. When the machine fails hard, backups are lost ,together with the original backup. So timeshift alone is not really a solution. Also: test your backups.
I really like kopia
Check out Pika backup. It's a beautiful frontend for Borg. And Borg is the shit.
Used to use Duplicati but it was buggy and would often need manual intervention to repair corruption. I gave up on it.
Now use Restic to Backblaze B2. I've been very happy.
I’ve used restic in the past; it’s good but requires a great deal of setup if memory serves me correctly. I’m currently using Duplicati on both Ubuntu and Windows and I’ve never had any issues. Thanks for sharing your experience though; I’ll be vigilant.
I rotate between a few computers. Everything is synced between them and they all have automatic btrfs snapshots. So I have several physical points to roll back from.
For a worst case scenario everything is also synced offsite weekly to a pCloud share. I have a little script that mounts it with pcloudfs, encfs and then rsyncs any updates.
BTRFS filesystem, Snapper for taking periodic snapshots and snap-sync for saving one to an external drive every now and then.
BTRFS is what makes everything incremental.
Get a Mac, use Time Machine. Go all in on the eco system. phone, watch, iPad, tv. I resisted for years but it's so good man and the apple silicon is just leaps beyond everything else.
Someone asking for Linux backup solution may prefer to avoid Apple 'ecosystem'.
Time Machine is not a backup, it is unreliable. I've had corrupted time machine backups and its backups are non-portable: You can only read the backups using an Apple machine. Apple Silicon is also not leaps beyond everything else, a 7000-series AMD chip will trade blows on performance per watt given the same power target. (source: I measured it, 60 watt power limit on a 7950X will closely match a M1 ultra given the same 60 watts of power)
Sure their laptops are tuned better out of the box and have great battery life, but that's not because of the Apple Silicon. Apple had good battery life before, even when their laptops had the same Intel chip as any other laptop. Why? Because of software.
Like before, their new M-chips are nothing special. Apple Silicon chips are great, but so are other modern chips. Apple Silicon is not "leaps beyond everything else".
If you look past their shiny fanboy-bait chips, you realize you pay **huge ** markups on RAM and storage. Apple's RAM and storage isn't anything special, but they're a lot more expensive than any other high-end RAM and storage modules, and it's not like their RAM or storage is better because, again, an AMD chip can just use regular RAM modules and an NVME SSD and it will match the M-chip performance given the same power target. Except you can replace the RAM modules and the SSD on the AMD chipset for reasonable prices.
In the end, a macbook is a great product and there's no other laptop that really gets close to its performance given its size. But that's it, that's where Apple's advantage ends. Past their ultra-light macbooks, you get overpriced hardware, crazy expensive upgrades, with an OS that isn't better, more reliable or more stable than Windows 11 (source: I use macOS and Windows 11 daily). You can buy a slightly thicker laptop (and it will still be thin and light) with replacable RAM and SSD and it will easily match the performance of the magic M1 chip with only a slight reduction in potential battery life. But guess what: If you actually USE your laptop for anything, the battery life of any laptop will quickly drop to 2-3 hours at best.
And that's just laptops. If you want actual work done, you get a desktop, and for the price of any Apple desktop you can easily get any PC to outperform it. In some cases, you can buy a PC to outperform the Apple desktop AND buy a macbook for on the go, and still have money left over. Except for power consumption ofcourse, but who cares about power consumption on a work machine? Only Apple fanboys care about that, because that's the only thing they got going for them. My time is more expensive than my power bill.
Time Machine is such a neglected product. Time Shift is worlds beyond it.
by the way, syncthing is great if you need bi-directional sync.
not exactly what you're looking for (sth like Duplicacy?) but you should probably know about it as it's a great tool.
Git projects and system configs are on GitHub (see etckeeper), the reset is synced to my self-hosted Nextcloud instance using their desktop client. There I have periodic backup using Borg for both the files and Nextcloud database.
Restic since 2018, both to locally hosted storage and to remote over ssh. I've "stuff I care about" and "stuff that can be relatively easily replaced" fairly well separated so my filtering rules are not too complicated. I used duplicity for many years before that and afbackup to DLT IV tapes prior to that.
I use rsync+btrfs snapshot solution.
- Use rsync to incrementally collect all data into a btrfs subvolume
- Deduplicate using
duperemove
- Create a read-only snapshot of the subvolume
I don't have a backup server, just an external drive that I only connect during backup.
Deduplication is mediocre, I am still looking for snapshot aware duperemove
replacement.
I'm not trying to start a flame war, but I'm genuinely curious. Why do people like btrfs over zfs? Btrfs seems very much so "not ready for prime time".
btrfs is included in the linux kernel, zfs is not on most distros
the tiny chance that an externel kernel module borking with a kernel upgrade happens sometimes and is probably scary enough for a lot of people
Features necessary for most btrfs use cases are all stable, plus btrfs is readily available in Linux kernel whereas for zfs you need additional kernel module. The availability advantage of btrfs is a big plus in case of a disaster. i.e. no additional work is required to recover your files.
(All the above only applies if your primary OS is Linux, if you use Solaris then zfs might be better.)
All my devices use Syncthing via Tailscale to get my data to my server.
From there, my server backs up nightly to rsync.net via BorgBackup.
I then have Zabbix monitoring my backups to make sure a daily is always uploaded.
DejaDup on one computer. Another is using Syncthing, another I do a manual Grsync. i really should have a better plan. lol
I do a periodic backup with Vorta towards my server. The server does a daily backup to an S3 service with Restic
I just run my own nextcloud instance. Everything important is synced to that with the nextcloud desktop client, and the server keeps a month's worth of backups on my NAS via rsync.
I use btrbk
to send btrfs snapshots to a local NAS. Consistent backups with no downtime. The only annoyance (for me at least) is that both send and receive ends must use the same SELinux policy or labels won't match.
timeshift with system files and manually my home folder
At the core it has always been rsync and Cron. Sure I add a NAS and things like rclone+cryptomator to have extra copies of synchronized data (mostly documents and media files) spread around, but it's always rsync+Cron at the core.
dont keep anything u would be upset to lose /s
Use synching on several devices to replicate data I want to keep backups of. Family photos, journals, important docs, etc. Works perfect and I run a relay node to give back to the community given I am on a unlimited data connection.
Do most of my work on nfs, with zfs backing on raidz2, send snapshots for offline backup.
Don't have a serious offsite setup yet, but it's coming.
Linux
From Wikipedia, the free encyclopedia
Linux is a family of open source Unix-like operating systems based on the Linux kernel, an operating system kernel first released on September 17, 1991 by Linus Torvalds. Linux is typically packaged in a Linux distribution (or distro for short).
Distributions include the Linux kernel and supporting system software and libraries, many of which are provided by the GNU Project. Many Linux distributions use the word "Linux" in their name, but the Free Software Foundation uses the name GNU/Linux to emphasize the importance of GNU software, causing some controversy.
Rules
- Posts must be relevant to operating systems running the Linux kernel. GNU/Linux or otherwise.
- No misinformation
- No NSFW content
- No hate speech, bigotry, etc
Related Communities
Community icon by Alpár-Etele Méder, licensed under CC BY 3.0