I’ve tried a few “common” packages, they all seem to fall down for me. I
have 500K+ files to warehouse, ranging in sizes from tiny up to 10G+. The >biggest problem seems to be RAM usage, and I’d like to have something that >works on a 3B, and ideally a 0W.
Does anyone have something that still scales up when scaling down to an RPi?
Thanks.
I’ve tried a few “common” packages, they all seem to fall down for me. I
have 500K+ files to warehouse, ranging in sizes from tiny up to 10G+.
The biggest problem seems to be RAM usage, and I’d like to have something that works on a 3B, and ideally a 0W.
Does rsync not get the job done ?
I’ve tried a few “common” packages, they all seem to fall down for me. I hav
files to warehouse, ranging in sizes from tiny up to 10G+. The biggest problem see
to be RAM usage, and I’d like to have something that works on a 3B, and ideally a
Does anyone have something that still scales up when scaling down to an RPi?
Thanks.
Does rsync not get the job done ?
I’ve tried a few “common” packages, they all seem to fall down for me. I
have 500K+ files to warehouse, ranging in sizes from tiny up to 10G+. The biggest problem seems to be RAM usage, and I’d like to have something that works on a 3B, and ideally a 0W.
Does anyone have something that still scales up when scaling down to an RPi?
Have you looked at rsnapshot?
I’ve tried a few “common” packages, they all seem to fall down for
me. I have 500K+ files to warehouse, ranging in sizes from tiny up
to 10G+. The biggest problem seems to be RAM usage, and I’d like to
have something that works on a 3B, and ideally a 0W.
Does anyone have something that still scales up when scaling down to
an RPi?
My problem with rsync-based solutions is that they seems to do *way* too
much processing in order to figure out how to do an efficient transfer.
I mean, to my mind, if I change 3 files out of 500K for a total of 3MB
out of 1TB of data, backing that up should be *fast*. I just don’t see that when I use rsync-based solutions.
DeepCore <devap@deepcore.eu> wrote:
Doc O'Leary wrote on 2022-04-12 at 20:54:
My problem with rsync-based solutions is that they seems to do *way* too >>> much processing in order to figure out how to do an efficient transfer.
I mean, to my mind, if I change 3 files out of 500K for a total of 3MB
out of 1TB of data, backing that up should be *fast*. I just don’t see >>> that when I use rsync-based solutions.
According to the following article on Arstechnica, rsync has to inspect
every file to determine which differences are to be sent over the
wire... that's the processing you experience.
By default rsync just looks at file metadata: are the file's length, date, attributes the same? If so, skip it. If they differ, go through the file and work out what changed, then send the changes. That means it has to inspect every inode, but not every file. You can make it checksum the file contents for every file, rather than just those with differing metadata, which of course is a lot slower.
Doc O'Leary wrote on 2022-04-12 at 20:54:
My problem with rsync-based solutions is that they seems to do *way* too much processing in order to figure out how to do an efficient transfer.
I mean, to my mind, if I change 3 files out of 500K for a total of 3MB
out of 1TB of data, backing that up should be *fast*. I just don’t see that when I use rsync-based solutions.
According to the following article on Arstechnica, rsync has to inspect
every file to determine which differences are to be sent over the
wire... that's the processing you experience.
I *do* already use rsync extensively for my current ???backup??? needs, but it just doesn???t have the smarts I???d like to see in a proper backup tool. Like I said, bup has features that are more in line with what I need, but
it seems to have trouble scaling down to an RPi sized server.
I make nightly rysnc differential backups of the SD cards of 15
Raspberry Pi's to SD card image files on a USB SSD drive connected to
one of the Pi 4s. The Pi's are of various generations, from 2 to 4 with
lots of Zero Ws, most are connected via WiFi, 3 on Ethernet a couple are
at remote sites. On average there is 961MB data in 4426 files
transferred, and the time to back up all of them sequentially is 14m26,
which I don't think is too bad.
If you want to avoid the inode inspection you need help from the filesystem to keep track of changes when you save them, rather than when you backup.
It's also possible to install
software that monitors file changes dynamically. That makes backups faster at the expense of making file accesses slower.
I’m pretty sure inotify support is a default part of the Linux kernel already. I’ve certainly been thinking of tapping into it to make my scripts even more efficient, but I really was hoping there was some
custom backup software that has already done the heavy lifting for me.
What do you mean by "have trouble scaling down to an RPi sized server".
Have you tried bup? What were the problems you had?
I've not come across bup before and am curious.
I make nightly rysnc differential backups of the SD cards of 15
Raspberry Pi's to SD card image files on a USB SSD drive connected to
one of the Pi 4s. The Pi's are of various generations, from 2 to 4 with
lots of Zero Ws, most are connected via WiFi, 3 on Ethernet a couple are
at remote sites. On average there is 961MB data in 4426 files
transferred, and the time to back up all of them sequentially is 14m26,
which I don't think is too bad.
Overgrive
Just appears to be a Google Drive frontend. Even if I did use Google
Drive, it isn’t clear to me it’s going to function any better/faster
than something (even rsync) replicating to a local drive. How do you
see it working better than dumb copies for backup?
On 12/04/2022 22:30, druck wrote:
I make nightly rysnc differential backups of the SD cards of 15
Raspberry Pi's to SD card image files on a USB SSD drive connected to
one of the Pi 4s. The Pi's are of various generations, from 2 to 4 with lots of Zero Ws, most are connected via WiFi, 3 on Ethernet a couple are
at remote sites. On average there is 961MB data in 4426 files
transferred, and the time to back up all of them sequentially is 14m26, which I don't think is too bad.
And it does work. I had a Pi fail to reboot yesterday, I found the the superblocks on the root partition been corrupted, so unfixable. So I
just used dd to overwrite it with last night's backup image, and it was working again within 5 minutes.
Am 13.04.2022 um 22:27 schrieb Doc O'Leary:
Just appears to be a Google Drive frontend. Even if I did use Google Drive, it isn’t clear to me it’s going to function any better/faster than something (even rsync) replicating to a local drive. How do you
see it working better than dumb copies for backup?
Once my flat burned down and killed all my backups.
druck <news@druck.org.uk> wrote:
On 12/04/2022 22:30, druck wrote:
I make nightly rysnc differential backups of the SD cards of 15
Raspberry Pi's to SD card image files
Do you loopback mount the target images, so they keep their ext4 partition format and bootability? That's quite a neat idea...
(if it were me I'd be tempted to keep a second copy of the files on the host's native FS, outside of the ext4 image. So if the target ext4 got corrupted in some way I could always recover the files)
Are you backing up a 1GB Pi using rsync *to* a remote system?
That way the
system with more performance and memory does all hard work of comparing
indexes.
But that really doesn’t take all that much CPU or RAM. I mean, having
more resources certainly *helps*, but that doesn’t mean a backup system
has to be architected in such a way to *require* 4GB of memory (or
more) to manage a data warehouse of 500K files totaling 2TB. I
wouldn’t even call that big data.
I do also use rsync for replication, but that’s just not the same as
having a backup system that archives data in perpetuity. It’s sounding like what I’m looking for doesn’t exist and I should just stick with my scripts.
I just use dump to dump (levels 0 and 3 only) almost the /entire/ fbsd
server onto one of the desktops. (which stood me in good stead this week
when I completely trashed /var; ooops :-{ )
Have you seen duplicity?
On Sat, 16 Apr 2022 12:53:23 +0100
Mike Scott <usenet.16@scottsonline.org.uk.invalid> wrote:
I just use dump to dump (levels 0 and 3 only) almost the /entire/ fbsd server onto one of the desktops. (which stood me in good stead this week when I completely trashed /var; ooops :-{ )
ZFS snapshots YKIMS.
Sysop: | Keyop |
---|---|
Location: | Huddersfield, West Yorkshire, UK |
Users: | 371 |
Nodes: | 16 (2 / 14) |
Uptime: | 37:32:20 |
Calls: | 7,932 |
Calls today: | 2 |
Files: | 12,998 |
Messages: | 5,805,631 |