-
-
Notifications
You must be signed in to change notification settings - Fork 753
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
try / document borg + ntfsclone #81
Comments
I tried it with system-rescuecd (includes borg, ntfsclone and other useful stuff), a MBR partitioned disk, windows on 2 partitions: Please verify the commands before you use them, especially the disk identifiers!
Note: mbr_gap contains the MBR (1 sector) plus the gap after it (which may be used or not). |
TODO: verify it, check dedup, maybe optimize chunker params, create nice rst docs. |
I just tried this and it works great. 585GB ntfsclone image, piped to borg, LZ4 compressed down to 475GB, then deduplicated down to 445GB. Took about 12 hours. |
@HomerSlated I had the impression that at least the first run is quite a lot slower than just writing it to disk. Maybe it gets better if one saves another version of that image to same repo. |
I'll test this when I make my first incremental, in a few days time. However, I believe this is expected behaviour from any deduplication system. Borg is very far from being the slowest. I'm currently doing an initial CrashPlan backup on exactly the same dataset, except half is excluded, and CrashPlan is predicting -> 3 DAYS <- for completion. That's 3 days to backup 250GB, vs Borg backing up the full 500GB in only 12 hours. I haven't tested OpenDedup yet, but I've heard that its even slower than CrashPlan. Also, CrashPlan's size estimate would suggest, bizarrely, that its final archive will actually be bigger than the input data, whereas Borg's was markedly smaller even on the initial run. I'll let you know the final results when they're ready. |
Here's the results. In summary, a full backup of ~550GB took ~10 hours, and the first incremental (three weeks later) took ~4 hours, deduplicating down to ~20GB.
I'm also testing UrBackup with its CBT (Changed Block Tracker) on the same dataset. So far it's predicting ~7 hours to back up (which seems wrong, unless the CBT isn't working properly), although I don't yet know the final size of the differential. I'm also planning to test burp2 at some point, for comparison. |
Hint: using more recent python (e.g. 3.5, if possible) may give better speed. The borg binary bundles latest python release (3.5.2 currently). |
@HomerSlated keep us updated about performance and other results of your comparison (I wanted to create and set up something to compare backup tools performance, but didn't come to it yet). |
Yes, a defined testing environment for controlled results would be useful. The problem is that at least one of those tests would require running from a different OS (Windows), thus breaking the control conditions (UrBackup imaging and CBT currently only works under Windows). Nonetheless, a "real-world" speed and size comparison is useful. |
Also Python 3.5 is currently masked unstable on Gentoo, and will cause all kinds of dependency issues if I emerge it, so I'll leave it for now.
|
Well it turned out that UrBackup took 222 minutes (3 hours, 42 mins) to complete, with an incremental size of just 4.94GB, and that's without CBT functioning correctly (would have reduced the time to maybe 5 minutes). |
15 MB/s is indeed rather slow (even for Borg 1.0). UrBackup's 40 MB/s is a lot better, but still kinda slow. I'm not sure how UrBackup works, it seems to be a classic full+delta backup system, so it probably uses much less CPU - likely the limiting factor in your case? CBT is indeed an interesting technology. It would seem to enable a similar behaviour as Borg's file metadata cache (which allows Borg to instantaneously backup an unchanged file, regardless of size), just for images, and not file systems -- quite interesting! |
I should add that the backup storage is on USB 3.0, which on my hardware benches at ~200 MB/s seq r/w, although clearly that's more than 40 MB/s, so that's probably not the bottleneck. Actually I'm not really bothered about the speed (except when the CBT client I paid for doesn't seem to work). I'm more interested in saving storage space, and both borg and UrBackup work well in that regard, although the latter seems to have the edge right now (in my tests so far). I just tried burp 2.x, but can't figure out how to get it to work. It seems I need to RTFM on OpenSSL certificates before I can even get started. |
@ThomasWaldmann |
Please note: If your deleted sectors contain random old data, deduplication will not be able to deduplicate everything and it will be still slow and large. 'dd' is not the best choice here. ntfsclone is designed to only copy the used sectors. Even without borg, you should benefit form using ntfsclone. Using borg+ntfsclone also seems to work without issues, giving even smaller backups. (just the discussed performance issues are left to be investigated) |
could be useful to clone windows systems (with ntfs filesystems)?
due to ntfsclone, would only save allocated blocks.
try it and document your results.
The text was updated successfully, but these errors were encountered: