5 KiB
5 KiB
backup2mdisc
Now you can enjoy a self-contained backup on each disc without chain-dependency across your entire multi TB backup set!:
- Independently decryptable (and restorable) archives on each M-Disc.
- Automatic ISO creation and optional disc burning in the same script.
- Fast compression via lz4.
Purpose:
- Scans all files in a source directory.
- Groups them into "chunks" so that each chunk is <= a specified size (default 100GB).
- Creates a TAR archive of each chunk, compresses it with
lz4
, and encrypts it with GPG (AES256). - Each
.tar.lz4.gpg
is fully independent (no other parts/discs needed to restore that chunk). - (Optional) Creates ISO images from each encrypted chunk if
--create-iso
is provided. - (Optional) Burns each chunk or ISO to M-Disc if
--burn
is provided.
How It Works
-
File Collection & Sorting
- The script uses
find
to list all files in yourSOURCE_DIR
with their sizes. - It sorts them in ascending order by size so it can pack smaller files first (you can remove
| sort -n
if you prefer a different method).
- The script uses
-
Chunk Accumulation
- It iterates over each file, summing up file sizes into a “current chunk.”
- If adding a new file would exceed
CHUNK_SIZE
(default 100GB), it finalizes the current chunk (creates.tar.lz4.gpg
) and starts a new one.
-
Archive, Compress, Encrypt
- For each chunk, it creates a
.tar.lz4.gpg
file. Specifically:tar -cf - -T $TMP_CHUNK_LIST
(archive of the files in that chunk)- Pipe into
lz4 -c
for fast compression - Pipe into
gpg --batch -c
(symmetric encrypt with AES256, using your passphrase)
- The result is a self-contained file like
chunk_001.tar.lz4.gpg
.
- For each chunk, it creates a
-
Checksums & Manifest
- It calculates the SHA-256 sum of each chunk archive and appends it to a manifest file along with the list of included files.
- That manifest is stored in
$WORK_DIR
.
-
Optional ISO Creation (
--create-iso
)- After each chunk is created, the script can build an ISO image containing just that
.tar.lz4.gpg
. - This step uses
genisoimage
(ormkisofs
). The resulting file ischunk_001.iso
, etc.
- After each chunk is created, the script can build an ISO image containing just that
-
Optional Burning (
--burn
)- If you specify
--burn
, the script will pause after creating each chunk/ISO and prompt you to insert a fresh M-Disc. - On Linux, it tries
growisofs
. - On macOS, it tries
hdiutil
(if creating an ISO). - If it doesn't find these commands, it'll instruct you to burn manually.
- If you specify
-
Repeat
- The script loops until all files have been placed into chunk(s).
Usage:
./backup2mdisc.sh /path/to/source /path/to/destination [CHUNK_SIZE] [--create-iso] [--burn]
Examples:
./backup2mdisc.sh /home/user/data /mnt/backup 100G --create-iso
./backup2mdisc.sh /data /backup 50G --burn
Dependencies:
bash
gpg
(for encryption)lz4
(for fast compression)tar
split
or file-based grouping approachsha256sum
(or 'shasum -a 256
' on macOS/FreeBSD)genisoimage
ormkisofs
(for creating ISOs if--create-iso
)growisofs
(Linux) orhdiutil
(macOS) for burning if--burn
Restoring Your Data
-
Disc is self-contained: If you have disc #4 containing
chunk_004.tar.lz4.gpg
, you can restore it independently of the others. -
Decrypt & Extract:
gpg --decrypt chunk_004.tar.lz4.gpg | lz4 -d | tar -xvf -
This will prompt for the passphrase you used during backup.
-
If one disc is lost, you only lose the files in that chunk; all other chunks remain restorable.
Why lz4?
- Speed:
lz4
is extremely fast at both compression and decompression. - Less compression ratio than xz, but if your priority is speed (and 100GB disc space is enough),
lz4
is a great choice. - For maximum compression at the cost of time, you could replace
lz4
withxz -9
, but expect slower backups and restores.
Tips & Caveats
-
Large Files
- A single file larger than your chunk size (e.g., 101GB file with a 100GB chunk limit) won't fit. This script doesn't handle that gracefully. You'd need to split such a file (e.g., with
split
) before archiving or use a backup tool that supports partial file splitting.
- A single file larger than your chunk size (e.g., 101GB file with a 100GB chunk limit) won't fit. This script doesn't handle that gracefully. You'd need to split such a file (e.g., with
-
Verification
- Always verify your discs after burning. Mount them and compare the chunk's SHA-256 with the manifest to ensure data integrity.
-
Incremental or Deduplicated Backups
- For advanced features (incremental, deduplication, partial-chunk checksums), consider specialized backup programs (like Borg, restic, Duplicati). However, they usually produce multi-volume archives that need all volumes to restore.
-
Cross-Platform
- On FreeBSD or macOS, you might need to tweak the commands for hashing (
sha256sum
vs.shasum -a 256
) or ISO creation (mkisofs
vs.genisoimage
). - For burning, Linux uses
growisofs
, macOS useshdiutil
, and FreeBSD may requirecdrecord
or another tool.
- On FreeBSD or macOS, you might need to tweak the commands for hashing (