# SYNOPSIS
bup split [-r *host*:*path*] <-b|-t|-c|-n *name*> [-v] [-q]
- [--bench] [--max-pack-size=*bytes*]
- [--max-pack-objects=*n*] [--fanout=*count] [filenames...]
+ [--bench] [--max-pack-size=*bytes*] [-#]
+ [--max-pack-objects=*n*] [--fanout=*count]
+ [--git-ids] [--keep-boundaries] [filenames...]
# DESCRIPTION
the content into chunks of around 8k using a rolling
checksum algorithm, and saves the chunks into a bup
repository. Chunks which have previously been stored are
-not stored again (ie. they are "deduplicated").
+not stored again (ie. they are 'deduplicated').
Because of the way the rolling checksum works, chunks
tend to be very stable across changes to a given file,
-r, --remote=*host*:*path*
: save the backup set to the given remote server. If
*path* is omitted, uses the default path on the remote
- server (you still need to include the ':')
-
+ server (you still need to include the ':'). The connection to the
+ remote server is made with SSH. If you'd like to specify which port, user
+ or private key to use for the SSH connection, we recommend you use the
+ `~/.ssh/config` file.
+
-b, --blobs
: output a series of git blob ids that correspond to the
chunks in the dataset.
-v, --verbose
: increase verbosity (can be used more than once).
+--git-ids
+: stdin is a list of git object ids instead of raw data.
+ `bup split` will read the contents of each named git
+ object (if it exists in the bup repository) and split
+ it. This might be useful for converting a git
+ repository with large binary files to use bup-style
+ hashsplitting instead. This option is probably most
+ useful when combined with `--keep-boundaries`.
+
+--keep-boundaries
+: if multiple filenames are given on the command line,
+ they are normally concatenated together as if the
+ content all came from a single file. That is, the
+ set of blobs/trees produced is identical to what it
+ would have been if there had been a single input file.
+ However, if you use `--keep-boundaries`, each file is
+ split separately. You still only get a single tree or
+ commit or series of blobs, but each blob comes from
+ only one of the files; the end of one of the input
+ files always ends a blob.
+
--noop
: read the data and split it into blocks based on the "bupsplit"
rolling checksum algorithm, but don't do anything with
like k, M, or G to specify multiples of 1024,
1024*1024, 1024*1024*1024 respectively.
+-*#*, --compress=*#*
+: set the compression level to # (a value from 0-9, where
+ 9 is the highest and 0 is no compression). The default
+ is 1 (fast, loose compression)
+
# EXAMPLE
# SEE ALSO
-`bup-join`(1), `bup-index`(1), `bup-save`(1)
+`bup-join`(1), `bup-index`(1), `bup-save`(1), `bup-on`(1), `ssh_config`(5)
# BUP