rclone mount


rclone mount

Mount the remote as a mountpoint. EXPERIMENTAL

Synopsis

rclone mount allows Linux, FreeBSD, macOS and Windows to mount any of Rclone’s cloud storage systems as a file system with FUSE.

This is EXPERIMENTAL - use with care.

First set up your remote using rclone config. Check it works with rclone ls etc.

Start the mount like this

rclone mount remote:path/to/files /path/to/local/mount

Or on Windows like this where X: is an unused drive letter

rclone mount remote:path/to/files X:

When the program ends, either via Ctrl+C or receiving a SIGINT or SIGTERM signal, the mount is automatically stopped.

The umount operation can fail, for example when the mountpoint is busy. When that happens, it is the user’s responsibility to stop the mount manually with

# Linux
fusermount -u /path/to/local/mount
# OS X
umount /path/to/local/mount

Installing on Windows

To run rclone mount on Windows, you will need to download and install WinFsp.

WinFsp is an open source Windows File System Proxy which makes it easy to write user space file systems for Windows. It provides a FUSE emulation layer which rclone uses combination with cgofuse. Both of these packages are by Bill Zissimopoulos who was very helpful during the implementation of rclone mount for Windows.

Windows caveats

Note that drives created as Administrator are not visible by other accounts (including the account that was elevated as Administrator). So if you start a Windows drive from an Administrative Command Prompt and then try to access the same drive from Explorer (which does not run as Administrator), you will not be able to see the new drive.

The easiest way around this is to start the drive from a normal command prompt. It is also possible to start a drive from the SYSTEM account (using the WinFsp.Launcher infrastructure) which creates drives accessible for everyone on the system or alternatively using the nssm service manager.

Limitations

Without the use of “--vfs-cache-mode” this can only write files sequentially, it can only seek when reading. This means that many applications won’t work with their files on an rclone mount without “--vfs-cache-mode writes” or “--vfs-cache-mode full”. See the File Caching section for more info.

The bucket based remotes (eg Swift, S3, Google Compute Storage, B2, Hubic) won’t work from the root - you will need to specify a bucket, or a path within the bucket. So swift: won’t work whereas swift:bucket will as will swift:bucket/path. None of these support the concept of directories, so empty directories will have a tendency to disappear once they fall out of the directory cache.

Only supported on Linux, FreeBSD, OS X and Windows at the moment.

rclone mount vs rclone sync/copy

File systems expect things to be 100% reliable, whereas cloud storage systems are a long way from 100% reliable. The rclone sync/copy commands cope with this with lots of retries. However rclone mount can’t use retries in the same way without making local copies of the uploads. Look at the EXPERIMENTAL file caching for solutions to make mount mount more reliable.

Attribute caching

You can use the flag --attr-timeout to set the time the kernel caches the attributes (size, modification time etc) for directory entries.

The default is “1s” which caches files just long enough to avoid too many callbacks to rclone from the kernel.

In theory 0s should be the correct value for filesystems which can change outside the control of the kernel. However this causes quite a few problems such as rclone using too much memory, rclone not serving files to samba and excessive time listing directories.

The kernel can cache the info about a file for the time given by “--attr-timeout”. You may see corruption if the remote file changes length during this window. It will show up as either a truncated file or a file with garbage on the end. With “--attr-timeout 1s” this is very unlikely but not impossible. The higher you set “--attr-timeout” the more likely it is. The default setting of “1s” is the lowest setting which mitigates the problems above.

If you set it higher (‘10s’ or ‘1m’ say) then the kernel will call back to rclone less often making it more efficient, however there is more chance of the corruption issue above.

If files don’t change on the remote outside of the control of rclone then there is no chance of corruption.

This is the same as setting the attr_timeout option in mount.fuse.

Filters

Note that all the rclone filters can be used to select a subset of the files to be visible in the mount.

systemd

When running rclone mount as a systemd service, it is possible to use Type=notify. In this case the service will enter the started state after the mountpoint has been successfully set up. Units having the rclone mount service specified as a requirement will see all files and folders immediately in this mode.

chunked reading

--vfs-read-chunk-size will enable reading the source objects in parts. This can reduce the used download quota for some remotes by requesting only chunks from the remote that are actually read at the cost of an increased number of requests.

When --vfs-read-chunk-size-limit is also specified and greater than --vfs-read-chunk-size, the chunk size for each open file will get doubled for each chunk read, until the specified value is reached. A value of -1 will disable the limit and the chunk size will grow indefinitely.

With --vfs-read-chunk-size 100M and --vfs-read-chunk-size-limit 0 the following parts will be downloaded: 0-100M, 100M-200M, 200M-300M, 300M-400M and so on. When --vfs-read-chunk-size-limit 500M is specified, the result would be 0-100M, 100M-300M, 300M-700M, 700M-1200M, 1200M-1700M and so on.

Chunked reading will only work with --vfs-cache-mode < full, as the file will always be copied to the vfs cache before opening with --vfs-cache-mode full.

Directory Cache

Using the --dir-cache-time flag, you can set how long a directory should be considered up to date and not refreshed from the backend. Changes made locally in the mount may appear immediately or invalidate the cache. However, changes done on the remote will only be picked up once the cache expires.

Alternatively, you can send a SIGHUP signal to rclone for it to flush all directory caches, regardless of how old they are. Assuming only one rclone instance is running, you can reset the cache like this:

kill -SIGHUP $(pidof rclone)

If you configure rclone with a remote control then you can use rclone rc to flush the whole directory cache:

rclone rc vfs/forget

Or individual files or directories:

rclone rc vfs/forget file=path/to/file dir=path/to/dir

File Buffering

The --buffer-size flag determines the amount of memory, that will be used to buffer data in advance.

Each open file descriptor will try to keep the specified amount of data in memory at all times. The buffered data is bound to one file descriptor and won’t be shared between multiple open file descriptors of the same file.

This flag is a upper limit for the used memory per file descriptor. The buffer will only use memory for data that is downloaded but not not yet read. If the buffer is empty, only a small amount of memory will be used. The maximum memory used by rclone for buffering can be up to --buffer-size * open files.

File Caching

NB File caching is EXPERIMENTAL - use with care!

These flags control the VFS file caching options. The VFS layer is used by rclone mount to make a cloud storage system work more like a normal file system.

You’ll need to enable VFS caching if you want, for example, to read and write simultaneously to a file. See below for more details.

Note that the VFS cache works in addition to the cache backend and you may find that you need one or the other or both.

--cache-dir string                   Directory rclone will use for caching.
--vfs-cache-max-age duration         Max age of objects in the cache. (default 1h0m0s)
--vfs-cache-mode string              Cache mode off|minimal|writes|full (default "off")
--vfs-cache-poll-interval duration   Interval to poll the cache for stale objects. (default 1m0s)

If run with -vv rclone will print the location of the file cache. The files are stored in the user cache file area which is OS dependent but can be controlled with --cache-dir or setting the appropriate environment variable.

The cache has 4 different modes selected by --vfs-cache-mode. The higher the cache mode the more compatible rclone becomes at the cost of using disk space.

Note that files are written back to the remote only when they are closed so if rclone is quit or dies with open files then these won’t get written back to the remote. However they will still be in the on disk cache.

--vfs-cache-mode off

In this mode the cache will read directly from the remote and write directly to the remote without caching anything on disk.

This will mean some operations are not possible

  • Files can’t be opened for both read AND write
  • Files opened for write can’t be seeked
  • Existing files opened for write must have O_TRUNC set
  • Files open for read with O_TRUNC will be opened write only
  • Files open for write only will behave as if O_TRUNC was supplied
  • Open modes O_APPEND, O_TRUNC are ignored
  • If an upload fails it can’t be retried

--vfs-cache-mode minimal

This is very similar to “off” except that files opened for read AND write will be buffered to disks. This means that files opened for write will be a lot more compatible, but uses the minimal disk space.

These operations are not possible

  • Files opened for write only can’t be seeked
  • Existing files opened for write must have O_TRUNC set
  • Files opened for write only will ignore O_APPEND, O_TRUNC
  • If an upload fails it can’t be retried

--vfs-cache-mode writes

In this mode files opened for read only are still read directly from the remote, write only and read/write files are buffered to disk first.

This mode should support all normal file system operations.

If an upload fails it will be retried up to --low-level-retries times.

--vfs-cache-mode full

In this mode all reads and writes are buffered to and from disk. When a file is opened for read it will be downloaded in its entirety first.

This may be appropriate for your needs, or you may prefer to look at the cache backend which does a much more sophisticated job of caching, including caching directory hierarchies and chunks of files.

In this mode, unlike the others, when a file is written to the disk, it will be kept on the disk after it is written to the remote. It will be purged on a schedule according to --vfs-cache-max-age.

This mode should support all normal file system operations.

If an upload or download fails it will be retried up to --low-level-retries times.

rclone mount remote:path /path/to/mountpoint [flags]

Options

      --allow-non-empty                    Allow mounting over a non-empty directory.
      --allow-other                        Allow access to other users.
      --allow-root                         Allow access to root user.
      --attr-timeout duration              Time for which file/directory attributes are cached. (default 1s)
      --daemon                             Run mount as a daemon (background mode).
      --daemon-timeout duration            Time limit for rclone to respond to kernel (not supported by all OSes).
      --debug-fuse                         Debug the FUSE internals - needs -v.
      --default-permissions                Makes kernel enforce access control based on the file mode.
      --dir-cache-time duration            Time to cache directory entries for. (default 5m0s)
      --fuse-flag stringArray              Flags or arguments to be passed direct to libfuse/WinFsp. Repeat if required.
      --gid uint32                         Override the gid field set by the filesystem. (default 1001)
  -h, --help                               help for mount
      --max-read-ahead int                 The number of bytes that can be prefetched for sequential reads. (default 128k)
      --no-checksum                        Don't compare checksums on up/download.
      --no-modtime                         Don't read/write the modification time (can speed things up).
      --no-seek                            Don't allow seeking in files.
  -o, --option stringArray                 Option for libfuse/WinFsp. Repeat if required.
      --poll-interval duration             Time to wait between polling for changes. Must be smaller than dir-cache-time. Only on supported remotes. Set to 0 to disable. (default 1m0s)
      --read-only                          Mount read-only.
      --uid uint32                         Override the uid field set by the filesystem. (default 1001)
      --umask int                          Override the permission bits set by the filesystem.
      --vfs-cache-max-age duration         Max age of objects in the cache. (default 1h0m0s)
      --vfs-cache-mode string              Cache mode off|minimal|writes|full (default "off")
      --vfs-cache-poll-interval duration   Interval to poll the cache for stale objects. (default 1m0s)
      --vfs-read-chunk-size int            Read the source objects in chunks. (default 128M)
      --vfs-read-chunk-size-limit int      If greater than --vfs-read-chunk-size, double the chunk size after each chunk read, until the limit is reached. 'off' is unlimited. (default off)
      --volname string                     Set the volume name (not supported by all OSes).
      --write-back-cache                   Makes kernel buffer writes before sending them to rclone. Without this, writethrough caching is used.

Options inherited from parent commands

      --acd-auth-url string                   Auth server URL.
      --acd-client-id string                  Amazon Application Client ID.
      --acd-client-secret string              Amazon Application Client Secret.
      --acd-templink-threshold SizeSuffix     Files >= this size will be downloaded via their tempLink. (default 9G)
      --acd-token-url string                  Token server url.
      --acd-upload-wait-per-gb Duration       Additional time per GB to wait after a failed complete upload to see if it appears. (default 3m0s)
      --alias-remote string                   Remote or path to alias.
      --ask-password                          Allow prompt for password for encrypted configuration. (default true)
      --auto-confirm                          If enabled, do not request console confirmation.
      --azureblob-account string              Storage Account Name (leave blank to use connection string or SAS URL)
      --azureblob-chunk-size SizeSuffix       Upload chunk size. Must fit in memory. (default 4M)
      --azureblob-endpoint string             Endpoint for the service
      --azureblob-key string                  Storage Account Key (leave blank to use connection string or SAS URL)
      --azureblob-sas-url string              SAS URL for container level access only
      --azureblob-upload-cutoff SizeSuffix    Cutoff for switching to chunked upload. (default 256M)
      --b2-account string                     Account ID or Application Key ID
      --b2-chunk-size SizeSuffix              Upload chunk size. Must fit in memory. (default 96M)
      --b2-endpoint string                    Endpoint for the service.
      --b2-hard-delete                        Permanently delete files on remote removal, otherwise hide files.
      --b2-key string                         Application Key
      --b2-test-mode string                   A flag string for X-Bz-Test-Mode header for debugging.
      --b2-upload-cutoff SizeSuffix           Cutoff for switching to chunked upload. (default 190.735M)
      --b2-versions                           Include old versions in directory listings.
      --backup-dir string                     Make backups into hierarchy based in DIR.
      --bind string                           Local address to bind to for outgoing connections, IPv4, IPv6 or name.
      --box-client-id string                  Box App Client Id.
      --box-client-secret string              Box App Client Secret
      --box-commit-retries int                Max number of times to try committing a multipart file. (default 100)
      --box-upload-cutoff SizeSuffix          Cutoff for switching to multipart upload. (default 50M)
      --buffer-size int                       In memory buffer size when reading files for each --transfer. (default 16M)
      --bwlimit BwTimetable                   Bandwidth limit in kBytes/s, or use suffix b|k|M|G or a full timetable.
      --cache-chunk-clean-interval Duration   Interval at which chunk cleanup runs (default 1m0s)
      --cache-chunk-no-memory                 Disable the in-memory cache for storing chunks during streaming
      --cache-chunk-path string               Directory to cache chunk files (default "/home/rclone/.cache/rclone/cache-backend")
      --cache-chunk-size SizeSuffix           The size of a chunk. Lower value good for slow connections but can affect seamless reading. (default 5M)
      --cache-chunk-total-size SizeSuffix     The maximum size of stored chunks. When the storage grows beyond this size, the oldest chunks will be deleted. (default 10G)
      --cache-db-path string                  Directory to cache DB (default "/home/rclone/.cache/rclone/cache-backend")
      --cache-db-purge                        Purge the cache DB before
      --cache-db-wait-time Duration           How long to wait for the DB to be available - 0 is unlimited (default 1s)
      --cache-dir string                      Directory rclone will use for caching. (default "/home/rclone/.cache/rclone")
      --cache-info-age Duration               How much time should object info (file size, file hashes etc) be stored in cache. (default 6h0m0s)
      --cache-plex-password string            The password of the Plex user
      --cache-plex-url string                 The URL of the Plex server
      --cache-plex-username string            The username of the Plex user
      --cache-read-retries int                How many times to retry a read from a cache storage (default 10)
      --cache-remote string                   Remote to cache.
      --cache-rps int                         Limits the number of requests per second to the source FS. -1 disables the rate limiter (default -1)
      --cache-tmp-upload-path string          Directory to keep temporary files until they are uploaded to the cloud storage
      --cache-tmp-wait-time Duration          How long should files be stored in local cache before being uploaded (default 15s)
      --cache-workers int                     How many workers should run in parallel to download chunks (default 4)
      --cache-writes                          Will cache file data on writes through the FS
      --checkers int                          Number of checkers to run in parallel. (default 8)
  -c, --checksum                              Skip based on checksum & size, not mod-time & size
      --config string                         Config file. (default "/home/rclone/.rclone.conf")
      --contimeout duration                   Connect timeout (default 1m0s)
  -L, --copy-links                            Follow symlinks and copy the pointed to item.
      --cpuprofile string                     Write cpu profile to file
      --crypt-directory-name-encryption       Option to either encrypt directory names or leave them intact. (default true)
      --crypt-filename-encryption string      How to encrypt the filenames. (default "standard")
      --crypt-password string                 Password or pass phrase for encryption.
      --crypt-password2 string                Password or pass phrase for salt. Optional but recommended.
      --crypt-remote string                   Remote to encrypt/decrypt.
      --crypt-show-mapping                    For all files listed show how the names encrypt.
      --delete-after                          When synchronizing, delete files on destination after transfering (default)
      --delete-before                         When synchronizing, delete files on destination before transfering
      --delete-during                         When synchronizing, delete files during transfer
      --delete-excluded                       Delete files on dest excluded from sync
      --disable string                        Disable a comma separated list of features.  Use help to see a list.
      --drive-acknowledge-abuse               Set to allow files which return cannotDownloadAbusiveFile to be downloaded.
      --drive-alternate-export                Use alternate export URLs for google documents export.
      --drive-auth-owner-only                 Only consider files owned by the authenticated user.
      --drive-chunk-size SizeSuffix           Upload chunk size. Must a power of 2 >= 256k. (default 8M)
      --drive-client-id string                Google Application Client Id
      --drive-client-secret string            Google Application Client Secret
      --drive-formats string                  Comma separated list of preferred formats for downloading Google docs. (default "docx,xlsx,pptx,svg")
      --drive-impersonate string              Impersonate this user when using a service account.
      --drive-keep-revision-forever           Keep new head revision forever.
      --drive-list-chunk int                  Size of listing chunk 100-1000. 0 to disable. (default 1000)
      --drive-root-folder-id string           ID of the root folder
      --drive-scope string                    Scope that rclone should use when requesting access from drive.
      --drive-service-account-file string     Service Account Credentials JSON file path
      --drive-shared-with-me                  Only show files that are shared with me
      --drive-skip-gdocs                      Skip google documents in all listings.
      --drive-trashed-only                    Only show files that are in the trash
      --drive-upload-cutoff SizeSuffix        Cutoff for switching to chunked upload (default 8M)
      --drive-use-created-date                Use created date instead of modified date.
      --drive-use-trash                       Send files to the trash instead of deleting permanently. (default true)
      --dropbox-chunk-size SizeSuffix         Upload chunk size. Max 150M. (default 48M)
      --dropbox-client-id string              Dropbox App Client Id
      --dropbox-client-secret string          Dropbox App Client Secret
  -n, --dry-run                               Do a trial run with no permanent changes
      --dump string                           List of items to dump from: headers,bodies,requests,responses,auth,filters,goroutines,openfiles
      --dump-bodies                           Dump HTTP headers and bodies - may contain sensitive info
      --dump-headers                          Dump HTTP bodies - may contain sensitive info
      --exclude stringArray                   Exclude files matching pattern
      --exclude-from stringArray              Read exclude patterns from file
      --exclude-if-present string             Exclude directories if filename is present
      --fast-list                             Use recursive list if available. Uses more memory but fewer transactions.
      --files-from stringArray                Read list of source-file names from file
  -f, --filter stringArray                    Add a file-filtering rule
      --filter-from stringArray               Read filtering patterns from a file
      --ftp-host string                       FTP host to connect to
      --ftp-pass string                       FTP password
      --ftp-port string                       FTP port, leave blank to use default (21)
      --ftp-user string                       FTP username, leave blank for current username,
      --gcs-bucket-acl string                 Access Control List for new buckets.
      --gcs-client-id string                  Google Application Client Id
      --gcs-client-secret string              Google Application Client Secret
      --gcs-location string                   Location for the newly created buckets.
      --gcs-object-acl string                 Access Control List for new objects.
      --gcs-project-number string             Project number.
      --gcs-service-account-file string       Service Account Credentials JSON file path
      --gcs-storage-class string              The storage class to use when storing objects in Google Cloud Storage.
      --http-url string                       URL of http host to connect to
      --hubic-client-id string                Hubic Client Id
      --hubic-client-secret string            Hubic Client Secret
      --ignore-checksum                       Skip post copy check of checksums.
      --ignore-errors                         delete even if there are I/O errors
      --ignore-existing                       Skip all files that exist on destination
      --ignore-size                           Ignore size when skipping use mod-time or checksum.
  -I, --ignore-times                          Don't skip files that match size and time - transfer all files
      --immutable                             Do not modify files. Fail if existing files have been modified.
      --include stringArray                   Include files matching pattern
      --include-from stringArray              Read include patterns from file
      --jottacloud-mountpoint string          The mountpoint to use.
      --jottacloud-pass string                Password.
      --jottacloud-user string                User Name
      --local-no-check-updated                Don't check to see if the files change during upload
      --local-no-unicode-normalization        Don't apply unicode normalization to paths and filenames
      --local-nounc string                    Disable UNC (long path names) conversion on Windows
      --log-file string                       Log everything to this file
      --log-level string                      Log level DEBUG|INFO|NOTICE|ERROR (default "NOTICE")
      --low-level-retries int                 Number of low level retries to do. (default 10)
      --max-age duration                      Only transfer files younger than this in s or suffix ms|s|m|h|d|w|M|y (default off)
      --max-backlog int                       Maximum number of objects in sync or check backlog. (default 10000)
      --max-delete int                        When synchronizing, limit the number of deletes (default -1)
      --max-depth int                         If set limits the recursion depth to this. (default -1)
      --max-size int                          Only transfer files smaller than this in k or suffix b|k|M|G (default off)
      --max-transfer int                      Maximum size of data to transfer. (default off)
      --mega-debug                            Output more debug from Mega.
      --mega-hard-delete                      Delete files permanently rather than putting them into the trash.
      --mega-pass string                      Password.
      --mega-user string                      User name
      --memprofile string                     Write memory profile to file
      --min-age duration                      Only transfer files older than this in s or suffix ms|s|m|h|d|w|M|y (default off)
      --min-size int                          Only transfer files bigger than this in k or suffix b|k|M|G (default off)
      --modify-window duration                Max time diff to be considered the same (default 1ns)
      --no-check-certificate                  Do not verify the server SSL certificate. Insecure.
      --no-gzip-encoding                      Don't set Accept-Encoding: gzip.
      --no-traverse                           Obsolete - does nothing.
      --no-update-modtime                     Don't update destination mod-time if files identical.
  -x, --one-file-system                       Don't cross filesystem boundaries (unix/macOS only).
      --onedrive-chunk-size SizeSuffix        Chunk size to upload files with - must be multiple of 320k. (default 10M)
      --onedrive-client-id string             Microsoft App Client Id
      --onedrive-client-secret string         Microsoft App Client Secret
      --opendrive-password string             Password.
      --opendrive-username string             Username
      --pcloud-client-id string               Pcloud App Client Id
      --pcloud-client-secret string           Pcloud App Client Secret
      --qingstor-access-key-id string         QingStor Access Key ID
      --qingstor-connection-retries int       Number of connnection retries. (default 3)
      --qingstor-endpoint string              Enter a endpoint URL to connection QingStor API.
      --qingstor-env-auth                     Get QingStor credentials from runtime. Only applies if access_key_id and secret_access_key is blank.
      --qingstor-secret-access-key string     QingStor Secret Access Key (password)
      --qingstor-zone string                  Zone to connect to.
  -q, --quiet                                 Print as little stuff as possible
      --rc                                    Enable the remote control server.
      --rc-addr string                        IPaddress:Port or :Port to bind server to. (default "localhost:5572")
      --rc-cert string                        SSL PEM key (concatenation of certificate and CA certificate)
      --rc-client-ca string                   Client certificate authority to verify clients with
      --rc-htpasswd string                    htpasswd file - if not provided no authentication is done
      --rc-key string                         SSL PEM Private key
      --rc-max-header-bytes int               Maximum size of request header (default 4096)
      --rc-pass string                        Password for authentication.
      --rc-realm string                       realm for authentication (default "rclone")
      --rc-server-read-timeout duration       Timeout for server reading data (default 1h0m0s)
      --rc-server-write-timeout duration      Timeout for server writing data (default 1h0m0s)
      --rc-user string                        User name for authentication.
      --retries int                           Retry operations this many times if they fail (default 3)
      --retries-sleep duration                Interval between retrying operations if they fail, e.g 500ms, 60s, 5m. (0 to disable)
      --s3-access-key-id string               AWS Access Key ID.
      --s3-acl string                         Canned ACL used when creating buckets and/or storing objects in S3.
      --s3-chunk-size SizeSuffix              Chunk size to use for uploading (default 5M)
      --s3-disable-checksum                   Don't store MD5 checksum with object metadata
      --s3-endpoint string                    Endpoint for S3 API.
      --s3-env-auth                           Get AWS credentials from runtime (environment variables or EC2/ECS meta data if no env vars).
      --s3-force-path-style                   If true use path style access if false use virtual hosted style. (default true)
      --s3-location-constraint string         Location constraint - must be set to match the Region.
      --s3-provider string                    Choose your S3 provider.
      --s3-region string                      Region to connect to.
      --s3-secret-access-key string           AWS Secret Access Key (password)
      --s3-server-side-encryption string      The server-side encryption algorithm used when storing this object in S3.
      --s3-storage-class string               The storage class to use when storing objects in S3.
      --s3-upload-concurrency int             Concurrency for multipart uploads. (default 2)
      --sftp-ask-password                     Allow asking for SFTP password when needed.
      --sftp-disable-hashcheck                Disable the execution of SSH commands to determine if remote file hashing is available.
      --sftp-host string                      SSH host to connect to
      --sftp-key-file string                  Path to unencrypted PEM-encoded private key file, leave blank to use ssh-agent.
      --sftp-pass string                      SSH password, leave blank to use ssh-agent.
      --sftp-path-override string             Override path used by SSH connection.
      --sftp-port string                      SSH port, leave blank to use default (22)
      --sftp-set-modtime                      Set the modified time on the remote if set. (default true)
      --sftp-use-insecure-cipher              Enable the use of the aes128-cbc cipher. This cipher is insecure and may allow plaintext data to be recovered by an attacker.
      --sftp-user string                      SSH username, leave blank for current username, rclone
      --size-only                             Skip based on size only, not mod-time or checksum
      --skip-links                            Don't warn about skipped symlinks.
      --stats duration                        Interval between printing stats, e.g 500ms, 60s, 5m. (0 to disable) (default 1m0s)
      --stats-file-name-length int            Max file name length in stats. 0 for no limit (default 40)
      --stats-log-level string                Log level to show --stats output DEBUG|INFO|NOTICE|ERROR (default "INFO")
      --stats-one-line                        Make the stats fit on one line.
      --stats-unit string                     Show data rate in stats as either 'bits' or 'bytes'/s (default "bytes")
      --streaming-upload-cutoff int           Cutoff for switching to chunked upload if file size is unknown. Upload starts after reaching cutoff or when file ends. (default 100k)
      --suffix string                         Suffix for use with --backup-dir.
      --swift-auth string                     Authentication URL for server (OS_AUTH_URL).
      --swift-auth-token string               Auth Token from alternate authentication - optional (OS_AUTH_TOKEN)
      --swift-auth-version int                AuthVersion - optional - set to (1,2,3) if your auth URL has no version (ST_AUTH_VERSION)
      --swift-chunk-size SizeSuffix           Above this size files will be chunked into a _segments container. (default 5G)
      --swift-domain string                   User domain - optional (v3 auth) (OS_USER_DOMAIN_NAME)
      --swift-endpoint-type string            Endpoint type to choose from the service catalogue (OS_ENDPOINT_TYPE) (default "public")
      --swift-env-auth                        Get swift credentials from environment variables in standard OpenStack form.
      --swift-key string                      API key or password (OS_PASSWORD).
      --swift-region string                   Region name - optional (OS_REGION_NAME)
      --swift-storage-policy string           The storage policy to use when creating a new container
      --swift-storage-url string              Storage URL - optional (OS_STORAGE_URL)
      --swift-tenant string                   Tenant name - optional for v1 auth, this or tenant_id required otherwise (OS_TENANT_NAME or OS_PROJECT_NAME)
      --swift-tenant-domain string            Tenant domain - optional (v3 auth) (OS_PROJECT_DOMAIN_NAME)
      --swift-tenant-id string                Tenant ID - optional for v1 auth, this or tenant required otherwise (OS_TENANT_ID)
      --swift-user string                     User name to log in (OS_USERNAME).
      --swift-user-id string                  User ID to log in - optional - most swift systems use user and leave this blank (v3 auth) (OS_USER_ID).
      --syslog                                Use Syslog for logging
      --syslog-facility string                Facility for syslog, eg KERN,USER,... (default "DAEMON")
      --timeout duration                      IO idle timeout (default 5m0s)
      --tpslimit float                        Limit HTTP transactions per second to this.
      --tpslimit-burst int                    Max burst of transactions for --tpslimit. (default 1)
      --track-renames                         When synchronizing, track file renames and do a server side move if possible
      --transfers int                         Number of file transfers to run in parallel. (default 4)
  -u, --update                                Skip files that are newer on the destination.
      --use-server-modtime                    Use server modified time instead of object metadata
      --user-agent string                     Set the user-agent to a specified string. The default is rclone/ version (default "rclone/v1.42-112-g7c911bf2")
  -v, --verbose count                         Print lots more stuff (repeat for more)
      --webdav-bearer-token string            Bearer token instead of user/pass (eg a Macaroon)
      --webdav-pass string                    Password.
      --webdav-url string                     URL of http host to connect to
      --webdav-user string                    User name
      --webdav-vendor string                  Name of the Webdav site/service/software you are using
      --yandex-client-id string               Yandex Client Id
      --yandex-client-secret string           Yandex Client Secret

SEE ALSO

  • rclone - Sync files and directories to and from local and remote object stores - v1.42-112-g7c911bf2
Auto generated by spf13/cobra on 20-Aug-2018

Share and Enjoy.