rclone cat


rclone cat

Concatenates any files and sends them to stdout.

Synopsis

rclone cat sends any files to standard output.

You can use it like this to output a single file

rclone cat remote:path/to/file

Or like this to output any file in dir or subdirectories.

rclone cat remote:path/to/dir

Or like this to output any .txt files in dir or subdirectories.

rclone --include "*.txt" cat remote:path/to/dir

Use the --head flag to print characters only at the start, --tail for the end and --offset and --count to print a section in the middle. Note that if offset is negative it will count from the end, so --offset -1 --count 1 is equivalent to --tail 1.

rclone cat remote:path [flags]

Options

      --count int    Only print N characters. (default -1)
      --discard      Discard the output instead of printing.
      --head int     Only print the first N characters.
  -h, --help         help for cat
      --offset int   Start printing at offset N (or from end if -ve).
      --tail int     Only print the last N characters.

Options inherited from parent commands

      --acd-auth-url string                   Auth server URL.
      --acd-client-id string                  Amazon Application Client ID.
      --acd-client-secret string              Amazon Application Client Secret.
      --acd-templink-threshold SizeSuffix     Files >= this size will be downloaded via their tempLink. (default 9G)
      --acd-token-url string                  Token server url.
      --acd-upload-wait-per-gb Duration       Additional time per GB to wait after a failed complete upload to see if it appears. (default 3m0s)
      --alias-remote string                   Remote or path to alias.
      --ask-password                          Allow prompt for password for encrypted configuration. (default true)
      --auto-confirm                          If enabled, do not request console confirmation.
      --azureblob-account string              Storage Account Name (leave blank to use connection string or SAS URL)
      --azureblob-chunk-size SizeSuffix       Upload chunk size. Must fit in memory. (default 4M)
      --azureblob-endpoint string             Endpoint for the service
      --azureblob-key string                  Storage Account Key (leave blank to use connection string or SAS URL)
      --azureblob-sas-url string              SAS URL for container level access only
      --azureblob-upload-cutoff SizeSuffix    Cutoff for switching to chunked upload. (default 256M)
      --b2-account string                     Account ID or Application Key ID
      --b2-chunk-size SizeSuffix              Upload chunk size. Must fit in memory. (default 96M)
      --b2-endpoint string                    Endpoint for the service.
      --b2-hard-delete                        Permanently delete files on remote removal, otherwise hide files.
      --b2-key string                         Application Key
      --b2-test-mode string                   A flag string for X-Bz-Test-Mode header for debugging.
      --b2-upload-cutoff SizeSuffix           Cutoff for switching to chunked upload. (default 190.735M)
      --b2-versions                           Include old versions in directory listings.
      --backup-dir string                     Make backups into hierarchy based in DIR.
      --bind string                           Local address to bind to for outgoing connections, IPv4, IPv6 or name.
      --box-client-id string                  Box App Client Id.
      --box-client-secret string              Box App Client Secret
      --box-commit-retries int                Max number of times to try committing a multipart file. (default 100)
      --box-upload-cutoff SizeSuffix          Cutoff for switching to multipart upload. (default 50M)
      --buffer-size int                       In memory buffer size when reading files for each --transfer. (default 16M)
      --bwlimit BwTimetable                   Bandwidth limit in kBytes/s, or use suffix b|k|M|G or a full timetable.
      --cache-chunk-clean-interval Duration   Interval at which chunk cleanup runs (default 1m0s)
      --cache-chunk-no-memory                 Disable the in-memory cache for storing chunks during streaming
      --cache-chunk-path string               Directory to cache chunk files (default "/home/rclone/.cache/rclone/cache-backend")
      --cache-chunk-size SizeSuffix           The size of a chunk. Lower value good for slow connections but can affect seamless reading. (default 5M)
      --cache-chunk-total-size SizeSuffix     The maximum size of stored chunks. When the storage grows beyond this size, the oldest chunks will be deleted. (default 10G)
      --cache-db-path string                  Directory to cache DB (default "/home/rclone/.cache/rclone/cache-backend")
      --cache-db-purge                        Purge the cache DB before
      --cache-db-wait-time Duration           How long to wait for the DB to be available - 0 is unlimited (default 1s)
      --cache-dir string                      Directory rclone will use for caching. (default "/home/rclone/.cache/rclone")
      --cache-info-age Duration               How much time should object info (file size, file hashes etc) be stored in cache. (default 6h0m0s)
      --cache-plex-password string            The password of the Plex user
      --cache-plex-url string                 The URL of the Plex server
      --cache-plex-username string            The username of the Plex user
      --cache-read-retries int                How many times to retry a read from a cache storage (default 10)
      --cache-remote string                   Remote to cache.
      --cache-rps int                         Limits the number of requests per second to the source FS. -1 disables the rate limiter (default -1)
      --cache-tmp-upload-path string          Directory to keep temporary files until they are uploaded to the cloud storage
      --cache-tmp-wait-time Duration          How long should files be stored in local cache before being uploaded (default 15s)
      --cache-workers int                     How many workers should run in parallel to download chunks (default 4)
      --cache-writes                          Will cache file data on writes through the FS
      --checkers int                          Number of checkers to run in parallel. (default 8)
  -c, --checksum                              Skip based on checksum & size, not mod-time & size
      --config string                         Config file. (default "/home/rclone/.rclone.conf")
      --contimeout duration                   Connect timeout (default 1m0s)
  -L, --copy-links                            Follow symlinks and copy the pointed to item.
      --cpuprofile string                     Write cpu profile to file
      --crypt-directory-name-encryption       Option to either encrypt directory names or leave them intact. (default true)
      --crypt-filename-encryption string      How to encrypt the filenames. (default "standard")
      --crypt-password string                 Password or pass phrase for encryption.
      --crypt-password2 string                Password or pass phrase for salt. Optional but recommended.
      --crypt-remote string                   Remote to encrypt/decrypt.
      --crypt-show-mapping                    For all files listed show how the names encrypt.
      --delete-after                          When synchronizing, delete files on destination after transfering (default)
      --delete-before                         When synchronizing, delete files on destination before transfering
      --delete-during                         When synchronizing, delete files during transfer
      --delete-excluded                       Delete files on dest excluded from sync
      --disable string                        Disable a comma separated list of features.  Use help to see a list.
      --drive-acknowledge-abuse               Set to allow files which return cannotDownloadAbusiveFile to be downloaded.
      --drive-alternate-export                Use alternate export URLs for google documents export.
      --drive-auth-owner-only                 Only consider files owned by the authenticated user.
      --drive-chunk-size SizeSuffix           Upload chunk size. Must a power of 2 >= 256k. (default 8M)
      --drive-client-id string                Google Application Client Id
      --drive-client-secret string            Google Application Client Secret
      --drive-formats string                  Comma separated list of preferred formats for downloading Google docs. (default "docx,xlsx,pptx,svg")
      --drive-impersonate string              Impersonate this user when using a service account.
      --drive-keep-revision-forever           Keep new head revision forever.
      --drive-list-chunk int                  Size of listing chunk 100-1000. 0 to disable. (default 1000)
      --drive-root-folder-id string           ID of the root folder
      --drive-scope string                    Scope that rclone should use when requesting access from drive.
      --drive-service-account-file string     Service Account Credentials JSON file path
      --drive-shared-with-me                  Only show files that are shared with me
      --drive-skip-gdocs                      Skip google documents in all listings.
      --drive-trashed-only                    Only show files that are in the trash
      --drive-upload-cutoff SizeSuffix        Cutoff for switching to chunked upload (default 8M)
      --drive-use-created-date                Use created date instead of modified date.
      --drive-use-trash                       Send files to the trash instead of deleting permanently. (default true)
      --dropbox-chunk-size SizeSuffix         Upload chunk size. Max 150M. (default 48M)
      --dropbox-client-id string              Dropbox App Client Id
      --dropbox-client-secret string          Dropbox App Client Secret
  -n, --dry-run                               Do a trial run with no permanent changes
      --dump string                           List of items to dump from: headers,bodies,requests,responses,auth,filters,goroutines,openfiles
      --dump-bodies                           Dump HTTP headers and bodies - may contain sensitive info
      --dump-headers                          Dump HTTP bodies - may contain sensitive info
      --exclude stringArray                   Exclude files matching pattern
      --exclude-from stringArray              Read exclude patterns from file
      --exclude-if-present string             Exclude directories if filename is present
      --fast-list                             Use recursive list if available. Uses more memory but fewer transactions.
      --files-from stringArray                Read list of source-file names from file
  -f, --filter stringArray                    Add a file-filtering rule
      --filter-from stringArray               Read filtering patterns from a file
      --ftp-host string                       FTP host to connect to
      --ftp-pass string                       FTP password
      --ftp-port string                       FTP port, leave blank to use default (21)
      --ftp-user string                       FTP username, leave blank for current username,
      --gcs-bucket-acl string                 Access Control List for new buckets.
      --gcs-client-id string                  Google Application Client Id
      --gcs-client-secret string              Google Application Client Secret
      --gcs-location string                   Location for the newly created buckets.
      --gcs-object-acl string                 Access Control List for new objects.
      --gcs-project-number string             Project number.
      --gcs-service-account-file string       Service Account Credentials JSON file path
      --gcs-storage-class string              The storage class to use when storing objects in Google Cloud Storage.
      --http-url string                       URL of http host to connect to
      --hubic-client-id string                Hubic Client Id
      --hubic-client-secret string            Hubic Client Secret
      --ignore-checksum                       Skip post copy check of checksums.
      --ignore-errors                         delete even if there are I/O errors
      --ignore-existing                       Skip all files that exist on destination
      --ignore-size                           Ignore size when skipping use mod-time or checksum.
  -I, --ignore-times                          Don't skip files that match size and time - transfer all files
      --immutable                             Do not modify files. Fail if existing files have been modified.
      --include stringArray                   Include files matching pattern
      --include-from stringArray              Read include patterns from file
      --jottacloud-mountpoint string          The mountpoint to use.
      --jottacloud-pass string                Password.
      --jottacloud-user string                User Name
      --local-no-check-updated                Don't check to see if the files change during upload
      --local-no-unicode-normalization        Don't apply unicode normalization to paths and filenames
      --local-nounc string                    Disable UNC (long path names) conversion on Windows
      --log-file string                       Log everything to this file
      --log-level string                      Log level DEBUG|INFO|NOTICE|ERROR (default "NOTICE")
      --low-level-retries int                 Number of low level retries to do. (default 10)
      --max-age duration                      Only transfer files younger than this in s or suffix ms|s|m|h|d|w|M|y (default off)
      --max-backlog int                       Maximum number of objects in sync or check backlog. (default 10000)
      --max-delete int                        When synchronizing, limit the number of deletes (default -1)
      --max-depth int                         If set limits the recursion depth to this. (default -1)
      --max-size int                          Only transfer files smaller than this in k or suffix b|k|M|G (default off)
      --max-transfer int                      Maximum size of data to transfer. (default off)
      --mega-debug                            Output more debug from Mega.
      --mega-hard-delete                      Delete files permanently rather than putting them into the trash.
      --mega-pass string                      Password.
      --mega-user string                      User name
      --memprofile string                     Write memory profile to file
      --min-age duration                      Only transfer files older than this in s or suffix ms|s|m|h|d|w|M|y (default off)
      --min-size int                          Only transfer files bigger than this in k or suffix b|k|M|G (default off)
      --modify-window duration                Max time diff to be considered the same (default 1ns)
      --no-check-certificate                  Do not verify the server SSL certificate. Insecure.
      --no-gzip-encoding                      Don't set Accept-Encoding: gzip.
      --no-traverse                           Obsolete - does nothing.
      --no-update-modtime                     Don't update destination mod-time if files identical.
  -x, --one-file-system                       Don't cross filesystem boundaries (unix/macOS only).
      --onedrive-chunk-size SizeSuffix        Chunk size to upload files with - must be multiple of 320k. (default 10M)
      --onedrive-client-id string             Microsoft App Client Id
      --onedrive-client-secret string         Microsoft App Client Secret
      --opendrive-password string             Password.
      --opendrive-username string             Username
      --pcloud-client-id string               Pcloud App Client Id
      --pcloud-client-secret string           Pcloud App Client Secret
      --qingstor-access-key-id string         QingStor Access Key ID
      --qingstor-connection-retries int       Number of connnection retries. (default 3)
      --qingstor-endpoint string              Enter a endpoint URL to connection QingStor API.
      --qingstor-env-auth                     Get QingStor credentials from runtime. Only applies if access_key_id and secret_access_key is blank.
      --qingstor-secret-access-key string     QingStor Secret Access Key (password)
      --qingstor-zone string                  Zone to connect to.
  -q, --quiet                                 Print as little stuff as possible
      --rc                                    Enable the remote control server.
      --rc-addr string                        IPaddress:Port or :Port to bind server to. (default "localhost:5572")
      --rc-cert string                        SSL PEM key (concatenation of certificate and CA certificate)
      --rc-client-ca string                   Client certificate authority to verify clients with
      --rc-htpasswd string                    htpasswd file - if not provided no authentication is done
      --rc-key string                         SSL PEM Private key
      --rc-max-header-bytes int               Maximum size of request header (default 4096)
      --rc-pass string                        Password for authentication.
      --rc-realm string                       realm for authentication (default "rclone")
      --rc-server-read-timeout duration       Timeout for server reading data (default 1h0m0s)
      --rc-server-write-timeout duration      Timeout for server writing data (default 1h0m0s)
      --rc-user string                        User name for authentication.
      --retries int                           Retry operations this many times if they fail (default 3)
      --retries-sleep duration                Interval between retrying operations if they fail, e.g 500ms, 60s, 5m. (0 to disable)
      --s3-access-key-id string               AWS Access Key ID.
      --s3-acl string                         Canned ACL used when creating buckets and/or storing objects in S3.
      --s3-chunk-size SizeSuffix              Chunk size to use for uploading (default 5M)
      --s3-disable-checksum                   Don't store MD5 checksum with object metadata
      --s3-endpoint string                    Endpoint for S3 API.
      --s3-env-auth                           Get AWS credentials from runtime (environment variables or EC2/ECS meta data if no env vars).
      --s3-force-path-style                   If true use path style access if false use virtual hosted style. (default true)
      --s3-location-constraint string         Location constraint - must be set to match the Region.
      --s3-provider string                    Choose your S3 provider.
      --s3-region string                      Region to connect to.
      --s3-secret-access-key string           AWS Secret Access Key (password)
      --s3-server-side-encryption string      The server-side encryption algorithm used when storing this object in S3.
      --s3-storage-class string               The storage class to use when storing objects in S3.
      --s3-upload-concurrency int             Concurrency for multipart uploads. (default 2)
      --sftp-ask-password                     Allow asking for SFTP password when needed.
      --sftp-disable-hashcheck                Disable the execution of SSH commands to determine if remote file hashing is available.
      --sftp-host string                      SSH host to connect to
      --sftp-key-file string                  Path to unencrypted PEM-encoded private key file, leave blank to use ssh-agent.
      --sftp-pass string                      SSH password, leave blank to use ssh-agent.
      --sftp-path-override string             Override path used by SSH connection.
      --sftp-port string                      SSH port, leave blank to use default (22)
      --sftp-set-modtime                      Set the modified time on the remote if set. (default true)
      --sftp-use-insecure-cipher              Enable the use of the aes128-cbc cipher. This cipher is insecure and may allow plaintext data to be recovered by an attacker.
      --sftp-user string                      SSH username, leave blank for current username, rclone
      --size-only                             Skip based on size only, not mod-time or checksum
      --skip-links                            Don't warn about skipped symlinks.
      --stats duration                        Interval between printing stats, e.g 500ms, 60s, 5m. (0 to disable) (default 1m0s)
      --stats-file-name-length int            Max file name length in stats. 0 for no limit (default 40)
      --stats-log-level string                Log level to show --stats output DEBUG|INFO|NOTICE|ERROR (default "INFO")
      --stats-one-line                        Make the stats fit on one line.
      --stats-unit string                     Show data rate in stats as either 'bits' or 'bytes'/s (default "bytes")
      --streaming-upload-cutoff int           Cutoff for switching to chunked upload if file size is unknown. Upload starts after reaching cutoff or when file ends. (default 100k)
      --suffix string                         Suffix for use with --backup-dir.
      --swift-auth string                     Authentication URL for server (OS_AUTH_URL).
      --swift-auth-token string               Auth Token from alternate authentication - optional (OS_AUTH_TOKEN)
      --swift-auth-version int                AuthVersion - optional - set to (1,2,3) if your auth URL has no version (ST_AUTH_VERSION)
      --swift-chunk-size SizeSuffix           Above this size files will be chunked into a _segments container. (default 5G)
      --swift-domain string                   User domain - optional (v3 auth) (OS_USER_DOMAIN_NAME)
      --swift-endpoint-type string            Endpoint type to choose from the service catalogue (OS_ENDPOINT_TYPE) (default "public")
      --swift-env-auth                        Get swift credentials from environment variables in standard OpenStack form.
      --swift-key string                      API key or password (OS_PASSWORD).
      --swift-region string                   Region name - optional (OS_REGION_NAME)
      --swift-storage-policy string           The storage policy to use when creating a new container
      --swift-storage-url string              Storage URL - optional (OS_STORAGE_URL)
      --swift-tenant string                   Tenant name - optional for v1 auth, this or tenant_id required otherwise (OS_TENANT_NAME or OS_PROJECT_NAME)
      --swift-tenant-domain string            Tenant domain - optional (v3 auth) (OS_PROJECT_DOMAIN_NAME)
      --swift-tenant-id string                Tenant ID - optional for v1 auth, this or tenant required otherwise (OS_TENANT_ID)
      --swift-user string                     User name to log in (OS_USERNAME).
      --swift-user-id string                  User ID to log in - optional - most swift systems use user and leave this blank (v3 auth) (OS_USER_ID).
      --syslog                                Use Syslog for logging
      --syslog-facility string                Facility for syslog, eg KERN,USER,... (default "DAEMON")
      --timeout duration                      IO idle timeout (default 5m0s)
      --tpslimit float                        Limit HTTP transactions per second to this.
      --tpslimit-burst int                    Max burst of transactions for --tpslimit. (default 1)
      --track-renames                         When synchronizing, track file renames and do a server side move if possible
      --transfers int                         Number of file transfers to run in parallel. (default 4)
  -u, --update                                Skip files that are newer on the destination.
      --use-server-modtime                    Use server modified time instead of object metadata
      --user-agent string                     Set the user-agent to a specified string. The default is rclone/ version (default "rclone/v1.42-112-g7c911bf2")
  -v, --verbose count                         Print lots more stuff (repeat for more)
      --webdav-bearer-token string            Bearer token instead of user/pass (eg a Macaroon)
      --webdav-pass string                    Password.
      --webdav-url string                     URL of http host to connect to
      --webdav-user string                    User name
      --webdav-vendor string                  Name of the Webdav site/service/software you are using
      --yandex-client-id string               Yandex Client Id
      --yandex-client-secret string           Yandex Client Secret

SEE ALSO

  • rclone - Sync files and directories to and from local and remote object stores - v1.42-112-g7c911bf2
Auto generated by spf13/cobra on 20-Aug-2018

Share and Enjoy.