Version v1.53.0

This commit is contained in:
Nick Craig-Wood 2020-09-02 16:59:04 +01:00
parent 358e2b2665
commit 510ac341e1
51 changed files with 39713 additions and 35620 deletions

19840
MANUAL.html generated

File diff suppressed because one or more lines are too long

3375
MANUAL.md generated

File diff suppressed because it is too large Load Diff

21661
MANUAL.txt generated

File diff suppressed because it is too large Load Diff

View File

@ -165,7 +165,8 @@ Here are the standard options specific to amazon cloud drive (Amazon Drive).
#### --acd-client-id
Amazon Application Client ID.
OAuth Client Id
Leave blank normally.
- Config: client_id
- Env Var: RCLONE_ACD_CLIENT_ID
@ -174,7 +175,8 @@ Amazon Application Client ID.
#### --acd-client-secret
Amazon Application Client Secret.
OAuth Client Secret
Leave blank normally.
- Config: client_secret
- Env Var: RCLONE_ACD_CLIENT_SECRET
@ -185,10 +187,19 @@ Amazon Application Client Secret.
Here are the advanced options specific to amazon cloud drive (Amazon Drive).
#### --acd-token
OAuth Access Token as a JSON blob.
- Config: token
- Env Var: RCLONE_ACD_TOKEN
- Type: string
- Default: ""
#### --acd-auth-url
Auth server URL.
Leave blank to use Amazon's.
Leave blank to use the provider defaults.
- Config: auth_url
- Env Var: RCLONE_ACD_AUTH_URL
@ -198,7 +209,7 @@ Leave blank to use Amazon's.
#### --acd-token-url
Token server url.
leave blank to use Amazon's.
Leave blank to use the provider defaults.
- Config: token_url
- Env Var: RCLONE_ACD_TOKEN_URL

View File

@ -413,6 +413,20 @@ This value should be set no larger than 4.657GiB (== 5GB).
- Type: SizeSuffix
- Default: 200M
#### --b2-copy-cutoff
Cutoff for switching to multipart copy
Any files larger than this that need to be server side copied will be
copied in chunks of this size.
The minimum is 0 and the maximum is 4.6GB.
- Config: copy_cutoff
- Env Var: RCLONE_B2_COPY_CUTOFF
- Type: SizeSuffix
- Default: 4G
#### --b2-chunk-size
Upload chunk size. Must fit in memory.
@ -467,6 +481,26 @@ The minimum value is 1 second. The maximum value is one week.
- Type: Duration
- Default: 1w
#### --b2-memory-pool-flush-time
How often internal memory buffer pools will be flushed.
Uploads which requires additional buffers (f.e multipart) will use memory pool for allocations.
This option controls how often unused buffers will be removed from the pool.
- Config: memory_pool_flush_time
- Env Var: RCLONE_B2_MEMORY_POOL_FLUSH_TIME
- Type: Duration
- Default: 1m0s
#### --b2-memory-pool-use-mmap
Whether to use mmap buffers in internal memory pool.
- Config: memory_pool_use_mmap
- Env Var: RCLONE_B2_MEMORY_POOL_USE_MMAP
- Type: bool
- Default: false
#### --b2-encoding
This sets the encoding for the backend.

View File

@ -270,7 +270,7 @@ Here are the standard options specific to box (Box).
#### --box-client-id
Box App Client Id.
OAuth Client Id
Leave blank normally.
- Config: client_id
@ -280,7 +280,7 @@ Leave blank normally.
#### --box-client-secret
Box App Client Secret
OAuth Client Secret
Leave blank normally.
- Config: client_secret
@ -293,11 +293,24 @@ Leave blank normally.
Box App config.json location
Leave blank normally.
Leading `~` will be expanded in the file name as will environment variables such as `${RCLONE_CONFIG_DIR}`.
- Config: box_config_file
- Env Var: RCLONE_BOX_BOX_CONFIG_FILE
- Type: string
- Default: ""
#### --box-access-token
Box App Primary Access Token
Leave blank normally.
- Config: access_token
- Env Var: RCLONE_BOX_ACCESS_TOKEN
- Type: string
- Default: ""
#### --box-box-sub-type
@ -316,6 +329,35 @@ Leave blank normally.
Here are the advanced options specific to box (Box).
#### --box-token
OAuth Access Token as a JSON blob.
- Config: token
- Env Var: RCLONE_BOX_TOKEN
- Type: string
- Default: ""
#### --box-auth-url
Auth server URL.
Leave blank to use the provider defaults.
- Config: auth_url
- Env Var: RCLONE_BOX_AUTH_URL
- Type: string
- Default: ""
#### --box-token-url
Token server url.
Leave blank to use the provider defaults.
- Config: token_url
- Env Var: RCLONE_BOX_TOKEN_URL
- Type: string
- Default: ""
#### --box-root-folder-id
Fill in for rclone to use a non root folder as its starting point.

View File

@ -343,6 +343,8 @@ The username of the Plex user
The password of the Plex user
**NB** Input to this must be obscured - see [rclone obscure](/commands/rclone_obscure/).
- Config: plex_password
- Env Var: RCLONE_CACHE_PLEX_PASSWORD
- Type: string

View File

@ -5,6 +5,183 @@ description: "Rclone Changelog"
# Changelog
## v1.53.0 - 2020-09-02
[See commits](https://github.com/rclone/rclone/compare/v1.52.0...v1.53.0)
* New Features
* The [VFS layer](/commands/rclone_mount/#vfs-virtual-file-system) was heavily reworked for this release - see below for more details
* Interactive mode [-i/--interactive](/docs/#interactive) for destructive operations (fishbullet)
* Add [--bwlimit-file](/docs/#bwlimit-file-bandwidth-spec) flag to limit speeds of individual file transfers (Nick Craig-Wood)
* Transfers are sorted by start time in the stats and progress output (Max Sum)
* Make sure backends expand `~` and environment vars in file names they use (Nick Craig-Wood)
* Add [--refresh-times](/docs/#refresh-times) flag to set modtimes on hashless backends (Nick Craig-Wood)
* build
* Remove vendor directory in favour of Go modules (Nick Craig-Wood)
* Build with go1.15.x by default (Nick Craig-Wood)
* Drop macOS 386 build as it is no longer supported by go1.15 (Nick Craig-Wood)
* Add ARMv7 to the supported builds (Nick Craig-Wood)
* Enable `rclone cmount` on macOS (Nick Craig-Wood)
* Make rclone build with gccgo (Nick Craig-Wood)
* Make rclone build with wasm (Nick Craig-Wood)
* Change beta numbering to be semver compatible (Nick Craig-Wood)
* Add file properties and icon to Windows executable (albertony)
* Add experimental interface for integrating rclone into browsers (Nick Craig-Wood)
* lib: Add file name compression (Klaus Post)
* rc
* Allow installation and use of plugins and test plugins with rclone-webui (Chaitanya Bankanhal)
* Add reverse proxy pluginsHandler for serving plugins (Chaitanya Bankanhal)
* Add `mount/listmounts` option for listing current mounts (Chaitanya Bankanhal)
* Add `operations/uploadfile` to upload a file through rc using encoding multipart/form-data (Chaitanya Bankanhal)
* Add `core/copmmand` to execute rclone terminal commands. (Chaitanya Bankanhal)
* `rclone check`
* Add reporting of filenames for same/missing/changed (Nick Craig-Wood)
* Make check command obey `--dry-run`/`-i`/`--interactive` (Nick Craig-Wood)
* Make check do `--checkers` files concurrently (Nick Craig-Wood)
* Retry downloads if they fail when using the `--download` flag (Nick Craig-Wood)
* Make it show stats by default (Nick Craig-Wood)
* `rclone obscure`: Allow obscure command to accept password on STDIN (David Ibarra)
* `rclone config`
* Set RCLONE_CONFIG_DIR for use in config files and subprocesses (Nick Craig-Wood)
* Reject remote names starting with a dash. (jtagcat)
* `rclone cryptcheck`: Add reporting of filenames for same/missing/changed (Nick Craig-Wood)
* `rclone dedupe`: Make it obey the `--size-only` flag for duplicate detection (Nick Craig-Wood)
* `rclone link`: Add `--expire` and `--unlink` flags (Roman Kredentser)
* `rclone mkdir`: Warn when using mkdir on remotes which can't have empty directories (Nick Craig-Wood)
* `rclone rc`: Allow JSON parameters to simplify command line usage (Nick Craig-Wood)
* `rclone serve ftp`
* Don't compile on < go1.13 after dependency update (Nick Craig-Wood)
* Add error message if auth proxy fails (Nick Craig-Wood)
* Use refactored goftp.io/server library for binary shrink (Nick Craig-Wood)
* `rclone serve restic`: Expose interfaces so that rclone can be used as a library from within restic (Jack)
* `rclone sync`: Add `--track-renames-strategy leaf` (Nick Craig-Wood)
* `rclone touch`: Add ability to set nanosecond resolution times (Nick Craig-Wood)
* `rclone tree`: Remove `-i` shorthand for `--noindent` as it conflicts with `-i`/`--interactive` (Nick Craig-Wood)
* Bug Fixes
* accounting
* Fix documentation for `speed`/`speedAvg` (Nick Craig-Wood)
* Fix elapsed time not show actual time since beginning (Chaitanya Bankanhal)
* Fix deadlock in stats printing (Nick Craig-Wood)
* build
* Fix file handle leak in GitHub release tool (Garrett Squire)
* `rclone check`: Fix successful retries with `--download` counting errors (Nick Craig-Wood)
* `rclone dedupe`: Fix logging to be easier to understand (Nick Craig-Wood)
* Mount
* Warn macOS users that mount implementation is changing (Nick Craig-Wood)
* to test the new implementation use `rclone cmount` instead of `rclone mount`
* this is because the library rclone uses has dropped macOS support
* rc interface
* Add call for unmount all (Chaitanya Bankanhal)
* Make `mount/mount` remote control take vfsOpt option (Nick Craig-Wood)
* Add mountOpt to `mount/mount` (Nick Craig-Wood)
* Add VFS and Mount options to `mount/listmounts` (Nick Craig-Wood)
* Catch panics in cgofuse initialization and turn into error messages (Nick Craig-Wood)
* Always supply stat information in Readdir (Nick Craig-Wood)
* Add support for reading unknown length files using direct IO (Windows) (Nick Craig-Wood)
* Fix On Windows don't add `-o uid/gid=-1` if user supplies `-o uid/gid`. (Nick Craig-Wood)
* Fix macOS losing directory contents in cmount (Nick Craig-Wood)
* Fix volume name broken in recent refactor (Nick Craig-Wood)
* VFS
* Implement partial reads for `--vfs-cache-mode full` (Nick Craig-Wood)
* Add `--vfs-writeback` option to delay writes back to cloud storage (Nick Craig-Wood)
* Add `--vfs-read-ahead` parameter for use with `--vfs-cache-mode full` (Nick Craig-Wood)
* Restart pending uploads on restart of the cache (Nick Craig-Wood)
* Support synchronous cache space recovery upon ENOSPC (Leo Luan)
* Allow ReadAt and WriteAt to run concurrently with themselves (Nick Craig-Wood)
* Change modtime of file before upload to current (Rob Calistri)
* Recommend `--vfs-cache-modes writes` on backends which can't stream (Nick Craig-Wood)
* Add an optional `fs` parameter to vfs rc methods (Nick Craig-Wood)
* Fix errors when using > 260 char files in the cache in Windows (Nick Craig-Wood)
* Fix renaming of items while they are being uploaded (Nick Craig-Wood)
* Fix very high load caused by slow directory listings (Nick Craig-Wood)
* Fix renamed files not being uploaded with `--vfs-cache-mode minimal` (Nick Craig-Wood)
* Fix directory locking caused by slow directory listings (Nick Craig-Wood)
* Fix saving from chrome without `--vfs-cache-mode writes` (Nick Craig-Wood)
* Local
* Add `--local-no-updated` to provide a consistent view of changing objects (Nick Craig-Wood)
* Add `--local-no-set-modtime` option to prevent modtime changes (tyhuber1)
* Fix race conditions updating and reading Object metadata (Nick Craig-Wood)
* Cache
* Make any created backends be cached to fix rc problems (Nick Craig-Wood)
* Fix dedupe on caches wrapping drives (Nick Craig-Wood)
* Crypt
* Add `--crypt-server-side-across-configs` flag (Nick Craig-Wood)
* Make any created backends be cached to fix rc problems (Nick Craig-Wood)
* Alias
* Make any created backends be cached to fix rc problems (Nick Craig-Wood)
* Azure Blob
* Don't compile on < go1.13 after dependency update (Nick Craig-Wood)
* B2
* Implement server side copy for files > 5GB (Nick Craig-Wood)
* Cancel in progress multipart uploads and copies on rclone exit (Nick Craig-Wood)
* Note that b2's encoding now allows \ but rclone's hasn't changed (Nick Craig-Wood)
* Fix transfers when using download_url (Nick Craig-Wood)
* Box
* Implement rclone cleanup (buengese)
* Cancel in progress multipart uploads and copies on rclone exit (Nick Craig-Wood)
* Allow authentication with access token (David)
* Chunker
* Make any created backends be cached to fix rc problems (Nick Craig-Wood)
* Drive
* Add `rclone backend drives` to list shared drives (teamdrives) (Nick Craig-Wood)
* Implement `rclone backend untrash` (Nick Craig-Wood)
* Work around drive bug which didn't set modtime of copied docs (Nick Craig-Wood)
* Added `--drive-starred-only` to only show starred files (Jay McEntire)
* Deprecate `--drive-alternate-export` as it is no longer needed (themylogin)
* Fix duplication of Google docs on server side copy (Nick Craig-Wood)
* Fix "panic: send on closed channel" when recycling dir entries (Nick Craig-Wood)
* Dropbox
* Add copyright detector info in limitations section in the docs (Alex Guerrero)
* Fix `rclone link` by removing expires parameter (Nick Craig-Wood)
* Fichier
* Detect Flood detected: IP Locked error and sleep for 30s (Nick Craig-Wood)
* FTP
* Add explicit TLS support (Heiko Bornholdt)
* Add support for `--dump bodies` and `--dump auth` for debugging (Nick Craig-Wood)
* Fix interoperation with pure-ftpd (Nick Craig-Wood)
* Google Cloud Storage
* Add support for anonymous access (Kai Lüke)
* Jottacloud
* Bring back legacy authentification for use with whitelabel versions (buengese)
* Switch to new api root - also implement a very ugly workaround for the DirMove failures (buengese)
* Onedrive
* Rework cancel of multipart uploads on rclone exit (Nick Craig-Wood)
* Implement rclone cleanup (Nick Craig-Wood)
* Add `--onedrive-no-versions` flag to remove old versions (Nick Craig-Wood)
* Pcloud
* Implement `rclone link` for public link creation (buengese)
* Qingstor
* Cancel in progress multipart uploads on rclone exit (Nick Craig-Wood)
* S3
* Preserve metadata when doing multipart copy (Nick Craig-Wood)
* Cancel in progress multipart uploads and copies on rclone exit (Nick Craig-Wood)
* Add `rclone link` for public link sharing (Roman Kredentser)
* Add `rclone backend restore` command to restore objects from GLACIER (Nick Craig-Wood)
* Add `rclone cleanup` and `rclone backend cleanup` to clean unfinished multipart uploads (Nick Craig-Wood)
* Add `rclone backend list-multipart-uploads` to list unfinished multipart uploads (Nick Craig-Wood)
* Add `--s3-max-upload-parts` support (Kamil Trzciński)
* Add `--s3-no-check-bucket` for minimising rclone transactions and perms (Nick Craig-Wood)
* Add `--s3-profile` and `--s3-shared-credentials-file` options (Nick Craig-Wood)
* Use regional s3 us-east-1 endpoint (David)
* Add Scaleway provider (Vincent Feltz)
* Update IBM COS endpoints (Egor Margineanu)
* Reduce the default `--s3-copy-cutoff` to < 5GB for Backblaze S3 compatibility (Nick Craig-Wood)
* Fix detection of bucket existing (Nick Craig-Wood)
* SFTP
* Use the absolute path instead of the relative path for listing for improved compatibility (Nick Craig-Wood)
* Add `--sftp-subsystem` and `--sftp-server-command` options (aus)
* Swift
* Fix dangling large objects breaking the listing (Nick Craig-Wood)
* Fix purge not deleting directory markers (Nick Craig-Wood)
* Fix update multipart object removing all of its own parts (Nick Craig-Wood)
* Fix missing hash from object returned from upload (Nick Craig-Wood)
* Tardigrade
* Upgrade to uplink v1.2.0 (Kaloyan Raev)
* Union
* Fix writing with the all policy (Nick Craig-Wood)
* WebDAV
* Fix directory creation with 4shared (Nick Craig-Wood)
## v1.52.3 - 2020-08-07
[See commits](https://github.com/rclone/rclone/compare/v1.52.2...v1.52.3)

View File

@ -46,7 +46,7 @@ See the [global flags page](/flags/) for global options not listed here.
* [rclone copyurl](/commands/rclone_copyurl/) - Copy url content to dest.
* [rclone cryptcheck](/commands/rclone_cryptcheck/) - Cryptcheck checks the integrity of a crypted remote.
* [rclone cryptdecode](/commands/rclone_cryptdecode/) - Cryptdecode returns unencrypted file names.
* [rclone dedupe](/commands/rclone_dedupe/) - Interactively find duplicate files and delete/rename them.
* [rclone dedupe](/commands/rclone_dedupe/) - Interactively find duplicate filenames and delete/rename them.
* [rclone delete](/commands/rclone_delete/) - Remove the contents of path.
* [rclone deletefile](/commands/rclone_deletefile/) - Remove a single file from remote.
* [rclone genautocomplete](/commands/rclone_genautocomplete/) - Output completion script for a given shell.
@ -65,7 +65,7 @@ See the [global flags page](/flags/) for global options not listed here.
* [rclone move](/commands/rclone_move/) - Move files from source to dest.
* [rclone moveto](/commands/rclone_moveto/) - Move file or directory from source to dest.
* [rclone ncdu](/commands/rclone_ncdu/) - Explore a remote with a text based user interface.
* [rclone obscure](/commands/rclone_obscure/) - Obscure password for use in the rclone.conf
* [rclone obscure](/commands/rclone_obscure/) - Obscure password for use in the rclone config file
* [rclone purge](/commands/rclone_purge/) - Remove the path and all of its contents.
* [rclone rc](/commands/rclone_rc/) - Run a command against a running rclone.
* [rclone rcat](/commands/rclone_rcat/) - Copies standard input to file on remote.

View File

@ -24,9 +24,26 @@ both remotes and check them against each other on the fly. This can
be useful for remotes that don't support hashes or if you really want
to check all the data.
If you supply the --one-way flag, it will only check that files in source
match the files in destination, not the other way around. Meaning extra files in
destination that are not in the source will not trigger an error.
If you supply the `--one-way` flag, it will only check that files in
the source match the files in the destination, not the other way
around. This means that extra files in the destination that are not in
the source will not be detected.
The `--differ`, `--missing-on-dst`, `--missing-on-src`, `--src-only`
and `--error` flags write paths, one per line, to the file name (or
stdout if it is `-`) supplied. What they write is described in the
help below. For example `--differ` will write all paths which are
present on both the source and destination but different.
The `--combined` flag will write a file (or stdout) which contains all
file paths with a symbol and then a space and then the path to tell
you what happened to it. These are reminiscent of diff files.
- `= path` means path was found in source and destination and was identical
- `- path` means path was missing on the source, so only in the destination
- `+ path` means path was missing on the destination, so only in the source
- `* path` means path was present in source and destination but different.
- `! path` means there was an error reading or hashing the source or dest.
```
@ -36,9 +53,14 @@ rclone check source:path dest:path [flags]
## Options
```
--download Check by downloading rather than with hash.
-h, --help help for check
--one-way Check one way only, source files must exist on remote
--combined string Make a combined report of changes to this file
--differ string Report all non-matching files to this file
--error string Report all files with errors (hashing or reading) to this file
-h, --help help for check
--match string Report all matching files to this file
--missing-on-dst string Report all files missing from the destination to this file
--missing-on-src string Report all files missing from the source to this file
--one-way Check one way only, source files must exist on remote
```
See the [global flags page](/flags/) for global options not listed here.

View File

@ -59,7 +59,9 @@ recently very efficiently like this:
rclone copy --max-age 24h --no-traverse /path/to/src remote:
**Note**: Use the `-P`/`--progress` flag to view real-time transfer statistics
**Note**: Use the `-P`/`--progress` flag to view real-time transfer statistics.
**Note**: Use the `--dry-run` or the `--interactive`/`-i` flag to test without copying anything.
```

View File

@ -35,9 +35,26 @@ the files in remote:path.
After it has run it will log the status of the encryptedremote:.
If you supply the --one-way flag, it will only check that files in source
match the files in destination, not the other way around. Meaning extra files in
destination that are not in the source will not trigger an error.
If you supply the `--one-way` flag, it will only check that files in
the source match the files in the destination, not the other way
around. This means that extra files in the destination that are not in
the source will not be detected.
The `--differ`, `--missing-on-dst`, `--missing-on-src`, `--src-only`
and `--error` flags write paths, one per line, to the file name (or
stdout if it is `-`) supplied. What they write is described in the
help below. For example `--differ` will write all paths which are
present on both the source and destination but different.
The `--combined` flag will write a file (or stdout) which contains all
file paths with a symbol and then a space and then the path to tell
you what happened to it. These are reminiscent of diff files.
- `= path` means path was found in source and destination and was identical
- `- path` means path was missing on the source, so only in the destination
- `+ path` means path was missing on the destination, so only in the source
- `* path` means path was present in source and destination but different.
- `! path` means there was an error reading or hashing the source or dest.
```
@ -47,8 +64,14 @@ rclone cryptcheck remote:path cryptedremote:path [flags]
## Options
```
-h, --help help for cryptcheck
--one-way Check one way only, source files must exist on destination
--combined string Make a combined report of changes to this file
--differ string Report all non-matching files to this file
--error string Report all files with errors (hashing or reading) to this file
-h, --help help for cryptcheck
--match string Report all matching files to this file
--missing-on-dst string Report all files missing from the destination to this file
--missing-on-src string Report all files missing from the source to this file
--one-way Check one way only, source files must exist on remote
```
See the [global flags page](/flags/) for global options not listed here.

View File

@ -1,29 +1,44 @@
---
title: "rclone dedupe"
description: "Interactively find duplicate files and delete/rename them."
description: "Interactively find duplicate filenames and delete/rename them."
slug: rclone_dedupe
url: /commands/rclone_dedupe/
# autogenerated - DO NOT EDIT, instead edit the source code in cmd/dedupe/ and as part of making a release run "make commanddocs"
---
# rclone dedupe
Interactively find duplicate files and delete/rename them.
Interactively find duplicate filenames and delete/rename them.
## Synopsis
By default `dedupe` interactively finds duplicate files and offers to
delete all but one or rename them to be different. Only useful with
Google Drive which can have duplicate file names.
By default `dedupe` interactively finds files with duplicate
names and offers to delete all but one or rename them to be
different.
This is only useful with backends like Google Drive which can have
duplicate file names. It can be run on wrapping backends (eg crypt) if
they wrap a backend which supports duplicate file names.
In the first pass it will merge directories with the same name. It
will do this iteratively until all the identical directories have been
merged.
will do this iteratively until all the identically named directories
have been merged.
The `dedupe` command will delete all but one of any identical (same
md5sum) files it finds without confirmation. This means that for most
duplicated files the `dedupe` command will not be interactive. You
can use `--dry-run` to see what would happen without doing anything.
In the second pass, for every group of duplicate file names, it will
delete all but one identical files it finds without confirmation.
This means that for most duplicated files the `dedupe`
command will not be interactive.
`dedupe` considers files to be identical if they have the
same hash. If the backend does not support hashes (eg crypt wrapping
Google Drive) then they will never be found to be identical. If you
use the `--size-only` flag then files will be considered
identical if they have the same size (any hash will be ignored). This
can be useful on crypt backends which do not support hashes.
**Important**: Since this can cause data loss, test first with the
`--dry-run` or the `--interactive`/`-i` flag.
Here is an example run.
@ -42,22 +57,22 @@ Now the `dedupe` session
$ rclone dedupe drive:dupes
2016/03/05 16:24:37 Google drive root 'dupes': Looking for duplicates using interactive mode.
one.txt: Found 4 duplicates - deleting identical copies
one.txt: Deleting 2/3 identical duplicates (md5sum "1eedaa9fe86fd4b8632e2ac549403b36")
one.txt: Found 4 files with duplicate names
one.txt: Deleting 2/3 identical duplicates (MD5 "1eedaa9fe86fd4b8632e2ac549403b36")
one.txt: 2 duplicates remain
1: 6048320 bytes, 2016-03-05 16:23:16.798000000, md5sum 1eedaa9fe86fd4b8632e2ac549403b36
2: 564374 bytes, 2016-03-05 16:23:06.731000000, md5sum 7594e7dc9fc28f727c42ee3e0749de81
1: 6048320 bytes, 2016-03-05 16:23:16.798000000, MD5 1eedaa9fe86fd4b8632e2ac549403b36
2: 564374 bytes, 2016-03-05 16:23:06.731000000, MD5 7594e7dc9fc28f727c42ee3e0749de81
s) Skip and do nothing
k) Keep just one (choose which in next step)
r) Rename all to be different (by changing file.jpg to file-1.jpg)
s/k/r> k
Enter the number of the file to keep> 1
one.txt: Deleted 1 extra copies
two.txt: Found 3 duplicates - deleting identical copies
two.txt: Found 3 files with duplicates names
two.txt: 3 duplicates remain
1: 564374 bytes, 2016-03-05 16:22:52.118000000, md5sum 7594e7dc9fc28f727c42ee3e0749de81
2: 6048320 bytes, 2016-03-05 16:22:46.185000000, md5sum 1eedaa9fe86fd4b8632e2ac549403b36
3: 1744073 bytes, 2016-03-05 16:22:38.104000000, md5sum 851957f7fb6f0bc4ce76be966d336802
1: 564374 bytes, 2016-03-05 16:22:52.118000000, MD5 7594e7dc9fc28f727c42ee3e0749de81
2: 6048320 bytes, 2016-03-05 16:22:46.185000000, MD5 1eedaa9fe86fd4b8632e2ac549403b36
3: 1744073 bytes, 2016-03-05 16:22:38.104000000, MD5 851957f7fb6f0bc4ce76be966d336802
s) Skip and do nothing
k) Keep just one (choose which in next step)
r) Rename all to be different (by changing file.jpg to file-1.jpg)

View File

@ -35,6 +35,9 @@ Then delete
That reads "delete everything with a minimum size of 100 MB", hence
delete all files bigger than 100MBytes.
**Important**: Since this can cause data loss, test first with the
`--dry-run` or the `--interactive`/`-i` flag.
```
rclone delete remote:path [flags]

View File

@ -11,16 +11,27 @@ Generate public link to file/folder.
## Synopsis
rclone link will create or retrieve a public link to the given file or folder.
rclone link will create, retrieve or remove a public link to the given
file or folder.
rclone link remote:path/to/file
rclone link remote:path/to/folder/
rclone link --unlink remote:path/to/folder/
rclone link --expire 1d remote:path/to/file
If successful, the last line of the output will contain the link. Exact
capabilities depend on the remote, but the link will always be created with
the least constraints e.g. no expiry, no password protection, accessible
without account.
If you supply the --expire flag, it will set the expiration time
otherwise it will use the default (100 years). **Note** not all
backends support the --expire flag - if the backend doesn't support it
then the link returned won't expire.
Use the --unlink flag to remove existing public links to the file or
folder. **Note** not all backends support "--unlink" flag - those that
don't will just ignore it.
If successful, the last line of the output will contain the
link. Exact capabilities depend on the remote, but the link will
always by default be created with the least constraints e.g. no
expiry, no password protection, accessible without account.
```
@ -30,7 +41,9 @@ rclone link remote:path [flags]
## Options
```
-h, --help help for link
--expire Duration The amount of time that the link will be valid (default 100y)
-h, --help help for link
--unlink Remove existing public link to file/folder
```
See the [global flags page](/flags/) for global options not listed here.

View File

@ -194,20 +194,39 @@ When --vfs-read-chunk-size-limit 500M is specified, the result would be
Chunked reading will only work with --vfs-cache-mode < full, as the file will always
be copied to the vfs cache before opening with --vfs-cache-mode full.
## Directory Cache
## VFS - Virtual File System
Using the `--dir-cache-time` flag, you can set how long a
This command uses the VFS layer. This adapts the cloud storage objects
that rclone uses into something which looks much more like a disk
filing system.
Cloud storage objects have lots of properties which aren't like disk
files - you can't extend them or write to the middle of them, so the
VFS layer has to deal with that. Because there is no one right way of
doing this there are various options explained below.
The VFS layer also implements a directory cache - this caches info
about files and directories (but not the data) in memory.
## VFS Directory Cache
Using the `--dir-cache-time` flag, you can control how long a
directory should be considered up to date and not refreshed from the
backend. Changes made locally in the mount may appear immediately or
invalidate the cache. However, changes done on the remote will only
be picked up once the cache expires if the backend configured does not
support polling for changes. If the backend supports polling, changes
will be picked up on within the polling interval.
backend. Changes made through the mount will appear immediately or
invalidate the cache.
Alternatively, you can send a `SIGHUP` signal to rclone for
it to flush all directory caches, regardless of how old they are.
Assuming only one rclone instance is running, you can reset the cache
like this:
--dir-cache-time duration Time to cache directory entries for. (default 5m0s)
--poll-interval duration Time to wait between polling for changes.
However, changes made directly on the cloud storage by the web
interface or a different copy of rclone will only be picked up once
the directory cache expires if the backend configured does not support
polling for changes. If the backend supports polling, changes will be
picked up within the polling interval.
You can send a `SIGHUP` signal to rclone for it to flush all
directory caches, regardless of how old they are. Assuming only one
rclone instance is running, you can reset the cache like this:
kill -SIGHUP $(pidof rclone)
@ -220,40 +239,41 @@ Or individual files or directories:
rclone rc vfs/forget file=path/to/file dir=path/to/dir
## File Buffering
## VFS File Buffering
The `--buffer-size` flag determines the amount of memory,
that will be used to buffer data in advance.
Each open file descriptor will try to keep the specified amount of
data in memory at all times. The buffered data is bound to one file
descriptor and won't be shared between multiple open file descriptors
of the same file.
Each open file will try to keep the specified amount of data in memory
at all times. The buffered data is bound to one open file and won't be
shared.
This flag is a upper limit for the used memory per open file. The
buffer will only use memory for data that is downloaded but not not
yet read. If the buffer is empty, only a small amount of memory will
be used.
This flag is a upper limit for the used memory per file descriptor.
The buffer will only use memory for data that is downloaded but not
not yet read. If the buffer is empty, only a small amount of memory
will be used.
The maximum memory used by rclone for buffering can be up to
`--buffer-size * open files`.
## File Caching
## VFS File Caching
These flags control the VFS file caching options. The VFS layer is
used by rclone mount to make a cloud storage system work more like a
normal file system.
These flags control the VFS file caching options. File caching is
necessary to make the VFS layer appear compatible with a normal file
system. It can be disabled at the cost of some compatibility.
You'll need to enable VFS caching if you want, for example, to read
and write simultaneously to a file. See below for more details.
For example you'll need to enable VFS caching if you want to read and
write simultaneously to a file. See below for more details.
Note that the VFS cache works in addition to the cache backend and you
may find that you need one or the other or both.
Note that the VFS cache is separate from the cache backend and you may
find that you need one or the other or both.
--cache-dir string Directory rclone will use for caching.
--vfs-cache-mode CacheMode Cache mode off|minimal|writes|full (default off)
--vfs-cache-max-age duration Max age of objects in the cache. (default 1h0m0s)
--vfs-cache-mode string Cache mode off|minimal|writes|full (default "off")
--vfs-cache-max-size SizeSuffix Max total size of objects in the cache. (default off)
--vfs-cache-poll-interval duration Interval to poll the cache for stale objects. (default 1m0s)
--vfs-cache-max-size int Max total size of objects in the cache. (default off)
--vfs-write-back duration Time to writeback files after last use when using cache. (default 5s)
If run with `-vv` rclone will print the location of the file cache. The
files are stored in the user cache file area which is OS dependent but
@ -265,9 +285,10 @@ The higher the cache mode the more compatible rclone becomes at the
cost of using disk space.
Note that files are written back to the remote only when they are
closed so if rclone is quit or dies with open files then these won't
get written back to the remote. However they will still be in the on
disk cache.
closed and if they haven't been accessed for --vfs-write-back
second. If rclone is quit or dies with files that haven't been
uploaded, these will be uploaded next time rclone is run with the same
flags.
If using --vfs-cache-max-size note that the cache may exceed this size
for two reasons. Firstly because it is only checked every
@ -276,7 +297,7 @@ evicted from the cache.
### --vfs-cache-mode off
In this mode the cache will read directly from the remote and write
In this mode (the default) the cache will read directly from the remote and write
directly to the remote without caching anything on disk.
This will mean some operations are not possible
@ -292,7 +313,7 @@ This will mean some operations are not possible
### --vfs-cache-mode minimal
This is very similar to "off" except that files opened for read AND
write will be buffered to disks. This means that files opened for
write will be buffered to disk. This means that files opened for
write will be a lot more compatible, but uses the minimal disk space.
These operations are not possible
@ -310,32 +331,72 @@ first.
This mode should support all normal file system operations.
If an upload fails it will be retried up to --low-level-retries times.
If an upload fails it will be retried at exponentially increasing
intervals up to 1 minute.
### --vfs-cache-mode full
In this mode all reads and writes are buffered to and from disk. When
a file is opened for read it will be downloaded in its entirety first.
In this mode all reads and writes are buffered to and from disk. When
data is read from the remote this is buffered to disk as well.
This may be appropriate for your needs, or you may prefer to look at
the cache backend which does a much more sophisticated job of caching,
including caching directory hierarchies and chunks of files.
In this mode the files in the cache will be sparse files and rclone
will keep track of which bits of the files it has dowloaded.
In this mode, unlike the others, when a file is written to the disk,
it will be kept on the disk after it is written to the remote. It
will be purged on a schedule according to `--vfs-cache-max-age`.
So if an application only reads the starts of each file, then rclone
will only buffer the start of the file. These files will appear to be
their full size in the cache, but they will be sparse files with only
the data that has been downloaded present in them.
This mode should support all normal file system operations.
This mode should support all normal file system operations and is
otherwise identical to --vfs-cache-mode writes.
If an upload or download fails it will be retried up to
--low-level-retries times.
When reading a file rclone will read --buffer-size plus
--vfs-read-ahead bytes ahead. The --buffer-size is buffered in memory
whereas the --vfs-read-ahead is buffered on disk.
## Case Sensitivity
When using this mode it is recommended that --buffer-size is not set
too big and --vfs-read-ahead is set large if required.
## VFS Performance
These flags may be used to enable/disable features of the VFS for
performance or other reasons.
In particular S3 and Swift benefit hugely from the --no-modtime flag
(or use --use-server-modtime for a slightly different effect) as each
read of the modification time takes a transaction.
--no-checksum Don't compare checksums on up/download.
--no-modtime Don't read/write the modification time (can speed things up).
--no-seek Don't allow seeking in files.
--read-only Mount read-only.
When rclone reads files from a remote it reads them in chunks. This
means that rather than requesting the whole file rclone reads the
chunk specified. This is advantageous because some cloud providers
account for reads being all the data requested, not all the data
delivered.
Rclone will keep doubling the chunk size requested starting at
--vfs-read-chunk-size with a maximum of --vfs-read-chunk-size-limit
unless it is set to "off" in which case there will be no limit.
--vfs-read-chunk-size SizeSuffix Read the source objects in chunks. (default 128M)
--vfs-read-chunk-size-limit SizeSuffix Max chunk doubling size (default "off")
Sometimes rclone is delivered reads or writes out of order. Rather
than seeking rclone will wait a short time for the in sequence read or
write to come in. These flags only come into effect when not using an
on disk cache file.
--vfs-read-wait duration Time to wait for in-sequence read before seeking. (default 20ms)
--vfs-write-wait duration Time to wait for in-sequence write before giving error. (default 1s)
## VFS Case Sensitivity
Linux file systems are case-sensitive: two files can differ only
by case, and the exact case must be used when opening a file.
Windows is not like most other operating systems supported by rclone.
File systems in modern Windows are case-insensitive but case-preserving:
although existing files can be opened using any case, the exact case used
to create the file is preserved and available for programs to query.
@ -346,7 +407,7 @@ file systems case-sensitive but that is not the default
The "--vfs-case-insensitive" mount flag controls how rclone handles these
two cases. If its value is "false", rclone passes file names to the mounted
file system as is. If the flag is "true" (or appears without a value on
file system as-is. If the flag is "true" (or appears without a value on
command line), rclone may perform a "fixup" as explained below.
The user may specify a file name to open/delete/rename/etc with a case
@ -362,7 +423,7 @@ Note that case sensitivity of the operating system running rclone (the target)
may differ from case sensitivity of a file system mounted by rclone (the source).
The flag controls whether "fixup" is performed to satisfy the target.
If the flag is not provided on command line, then its default value depends
If the flag is not provided on the command line, then its default value depends
on the operating system where rclone runs: "true" on Windows and macOS, "false"
otherwise. If the flag is provided without a value, then it is "true".
@ -403,9 +464,11 @@ rclone mount remote:path /path/to/mountpoint [flags]
--vfs-cache-mode CacheMode Cache mode off|minimal|writes|full (default off)
--vfs-cache-poll-interval duration Interval to poll the cache for stale objects. (default 1m0s)
--vfs-case-insensitive If a file name not found, find a case insensitive match.
--vfs-read-ahead SizeSuffix Extra read ahead over --buffer-size when using cache-mode full.
--vfs-read-chunk-size SizeSuffix Read the source objects in chunks. (default 128M)
--vfs-read-chunk-size-limit SizeSuffix If greater than --vfs-read-chunk-size, double the chunk size after each chunk read, until the limit is reached. 'off' is unlimited. (default off)
--vfs-read-wait duration Time to wait for in-sequence read before seeking. (default 20ms)
--vfs-write-back duration Time to writeback files after last use when using cache. (default 5s)
--vfs-write-wait duration Time to wait for in-sequence write before giving error. (default 1s)
--volname string Set the volume name (not supported by all OSes).
--write-back-cache Makes kernel buffer writes before sending them to rclone. Without this, writethrough caching is used.

View File

@ -34,7 +34,7 @@ option when moving a small number of files into a large destination
can speed transfers up greatly.
**Important**: Since this can cause data loss, test first with the
--dry-run flag.
`--dry-run` or the `--interactive`/`-i` flag.
**Note**: Use the `-P`/`--progress` flag to view real-time transfer statistics.

View File

@ -39,7 +39,7 @@ modification time or MD5SUM. src will be deleted on successful
transfer.
**Important**: Since this can cause data loss, test first with the
--dry-run flag.
`--dry-run` or the `--interactive`/`-i` flag.
**Note**: Use the `-P`/`--progress` flag to view real-time transfer statistics.

View File

@ -1,17 +1,38 @@
---
title: "rclone obscure"
description: "Obscure password for use in the rclone.conf"
description: "Obscure password for use in the rclone config file"
slug: rclone_obscure
url: /commands/rclone_obscure/
# autogenerated - DO NOT EDIT, instead edit the source code in cmd/obscure/ and as part of making a release run "make commanddocs"
---
# rclone obscure
Obscure password for use in the rclone.conf
Obscure password for use in the rclone config file
## Synopsis
Obscure password for use in the rclone.conf
In the rclone config file, human readable passwords are
obscured. Obscuring them is done by encrypting them and writing them
out in base64. This is **not** a secure way of encrypting these
passwords as rclone can decrypt them - it is to prevent "eyedropping"
- namely someone seeing a password in the rclone config file by
accident.
Many equally important things (like access tokens) are not obscured in
the config file. However it is very hard to shoulder surf a 64
character hex token.
This command can also accept a password through STDIN instead of an
argument by passing a hyphen as an argument. Example:
echo "secretpassword" | rclone obscure -
If there is no data on STDIN to read, rclone obscure will default to
obfuscating the hyphen itself.
If you want to encrypt the config file then please use config file
encryption - see [rclone config](/commands/rclone_config/) for more
info.
```
rclone obscure password [flags]

View File

@ -16,6 +16,9 @@ Remove the path and all of its contents. Note that this does not obey
include/exclude filters - everything will be removed. Use `delete` if
you want to selectively delete files.
**Important**: Since this can cause data loss, test first with the
`--dry-run` or the `--interactive`/`-i` flag.
```
rclone purge remote:path [flags]

View File

@ -23,30 +23,49 @@ players might show files that they are not able to play back correctly.
## Server options
Use --addr to specify which IP address and port the server should
listen on, eg --addr 1.2.3.4:8000 or --addr :8080 to listen to all
Use `--addr` to specify which IP address and port the server should
listen on, eg `--addr 1.2.3.4:8000` or `--addr :8080` to listen to all
IPs.
Use --name to choose the friendly server name, which is by
Use `--name` to choose the friendly server name, which is by
default "rclone (hostname)".
Use --log-trace in conjunction with -vv to enable additional debug
Use `--log-trace` in conjunction with `-vv` to enable additional debug
logging of all UPNP traffic.
## Directory Cache
## VFS - Virtual File System
Using the `--dir-cache-time` flag, you can set how long a
This command uses the VFS layer. This adapts the cloud storage objects
that rclone uses into something which looks much more like a disk
filing system.
Cloud storage objects have lots of properties which aren't like disk
files - you can't extend them or write to the middle of them, so the
VFS layer has to deal with that. Because there is no one right way of
doing this there are various options explained below.
The VFS layer also implements a directory cache - this caches info
about files and directories (but not the data) in memory.
## VFS Directory Cache
Using the `--dir-cache-time` flag, you can control how long a
directory should be considered up to date and not refreshed from the
backend. Changes made locally in the mount may appear immediately or
invalidate the cache. However, changes done on the remote will only
be picked up once the cache expires if the backend configured does not
support polling for changes. If the backend supports polling, changes
will be picked up on within the polling interval.
backend. Changes made through the mount will appear immediately or
invalidate the cache.
Alternatively, you can send a `SIGHUP` signal to rclone for
it to flush all directory caches, regardless of how old they are.
Assuming only one rclone instance is running, you can reset the cache
like this:
--dir-cache-time duration Time to cache directory entries for. (default 5m0s)
--poll-interval duration Time to wait between polling for changes.
However, changes made directly on the cloud storage by the web
interface or a different copy of rclone will only be picked up once
the directory cache expires if the backend configured does not support
polling for changes. If the backend supports polling, changes will be
picked up within the polling interval.
You can send a `SIGHUP` signal to rclone for it to flush all
directory caches, regardless of how old they are. Assuming only one
rclone instance is running, you can reset the cache like this:
kill -SIGHUP $(pidof rclone)
@ -59,40 +78,41 @@ Or individual files or directories:
rclone rc vfs/forget file=path/to/file dir=path/to/dir
## File Buffering
## VFS File Buffering
The `--buffer-size` flag determines the amount of memory,
that will be used to buffer data in advance.
Each open file descriptor will try to keep the specified amount of
data in memory at all times. The buffered data is bound to one file
descriptor and won't be shared between multiple open file descriptors
of the same file.
Each open file will try to keep the specified amount of data in memory
at all times. The buffered data is bound to one open file and won't be
shared.
This flag is a upper limit for the used memory per open file. The
buffer will only use memory for data that is downloaded but not not
yet read. If the buffer is empty, only a small amount of memory will
be used.
This flag is a upper limit for the used memory per file descriptor.
The buffer will only use memory for data that is downloaded but not
not yet read. If the buffer is empty, only a small amount of memory
will be used.
The maximum memory used by rclone for buffering can be up to
`--buffer-size * open files`.
## File Caching
## VFS File Caching
These flags control the VFS file caching options. The VFS layer is
used by rclone mount to make a cloud storage system work more like a
normal file system.
These flags control the VFS file caching options. File caching is
necessary to make the VFS layer appear compatible with a normal file
system. It can be disabled at the cost of some compatibility.
You'll need to enable VFS caching if you want, for example, to read
and write simultaneously to a file. See below for more details.
For example you'll need to enable VFS caching if you want to read and
write simultaneously to a file. See below for more details.
Note that the VFS cache works in addition to the cache backend and you
may find that you need one or the other or both.
Note that the VFS cache is separate from the cache backend and you may
find that you need one or the other or both.
--cache-dir string Directory rclone will use for caching.
--vfs-cache-mode CacheMode Cache mode off|minimal|writes|full (default off)
--vfs-cache-max-age duration Max age of objects in the cache. (default 1h0m0s)
--vfs-cache-mode string Cache mode off|minimal|writes|full (default "off")
--vfs-cache-max-size SizeSuffix Max total size of objects in the cache. (default off)
--vfs-cache-poll-interval duration Interval to poll the cache for stale objects. (default 1m0s)
--vfs-cache-max-size int Max total size of objects in the cache. (default off)
--vfs-write-back duration Time to writeback files after last use when using cache. (default 5s)
If run with `-vv` rclone will print the location of the file cache. The
files are stored in the user cache file area which is OS dependent but
@ -104,9 +124,10 @@ The higher the cache mode the more compatible rclone becomes at the
cost of using disk space.
Note that files are written back to the remote only when they are
closed so if rclone is quit or dies with open files then these won't
get written back to the remote. However they will still be in the on
disk cache.
closed and if they haven't been accessed for --vfs-write-back
second. If rclone is quit or dies with files that haven't been
uploaded, these will be uploaded next time rclone is run with the same
flags.
If using --vfs-cache-max-size note that the cache may exceed this size
for two reasons. Firstly because it is only checked every
@ -115,7 +136,7 @@ evicted from the cache.
### --vfs-cache-mode off
In this mode the cache will read directly from the remote and write
In this mode (the default) the cache will read directly from the remote and write
directly to the remote without caching anything on disk.
This will mean some operations are not possible
@ -131,7 +152,7 @@ This will mean some operations are not possible
### --vfs-cache-mode minimal
This is very similar to "off" except that files opened for read AND
write will be buffered to disks. This means that files opened for
write will be buffered to disk. This means that files opened for
write will be a lot more compatible, but uses the minimal disk space.
These operations are not possible
@ -149,32 +170,72 @@ first.
This mode should support all normal file system operations.
If an upload fails it will be retried up to --low-level-retries times.
If an upload fails it will be retried at exponentially increasing
intervals up to 1 minute.
### --vfs-cache-mode full
In this mode all reads and writes are buffered to and from disk. When
a file is opened for read it will be downloaded in its entirety first.
In this mode all reads and writes are buffered to and from disk. When
data is read from the remote this is buffered to disk as well.
This may be appropriate for your needs, or you may prefer to look at
the cache backend which does a much more sophisticated job of caching,
including caching directory hierarchies and chunks of files.
In this mode the files in the cache will be sparse files and rclone
will keep track of which bits of the files it has dowloaded.
In this mode, unlike the others, when a file is written to the disk,
it will be kept on the disk after it is written to the remote. It
will be purged on a schedule according to `--vfs-cache-max-age`.
So if an application only reads the starts of each file, then rclone
will only buffer the start of the file. These files will appear to be
their full size in the cache, but they will be sparse files with only
the data that has been downloaded present in them.
This mode should support all normal file system operations.
This mode should support all normal file system operations and is
otherwise identical to --vfs-cache-mode writes.
If an upload or download fails it will be retried up to
--low-level-retries times.
When reading a file rclone will read --buffer-size plus
--vfs-read-ahead bytes ahead. The --buffer-size is buffered in memory
whereas the --vfs-read-ahead is buffered on disk.
## Case Sensitivity
When using this mode it is recommended that --buffer-size is not set
too big and --vfs-read-ahead is set large if required.
## VFS Performance
These flags may be used to enable/disable features of the VFS for
performance or other reasons.
In particular S3 and Swift benefit hugely from the --no-modtime flag
(or use --use-server-modtime for a slightly different effect) as each
read of the modification time takes a transaction.
--no-checksum Don't compare checksums on up/download.
--no-modtime Don't read/write the modification time (can speed things up).
--no-seek Don't allow seeking in files.
--read-only Mount read-only.
When rclone reads files from a remote it reads them in chunks. This
means that rather than requesting the whole file rclone reads the
chunk specified. This is advantageous because some cloud providers
account for reads being all the data requested, not all the data
delivered.
Rclone will keep doubling the chunk size requested starting at
--vfs-read-chunk-size with a maximum of --vfs-read-chunk-size-limit
unless it is set to "off" in which case there will be no limit.
--vfs-read-chunk-size SizeSuffix Read the source objects in chunks. (default 128M)
--vfs-read-chunk-size-limit SizeSuffix Max chunk doubling size (default "off")
Sometimes rclone is delivered reads or writes out of order. Rather
than seeking rclone will wait a short time for the in sequence read or
write to come in. These flags only come into effect when not using an
on disk cache file.
--vfs-read-wait duration Time to wait for in-sequence read before seeking. (default 20ms)
--vfs-write-wait duration Time to wait for in-sequence write before giving error. (default 1s)
## VFS Case Sensitivity
Linux file systems are case-sensitive: two files can differ only
by case, and the exact case must be used when opening a file.
Windows is not like most other operating systems supported by rclone.
File systems in modern Windows are case-insensitive but case-preserving:
although existing files can be opened using any case, the exact case used
to create the file is preserved and available for programs to query.
@ -185,7 +246,7 @@ file systems case-sensitive but that is not the default
The "--vfs-case-insensitive" mount flag controls how rclone handles these
two cases. If its value is "false", rclone passes file names to the mounted
file system as is. If the flag is "true" (or appears without a value on
file system as-is. If the flag is "true" (or appears without a value on
command line), rclone may perform a "fixup" as explained below.
The user may specify a file name to open/delete/rename/etc with a case
@ -201,7 +262,7 @@ Note that case sensitivity of the operating system running rclone (the target)
may differ from case sensitivity of a file system mounted by rclone (the source).
The flag controls whether "fixup" is performed to satisfy the target.
If the flag is not provided on command line, then its default value depends
If the flag is not provided on the command line, then its default value depends
on the operating system where rclone runs: "true" on Windows and macOS, "false"
otherwise. If the flag is provided without a value, then it is "true".
@ -233,9 +294,11 @@ rclone serve dlna remote:path [flags]
--vfs-cache-mode CacheMode Cache mode off|minimal|writes|full (default off)
--vfs-cache-poll-interval duration Interval to poll the cache for stale objects. (default 1m0s)
--vfs-case-insensitive If a file name not found, find a case insensitive match.
--vfs-read-ahead SizeSuffix Extra read ahead over --buffer-size when using cache-mode full.
--vfs-read-chunk-size SizeSuffix Read the source objects in chunks. (default 128M)
--vfs-read-chunk-size-limit SizeSuffix If greater than --vfs-read-chunk-size, double the chunk size after each chunk read, until the limit is reached. 'off' is unlimited. (default off)
--vfs-read-wait duration Time to wait for in-sequence read before seeking. (default 20ms)
--vfs-write-back duration Time to writeback files after last use when using cache. (default 5s)
--vfs-write-wait duration Time to wait for in-sequence write before giving error. (default 1s)
```

View File

@ -32,20 +32,39 @@ By default this will serve files without needing a login.
You can set a single username and password with the --user and --pass flags.
## Directory Cache
## VFS - Virtual File System
Using the `--dir-cache-time` flag, you can set how long a
This command uses the VFS layer. This adapts the cloud storage objects
that rclone uses into something which looks much more like a disk
filing system.
Cloud storage objects have lots of properties which aren't like disk
files - you can't extend them or write to the middle of them, so the
VFS layer has to deal with that. Because there is no one right way of
doing this there are various options explained below.
The VFS layer also implements a directory cache - this caches info
about files and directories (but not the data) in memory.
## VFS Directory Cache
Using the `--dir-cache-time` flag, you can control how long a
directory should be considered up to date and not refreshed from the
backend. Changes made locally in the mount may appear immediately or
invalidate the cache. However, changes done on the remote will only
be picked up once the cache expires if the backend configured does not
support polling for changes. If the backend supports polling, changes
will be picked up on within the polling interval.
backend. Changes made through the mount will appear immediately or
invalidate the cache.
Alternatively, you can send a `SIGHUP` signal to rclone for
it to flush all directory caches, regardless of how old they are.
Assuming only one rclone instance is running, you can reset the cache
like this:
--dir-cache-time duration Time to cache directory entries for. (default 5m0s)
--poll-interval duration Time to wait between polling for changes.
However, changes made directly on the cloud storage by the web
interface or a different copy of rclone will only be picked up once
the directory cache expires if the backend configured does not support
polling for changes. If the backend supports polling, changes will be
picked up within the polling interval.
You can send a `SIGHUP` signal to rclone for it to flush all
directory caches, regardless of how old they are. Assuming only one
rclone instance is running, you can reset the cache like this:
kill -SIGHUP $(pidof rclone)
@ -58,40 +77,41 @@ Or individual files or directories:
rclone rc vfs/forget file=path/to/file dir=path/to/dir
## File Buffering
## VFS File Buffering
The `--buffer-size` flag determines the amount of memory,
that will be used to buffer data in advance.
Each open file descriptor will try to keep the specified amount of
data in memory at all times. The buffered data is bound to one file
descriptor and won't be shared between multiple open file descriptors
of the same file.
Each open file will try to keep the specified amount of data in memory
at all times. The buffered data is bound to one open file and won't be
shared.
This flag is a upper limit for the used memory per open file. The
buffer will only use memory for data that is downloaded but not not
yet read. If the buffer is empty, only a small amount of memory will
be used.
This flag is a upper limit for the used memory per file descriptor.
The buffer will only use memory for data that is downloaded but not
not yet read. If the buffer is empty, only a small amount of memory
will be used.
The maximum memory used by rclone for buffering can be up to
`--buffer-size * open files`.
## File Caching
## VFS File Caching
These flags control the VFS file caching options. The VFS layer is
used by rclone mount to make a cloud storage system work more like a
normal file system.
These flags control the VFS file caching options. File caching is
necessary to make the VFS layer appear compatible with a normal file
system. It can be disabled at the cost of some compatibility.
You'll need to enable VFS caching if you want, for example, to read
and write simultaneously to a file. See below for more details.
For example you'll need to enable VFS caching if you want to read and
write simultaneously to a file. See below for more details.
Note that the VFS cache works in addition to the cache backend and you
may find that you need one or the other or both.
Note that the VFS cache is separate from the cache backend and you may
find that you need one or the other or both.
--cache-dir string Directory rclone will use for caching.
--vfs-cache-mode CacheMode Cache mode off|minimal|writes|full (default off)
--vfs-cache-max-age duration Max age of objects in the cache. (default 1h0m0s)
--vfs-cache-mode string Cache mode off|minimal|writes|full (default "off")
--vfs-cache-max-size SizeSuffix Max total size of objects in the cache. (default off)
--vfs-cache-poll-interval duration Interval to poll the cache for stale objects. (default 1m0s)
--vfs-cache-max-size int Max total size of objects in the cache. (default off)
--vfs-write-back duration Time to writeback files after last use when using cache. (default 5s)
If run with `-vv` rclone will print the location of the file cache. The
files are stored in the user cache file area which is OS dependent but
@ -103,9 +123,10 @@ The higher the cache mode the more compatible rclone becomes at the
cost of using disk space.
Note that files are written back to the remote only when they are
closed so if rclone is quit or dies with open files then these won't
get written back to the remote. However they will still be in the on
disk cache.
closed and if they haven't been accessed for --vfs-write-back
second. If rclone is quit or dies with files that haven't been
uploaded, these will be uploaded next time rclone is run with the same
flags.
If using --vfs-cache-max-size note that the cache may exceed this size
for two reasons. Firstly because it is only checked every
@ -114,7 +135,7 @@ evicted from the cache.
### --vfs-cache-mode off
In this mode the cache will read directly from the remote and write
In this mode (the default) the cache will read directly from the remote and write
directly to the remote without caching anything on disk.
This will mean some operations are not possible
@ -130,7 +151,7 @@ This will mean some operations are not possible
### --vfs-cache-mode minimal
This is very similar to "off" except that files opened for read AND
write will be buffered to disks. This means that files opened for
write will be buffered to disk. This means that files opened for
write will be a lot more compatible, but uses the minimal disk space.
These operations are not possible
@ -148,32 +169,72 @@ first.
This mode should support all normal file system operations.
If an upload fails it will be retried up to --low-level-retries times.
If an upload fails it will be retried at exponentially increasing
intervals up to 1 minute.
### --vfs-cache-mode full
In this mode all reads and writes are buffered to and from disk. When
a file is opened for read it will be downloaded in its entirety first.
In this mode all reads and writes are buffered to and from disk. When
data is read from the remote this is buffered to disk as well.
This may be appropriate for your needs, or you may prefer to look at
the cache backend which does a much more sophisticated job of caching,
including caching directory hierarchies and chunks of files.
In this mode the files in the cache will be sparse files and rclone
will keep track of which bits of the files it has dowloaded.
In this mode, unlike the others, when a file is written to the disk,
it will be kept on the disk after it is written to the remote. It
will be purged on a schedule according to `--vfs-cache-max-age`.
So if an application only reads the starts of each file, then rclone
will only buffer the start of the file. These files will appear to be
their full size in the cache, but they will be sparse files with only
the data that has been downloaded present in them.
This mode should support all normal file system operations.
This mode should support all normal file system operations and is
otherwise identical to --vfs-cache-mode writes.
If an upload or download fails it will be retried up to
--low-level-retries times.
When reading a file rclone will read --buffer-size plus
--vfs-read-ahead bytes ahead. The --buffer-size is buffered in memory
whereas the --vfs-read-ahead is buffered on disk.
## Case Sensitivity
When using this mode it is recommended that --buffer-size is not set
too big and --vfs-read-ahead is set large if required.
## VFS Performance
These flags may be used to enable/disable features of the VFS for
performance or other reasons.
In particular S3 and Swift benefit hugely from the --no-modtime flag
(or use --use-server-modtime for a slightly different effect) as each
read of the modification time takes a transaction.
--no-checksum Don't compare checksums on up/download.
--no-modtime Don't read/write the modification time (can speed things up).
--no-seek Don't allow seeking in files.
--read-only Mount read-only.
When rclone reads files from a remote it reads them in chunks. This
means that rather than requesting the whole file rclone reads the
chunk specified. This is advantageous because some cloud providers
account for reads being all the data requested, not all the data
delivered.
Rclone will keep doubling the chunk size requested starting at
--vfs-read-chunk-size with a maximum of --vfs-read-chunk-size-limit
unless it is set to "off" in which case there will be no limit.
--vfs-read-chunk-size SizeSuffix Read the source objects in chunks. (default 128M)
--vfs-read-chunk-size-limit SizeSuffix Max chunk doubling size (default "off")
Sometimes rclone is delivered reads or writes out of order. Rather
than seeking rclone will wait a short time for the in sequence read or
write to come in. These flags only come into effect when not using an
on disk cache file.
--vfs-read-wait duration Time to wait for in-sequence read before seeking. (default 20ms)
--vfs-write-wait duration Time to wait for in-sequence write before giving error. (default 1s)
## VFS Case Sensitivity
Linux file systems are case-sensitive: two files can differ only
by case, and the exact case must be used when opening a file.
Windows is not like most other operating systems supported by rclone.
File systems in modern Windows are case-insensitive but case-preserving:
although existing files can be opened using any case, the exact case used
to create the file is preserved and available for programs to query.
@ -184,7 +245,7 @@ file systems case-sensitive but that is not the default
The "--vfs-case-insensitive" mount flag controls how rclone handles these
two cases. If its value is "false", rclone passes file names to the mounted
file system as is. If the flag is "true" (or appears without a value on
file system as-is. If the flag is "true" (or appears without a value on
command line), rclone may perform a "fixup" as explained below.
The user may specify a file name to open/delete/rename/etc with a case
@ -200,7 +261,7 @@ Note that case sensitivity of the operating system running rclone (the target)
may differ from case sensitivity of a file system mounted by rclone (the source).
The flag controls whether "fixup" is performed to satisfy the target.
If the flag is not provided on command line, then its default value depends
If the flag is not provided on the command line, then its default value depends
on the operating system where rclone runs: "true" on Windows and macOS, "false"
otherwise. If the flag is provided without a value, then it is "true".
@ -316,9 +377,11 @@ rclone serve ftp remote:path [flags]
--vfs-cache-mode CacheMode Cache mode off|minimal|writes|full (default off)
--vfs-cache-poll-interval duration Interval to poll the cache for stale objects. (default 1m0s)
--vfs-case-insensitive If a file name not found, find a case insensitive match.
--vfs-read-ahead SizeSuffix Extra read ahead over --buffer-size when using cache-mode full.
--vfs-read-chunk-size SizeSuffix Read the source objects in chunks. (default 128M)
--vfs-read-chunk-size-limit SizeSuffix If greater than --vfs-read-chunk-size, double the chunk size after each chunk read, until the limit is reached. 'off' is unlimited. (default off)
--vfs-read-wait duration Time to wait for in-sequence read before seeking. (default 20ms)
--vfs-write-back duration Time to writeback files after last use when using cache. (default 5s)
--vfs-write-wait duration Time to wait for in-sequence write before giving error. (default 1s)
```

View File

@ -104,20 +104,39 @@ of that with the CA certificate. --key should be the PEM encoded
private key and --client-ca should be the PEM encoded client
certificate authority certificate.
## Directory Cache
## VFS - Virtual File System
Using the `--dir-cache-time` flag, you can set how long a
This command uses the VFS layer. This adapts the cloud storage objects
that rclone uses into something which looks much more like a disk
filing system.
Cloud storage objects have lots of properties which aren't like disk
files - you can't extend them or write to the middle of them, so the
VFS layer has to deal with that. Because there is no one right way of
doing this there are various options explained below.
The VFS layer also implements a directory cache - this caches info
about files and directories (but not the data) in memory.
## VFS Directory Cache
Using the `--dir-cache-time` flag, you can control how long a
directory should be considered up to date and not refreshed from the
backend. Changes made locally in the mount may appear immediately or
invalidate the cache. However, changes done on the remote will only
be picked up once the cache expires if the backend configured does not
support polling for changes. If the backend supports polling, changes
will be picked up on within the polling interval.
backend. Changes made through the mount will appear immediately or
invalidate the cache.
Alternatively, you can send a `SIGHUP` signal to rclone for
it to flush all directory caches, regardless of how old they are.
Assuming only one rclone instance is running, you can reset the cache
like this:
--dir-cache-time duration Time to cache directory entries for. (default 5m0s)
--poll-interval duration Time to wait between polling for changes.
However, changes made directly on the cloud storage by the web
interface or a different copy of rclone will only be picked up once
the directory cache expires if the backend configured does not support
polling for changes. If the backend supports polling, changes will be
picked up within the polling interval.
You can send a `SIGHUP` signal to rclone for it to flush all
directory caches, regardless of how old they are. Assuming only one
rclone instance is running, you can reset the cache like this:
kill -SIGHUP $(pidof rclone)
@ -130,40 +149,41 @@ Or individual files or directories:
rclone rc vfs/forget file=path/to/file dir=path/to/dir
## File Buffering
## VFS File Buffering
The `--buffer-size` flag determines the amount of memory,
that will be used to buffer data in advance.
Each open file descriptor will try to keep the specified amount of
data in memory at all times. The buffered data is bound to one file
descriptor and won't be shared between multiple open file descriptors
of the same file.
Each open file will try to keep the specified amount of data in memory
at all times. The buffered data is bound to one open file and won't be
shared.
This flag is a upper limit for the used memory per open file. The
buffer will only use memory for data that is downloaded but not not
yet read. If the buffer is empty, only a small amount of memory will
be used.
This flag is a upper limit for the used memory per file descriptor.
The buffer will only use memory for data that is downloaded but not
not yet read. If the buffer is empty, only a small amount of memory
will be used.
The maximum memory used by rclone for buffering can be up to
`--buffer-size * open files`.
## File Caching
## VFS File Caching
These flags control the VFS file caching options. The VFS layer is
used by rclone mount to make a cloud storage system work more like a
normal file system.
These flags control the VFS file caching options. File caching is
necessary to make the VFS layer appear compatible with a normal file
system. It can be disabled at the cost of some compatibility.
You'll need to enable VFS caching if you want, for example, to read
and write simultaneously to a file. See below for more details.
For example you'll need to enable VFS caching if you want to read and
write simultaneously to a file. See below for more details.
Note that the VFS cache works in addition to the cache backend and you
may find that you need one or the other or both.
Note that the VFS cache is separate from the cache backend and you may
find that you need one or the other or both.
--cache-dir string Directory rclone will use for caching.
--vfs-cache-mode CacheMode Cache mode off|minimal|writes|full (default off)
--vfs-cache-max-age duration Max age of objects in the cache. (default 1h0m0s)
--vfs-cache-mode string Cache mode off|minimal|writes|full (default "off")
--vfs-cache-max-size SizeSuffix Max total size of objects in the cache. (default off)
--vfs-cache-poll-interval duration Interval to poll the cache for stale objects. (default 1m0s)
--vfs-cache-max-size int Max total size of objects in the cache. (default off)
--vfs-write-back duration Time to writeback files after last use when using cache. (default 5s)
If run with `-vv` rclone will print the location of the file cache. The
files are stored in the user cache file area which is OS dependent but
@ -175,9 +195,10 @@ The higher the cache mode the more compatible rclone becomes at the
cost of using disk space.
Note that files are written back to the remote only when they are
closed so if rclone is quit or dies with open files then these won't
get written back to the remote. However they will still be in the on
disk cache.
closed and if they haven't been accessed for --vfs-write-back
second. If rclone is quit or dies with files that haven't been
uploaded, these will be uploaded next time rclone is run with the same
flags.
If using --vfs-cache-max-size note that the cache may exceed this size
for two reasons. Firstly because it is only checked every
@ -186,7 +207,7 @@ evicted from the cache.
### --vfs-cache-mode off
In this mode the cache will read directly from the remote and write
In this mode (the default) the cache will read directly from the remote and write
directly to the remote without caching anything on disk.
This will mean some operations are not possible
@ -202,7 +223,7 @@ This will mean some operations are not possible
### --vfs-cache-mode minimal
This is very similar to "off" except that files opened for read AND
write will be buffered to disks. This means that files opened for
write will be buffered to disk. This means that files opened for
write will be a lot more compatible, but uses the minimal disk space.
These operations are not possible
@ -220,32 +241,72 @@ first.
This mode should support all normal file system operations.
If an upload fails it will be retried up to --low-level-retries times.
If an upload fails it will be retried at exponentially increasing
intervals up to 1 minute.
### --vfs-cache-mode full
In this mode all reads and writes are buffered to and from disk. When
a file is opened for read it will be downloaded in its entirety first.
In this mode all reads and writes are buffered to and from disk. When
data is read from the remote this is buffered to disk as well.
This may be appropriate for your needs, or you may prefer to look at
the cache backend which does a much more sophisticated job of caching,
including caching directory hierarchies and chunks of files.
In this mode the files in the cache will be sparse files and rclone
will keep track of which bits of the files it has dowloaded.
In this mode, unlike the others, when a file is written to the disk,
it will be kept on the disk after it is written to the remote. It
will be purged on a schedule according to `--vfs-cache-max-age`.
So if an application only reads the starts of each file, then rclone
will only buffer the start of the file. These files will appear to be
their full size in the cache, but they will be sparse files with only
the data that has been downloaded present in them.
This mode should support all normal file system operations.
This mode should support all normal file system operations and is
otherwise identical to --vfs-cache-mode writes.
If an upload or download fails it will be retried up to
--low-level-retries times.
When reading a file rclone will read --buffer-size plus
--vfs-read-ahead bytes ahead. The --buffer-size is buffered in memory
whereas the --vfs-read-ahead is buffered on disk.
## Case Sensitivity
When using this mode it is recommended that --buffer-size is not set
too big and --vfs-read-ahead is set large if required.
## VFS Performance
These flags may be used to enable/disable features of the VFS for
performance or other reasons.
In particular S3 and Swift benefit hugely from the --no-modtime flag
(or use --use-server-modtime for a slightly different effect) as each
read of the modification time takes a transaction.
--no-checksum Don't compare checksums on up/download.
--no-modtime Don't read/write the modification time (can speed things up).
--no-seek Don't allow seeking in files.
--read-only Mount read-only.
When rclone reads files from a remote it reads them in chunks. This
means that rather than requesting the whole file rclone reads the
chunk specified. This is advantageous because some cloud providers
account for reads being all the data requested, not all the data
delivered.
Rclone will keep doubling the chunk size requested starting at
--vfs-read-chunk-size with a maximum of --vfs-read-chunk-size-limit
unless it is set to "off" in which case there will be no limit.
--vfs-read-chunk-size SizeSuffix Read the source objects in chunks. (default 128M)
--vfs-read-chunk-size-limit SizeSuffix Max chunk doubling size (default "off")
Sometimes rclone is delivered reads or writes out of order. Rather
than seeking rclone will wait a short time for the in sequence read or
write to come in. These flags only come into effect when not using an
on disk cache file.
--vfs-read-wait duration Time to wait for in-sequence read before seeking. (default 20ms)
--vfs-write-wait duration Time to wait for in-sequence write before giving error. (default 1s)
## VFS Case Sensitivity
Linux file systems are case-sensitive: two files can differ only
by case, and the exact case must be used when opening a file.
Windows is not like most other operating systems supported by rclone.
File systems in modern Windows are case-insensitive but case-preserving:
although existing files can be opened using any case, the exact case used
to create the file is preserved and available for programs to query.
@ -256,7 +317,7 @@ file systems case-sensitive but that is not the default
The "--vfs-case-insensitive" mount flag controls how rclone handles these
two cases. If its value is "false", rclone passes file names to the mounted
file system as is. If the flag is "true" (or appears without a value on
file system as-is. If the flag is "true" (or appears without a value on
command line), rclone may perform a "fixup" as explained below.
The user may specify a file name to open/delete/rename/etc with a case
@ -272,7 +333,7 @@ Note that case sensitivity of the operating system running rclone (the target)
may differ from case sensitivity of a file system mounted by rclone (the source).
The flag controls whether "fixup" is performed to satisfy the target.
If the flag is not provided on command line, then its default value depends
If the flag is not provided on the command line, then its default value depends
on the operating system where rclone runs: "true" on Windows and macOS, "false"
otherwise. If the flag is provided without a value, then it is "true".
@ -314,9 +375,11 @@ rclone serve http remote:path [flags]
--vfs-cache-mode CacheMode Cache mode off|minimal|writes|full (default off)
--vfs-cache-poll-interval duration Interval to poll the cache for stale objects. (default 1m0s)
--vfs-case-insensitive If a file name not found, find a case insensitive match.
--vfs-read-ahead SizeSuffix Extra read ahead over --buffer-size when using cache-mode full.
--vfs-read-chunk-size SizeSuffix Read the source objects in chunks. (default 128M)
--vfs-read-chunk-size-limit SizeSuffix If greater than --vfs-read-chunk-size, double the chunk size after each chunk read, until the limit is reached. 'off' is unlimited. (default off)
--vfs-read-wait duration Time to wait for in-sequence read before seeking. (default 20ms)
--vfs-write-back duration Time to writeback files after last use when using cache. (default 5s)
--vfs-write-wait duration Time to wait for in-sequence write before giving error. (default 1s)
```

View File

@ -43,20 +43,39 @@ Note that the default of "--vfs-cache-mode off" is fine for the rclone
sftp backend, but it may not be with other SFTP clients.
## Directory Cache
## VFS - Virtual File System
Using the `--dir-cache-time` flag, you can set how long a
This command uses the VFS layer. This adapts the cloud storage objects
that rclone uses into something which looks much more like a disk
filing system.
Cloud storage objects have lots of properties which aren't like disk
files - you can't extend them or write to the middle of them, so the
VFS layer has to deal with that. Because there is no one right way of
doing this there are various options explained below.
The VFS layer also implements a directory cache - this caches info
about files and directories (but not the data) in memory.
## VFS Directory Cache
Using the `--dir-cache-time` flag, you can control how long a
directory should be considered up to date and not refreshed from the
backend. Changes made locally in the mount may appear immediately or
invalidate the cache. However, changes done on the remote will only
be picked up once the cache expires if the backend configured does not
support polling for changes. If the backend supports polling, changes
will be picked up on within the polling interval.
backend. Changes made through the mount will appear immediately or
invalidate the cache.
Alternatively, you can send a `SIGHUP` signal to rclone for
it to flush all directory caches, regardless of how old they are.
Assuming only one rclone instance is running, you can reset the cache
like this:
--dir-cache-time duration Time to cache directory entries for. (default 5m0s)
--poll-interval duration Time to wait between polling for changes.
However, changes made directly on the cloud storage by the web
interface or a different copy of rclone will only be picked up once
the directory cache expires if the backend configured does not support
polling for changes. If the backend supports polling, changes will be
picked up within the polling interval.
You can send a `SIGHUP` signal to rclone for it to flush all
directory caches, regardless of how old they are. Assuming only one
rclone instance is running, you can reset the cache like this:
kill -SIGHUP $(pidof rclone)
@ -69,40 +88,41 @@ Or individual files or directories:
rclone rc vfs/forget file=path/to/file dir=path/to/dir
## File Buffering
## VFS File Buffering
The `--buffer-size` flag determines the amount of memory,
that will be used to buffer data in advance.
Each open file descriptor will try to keep the specified amount of
data in memory at all times. The buffered data is bound to one file
descriptor and won't be shared between multiple open file descriptors
of the same file.
Each open file will try to keep the specified amount of data in memory
at all times. The buffered data is bound to one open file and won't be
shared.
This flag is a upper limit for the used memory per open file. The
buffer will only use memory for data that is downloaded but not not
yet read. If the buffer is empty, only a small amount of memory will
be used.
This flag is a upper limit for the used memory per file descriptor.
The buffer will only use memory for data that is downloaded but not
not yet read. If the buffer is empty, only a small amount of memory
will be used.
The maximum memory used by rclone for buffering can be up to
`--buffer-size * open files`.
## File Caching
## VFS File Caching
These flags control the VFS file caching options. The VFS layer is
used by rclone mount to make a cloud storage system work more like a
normal file system.
These flags control the VFS file caching options. File caching is
necessary to make the VFS layer appear compatible with a normal file
system. It can be disabled at the cost of some compatibility.
You'll need to enable VFS caching if you want, for example, to read
and write simultaneously to a file. See below for more details.
For example you'll need to enable VFS caching if you want to read and
write simultaneously to a file. See below for more details.
Note that the VFS cache works in addition to the cache backend and you
may find that you need one or the other or both.
Note that the VFS cache is separate from the cache backend and you may
find that you need one or the other or both.
--cache-dir string Directory rclone will use for caching.
--vfs-cache-mode CacheMode Cache mode off|minimal|writes|full (default off)
--vfs-cache-max-age duration Max age of objects in the cache. (default 1h0m0s)
--vfs-cache-mode string Cache mode off|minimal|writes|full (default "off")
--vfs-cache-max-size SizeSuffix Max total size of objects in the cache. (default off)
--vfs-cache-poll-interval duration Interval to poll the cache for stale objects. (default 1m0s)
--vfs-cache-max-size int Max total size of objects in the cache. (default off)
--vfs-write-back duration Time to writeback files after last use when using cache. (default 5s)
If run with `-vv` rclone will print the location of the file cache. The
files are stored in the user cache file area which is OS dependent but
@ -114,9 +134,10 @@ The higher the cache mode the more compatible rclone becomes at the
cost of using disk space.
Note that files are written back to the remote only when they are
closed so if rclone is quit or dies with open files then these won't
get written back to the remote. However they will still be in the on
disk cache.
closed and if they haven't been accessed for --vfs-write-back
second. If rclone is quit or dies with files that haven't been
uploaded, these will be uploaded next time rclone is run with the same
flags.
If using --vfs-cache-max-size note that the cache may exceed this size
for two reasons. Firstly because it is only checked every
@ -125,7 +146,7 @@ evicted from the cache.
### --vfs-cache-mode off
In this mode the cache will read directly from the remote and write
In this mode (the default) the cache will read directly from the remote and write
directly to the remote without caching anything on disk.
This will mean some operations are not possible
@ -141,7 +162,7 @@ This will mean some operations are not possible
### --vfs-cache-mode minimal
This is very similar to "off" except that files opened for read AND
write will be buffered to disks. This means that files opened for
write will be buffered to disk. This means that files opened for
write will be a lot more compatible, but uses the minimal disk space.
These operations are not possible
@ -159,32 +180,72 @@ first.
This mode should support all normal file system operations.
If an upload fails it will be retried up to --low-level-retries times.
If an upload fails it will be retried at exponentially increasing
intervals up to 1 minute.
### --vfs-cache-mode full
In this mode all reads and writes are buffered to and from disk. When
a file is opened for read it will be downloaded in its entirety first.
In this mode all reads and writes are buffered to and from disk. When
data is read from the remote this is buffered to disk as well.
This may be appropriate for your needs, or you may prefer to look at
the cache backend which does a much more sophisticated job of caching,
including caching directory hierarchies and chunks of files.
In this mode the files in the cache will be sparse files and rclone
will keep track of which bits of the files it has dowloaded.
In this mode, unlike the others, when a file is written to the disk,
it will be kept on the disk after it is written to the remote. It
will be purged on a schedule according to `--vfs-cache-max-age`.
So if an application only reads the starts of each file, then rclone
will only buffer the start of the file. These files will appear to be
their full size in the cache, but they will be sparse files with only
the data that has been downloaded present in them.
This mode should support all normal file system operations.
This mode should support all normal file system operations and is
otherwise identical to --vfs-cache-mode writes.
If an upload or download fails it will be retried up to
--low-level-retries times.
When reading a file rclone will read --buffer-size plus
--vfs-read-ahead bytes ahead. The --buffer-size is buffered in memory
whereas the --vfs-read-ahead is buffered on disk.
## Case Sensitivity
When using this mode it is recommended that --buffer-size is not set
too big and --vfs-read-ahead is set large if required.
## VFS Performance
These flags may be used to enable/disable features of the VFS for
performance or other reasons.
In particular S3 and Swift benefit hugely from the --no-modtime flag
(or use --use-server-modtime for a slightly different effect) as each
read of the modification time takes a transaction.
--no-checksum Don't compare checksums on up/download.
--no-modtime Don't read/write the modification time (can speed things up).
--no-seek Don't allow seeking in files.
--read-only Mount read-only.
When rclone reads files from a remote it reads them in chunks. This
means that rather than requesting the whole file rclone reads the
chunk specified. This is advantageous because some cloud providers
account for reads being all the data requested, not all the data
delivered.
Rclone will keep doubling the chunk size requested starting at
--vfs-read-chunk-size with a maximum of --vfs-read-chunk-size-limit
unless it is set to "off" in which case there will be no limit.
--vfs-read-chunk-size SizeSuffix Read the source objects in chunks. (default 128M)
--vfs-read-chunk-size-limit SizeSuffix Max chunk doubling size (default "off")
Sometimes rclone is delivered reads or writes out of order. Rather
than seeking rclone will wait a short time for the in sequence read or
write to come in. These flags only come into effect when not using an
on disk cache file.
--vfs-read-wait duration Time to wait for in-sequence read before seeking. (default 20ms)
--vfs-write-wait duration Time to wait for in-sequence write before giving error. (default 1s)
## VFS Case Sensitivity
Linux file systems are case-sensitive: two files can differ only
by case, and the exact case must be used when opening a file.
Windows is not like most other operating systems supported by rclone.
File systems in modern Windows are case-insensitive but case-preserving:
although existing files can be opened using any case, the exact case used
to create the file is preserved and available for programs to query.
@ -195,7 +256,7 @@ file systems case-sensitive but that is not the default
The "--vfs-case-insensitive" mount flag controls how rclone handles these
two cases. If its value is "false", rclone passes file names to the mounted
file system as is. If the flag is "true" (or appears without a value on
file system as-is. If the flag is "true" (or appears without a value on
command line), rclone may perform a "fixup" as explained below.
The user may specify a file name to open/delete/rename/etc with a case
@ -211,7 +272,7 @@ Note that case sensitivity of the operating system running rclone (the target)
may differ from case sensitivity of a file system mounted by rclone (the source).
The flag controls whether "fixup" is performed to satisfy the target.
If the flag is not provided on command line, then its default value depends
If the flag is not provided on the command line, then its default value depends
on the operating system where rclone runs: "true" on Windows and macOS, "false"
otherwise. If the flag is provided without a value, then it is "true".
@ -328,9 +389,11 @@ rclone serve sftp remote:path [flags]
--vfs-cache-mode CacheMode Cache mode off|minimal|writes|full (default off)
--vfs-cache-poll-interval duration Interval to poll the cache for stale objects. (default 1m0s)
--vfs-case-insensitive If a file name not found, find a case insensitive match.
--vfs-read-ahead SizeSuffix Extra read ahead over --buffer-size when using cache-mode full.
--vfs-read-chunk-size SizeSuffix Read the source objects in chunks. (default 128M)
--vfs-read-chunk-size-limit SizeSuffix If greater than --vfs-read-chunk-size, double the chunk size after each chunk read, until the limit is reached. 'off' is unlimited. (default off)
--vfs-read-wait duration Time to wait for in-sequence read before seeking. (default 20ms)
--vfs-write-back duration Time to writeback files after last use when using cache. (default 5s)
--vfs-write-wait duration Time to wait for in-sequence write before giving error. (default 1s)
```

View File

@ -112,20 +112,39 @@ of that with the CA certificate. --key should be the PEM encoded
private key and --client-ca should be the PEM encoded client
certificate authority certificate.
## Directory Cache
## VFS - Virtual File System
Using the `--dir-cache-time` flag, you can set how long a
This command uses the VFS layer. This adapts the cloud storage objects
that rclone uses into something which looks much more like a disk
filing system.
Cloud storage objects have lots of properties which aren't like disk
files - you can't extend them or write to the middle of them, so the
VFS layer has to deal with that. Because there is no one right way of
doing this there are various options explained below.
The VFS layer also implements a directory cache - this caches info
about files and directories (but not the data) in memory.
## VFS Directory Cache
Using the `--dir-cache-time` flag, you can control how long a
directory should be considered up to date and not refreshed from the
backend. Changes made locally in the mount may appear immediately or
invalidate the cache. However, changes done on the remote will only
be picked up once the cache expires if the backend configured does not
support polling for changes. If the backend supports polling, changes
will be picked up on within the polling interval.
backend. Changes made through the mount will appear immediately or
invalidate the cache.
Alternatively, you can send a `SIGHUP` signal to rclone for
it to flush all directory caches, regardless of how old they are.
Assuming only one rclone instance is running, you can reset the cache
like this:
--dir-cache-time duration Time to cache directory entries for. (default 5m0s)
--poll-interval duration Time to wait between polling for changes.
However, changes made directly on the cloud storage by the web
interface or a different copy of rclone will only be picked up once
the directory cache expires if the backend configured does not support
polling for changes. If the backend supports polling, changes will be
picked up within the polling interval.
You can send a `SIGHUP` signal to rclone for it to flush all
directory caches, regardless of how old they are. Assuming only one
rclone instance is running, you can reset the cache like this:
kill -SIGHUP $(pidof rclone)
@ -138,40 +157,41 @@ Or individual files or directories:
rclone rc vfs/forget file=path/to/file dir=path/to/dir
## File Buffering
## VFS File Buffering
The `--buffer-size` flag determines the amount of memory,
that will be used to buffer data in advance.
Each open file descriptor will try to keep the specified amount of
data in memory at all times. The buffered data is bound to one file
descriptor and won't be shared between multiple open file descriptors
of the same file.
Each open file will try to keep the specified amount of data in memory
at all times. The buffered data is bound to one open file and won't be
shared.
This flag is a upper limit for the used memory per open file. The
buffer will only use memory for data that is downloaded but not not
yet read. If the buffer is empty, only a small amount of memory will
be used.
This flag is a upper limit for the used memory per file descriptor.
The buffer will only use memory for data that is downloaded but not
not yet read. If the buffer is empty, only a small amount of memory
will be used.
The maximum memory used by rclone for buffering can be up to
`--buffer-size * open files`.
## File Caching
## VFS File Caching
These flags control the VFS file caching options. The VFS layer is
used by rclone mount to make a cloud storage system work more like a
normal file system.
These flags control the VFS file caching options. File caching is
necessary to make the VFS layer appear compatible with a normal file
system. It can be disabled at the cost of some compatibility.
You'll need to enable VFS caching if you want, for example, to read
and write simultaneously to a file. See below for more details.
For example you'll need to enable VFS caching if you want to read and
write simultaneously to a file. See below for more details.
Note that the VFS cache works in addition to the cache backend and you
may find that you need one or the other or both.
Note that the VFS cache is separate from the cache backend and you may
find that you need one or the other or both.
--cache-dir string Directory rclone will use for caching.
--vfs-cache-mode CacheMode Cache mode off|minimal|writes|full (default off)
--vfs-cache-max-age duration Max age of objects in the cache. (default 1h0m0s)
--vfs-cache-mode string Cache mode off|minimal|writes|full (default "off")
--vfs-cache-max-size SizeSuffix Max total size of objects in the cache. (default off)
--vfs-cache-poll-interval duration Interval to poll the cache for stale objects. (default 1m0s)
--vfs-cache-max-size int Max total size of objects in the cache. (default off)
--vfs-write-back duration Time to writeback files after last use when using cache. (default 5s)
If run with `-vv` rclone will print the location of the file cache. The
files are stored in the user cache file area which is OS dependent but
@ -183,9 +203,10 @@ The higher the cache mode the more compatible rclone becomes at the
cost of using disk space.
Note that files are written back to the remote only when they are
closed so if rclone is quit or dies with open files then these won't
get written back to the remote. However they will still be in the on
disk cache.
closed and if they haven't been accessed for --vfs-write-back
second. If rclone is quit or dies with files that haven't been
uploaded, these will be uploaded next time rclone is run with the same
flags.
If using --vfs-cache-max-size note that the cache may exceed this size
for two reasons. Firstly because it is only checked every
@ -194,7 +215,7 @@ evicted from the cache.
### --vfs-cache-mode off
In this mode the cache will read directly from the remote and write
In this mode (the default) the cache will read directly from the remote and write
directly to the remote without caching anything on disk.
This will mean some operations are not possible
@ -210,7 +231,7 @@ This will mean some operations are not possible
### --vfs-cache-mode minimal
This is very similar to "off" except that files opened for read AND
write will be buffered to disks. This means that files opened for
write will be buffered to disk. This means that files opened for
write will be a lot more compatible, but uses the minimal disk space.
These operations are not possible
@ -228,32 +249,72 @@ first.
This mode should support all normal file system operations.
If an upload fails it will be retried up to --low-level-retries times.
If an upload fails it will be retried at exponentially increasing
intervals up to 1 minute.
### --vfs-cache-mode full
In this mode all reads and writes are buffered to and from disk. When
a file is opened for read it will be downloaded in its entirety first.
In this mode all reads and writes are buffered to and from disk. When
data is read from the remote this is buffered to disk as well.
This may be appropriate for your needs, or you may prefer to look at
the cache backend which does a much more sophisticated job of caching,
including caching directory hierarchies and chunks of files.
In this mode the files in the cache will be sparse files and rclone
will keep track of which bits of the files it has dowloaded.
In this mode, unlike the others, when a file is written to the disk,
it will be kept on the disk after it is written to the remote. It
will be purged on a schedule according to `--vfs-cache-max-age`.
So if an application only reads the starts of each file, then rclone
will only buffer the start of the file. These files will appear to be
their full size in the cache, but they will be sparse files with only
the data that has been downloaded present in them.
This mode should support all normal file system operations.
This mode should support all normal file system operations and is
otherwise identical to --vfs-cache-mode writes.
If an upload or download fails it will be retried up to
--low-level-retries times.
When reading a file rclone will read --buffer-size plus
--vfs-read-ahead bytes ahead. The --buffer-size is buffered in memory
whereas the --vfs-read-ahead is buffered on disk.
## Case Sensitivity
When using this mode it is recommended that --buffer-size is not set
too big and --vfs-read-ahead is set large if required.
## VFS Performance
These flags may be used to enable/disable features of the VFS for
performance or other reasons.
In particular S3 and Swift benefit hugely from the --no-modtime flag
(or use --use-server-modtime for a slightly different effect) as each
read of the modification time takes a transaction.
--no-checksum Don't compare checksums on up/download.
--no-modtime Don't read/write the modification time (can speed things up).
--no-seek Don't allow seeking in files.
--read-only Mount read-only.
When rclone reads files from a remote it reads them in chunks. This
means that rather than requesting the whole file rclone reads the
chunk specified. This is advantageous because some cloud providers
account for reads being all the data requested, not all the data
delivered.
Rclone will keep doubling the chunk size requested starting at
--vfs-read-chunk-size with a maximum of --vfs-read-chunk-size-limit
unless it is set to "off" in which case there will be no limit.
--vfs-read-chunk-size SizeSuffix Read the source objects in chunks. (default 128M)
--vfs-read-chunk-size-limit SizeSuffix Max chunk doubling size (default "off")
Sometimes rclone is delivered reads or writes out of order. Rather
than seeking rclone will wait a short time for the in sequence read or
write to come in. These flags only come into effect when not using an
on disk cache file.
--vfs-read-wait duration Time to wait for in-sequence read before seeking. (default 20ms)
--vfs-write-wait duration Time to wait for in-sequence write before giving error. (default 1s)
## VFS Case Sensitivity
Linux file systems are case-sensitive: two files can differ only
by case, and the exact case must be used when opening a file.
Windows is not like most other operating systems supported by rclone.
File systems in modern Windows are case-insensitive but case-preserving:
although existing files can be opened using any case, the exact case used
to create the file is preserved and available for programs to query.
@ -264,7 +325,7 @@ file systems case-sensitive but that is not the default
The "--vfs-case-insensitive" mount flag controls how rclone handles these
two cases. If its value is "false", rclone passes file names to the mounted
file system as is. If the flag is "true" (or appears without a value on
file system as-is. If the flag is "true" (or appears without a value on
command line), rclone may perform a "fixup" as explained below.
The user may specify a file name to open/delete/rename/etc with a case
@ -280,7 +341,7 @@ Note that case sensitivity of the operating system running rclone (the target)
may differ from case sensitivity of a file system mounted by rclone (the source).
The flag controls whether "fixup" is performed to satisfy the target.
If the flag is not provided on command line, then its default value depends
If the flag is not provided on the command line, then its default value depends
on the operating system where rclone runs: "true" on Windows and macOS, "false"
otherwise. If the flag is provided without a value, then it is "true".
@ -406,9 +467,11 @@ rclone serve webdav remote:path [flags]
--vfs-cache-mode CacheMode Cache mode off|minimal|writes|full (default off)
--vfs-cache-poll-interval duration Interval to poll the cache for stale objects. (default 1m0s)
--vfs-case-insensitive If a file name not found, find a case insensitive match.
--vfs-read-ahead SizeSuffix Extra read ahead over --buffer-size when using cache-mode full.
--vfs-read-chunk-size SizeSuffix Read the source objects in chunks. (default 128M)
--vfs-read-chunk-size-limit SizeSuffix If greater than --vfs-read-chunk-size, double the chunk size after each chunk read, until the limit is reached. 'off' is unlimited. (default off)
--vfs-read-wait duration Time to wait for in-sequence read before seeking. (default 20ms)
--vfs-write-back duration Time to writeback files after last use when using cache. (default 5s)
--vfs-write-wait duration Time to wait for in-sequence write before giving error. (default 1s)
```

View File

@ -18,7 +18,9 @@ modification time or MD5SUM. Destination is updated to match
source, including deleting files if necessary.
**Important**: Since this can cause data loss, test first with the
`--dry-run` flag to see exactly what would be copied and deleted.
`--dry-run` or the `--interactive`/`-i` flag.
rclone sync -i SOURCE remote:DESTINATION
Note that files in the destination won't be deleted if there were any
errors at any point.

View File

@ -23,6 +23,7 @@ time instead of the current time. Times may be specified as one of:
- 'YYMMDD' - eg. 17.10.30
- 'YYYY-MM-DDTHH:MM:SS' - eg. 2006-01-02T15:04:05
- 'YYYY-MM-DDTHH:MM:SS.SSS' - eg. 2006-01-02T15:04:05.123456789
Note that --timestamp is in UTC if you want local time then add the
--localtime flag.

View File

@ -52,7 +52,7 @@ rclone tree remote:path [flags]
--human Print the size in a more human readable way.
--level int Descend only level directories deep.
-D, --modtime Print the date of last modification.
-i, --noindent Don't print indentation lines.
--noindent Don't print indentation lines.
--noreport Turn off file/directory count at end of tree listing.
-o, --output string Output to file instead of stdout.
-p, --protections Print the protections for each file.

View File

@ -328,6 +328,8 @@ NB If filename_encryption is "off" then this option will do nothing.
Password or pass phrase for encryption.
**NB** Input to this must be obscured - see [rclone obscure](/commands/rclone_obscure/).
- Config: password
- Env Var: RCLONE_CRYPT_PASSWORD
- Type: string
@ -338,6 +340,8 @@ Password or pass phrase for encryption.
Password or pass phrase for salt. Optional but recommended.
Should be different to the previous password.
**NB** Input to this must be obscured - see [rclone obscure](/commands/rclone_obscure/).
- Config: password2
- Env Var: RCLONE_CRYPT_PASSWORD2
- Type: string
@ -347,6 +351,24 @@ Should be different to the previous password.
Here are the advanced options specific to crypt (Encrypt/Decrypt a remote).
#### --crypt-server-side-across-configs
Allow server side operations (eg copy) to work across different crypt configs.
Normally this option is not what you want, but if you have two crypts
pointing to the same backend you can use it.
This can be used, for example, to change file name encryption type
without re-uploading all the data. Just make two crypt backends
pointing to two different directories with the single changed
parameter and use rclone move to move the files between the crypt
remotes.
- Config: server_side_across_configs
- Env Var: RCLONE_CRYPT_SERVER_SIDE_ACROSS_CONFIGS
- Type: bool
- Default: false
#### --crypt-show-mapping
For all files listed show how the names encrypt.

View File

@ -547,10 +547,8 @@ Here are the standard options specific to drive (Google Drive).
#### --drive-client-id
Google Application Client Id
Setting your own is recommended.
See https://rclone.org/drive/#making-your-own-client-id for how to create your own.
If you leave this blank, it will use an internal key which is low performance.
OAuth Client Id
Leave blank normally.
- Config: client_id
- Env Var: RCLONE_DRIVE_CLIENT_ID
@ -559,8 +557,8 @@ If you leave this blank, it will use an internal key which is low performance.
#### --drive-client-secret
Google Application Client Secret
Setting your own is recommended.
OAuth Client Secret
Leave blank normally.
- Config: client_secret
- Env Var: RCLONE_DRIVE_CLIENT_SECRET
@ -599,9 +597,6 @@ Leave blank normally.
Fill in to access "Computers" folders (see docs), or for rclone to use
a non root folder as its starting point.
Note that if this is blank, the first time rclone runs it will fill it
in with the ID of the root folder.
- Config: root_folder_id
- Env Var: RCLONE_DRIVE_ROOT_FOLDER_ID
@ -614,15 +609,56 @@ Service Account Credentials JSON file path
Leave blank normally.
Needed only if you want use SA instead of interactive login.
Leading `~` will be expanded in the file name as will environment variables such as `${RCLONE_CONFIG_DIR}`.
- Config: service_account_file
- Env Var: RCLONE_DRIVE_SERVICE_ACCOUNT_FILE
- Type: string
- Default: ""
#### --drive-alternate-export
Deprecated: no longer needed
- Config: alternate_export
- Env Var: RCLONE_DRIVE_ALTERNATE_EXPORT
- Type: bool
- Default: false
### Advanced Options
Here are the advanced options specific to drive (Google Drive).
#### --drive-token
OAuth Access Token as a JSON blob.
- Config: token
- Env Var: RCLONE_DRIVE_TOKEN
- Type: string
- Default: ""
#### --drive-auth-url
Auth server URL.
Leave blank to use the provider defaults.
- Config: auth_url
- Env Var: RCLONE_DRIVE_AUTH_URL
- Type: string
- Default: ""
#### --drive-token-url
Token server url.
Leave blank to use the provider defaults.
- Config: token_url
- Env Var: RCLONE_DRIVE_TOKEN_URL
- Type: string
- Default: ""
#### --drive-service-account-credentials
Service Account Credentials JSON blob
@ -719,6 +755,15 @@ This will show trashed files in their original directory structure.
- Type: bool
- Default: false
#### --drive-starred-only
Only show files that are starred.
- Config: starred_only
- Env Var: RCLONE_DRIVE_STARRED_ONLY
- Type: bool
- Default: false
#### --drive-formats
Deprecated: see export_formats
@ -1069,6 +1114,63 @@ Options:
- "target": optional target remote for the shortcut destination
#### drives
List the shared drives available to this account
rclone backend drives remote: [options] [<arguments>+]
This command lists the shared drives (teamdrives) available to this
account.
Usage:
rclone backend drives drive:
This will return a JSON list of objects like this
[
{
"id": "0ABCDEF-01234567890",
"kind": "drive#teamDrive",
"name": "My Drive"
},
{
"id": "0ABCDEFabcdefghijkl",
"kind": "drive#teamDrive",
"name": "Test Drive"
}
]
#### untrash
Untrash files and directories
rclone backend untrash remote: [options] [<arguments>+]
This command untrashes all the files and directories in the directory
passed in recursively.
Usage:
This takes an optional directory to trash which make this easier to
use via the API.
rclone backend untrash drive:directory
rclone backend -i untrash drive:directory subdir
Use the -i flag to see what would be restored before restoring it.
Result:
{
"Untrashed": 17,
"Errors": 0
}
{{< rem autogenerated options stop >}}
### Limitations ###

View File

@ -126,7 +126,7 @@ Here are the standard options specific to dropbox (Dropbox).
#### --dropbox-client-id
Dropbox App Client Id
OAuth Client Id
Leave blank normally.
- Config: client_id
@ -136,7 +136,7 @@ Leave blank normally.
#### --dropbox-client-secret
Dropbox App Client Secret
OAuth Client Secret
Leave blank normally.
- Config: client_secret
@ -148,6 +148,35 @@ Leave blank normally.
Here are the advanced options specific to dropbox (Dropbox).
#### --dropbox-token
OAuth Access Token as a JSON blob.
- Config: token
- Env Var: RCLONE_DROPBOX_TOKEN
- Type: string
- Default: ""
#### --dropbox-auth-url
Auth server URL.
Leave blank to use the provider defaults.
- Config: auth_url
- Env Var: RCLONE_DROPBOX_AUTH_URL
- Type: string
- Default: ""
#### --dropbox-token-url
Token server url.
Leave blank to use the provider defaults.
- Config: token_url
- Env Var: RCLONE_DROPBOX_TOKEN_URL
- Type: string
- Default: ""
#### --dropbox-chunk-size
Upload chunk size. (< 150M).

View File

@ -19,6 +19,7 @@ These flags are available for every command.
--bind string Local address to bind to for outgoing connections, IPv4, IPv6 or name.
--buffer-size SizeSuffix In memory buffer size when reading files for each --transfer. (default 16M)
--bwlimit BwTimetable Bandwidth limit in kBytes/s, or use suffix b|k|M|G or a full timetable.
--bwlimit-file BwTimetable Bandwidth limit per file in kBytes/s, or use suffix b|k|M|G or a full timetable.
--ca-cert string CA certificate used to verify servers
--cache-dir string Directory rclone will use for caching. (default "$HOME/.cache/rclone")
--check-first Do all the checks before starting transfers.
@ -64,6 +65,7 @@ These flags are available for every command.
--immutable Do not modify files. Fail if existing files have been modified.
--include stringArray Include files matching pattern
--include-from stringArray Read include patterns from file (use - to read from stdin)
-i, --interactive Enable interactive mode
--log-file string Log everything to this file
--log-format string Comma separated list of log format options (default "date,time")
--log-level string Log level DEBUG|INFO|NOTICE|ERROR (default "NOTICE")
@ -118,6 +120,7 @@ These flags are available for every command.
--rc-web-gui-force-update Force update to latest version of web gui
--rc-web-gui-no-open-browser Don't open the browser automatically
--rc-web-gui-update Check and update to latest version of web gui
--refresh-times Refresh the modtime of remote files.
--retries int Retry operations this many times if they fail (default 3)
--retries-sleep duration Interval between retrying operations if they fail, e.g 500ms, 60s, 5m. (0 to disable)
--size-only Skip based on size only, not mod-time or checksum
@ -137,14 +140,14 @@ These flags are available for every command.
--tpslimit float Limit HTTP transactions per second to this.
--tpslimit-burst int Max burst of transactions for --tpslimit. (default 1)
--track-renames When synchronizing, track file renames and do a server side move if possible
--track-renames-strategy string Strategies to use when synchronizing using track-renames hash|modtime (default "hash")
--track-renames-strategy string Strategies to use when synchronizing using track-renames hash|modtime|leaf (default "hash")
--transfers int Number of file transfers to run in parallel. (default 4)
-u, --update Skip files that are newer on the destination.
--use-cookies Enable session cookiejar.
--use-json-log Use json log format.
--use-mmap Use mmap allocator (see docs).
--use-server-modtime Use server modified time instead of object metadata
--user-agent string Set the user-agent to a specified string. The default is rclone/ version (default "rclone/v1.52.0")
--user-agent string Set the user-agent to a specified string. The default is rclone/ version (default "rclone/v1.53.0")
-v, --verbose count Print lots more stuff (repeat for more)
```
@ -155,10 +158,11 @@ and may be set in the config file.
```
--acd-auth-url string Auth server URL.
--acd-client-id string Amazon Application Client ID.
--acd-client-secret string Amazon Application Client Secret.
--acd-client-id string OAuth Client Id
--acd-client-secret string OAuth Client Secret
--acd-encoding MultiEncoder This sets the encoding for the backend. (default Slash,InvalidUtf8,Dot)
--acd-templink-threshold SizeSuffix Files >= this size will be downloaded via their tempLink. (default 9G)
--acd-token string OAuth Access Token as a JSON blob.
--acd-token-url string Token server url.
--acd-upload-wait-per-gb Duration Additional time per GB to wait after a failed complete upload to see if it appears. (default 3m0s)
--alias-remote string Remote or path to alias.
@ -177,6 +181,7 @@ and may be set in the config file.
--azureblob-use-emulator Uses local storage emulator if provided as 'true' (leave blank if using real azure storage endpoint)
--b2-account string Account ID or Application Key ID
--b2-chunk-size SizeSuffix Upload chunk size. Must fit in memory. (default 96M)
--b2-copy-cutoff SizeSuffix Cutoff for switching to multipart copy (default 4G)
--b2-disable-checksum Disable checksums for large (> upload cutoff) files
--b2-download-auth-duration Duration Time before the authorization token will expire in s or suffix ms|s|m|h|d. (default 1w)
--b2-download-url string Custom endpoint for downloads.
@ -184,16 +189,22 @@ and may be set in the config file.
--b2-endpoint string Endpoint for the service.
--b2-hard-delete Permanently delete files on remote removal, otherwise hide files.
--b2-key string Application Key
--b2-memory-pool-flush-time Duration How often internal memory buffer pools will be flushed. (default 1m0s)
--b2-memory-pool-use-mmap Whether to use mmap buffers in internal memory pool.
--b2-test-mode string A flag string for X-Bz-Test-Mode header for debugging.
--b2-upload-cutoff SizeSuffix Cutoff for switching to chunked upload. (default 200M)
--b2-versions Include old versions in directory listings.
--box-access-token string Box App Primary Access Token
--box-auth-url string Auth server URL.
--box-box-config-file string Box App config.json location
--box-box-sub-type string (default "user")
--box-client-id string Box App Client Id.
--box-client-secret string Box App Client Secret
--box-client-id string OAuth Client Id
--box-client-secret string OAuth Client Secret
--box-commit-retries int Max number of times to try committing a multipart file. (default 100)
--box-encoding MultiEncoder This sets the encoding for the backend. (default Slash,BackSlash,Del,Ctl,RightSpace,InvalidUtf8,Dot)
--box-root-folder-id string Fill in for rclone to use a non root folder as its starting point.
--box-token string OAuth Access Token as a JSON blob.
--box-token-url string Token server url.
--box-upload-cutoff SizeSuffix Cutoff for switching to multipart upload (>= 50MB). (default 50M)
--cache-chunk-clean-interval Duration How often should the cache perform cleanups of the chunk storage. (default 1m0s)
--cache-chunk-no-memory Disable the in-memory cache for storing chunks during streaming.
@ -205,7 +216,7 @@ and may be set in the config file.
--cache-db-wait-time Duration How long to wait for the DB to be available - 0 is unlimited (default 1s)
--cache-info-age Duration How long to cache file structure information (directory listings, file size, times etc). (default 6h0m0s)
--cache-plex-insecure string Skip all certificate verification when connecting to the Plex server
--cache-plex-password string The password of the Plex user
--cache-plex-password string The password of the Plex user (obscured)
--cache-plex-url string The URL of the Plex server
--cache-plex-username string The username of the Plex user
--cache-read-retries int How many times to retry a read from a cache storage. (default 10)
@ -225,17 +236,18 @@ and may be set in the config file.
-L, --copy-links Follow symlinks and copy the pointed to item.
--crypt-directory-name-encryption Option to either encrypt directory names or leave them intact. (default true)
--crypt-filename-encryption string How to encrypt the filenames. (default "standard")
--crypt-password string Password or pass phrase for encryption.
--crypt-password2 string Password or pass phrase for salt. Optional but recommended.
--crypt-password string Password or pass phrase for encryption. (obscured)
--crypt-password2 string Password or pass phrase for salt. Optional but recommended. (obscured)
--crypt-remote string Remote to encrypt/decrypt.
--crypt-server-side-across-configs Allow server side operations (eg copy) to work across different crypt configs.
--crypt-show-mapping For all files listed show how the names encrypt.
--drive-acknowledge-abuse Set to allow files which return cannotDownloadAbusiveFile to be downloaded.
--drive-allow-import-name-change Allow the filetype to change when uploading Google docs (e.g. file.doc to file.docx). This will confuse sync and reupload every time.
--drive-alternate-export Use alternate export URLs for google documents export.,
--drive-auth-owner-only Only consider files owned by the authenticated user.
--drive-auth-url string Auth server URL.
--drive-chunk-size SizeSuffix Upload chunk size. Must a power of 2 >= 256k. (default 8M)
--drive-client-id string Google Application Client Id
--drive-client-secret string Google Application Client Secret
--drive-client-id string OAuth Client Id
--drive-client-secret string OAuth Client Secret
--drive-disable-http2 Disable drive using http2 (default true)
--drive-encoding MultiEncoder This sets the encoding for the backend. (default InvalidUtf8)
--drive-export-formats string Comma separated list of preferred formats for downloading Google docs. (default "docx,xlsx,pptx,svg")
@ -256,65 +268,81 @@ and may be set in the config file.
--drive-skip-checksum-gphotos Skip MD5 checksum on Google photos and videos only.
--drive-skip-gdocs Skip google documents in all listings.
--drive-skip-shortcuts If set skip shortcut files
--drive-starred-only Only show files that are starred.
--drive-stop-on-upload-limit Make upload limit errors be fatal
--drive-team-drive string ID of the Team Drive
--drive-token string OAuth Access Token as a JSON blob.
--drive-token-url string Token server url.
--drive-trashed-only Only show files that are in the trash.
--drive-upload-cutoff SizeSuffix Cutoff for switching to chunked upload (default 8M)
--drive-use-created-date Use file created date instead of modified date.,
--drive-use-shared-date Use date file was shared instead of modified date.
--drive-use-trash Send files to the trash instead of deleting permanently. (default true)
--drive-v2-download-min-size SizeSuffix If Object's are greater, use drive v2 API to download. (default off)
--dropbox-auth-url string Auth server URL.
--dropbox-chunk-size SizeSuffix Upload chunk size. (< 150M). (default 48M)
--dropbox-client-id string Dropbox App Client Id
--dropbox-client-secret string Dropbox App Client Secret
--dropbox-client-id string OAuth Client Id
--dropbox-client-secret string OAuth Client Secret
--dropbox-encoding MultiEncoder This sets the encoding for the backend. (default Slash,BackSlash,Del,RightSpace,InvalidUtf8,Dot)
--dropbox-impersonate string Impersonate this user when using a business account.
--dropbox-token string OAuth Access Token as a JSON blob.
--dropbox-token-url string Token server url.
--fichier-api-key string Your API Key, get it from https://1fichier.com/console/params.pl
--fichier-encoding MultiEncoder This sets the encoding for the backend. (default Slash,LtGt,DoubleQuote,SingleQuote,BackQuote,Dollar,BackSlash,Del,Ctl,LeftSpace,RightSpace,InvalidUtf8,Dot)
--fichier-shared-folder string If you want to download a shared folder, add this parameter
--ftp-concurrency int Maximum number of FTP simultaneous connections, 0 for unlimited
--ftp-disable-epsv Disable using EPSV even if server advertises support
--ftp-encoding MultiEncoder This sets the encoding for the backend. (default Slash,Del,Ctl,RightSpace,Dot)
--ftp-explicit-tls Use FTP over TLS (Explicit)
--ftp-host string FTP host to connect to
--ftp-no-check-certificate Do not verify the TLS certificate of the server
--ftp-pass string FTP password
--ftp-pass string FTP password (obscured)
--ftp-port string FTP port, leave blank to use default (21)
--ftp-tls Use FTP over TLS (Implicit)
--ftp-tls Use FTPS over TLS (Implicit)
--ftp-user string FTP username, leave blank for current username, $USER
--gcs-anonymous Access public buckets and objects without credentials
--gcs-auth-url string Auth server URL.
--gcs-bucket-acl string Access Control List for new buckets.
--gcs-bucket-policy-only Access checks should use bucket-level IAM policies.
--gcs-client-id string Google Application Client Id
--gcs-client-secret string Google Application Client Secret
--gcs-client-id string OAuth Client Id
--gcs-client-secret string OAuth Client Secret
--gcs-encoding MultiEncoder This sets the encoding for the backend. (default Slash,CrLf,InvalidUtf8,Dot)
--gcs-location string Location for the newly created buckets.
--gcs-object-acl string Access Control List for new objects.
--gcs-project-number string Project number.
--gcs-service-account-file string Service Account Credentials JSON file path
--gcs-storage-class string The storage class to use when storing objects in Google Cloud Storage.
--gphotos-client-id string Google Application Client Id
--gphotos-client-secret string Google Application Client Secret
--gcs-token string OAuth Access Token as a JSON blob.
--gcs-token-url string Token server url.
--gphotos-auth-url string Auth server URL.
--gphotos-client-id string OAuth Client Id
--gphotos-client-secret string OAuth Client Secret
--gphotos-read-only Set to make the Google Photos backend read only.
--gphotos-read-size Set to read the size of media items.
--gphotos-start-year int Year limits the photos to be downloaded to those which are uploaded after the given year (default 2000)
--gphotos-token string OAuth Access Token as a JSON blob.
--gphotos-token-url string Token server url.
--http-headers CommaSepList Set HTTP headers for all transactions
--http-no-head Don't use HEAD requests to find file sizes in dir listing
--http-no-slash Set this if the site doesn't end directories with /
--http-url string URL of http host to connect to
--hubic-auth-url string Auth server URL.
--hubic-chunk-size SizeSuffix Above this size files will be chunked into a _segments container. (default 5G)
--hubic-client-id string Hubic Client Id
--hubic-client-secret string Hubic Client Secret
--hubic-client-id string OAuth Client Id
--hubic-client-secret string OAuth Client Secret
--hubic-encoding MultiEncoder This sets the encoding for the backend. (default Slash,InvalidUtf8)
--hubic-no-chunk Don't chunk files during streaming upload.
--hubic-token string OAuth Access Token as a JSON blob.
--hubic-token-url string Token server url.
--jottacloud-encoding MultiEncoder This sets the encoding for the backend. (default Slash,LtGt,DoubleQuote,Colon,Question,Asterisk,Pipe,Del,Ctl,InvalidUtf8,Dot)
--jottacloud-hard-delete Delete files permanently rather than putting them into the trash.
--jottacloud-md5-memory-limit SizeSuffix Files bigger than this will be cached on disk to calculate the MD5 if required. (default 10M)
--jottacloud-trashed-only Only show files that are in the trash.
--jottacloud-unlink Remove existing public link to file/folder with link command rather than creating.
--jottacloud-upload-resume-limit SizeSuffix Files bigger than this can be resumed if the upload fail's. (default 10M)
--koofr-encoding MultiEncoder This sets the encoding for the backend. (default Slash,BackSlash,Del,Ctl,InvalidUtf8,Dot)
--koofr-endpoint string The Koofr API endpoint to use (default "https://app.koofr.net")
--koofr-mountid string Mount ID of the mount to use. If omitted, the primary mount is used.
--koofr-password string Your Koofr password for rclone (generate one at https://app.koofr.net/app/admin/preferences/password)
--koofr-password string Your Koofr password for rclone (generate one at https://app.koofr.net/app/admin/preferences/password) (obscured)
--koofr-setmtime Does the backend support setting modification time. Set this to false if you use a mount ID that points to a Dropbox or Amazon Drive backend. (default true)
--koofr-user string Your Koofr user name
-l, --links Translate symlinks to/from regular files with a '.rclonelink' extension
@ -322,12 +350,13 @@ and may be set in the config file.
--local-case-sensitive Force the filesystem to report itself as case sensitive.
--local-encoding MultiEncoder This sets the encoding for the backend. (default Slash,Dot)
--local-no-check-updated Don't check to see if the files change during upload
--local-no-set-modtime Disable setting modtime
--local-no-sparse Disable sparse files for multi-thread downloads
--local-no-unicode-normalization Don't apply unicode normalization to paths and filenames (Deprecated)
--local-nounc string Disable UNC (long path names) conversion on Windows
--mailru-check-hash What should copy do if file checksum is mismatched or invalid (default true)
--mailru-encoding MultiEncoder This sets the encoding for the backend. (default Slash,LtGt,DoubleQuote,Colon,Question,Asterisk,Pipe,BackSlash,Del,Ctl,InvalidUtf8,Dot)
--mailru-pass string Password
--mailru-pass string Password (obscured)
--mailru-speedup-enable Skip full upload if there is another file with same data hash. (default true)
--mailru-speedup-file-patterns string Comma separated list of file name patterns eligible for speedup (put by hash). (default "*.mkv,*.avi,*.mp4,*.mp3,*.zip,*.gz,*.rar,*.pdf")
--mailru-speedup-max-disk SizeSuffix This option allows you to disable speedup (put by hash) for large files (default 3G)
@ -336,25 +365,33 @@ and may be set in the config file.
--mega-debug Output more debug from Mega.
--mega-encoding MultiEncoder This sets the encoding for the backend. (default Slash,InvalidUtf8,Dot)
--mega-hard-delete Delete files permanently rather than putting them into the trash.
--mega-pass string Password.
--mega-pass string Password. (obscured)
--mega-user string User name
-x, --one-file-system Don't cross filesystem boundaries (unix/macOS only).
--onedrive-auth-url string Auth server URL.
--onedrive-chunk-size SizeSuffix Chunk size to upload files with - must be multiple of 320k (327,680 bytes). (default 10M)
--onedrive-client-id string Microsoft App Client Id
--onedrive-client-secret string Microsoft App Client Secret
--onedrive-client-id string OAuth Client Id
--onedrive-client-secret string OAuth Client Secret
--onedrive-drive-id string The ID of the drive to use
--onedrive-drive-type string The type of the drive ( personal | business | documentLibrary )
--onedrive-encoding MultiEncoder This sets the encoding for the backend. (default Slash,LtGt,DoubleQuote,Colon,Question,Asterisk,Pipe,Hash,Percent,BackSlash,Del,Ctl,LeftSpace,LeftTilde,RightSpace,RightPeriod,InvalidUtf8,Dot)
--onedrive-expose-onenote-files Set to make OneNote files show up in directory listings.
--onedrive-no-versions Remove all versions on modifying operations
--onedrive-server-side-across-configs Allow server side operations (eg copy) to work across different onedrive configs.
--onedrive-token string OAuth Access Token as a JSON blob.
--onedrive-token-url string Token server url.
--opendrive-chunk-size SizeSuffix Files will be uploaded in chunks this size. (default 10M)
--opendrive-encoding MultiEncoder This sets the encoding for the backend. (default Slash,LtGt,DoubleQuote,Colon,Question,Asterisk,Pipe,BackSlash,LeftSpace,LeftCrLfHtVt,RightSpace,RightCrLfHtVt,InvalidUtf8,Dot)
--opendrive-password string Password.
--opendrive-password string Password. (obscured)
--opendrive-username string Username
--pcloud-client-id string Pcloud App Client Id
--pcloud-client-secret string Pcloud App Client Secret
--pcloud-auth-url string Auth server URL.
--pcloud-client-id string OAuth Client Id
--pcloud-client-secret string OAuth Client Secret
--pcloud-encoding MultiEncoder This sets the encoding for the backend. (default Slash,BackSlash,Del,Ctl,InvalidUtf8,Dot)
--pcloud-hostname string Hostname to connect to. (default "api.pcloud.com")
--pcloud-root-folder-id string Fill in for rclone to use a non root folder as its starting point. (default "d0")
--pcloud-token string OAuth Access Token as a JSON blob.
--pcloud-token-url string Token server url.
--premiumizeme-encoding MultiEncoder This sets the encoding for the backend. (default Slash,DoubleQuote,BackSlash,Del,Ctl,InvalidUtf8,Dot)
--putio-encoding MultiEncoder This sets the encoding for the backend. (default Slash,BackSlash,Del,Ctl,InvalidUtf8,Dot)
--qingstor-access-key-id string QingStor Access Key ID
@ -371,7 +408,7 @@ and may be set in the config file.
--s3-acl string Canned ACL used when creating buckets and storing or copying objects.
--s3-bucket-acl string Canned ACL used when creating buckets.
--s3-chunk-size SizeSuffix Chunk size to use for uploading. (default 5M)
--s3-copy-cutoff SizeSuffix Cutoff for switching to multipart copy (default 5G)
--s3-copy-cutoff SizeSuffix Cutoff for switching to multipart copy (default 4.656G)
--s3-disable-checksum Don't store MD5 checksum with object metadata
--s3-encoding MultiEncoder This sets the encoding for the backend. (default Slash,InvalidUtf8,Dot)
--s3-endpoint string Endpoint for S3 API.
@ -380,13 +417,17 @@ and may be set in the config file.
--s3-leave-parts-on-error If true avoid calling abort upload on a failure, leaving all successfully uploaded parts on S3 for manual recovery.
--s3-list-chunk int Size of listing chunk (response list for each ListObject S3 request). (default 1000)
--s3-location-constraint string Location constraint - must be set to match the Region.
--s3-max-upload-parts int Maximum number of parts in a multipart upload. (default 10000)
--s3-memory-pool-flush-time Duration How often internal memory buffer pools will be flushed. (default 1m0s)
--s3-memory-pool-use-mmap Whether to use mmap buffers in internal memory pool.
--s3-no-check-bucket If set don't attempt to check the bucket exists or create it
--s3-profile string Profile to use in the shared credentials file
--s3-provider string Choose your S3 provider.
--s3-region string Region to connect to.
--s3-secret-access-key string AWS Secret Access Key (password)
--s3-server-side-encryption string The server-side encryption algorithm used when storing this object in S3.
--s3-session-token string An AWS session token
--s3-shared-credentials-file string Path to the shared credentials file
--s3-sse-customer-algorithm string If using SSE-C, the server-side encryption algorithm used when storing this object in S3.
--s3-sse-customer-key string If using SSE-C you must provide the secret encryption key used to encrypt/decrypt your data.
--s3-sse-customer-key-md5 string If using SSE-C you must provide the secret encryption key MD5 checksum.
@ -400,24 +441,26 @@ and may be set in the config file.
--seafile-create-library Should rclone create a library if it doesn't exist
--seafile-encoding MultiEncoder This sets the encoding for the backend. (default Slash,DoubleQuote,BackSlash,Ctl,InvalidUtf8)
--seafile-library string Name of the library. Leave blank to access all non-encrypted libraries.
--seafile-library-key string Library password (for encrypted libraries only). Leave blank if you pass it through the command line.
--seafile-pass string Password
--seafile-library-key string Library password (for encrypted libraries only). Leave blank if you pass it through the command line. (obscured)
--seafile-pass string Password (obscured)
--seafile-url string URL of seafile host to connect to
--seafile-user string User name (usually email address)
--sftp-ask-password Allow asking for SFTP password when needed.
--sftp-disable-hashcheck Disable the execution of SSH commands to determine if remote file hashing is available.
--sftp-host string SSH host to connect to
--sftp-key-file string Path to PEM-encoded private key file, leave blank or set key-use-agent to use ssh-agent.
--sftp-key-file-pass string The passphrase to decrypt the PEM-encoded private key file.
--sftp-key-file-pass string The passphrase to decrypt the PEM-encoded private key file. (obscured)
--sftp-key-pem string Raw PEM-encoded private key, If specified, will override key_file parameter.
--sftp-key-use-agent When set forces the usage of the ssh-agent.
--sftp-md5sum-command string The command used to read md5 hashes. Leave blank for autodetect.
--sftp-pass string SSH password, leave blank to use ssh-agent.
--sftp-pass string SSH password, leave blank to use ssh-agent. (obscured)
--sftp-path-override string Override path used by SSH connection.
--sftp-port string SSH port, leave blank to use default (22)
--sftp-server-command string Specifies the path or command to run a sftp server on the remote host.
--sftp-set-modtime Set the modified time on the remote if set. (default true)
--sftp-sha1sum-command string The command used to read sha1 hashes. Leave blank for autodetect.
--sftp-skip-links Set to skip any symlinks and any other non regular files.
--sftp-subsystem string Specifies the SSH2 subsystem on the remote host. (default "sftp")
--sftp-use-insecure-cipher Enable the use of insecure ciphers and key exchange methods.
--sftp-user string SSH username, leave blank for current username, ncw
--sharefile-chunk-size SizeSuffix Upload chunk size. Must a power of 2 >= 256k. (default 64M)
@ -470,12 +513,14 @@ and may be set in the config file.
--union-upstreams string List of space separated upstreams.
--webdav-bearer-token string Bearer token instead of user/pass (eg a Macaroon)
--webdav-bearer-token-command string Command to run to get a bearer token
--webdav-pass string Password.
--webdav-pass string Password. (obscured)
--webdav-url string URL of http host to connect to
--webdav-user string User name
--webdav-vendor string Name of the Webdav site/service/software you are using
--yandex-client-id string Yandex Client Id
--yandex-client-secret string Yandex Client Secret
--yandex-auth-url string Auth server URL.
--yandex-client-id string OAuth Client Id
--yandex-client-secret string OAuth Client Secret
--yandex-encoding MultiEncoder This sets the encoding for the backend. (default Slash,Del,Ctl,InvalidUtf8,Dot)
--yandex-unlink Remove existing public link to file/folder with link command rather than creating.
--yandex-token string OAuth Access Token as a JSON blob.
--yandex-token-url string Token server url.
```

View File

@ -179,6 +179,8 @@ FTP port, leave blank to use default (21)
FTP password
**NB** Input to this must be obscured - see [rclone obscure](/commands/rclone_obscure/).
- Config: pass
- Env Var: RCLONE_FTP_PASS
- Type: string
@ -186,8 +188,7 @@ FTP password
#### --ftp-tls
Use FTP over TLS (Implicit)
Use FTPS over TLS (Implicit)
When using implicit FTP over TLS the client will connect using TLS
right from the start, which in turn breaks the compatibility with
non-TLS-aware servers. This is usually served over port 990 rather
@ -201,13 +202,12 @@ than port 21. Cannot be used in combination with explicit FTP.
#### --ftp-explicit-tls
Use FTP over TLS (Explicit)
When using explicit FTP over TLS the client explicitly request
security from the server in order to upgrade a plain text connection
to an encrypted one. Cannot be used in combination with implicit FTP.
- Config: explicit_tls
- Env Var: RCLONE_FTP_TLS
- Env Var: RCLONE_FTP_EXPLICIT_TLS
- Type: bool
- Default: false

View File

@ -264,7 +264,7 @@ Here are the standard options specific to google cloud storage (Google Cloud Sto
#### --gcs-client-id
Google Application Client Id
OAuth Client Id
Leave blank normally.
- Config: client_id
@ -274,7 +274,7 @@ Leave blank normally.
#### --gcs-client-secret
Google Application Client Secret
OAuth Client Secret
Leave blank normally.
- Config: client_secret
@ -298,6 +298,9 @@ Service Account Credentials JSON file path
Leave blank normally.
Needed only if you want use SA instead of interactive login.
Leading `~` will be expanded in the file name as will environment variables such as `${RCLONE_CONFIG_DIR}`.
- Config: service_account_file
- Env Var: RCLONE_GCS_SERVICE_ACCOUNT_FILE
- Type: string
@ -314,6 +317,16 @@ Needed only if you want use SA instead of interactive login.
- Type: string
- Default: ""
#### --gcs-anonymous
Access public buckets and objects without credentials
Set to 'true' if you just want to download files and don't configure credentials.
- Config: anonymous
- Env Var: RCLONE_GCS_ANONYMOUS
- Type: bool
- Default: false
#### --gcs-object-acl
Access Control List for new objects.
@ -455,6 +468,35 @@ The storage class to use when storing objects in Google Cloud Storage.
Here are the advanced options specific to google cloud storage (Google Cloud Storage (this is not Google Drive)).
#### --gcs-token
OAuth Access Token as a JSON blob.
- Config: token
- Env Var: RCLONE_GCS_TOKEN
- Type: string
- Default: ""
#### --gcs-auth-url
Auth server URL.
Leave blank to use the provider defaults.
- Config: auth_url
- Env Var: RCLONE_GCS_AUTH_URL
- Type: string
- Default: ""
#### --gcs-token-url
Token server url.
Leave blank to use the provider defaults.
- Config: token_url
- Env Var: RCLONE_GCS_TOKEN_URL
- Type: string
- Default: ""
#### --gcs-encoding
This sets the encoding for the backend.

View File

@ -315,7 +315,7 @@ Here are the standard options specific to google photos (Google Photos).
#### --gphotos-client-id
Google Application Client Id
OAuth Client Id
Leave blank normally.
- Config: client_id
@ -325,7 +325,7 @@ Leave blank normally.
#### --gphotos-client-secret
Google Application Client Secret
OAuth Client Secret
Leave blank normally.
- Config: client_secret
@ -349,6 +349,35 @@ to your photos, otherwise rclone will request full access.
Here are the advanced options specific to google photos (Google Photos).
#### --gphotos-token
OAuth Access Token as a JSON blob.
- Config: token
- Env Var: RCLONE_GPHOTOS_TOKEN
- Type: string
- Default: ""
#### --gphotos-auth-url
Auth server URL.
Leave blank to use the provider defaults.
- Config: auth_url
- Env Var: RCLONE_GPHOTOS_AUTH_URL
- Type: string
- Default: ""
#### --gphotos-token-url
Token server url.
Leave blank to use the provider defaults.
- Config: token_url
- Env Var: RCLONE_GPHOTOS_TOKEN_URL
- Type: string
- Default: ""
#### --gphotos-read-size
Set to read the size of media items.

View File

@ -112,7 +112,7 @@ Here are the standard options specific to hubic (Hubic).
#### --hubic-client-id
Hubic Client Id
OAuth Client Id
Leave blank normally.
- Config: client_id
@ -122,7 +122,7 @@ Leave blank normally.
#### --hubic-client-secret
Hubic Client Secret
OAuth Client Secret
Leave blank normally.
- Config: client_secret
@ -134,6 +134,35 @@ Leave blank normally.
Here are the advanced options specific to hubic (Hubic).
#### --hubic-token
OAuth Access Token as a JSON blob.
- Config: token
- Env Var: RCLONE_HUBIC_TOKEN
- Type: string
- Default: ""
#### --hubic-auth-url
Auth server URL.
Leave blank to use the provider defaults.
- Config: auth_url
- Env Var: RCLONE_HUBIC_AUTH_URL
- Type: string
- Default: ""
#### --hubic-token-url
Token server url.
Leave blank to use the provider defaults.
- Config: token_url
- Env Var: RCLONE_HUBIC_TOKEN_URL
- Type: string
- Default: ""
#### --hubic-chunk-size
Above this size files will be chunked into a _segments container.

View File

@ -220,16 +220,6 @@ Delete files permanently rather than putting them into the trash.
- Type: bool
- Default: false
#### --jottacloud-unlink
Remove existing public link to file/folder with link command rather than creating.
Default is false, meaning link command will create or retrieve public link.
- Config: unlink
- Env Var: RCLONE_JOTTACLOUD_UNLINK
- Type: bool
- Default: false
#### --jottacloud-upload-resume-limit
Files bigger than this can be resumed if the upload fail's.

View File

@ -115,6 +115,8 @@ Your Koofr user name
Your Koofr password for rclone (generate one at https://app.koofr.net/app/admin/preferences/password)
**NB** Input to this must be obscured - see [rclone obscure](/commands/rclone_obscure/).
- Config: password
- Env Var: RCLONE_KOOFR_PASSWORD
- Type: string

View File

@ -372,6 +372,23 @@ However on some file systems this modification time check may fail (eg
[Glusterfs #2206](https://github.com/rclone/rclone/issues/2206)) so this
check can be disabled with this flag.
If this flag is set, rclone will use its best efforts to transfer a
file which is being updated. If the file is only having things
appended to it (eg a log) then rclone will transfer the log file with
the size it had the first time rclone saw it.
If the file is being modified throughout (not just appended to) then
the transfer may fail with a hash check failure.
In detail, once the file has had stat() called on it for the first
time we:
- Only transfer the size that stat gave
- Only checksum the size that stat gave
- Don't update the stat info for the file
- Config: no_check_updated
- Env Var: RCLONE_LOCAL_NO_CHECK_UPDATED
- Type: bool
@ -426,6 +443,21 @@ cause disk fragmentation and can be slow to work with.
- Type: bool
- Default: false
#### --local-no-set-modtime
Disable setting modtime
Normally rclone updates modification time of files after they are done
uploading. This can cause permissions issues on Linux platforms when
the user rclone is running as does not own the file uploaded, such as
when copying to a CIFS mount owned by another user. If this option is
enabled, rclone will no longer update the modtime after copying a file.
- Config: no_set_modtime
- Env Var: RCLONE_LOCAL_NO_SET_MODTIME
- Type: bool
- Default: false
#### --local-encoding
This sets the encoding for the backend.

View File

@ -181,6 +181,8 @@ User name (usually email)
Password
**NB** Input to this must be obscured - see [rclone obscure](/commands/rclone_obscure/).
- Config: pass
- Env Var: RCLONE_MAILRU_PASS
- Type: string

View File

@ -168,6 +168,8 @@ User name
Password.
**NB** Input to this must be obscured - see [rclone obscure](/commands/rclone_obscure/).
- Config: pass
- Env Var: RCLONE_MEGA_PASS
- Type: string

View File

@ -199,7 +199,7 @@ Here are the standard options specific to onedrive (Microsoft OneDrive).
#### --onedrive-client-id
Microsoft App Client Id
OAuth Client Id
Leave blank normally.
- Config: client_id
@ -209,7 +209,7 @@ Leave blank normally.
#### --onedrive-client-secret
Microsoft App Client Secret
OAuth Client Secret
Leave blank normally.
- Config: client_secret
@ -221,6 +221,35 @@ Leave blank normally.
Here are the advanced options specific to onedrive (Microsoft OneDrive).
#### --onedrive-token
OAuth Access Token as a JSON blob.
- Config: token
- Env Var: RCLONE_ONEDRIVE_TOKEN
- Type: string
- Default: ""
#### --onedrive-auth-url
Auth server URL.
Leave blank to use the provider defaults.
- Config: auth_url
- Env Var: RCLONE_ONEDRIVE_AUTH_URL
- Type: string
- Default: ""
#### --onedrive-token-url
Token server url.
Leave blank to use the provider defaults.
- Config: token_url
- Env Var: RCLONE_ONEDRIVE_TOKEN_URL
- Type: string
- Default: ""
#### --onedrive-chunk-size
Chunk size to upload files with - must be multiple of 320k (327,680 bytes).
@ -281,6 +310,27 @@ configurations.
- Type: bool
- Default: false
#### --onedrive-no-versions
Remove all versions on modifying operations
Onedrive for business creates versions when rclone uploads new files
overwriting an existing one and when it sets the modification time.
These versions take up space out of the quota.
This flag checks for versions after file upload and setting
modification time and removes all but the last version.
**NB** Onedrive personal can't currently delete versions so don't use
this flag there.
- Config: no_versions
- Env Var: RCLONE_ONEDRIVE_NO_VERSIONS
- Type: bool
- Default: false
#### --onedrive-encoding
This sets the encoding for the backend.

View File

@ -116,6 +116,8 @@ Username
Password.
**NB** Input to this must be obscured - see [rclone obscure](/commands/rclone_obscure/).
- Config: password
- Env Var: RCLONE_OPENDRIVE_PASSWORD
- Type: string

View File

@ -140,7 +140,7 @@ Here are the standard options specific to pcloud (Pcloud).
#### --pcloud-client-id
Pcloud App Client Id
OAuth Client Id
Leave blank normally.
- Config: client_id
@ -150,7 +150,7 @@ Leave blank normally.
#### --pcloud-client-secret
Pcloud App Client Secret
OAuth Client Secret
Leave blank normally.
- Config: client_secret
@ -162,6 +162,35 @@ Leave blank normally.
Here are the advanced options specific to pcloud (Pcloud).
#### --pcloud-token
OAuth Access Token as a JSON blob.
- Config: token
- Env Var: RCLONE_PCLOUD_TOKEN
- Type: string
- Default: ""
#### --pcloud-auth-url
Auth server URL.
Leave blank to use the provider defaults.
- Config: auth_url
- Env Var: RCLONE_PCLOUD_AUTH_URL
- Type: string
- Default: ""
#### --pcloud-token-url
Token server url.
Leave blank to use the provider defaults.
- Config: token_url
- Env Var: RCLONE_PCLOUD_TOKEN_URL
- Type: string
- Default: ""
#### --pcloud-encoding
This sets the encoding for the backend.
@ -182,4 +211,23 @@ Fill in for rclone to use a non root folder as its starting point.
- Type: string
- Default: "d0"
#### --pcloud-hostname
Hostname to connect to.
This is normally set when rclone initially does the oauth connection,
however you will need to set it by hand if you are using remote config
with rclone authorize.
- Config: hostname
- Env Var: RCLONE_PCLOUD_HOSTNAME
- Type: string
- Default: "api.pcloud.com"
- Examples:
- "api.pcloud.com"
- Original/US region
- "eapi.pcloud.com"
- EU region
{{< rem autogenerated options stop >}}

View File

@ -504,6 +504,41 @@ except only one bandwidth may be specified.
In either case "rate" is returned as a human readable string, and
"bytesPerSecond" is returned as a number.
### core/command: Run a rclone terminal command over rc. {#core-command}
This takes the following parameters
- command - a string with the command name
- arg - a list of arguments for the backend command
- opt - a map of string to string of options
Returns
- result - result from the backend command
- error - set if rclone exits with an error code
- returnType - one of ("COMBINED_OUTPUT", "STREAM", "STREAM_ONLY_STDOUT". "STREAM_ONLY_STDERR")
For example
rclone rc core/command command=ls -a mydrive:/ -o max-depth=1
rclone rc core/command -a ls -a mydrive:/ -o max-depth=1
Returns
```
{
"error": false,
"result": "<Raw command line output>"
}
OR
{
"error": true,
"result": "<Raw command line output>"
}
**Authentication is required for this call.**
### core/gc: Runs a garbage collection. {#core-gc}
This tells the go runtime to do a garbage collection run. It isn't
@ -582,6 +617,7 @@ Returns the following values:
"transfers": number of transferred files,
"deletes" : number of deleted files,
"renames" : number of renamed files,
"transferTime" : total time spent on running jobs,
"elapsedTime": time in seconds since the start of the process,
"lastError": last occurred error,
"transferring": an array of currently active file transfers:
@ -591,8 +627,8 @@ Returns the following values:
"eta": estimated time in seconds until file transfer completion
"name": name of the file,
"percentage": progress of the file transfer in percent,
"speed": speed in bytes/sec,
"speedAvg": speed in bytes/sec as an exponentially weighted moving average,
"speed": average speed over the whole transfer in bytes/sec,
"speedAvg": current speed in bytes/sec as an exponentially weighted moving average,
"size": size of the file in bytes
}
],
@ -657,10 +693,10 @@ Returns the following values:
This shows the current version of go and the go runtime
- version - rclone version, eg "v1.44"
- decomposed - version number as [major, minor, patch, subpatch]
- note patch and subpatch will be 999 for a git compiled version
- version - rclone version, eg "v1.53.0"
- decomposed - version number as [major, minor, patch]
- isGit - boolean - true if this was compiled from the git version
- isBeta - boolean - true if this is a beta version
- os - OS in use as according to Go
- arch - cpu architecture in use according to Go
- goVersion - version of Go runtime in use
@ -736,6 +772,20 @@ Parameters
- jobid - id of the job (integer)
### mount/listmounts: Show current mount points {#mount-listmounts}
This shows currently mounted points, which can be used for performing an unmount
This takes no parameters and returns
- mountPoints: list of current mount points
Eg
rclone rc mount/listmounts
**Authentication is required for this call.**
### mount/mount: Create a new mount point {#mount-mount}
rclone allows Linux, FreeBSD, macOS and Windows to mount any of
@ -748,11 +798,19 @@ This takes the following parameters
- fs - a remote path to be mounted (required)
- mountPoint: valid path on the local machine (required)
- mountType: One of the values (mount, cmount, mount2) specifies the mount implementation to use
- mountOpt: a JSON object with Mount options in.
- vfsOpt: a JSON object with VFS options in.
Eg
rclone rc mount/mount fs=mydrive: mountPoint=/home/<user>/mountPoint
rclone rc mount/mount fs=mydrive: mountPoint=/home/<user>/mountPoint mountType=mount
rclone rc mount/mount fs=TestDrive: mountPoint=/mnt/tmp vfsOpt='{"CacheMode": 2}' mountOpt='{"AllowOther": true}'
The vfsOpt are as described in options/get and can be seen in the the
"vfs" section when running and the mountOpt can be seen in the "mount" section.
rclone rc options/get
**Authentication is required for this call.**
@ -773,7 +831,7 @@ Eg
**Authentication is required for this call.**
### mount/unmount: Unmount all active mounts {#mount-unmount}
### mount/unmount: Unmount selected active mount {#mount-unmount}
rclone allows Linux, FreeBSD, macOS and Windows to
mount any of Rclone's cloud storage systems as a file system with
@ -789,6 +847,18 @@ Eg
**Authentication is required for this call.**
### mount/unmountall: Show current mount points {#mount-unmountall}
This shows currently mounted points, which can be used for performing an unmount
This takes no parameters and returns error if unmount does not succeed.
Eg
rclone rc mount/unmountall
**Authentication is required for this call.**
### operations/about: Return the space used on the remote {#operations-about}
This takes the following parameters
@ -962,6 +1032,8 @@ This takes the following parameters
- fs - a remote name string eg "drive:"
- remote - a path within that remote eg "dir"
- unlink - boolean - if set removes the link rather than adding it (optional)
- expire - string - the expiry time of the link eg "1d" (optional)
Returns
@ -1020,6 +1092,17 @@ See the [size command](/commands/rclone_size/) command for more information on t
**Authentication is required for this call.**
### operations/uploadfile: Upload file using multiform/form-data {#operations-uploadfile}
This takes the following parameters
- fs - a remote name string eg "drive:"
- remote - a path within that remote eg "dir"
- each part in body represents a file to be uploaded
See the [uploadfile command](/commands/rclone_uploadfile/) command for more information on the above.
**Authentication is required for this call.**
### options/blocks: List all the option blocks {#options-blocks}
Returns
@ -1060,6 +1143,97 @@ And this sets NOTICE level logs (normal without -v)
rclone rc options/set --json '{"main": {"LogLevel": 6}}'
### pluginsctl/addPlugin: Add a plugin using url {#pluginsctl-addPlugin}
used for adding a plugin to the webgui
This takes the following parameters
- url: http url of the github repo where the plugin is hosted (http://github.com/rclone/rclone-webui-react)
Eg
rclone rc pluginsctl/addPlugin
**Authentication is required for this call.**
### pluginsctl/getPluginsForType: Get plugins with type criteria {#pluginsctl-getPluginsForType}
This shows all possible plugins by a mime type
This takes the following parameters
- type: supported mime type by a loaded plugin eg (video/mp4, audio/mp3)
- pluginType: filter plugins based on their type eg (DASHBOARD, FILE_HANDLER, TERMINAL)
and returns
- loadedPlugins: list of current production plugins
- testPlugins: list of temporarily loaded development plugins, usually running on a different server.
Eg
rclone rc pluginsctl/getPluginsForType type=video/mp4
**Authentication is required for this call.**
### pluginsctl/listPlugins: Get the list of currently loaded plugins {#pluginsctl-listPlugins}
This allows you to get the currently enabled plugins and their details.
This takes no parameters and returns
- loadedPlugins: list of current production plugins
- testPlugins: list of temporarily loaded development plugins, usually running on a different server.
Eg
rclone rc pluginsctl/listPlugins
**Authentication is required for this call.**
### pluginsctl/listTestPlugins: Show currently loaded test plugins {#pluginsctl-listTestPlugins}
allows listing of test plugins with the rclone.test set to true in package.json of the plugin
This takes no parameters and returns
- loadedTestPlugins: list of currently available test plugins
Eg
rclone rc pluginsctl/listTestPlugins
**Authentication is required for this call.**
### pluginsctl/removePlugin: Remove a loaded plugin {#pluginsctl-removePlugin}
This allows you to remove a plugin using it's name
This takes parameters
- name: name of the plugin in the format <author>/<plugin_name>
Eg
rclone rc pluginsctl/removePlugin name=rclone/video-plugin
**Authentication is required for this call.**
### pluginsctl/removeTestPlugin: Remove a test plugin {#pluginsctl-removeTestPlugin}
This allows you to remove a plugin using it's name
This takes the following parameters
- name: name of the plugin in the format <author>/<plugin_name>
Eg
rclone rc pluginsctl/removeTestPlugin name=rclone/rclone-webui-react
**Authentication is required for this call.**
### rc/error: This returns an error {#rc-error}
This returns an error with the input as part of its error string.
@ -1136,6 +1310,19 @@ parameter key starting with file will forget that file and any
starting with dir will forget that dir, eg
rclone rc vfs/forget file=hello file2=goodbye dir=home/junk
This command takes an "fs" parameter. If this parameter is not
supplied and if there is only one VFS in use then that VFS will be
used. If there is more than one VFS in use then the "fs" parameter
must be supplied.
### vfs/list: List active VFSes. {#vfs-list}
This lists the active VFSes.
It returns a list under the key "vfses" where the values are the VFS
names that could be passed to the other VFS commands in the "fs"
parameter.
### vfs/poll-interval: Get the status or update the value of the poll-interval option. {#vfs-poll-interval}
@ -1158,6 +1345,11 @@ not reached.
If poll-interval is updated or disabled temporarily, some changes
might not get picked up by the polling function, depending on the
used remote.
This command takes an "fs" parameter. If this parameter is not
supplied and if there is only one VFS in use then that VFS will be
used. If there is more than one VFS in use then the "fs" parameter
must be supplied.
### vfs/refresh: Refresh the directory cache. {#vfs-refresh}
@ -1175,6 +1367,11 @@ starting with dir will refresh that directory, eg
If the parameter recursive=true is given the whole directory tree
will get refreshed. This refresh will use --fast-list if enabled.
This command takes an "fs" parameter. If this parameter is not
supplied and if there is only one VFS in use then that VFS will be
used. If there is more than one VFS in use then the "fs" parameter
must be supplied.
{{< rem autogenerated stop >}}

View File

@ -482,6 +482,8 @@ Choose your S3 provider.
- Minio Object Storage
- "Netease"
- Netease Object Storage (NOS)
- "Scaleway"
- Scaleway Object Storage
- "StackPath"
- StackPath Object Storage
- "Wasabi"
@ -585,6 +587,20 @@ Region to connect to.
#### --s3-region
Region to connect to.
- Config: region
- Env Var: RCLONE_S3_REGION
- Type: string
- Default: ""
- Examples:
- "nl-ams"
- Amsterdam, The Netherlands
- "fr-par"
- Paris, France
#### --s3-region
Region to connect to.
Leave blank if you are using an S3 clone and you don't have a region.
@ -618,74 +634,130 @@ Specify if using an IBM COS On Premise.
- Type: string
- Default: ""
- Examples:
- "s3-api.us-geo.objectstorage.softlayer.net"
- "s3.us.cloud-object-storage.appdomain.cloud"
- US Cross Region Endpoint
- "s3-api.dal.us-geo.objectstorage.softlayer.net"
- "s3.dal.us.cloud-object-storage.appdomain.cloud"
- US Cross Region Dallas Endpoint
- "s3-api.wdc-us-geo.objectstorage.softlayer.net"
- "s3.wdc.us.cloud-object-storage.appdomain.cloud"
- US Cross Region Washington DC Endpoint
- "s3-api.sjc-us-geo.objectstorage.softlayer.net"
- "s3.sjc.us.cloud-object-storage.appdomain.cloud"
- US Cross Region San Jose Endpoint
- "s3-api.us-geo.objectstorage.service.networklayer.com"
- "s3.private.us.cloud-object-storage.appdomain.cloud"
- US Cross Region Private Endpoint
- "s3-api.dal-us-geo.objectstorage.service.networklayer.com"
- "s3.private.dal.us.cloud-object-storage.appdomain.cloud"
- US Cross Region Dallas Private Endpoint
- "s3-api.wdc-us-geo.objectstorage.service.networklayer.com"
- "s3.private.wdc.us.cloud-object-storage.appdomain.cloud"
- US Cross Region Washington DC Private Endpoint
- "s3-api.sjc-us-geo.objectstorage.service.networklayer.com"
- "s3.private.sjc.us.cloud-object-storage.appdomain.cloud"
- US Cross Region San Jose Private Endpoint
- "s3.us-east.objectstorage.softlayer.net"
- "s3.us-east.cloud-object-storage.appdomain.cloud"
- US Region East Endpoint
- "s3.us-east.objectstorage.service.networklayer.com"
- "s3.private.us-east.cloud-object-storage.appdomain.cloud"
- US Region East Private Endpoint
- "s3.us-south.objectstorage.softlayer.net"
- "s3.us-south.cloud-object-storage.appdomain.cloud"
- US Region South Endpoint
- "s3.us-south.objectstorage.service.networklayer.com"
- "s3.private.us-south.cloud-object-storage.appdomain.cloud"
- US Region South Private Endpoint
- "s3.eu-geo.objectstorage.softlayer.net"
- "s3.eu.cloud-object-storage.appdomain.cloud"
- EU Cross Region Endpoint
- "s3.fra-eu-geo.objectstorage.softlayer.net"
- "s3.fra.eu.cloud-object-storage.appdomain.cloud"
- EU Cross Region Frankfurt Endpoint
- "s3.mil-eu-geo.objectstorage.softlayer.net"
- "s3.mil.eu.cloud-object-storage.appdomain.cloud"
- EU Cross Region Milan Endpoint
- "s3.ams-eu-geo.objectstorage.softlayer.net"
- "s3.ams.eu.cloud-object-storage.appdomain.cloud"
- EU Cross Region Amsterdam Endpoint
- "s3.eu-geo.objectstorage.service.networklayer.com"
- "s3.private.eu.cloud-object-storage.appdomain.cloud"
- EU Cross Region Private Endpoint
- "s3.fra-eu-geo.objectstorage.service.networklayer.com"
- "s3.private.fra.eu.cloud-object-storage.appdomain.cloud"
- EU Cross Region Frankfurt Private Endpoint
- "s3.mil-eu-geo.objectstorage.service.networklayer.com"
- "s3.private.mil.eu.cloud-object-storage.appdomain.cloud"
- EU Cross Region Milan Private Endpoint
- "s3.ams-eu-geo.objectstorage.service.networklayer.com"
- "s3.private.ams.eu.cloud-object-storage.appdomain.cloud"
- EU Cross Region Amsterdam Private Endpoint
- "s3.eu-gb.objectstorage.softlayer.net"
- "s3.eu-gb.cloud-object-storage.appdomain.cloud"
- Great Britain Endpoint
- "s3.eu-gb.objectstorage.service.networklayer.com"
- "s3.private.eu-gb.cloud-object-storage.appdomain.cloud"
- Great Britain Private Endpoint
- "s3.ap-geo.objectstorage.softlayer.net"
- "s3.eu-de.cloud-object-storage.appdomain.cloud"
- EU Region DE Endpoint
- "s3.private.eu-de.cloud-object-storage.appdomain.cloud"
- EU Region DE Private Endpoint
- "s3.ap.cloud-object-storage.appdomain.cloud"
- APAC Cross Regional Endpoint
- "s3.tok-ap-geo.objectstorage.softlayer.net"
- "s3.tok.ap.cloud-object-storage.appdomain.cloud"
- APAC Cross Regional Tokyo Endpoint
- "s3.hkg-ap-geo.objectstorage.softlayer.net"
- "s3.hkg.ap.cloud-object-storage.appdomain.cloud"
- APAC Cross Regional HongKong Endpoint
- "s3.seo-ap-geo.objectstorage.softlayer.net"
- "s3.seo.ap.cloud-object-storage.appdomain.cloud"
- APAC Cross Regional Seoul Endpoint
- "s3.ap-geo.objectstorage.service.networklayer.com"
- "s3.private.ap.cloud-object-storage.appdomain.cloud"
- APAC Cross Regional Private Endpoint
- "s3.tok-ap-geo.objectstorage.service.networklayer.com"
- "s3.private.tok.ap.cloud-object-storage.appdomain.cloud"
- APAC Cross Regional Tokyo Private Endpoint
- "s3.hkg-ap-geo.objectstorage.service.networklayer.com"
- "s3.private.hkg.ap.cloud-object-storage.appdomain.cloud"
- APAC Cross Regional HongKong Private Endpoint
- "s3.seo-ap-geo.objectstorage.service.networklayer.com"
- "s3.private.seo.ap.cloud-object-storage.appdomain.cloud"
- APAC Cross Regional Seoul Private Endpoint
- "s3.mel01.objectstorage.softlayer.net"
- "s3.jp-tok.cloud-object-storage.appdomain.cloud"
- APAC Region Japan Endpoint
- "s3.private.jp-tok.cloud-object-storage.appdomain.cloud"
- APAC Region Japan Private Endpoint
- "s3.au-syd.cloud-object-storage.appdomain.cloud"
- APAC Region Australia Endpoint
- "s3.private.au-syd.cloud-object-storage.appdomain.cloud"
- APAC Region Australia Private Endpoint
- "s3.ams03.cloud-object-storage.appdomain.cloud"
- Amsterdam Single Site Endpoint
- "s3.private.ams03.cloud-object-storage.appdomain.cloud"
- Amsterdam Single Site Private Endpoint
- "s3.che01.cloud-object-storage.appdomain.cloud"
- Chennai Single Site Endpoint
- "s3.private.che01.cloud-object-storage.appdomain.cloud"
- Chennai Single Site Private Endpoint
- "s3.mel01.cloud-object-storage.appdomain.cloud"
- Melbourne Single Site Endpoint
- "s3.mel01.objectstorage.service.networklayer.com"
- "s3.private.mel01.cloud-object-storage.appdomain.cloud"
- Melbourne Single Site Private Endpoint
- "s3.tor01.objectstorage.softlayer.net"
- "s3.osl01.cloud-object-storage.appdomain.cloud"
- Oslo Single Site Endpoint
- "s3.private.osl01.cloud-object-storage.appdomain.cloud"
- Oslo Single Site Private Endpoint
- "s3.tor01.cloud-object-storage.appdomain.cloud"
- Toronto Single Site Endpoint
- "s3.tor01.objectstorage.service.networklayer.com"
- "s3.private.tor01.cloud-object-storage.appdomain.cloud"
- Toronto Single Site Private Endpoint
- "s3.seo01.cloud-object-storage.appdomain.cloud"
- Seoul Single Site Endpoint
- "s3.private.seo01.cloud-object-storage.appdomain.cloud"
- Seoul Single Site Private Endpoint
- "s3.mon01.cloud-object-storage.appdomain.cloud"
- Montreal Single Site Endpoint
- "s3.private.mon01.cloud-object-storage.appdomain.cloud"
- Montreal Single Site Private Endpoint
- "s3.mex01.cloud-object-storage.appdomain.cloud"
- Mexico Single Site Endpoint
- "s3.private.mex01.cloud-object-storage.appdomain.cloud"
- Mexico Single Site Private Endpoint
- "s3.sjc04.cloud-object-storage.appdomain.cloud"
- San Jose Single Site Endpoint
- "s3.private.sjc04.cloud-object-storage.appdomain.cloud"
- San Jose Single Site Private Endpoint
- "s3.mil01.cloud-object-storage.appdomain.cloud"
- Milan Single Site Endpoint
- "s3.private.mil01.cloud-object-storage.appdomain.cloud"
- Milan Single Site Private Endpoint
- "s3.hkg02.cloud-object-storage.appdomain.cloud"
- Hong Kong Single Site Endpoint
- "s3.private.hkg02.cloud-object-storage.appdomain.cloud"
- Hong Kong Single Site Private Endpoint
- "s3.par01.cloud-object-storage.appdomain.cloud"
- Paris Single Site Endpoint
- "s3.private.par01.cloud-object-storage.appdomain.cloud"
- Paris Single Site Private Endpoint
- "s3.sng01.cloud-object-storage.appdomain.cloud"
- Singapore Single Site Endpoint
- "s3.private.sng01.cloud-object-storage.appdomain.cloud"
- Singapore Single Site Private Endpoint
#### --s3-endpoint
@ -737,6 +809,20 @@ Endpoint for OSS API.
#### --s3-endpoint
Endpoint for Scaleway Object Storage.
- Config: endpoint
- Env Var: RCLONE_S3_ENDPOINT
- Type: string
- Default: ""
- Examples:
- "s3.nl-ams.scw.cloud"
- Amsterdam Endpoint
- "s3.fr-par.scw.cloud"
- Paris Endpoint
#### --s3-endpoint
Endpoint for StackPath Object Storage.
- Config: endpoint
@ -1018,6 +1104,22 @@ The storage class to use when storing new objects in OSS.
- "STANDARD_IA"
- Infrequent access storage mode.
#### --s3-storage-class
The storage class to use when storing new objects in S3.
- Config: storage_class
- Env Var: RCLONE_S3_STORAGE_CLASS
- Type: string
- Default: ""
- Examples:
- ""
- Default
- "STANDARD"
- The Standard class for any upload; suitable for on-demand content like streaming or CDN.
- "GLACIER"
- Archived storage; prices are lower, but it needs to be restored first to be accessed.
### Advanced Options
Here are the advanced options specific to s3 (Amazon S3 Compliant Storage Provider (AWS, Alibaba, Ceph, Digital Ocean, Dreamhost, IBM COS, Minio, etc)).
@ -1125,6 +1227,25 @@ larger files then you will need to increase chunk_size.
- Type: SizeSuffix
- Default: 5M
#### --s3-max-upload-parts
Maximum number of parts in a multipart upload.
This option defines the maximum number of multipart chunks to use
when doing a multipart upload.
This can be useful if a service does not support the AWS S3
specification of 10,000 chunks.
Rclone will automatically increase the chunk size when uploading a
large file of a known size to stay below this number of chunks limit.
- Config: max_upload_parts
- Env Var: RCLONE_S3_MAX_UPLOAD_PARTS
- Type: int
- Default: 10000
#### --s3-copy-cutoff
Cutoff for switching to multipart copy
@ -1137,7 +1258,7 @@ The minimum is 0 and the maximum is 5GB.
- Config: copy_cutoff
- Env Var: RCLONE_S3_COPY_CUTOFF
- Type: SizeSuffix
- Default: 5G
- Default: 4.656G
#### --s3-disable-checksum
@ -1153,6 +1274,41 @@ to start uploading.
- Type: bool
- Default: false
#### --s3-shared-credentials-file
Path to the shared credentials file
If env_auth = true then rclone can use a shared credentials file.
If this variable is empty rclone will look for the
"AWS_SHARED_CREDENTIALS_FILE" env variable. If the env value is empty
it will default to the current user's home directory.
Linux/OSX: "$HOME/.aws/credentials"
Windows: "%USERPROFILE%\.aws\credentials"
- Config: shared_credentials_file
- Env Var: RCLONE_S3_SHARED_CREDENTIALS_FILE
- Type: string
- Default: ""
#### --s3-profile
Profile to use in the shared credentials file
If env_auth = true then rclone can use a shared credentials file. This
variable controls which profile is used in that file.
If empty it will default to the environment variable "AWS_PROFILE" or
"default" if that environment variable is also not set.
- Config: profile
- Env Var: RCLONE_S3_PROFILE
- Type: string
- Default: ""
#### --s3-session-token
An AWS session token
@ -1250,6 +1406,19 @@ In Ceph, this can be increased with the "rgw list buckets max chunk" option.
- Type: int
- Default: 1000
#### --s3-no-check-bucket
If set don't attempt to check the bucket exists or create it
This can be useful when trying to minimise the number of transactions
rclone does if you know the bucket exists already.
- Config: no_check_bucket
- Env Var: RCLONE_S3_NO_CHECK_BUCKET
- Type: bool
- Default: false
#### --s3-encoding
This sets the encoding for the backend.
@ -1281,6 +1450,129 @@ Whether to use mmap buffers in internal memory pool.
- Type: bool
- Default: false
### Backend commands
Here are the commands specific to the s3 backend.
Run them with
rclone backend COMMAND remote:
The help below will explain what arguments each command takes.
See [the "rclone backend" command](/commands/rclone_backend/) for more
info on how to pass options and arguments.
These can be run on a running backend using the rc command
[backend/command](/rc/#backend/command).
#### restore
Restore objects from GLACIER to normal storage
rclone backend restore remote: [options] [<arguments>+]
This command can be used to restore one or more objects from GLACIER
to normal storage.
Usage Examples:
rclone backend restore s3:bucket/path/to/object [-o priority=PRIORITY] [-o lifetime=DAYS]
rclone backend restore s3:bucket/path/to/directory [-o priority=PRIORITY] [-o lifetime=DAYS]
rclone backend restore s3:bucket [-o priority=PRIORITY] [-o lifetime=DAYS]
This flag also obeys the filters. Test first with -i/--interactive or --dry-run flags
rclone -i backend restore --include "*.txt" s3:bucket/path -o priority=Standard
All the objects shown will be marked for restore, then
rclone backend restore --include "*.txt" s3:bucket/path -o priority=Standard
It returns a list of status dictionaries with Remote and Status
keys. The Status will be OK if it was successfull or an error message
if not.
[
{
"Status": "OK",
"Path": "test.txt"
},
{
"Status": "OK",
"Path": "test/file4.txt"
}
]
Options:
- "description": The optional description for the job.
- "lifetime": Lifetime of the active copy in days
- "priority": Priority of restore: Standard|Expedited|Bulk
#### list-multipart-uploads
List the unfinished multipart uploads
rclone backend list-multipart-uploads remote: [options] [<arguments>+]
This command lists the unfinished multipart uploads in JSON format.
rclone backend list-multipart s3:bucket/path/to/object
It returns a dictionary of buckets with values as lists of unfinished
multipart uploads.
You can call it with no bucket in which case it lists all bucket, with
a bucket or with a bucket and path.
{
"rclone": [
{
"Initiated": "2020-06-26T14:20:36Z",
"Initiator": {
"DisplayName": "XXX",
"ID": "arn:aws:iam::XXX:user/XXX"
},
"Key": "KEY",
"Owner": {
"DisplayName": null,
"ID": "XXX"
},
"StorageClass": "STANDARD",
"UploadId": "XXX"
}
],
"rclone-1000files": [],
"rclone-dst": []
}
#### cleanup
Remove unfinished multipart uploads.
rclone backend cleanup remote: [options] [<arguments>+]
This command removes unfinished multipart uploads of age greater than
max-age which defaults to 24 hours.
Note that you can use -i/--dry-run with this command to see what it
would do.
rclone backend cleanup s3:bucket/path/to/object
rclone backend cleanup -o max-age=7w s3:bucket/path/to/object
Durations are parsed as per the rest of rclone, 2h, 7d, 7w etc.
Options:
- "max-age": Max age of upload to delete
{{< rem autogenerated options stop >}}
### Anonymous access to public buckets ###

View File

@ -294,6 +294,8 @@ User name (usually email address)
Password
**NB** Input to this must be obscured - see [rclone obscure](/commands/rclone_obscure/).
- Config: pass
- Env Var: RCLONE_SEAFILE_PASS
- Type: string
@ -321,6 +323,8 @@ Name of the library. Leave blank to access all non-encrypted libraries.
Library password (for encrypted libraries only). Leave blank if you pass it through the command line.
**NB** Input to this must be obscured - see [rclone obscure](/commands/rclone_obscure/).
- Config: library_key
- Env Var: RCLONE_SEAFILE_LIBRARY_KEY
- Type: string

View File

@ -192,6 +192,8 @@ SSH port, leave blank to use default (22)
SSH password, leave blank to use ssh-agent.
**NB** Input to this must be obscured - see [rclone obscure](/commands/rclone_obscure/).
- Config: pass
- Env Var: RCLONE_SFTP_PASS
- Type: string
@ -210,6 +212,9 @@ Raw PEM-encoded private key, If specified, will override key_file parameter.
Path to PEM-encoded private key file, leave blank or set key-use-agent to use ssh-agent.
Leading `~` will be expanded in the file name as will environment variables such as `${RCLONE_CONFIG_DIR}`.
- Config: key_file
- Env Var: RCLONE_SFTP_KEY_FILE
- Type: string
@ -222,6 +227,8 @@ The passphrase to decrypt the PEM-encoded private key file.
Only PEM encrypted key files (old OpenSSH format) are supported. Encrypted keys
in the new OpenSSH format can't be used.
**NB** Input to this must be obscured - see [rclone obscure](/commands/rclone_obscure/).
- Config: key_file_pass
- Env Var: RCLONE_SFTP_KEY_FILE_PASS
- Type: string
@ -302,11 +309,11 @@ different. This issue affects among others Synology NAS boxes.
Shared folders can be found in directories representing volumes
rclone sync -i /home/local/directory remote:/directory --ssh-path-override /volume2/directory
rclone sync /home/local/directory remote:/directory --ssh-path-override /volume2/directory
Home directory can be found in a shared folder called "home"
rclone sync -i /home/local/directory remote:/home/directory --ssh-path-override /volume1/homes/USER/directory
rclone sync /home/local/directory remote:/home/directory --ssh-path-override /volume1/homes/USER/directory
- Config: path_override
- Env Var: RCLONE_SFTP_PATH_OVERRIDE
@ -349,6 +356,26 @@ Set to skip any symlinks and any other non regular files.
- Type: bool
- Default: false
#### --sftp-subsystem
Specifies the SSH2 subsystem on the remote host.
- Config: subsystem
- Env Var: RCLONE_SFTP_SUBSYSTEM
- Type: string
- Default: "sftp"
#### --sftp-server-command
Specifies the path or command to run a sftp server on the remote host.
The subsystem option is ignored when server_command is defined.
- Config: server_command
- Env Var: RCLONE_SFTP_SERVER_COMMAND
- Type: string
- Default: ""
{{< rem autogenerated options stop >}}
### Limitations ###

View File

@ -152,6 +152,8 @@ User name
Password.
**NB** Input to this must be obscured - see [rclone obscure](/commands/rclone_obscure/).
- Config: pass
- Env Var: RCLONE_WEBDAV_PASS
- Type: string

View File

@ -132,7 +132,7 @@ Here are the standard options specific to yandex (Yandex Disk).
#### --yandex-client-id
Yandex Client Id
OAuth Client Id
Leave blank normally.
- Config: client_id
@ -142,7 +142,7 @@ Leave blank normally.
#### --yandex-client-secret
Yandex Client Secret
OAuth Client Secret
Leave blank normally.
- Config: client_secret
@ -154,15 +154,34 @@ Leave blank normally.
Here are the advanced options specific to yandex (Yandex Disk).
#### --yandex-unlink
#### --yandex-token
Remove existing public link to file/folder with link command rather than creating.
Default is false, meaning link command will create or retrieve public link.
OAuth Access Token as a JSON blob.
- Config: unlink
- Env Var: RCLONE_YANDEX_UNLINK
- Type: bool
- Default: false
- Config: token
- Env Var: RCLONE_YANDEX_TOKEN
- Type: string
- Default: ""
#### --yandex-auth-url
Auth server URL.
Leave blank to use the provider defaults.
- Config: auth_url
- Env Var: RCLONE_YANDEX_AUTH_URL
- Type: string
- Default: ""
#### --yandex-token-url
Token server url.
Leave blank to use the provider defaults.
- Config: token_url
- Env Var: RCLONE_YANDEX_TOKEN_URL
- Type: string
- Default: ""
#### --yandex-encoding

27761
rclone.1 generated

File diff suppressed because it is too large Load Diff