Version v1.56.0

This commit is contained in:
Nick Craig-Wood 2021-07-20 19:45:41 +01:00
parent c67c1ab4ee
commit 37ff05a5fa
64 changed files with 11061 additions and 2081 deletions

2451
MANUAL.html generated

File diff suppressed because it is too large Load Diff

3001
MANUAL.md generated

File diff suppressed because it is too large Load Diff

3139
MANUAL.txt generated

File diff suppressed because it is too large Load Diff

View File

@ -2,3 +2,4 @@
<nick@raig-wood.com> <nick@raig-wood.com>
<anaghk.dos@gmail.com> <anaghk.dos@gmail.com>
<33207650+sp31415t1@users.noreply.github.com> <33207650+sp31415t1@users.noreply.github.com>
<unknown>

View File

@ -23,6 +23,7 @@ docs = [
"rc.md", "rc.md",
"overview.md", "overview.md",
"flags.md", "flags.md",
"docker.md",
# Keep these alphabetical by full name # Keep these alphabetical by full name
"fichier.md", "fichier.md",

View File

@ -3,8 +3,7 @@ title: "Alias"
description: "Remote Aliases" description: "Remote Aliases"
--- ---
{{< icon "fa fa-link" >}} Alias # {{< icon "fa fa-link" >}} Alias
-----------------------------------------
The `alias` remote provides a new name for another remote. The `alias` remote provides a new name for another remote.

View File

@ -3,8 +3,7 @@ title: "Amazon Drive"
description: "Rclone docs for Amazon Drive" description: "Rclone docs for Amazon Drive"
--- ---
{{< icon "fab fa-amazon" >}} Amazon Drive # {{< icon "fab fa-amazon" >}} Amazon Drive
-----------------------------------------
Amazon Drive, formerly known as Amazon Cloud Drive, is a cloud storage Amazon Drive, formerly known as Amazon Cloud Drive, is a cloud storage
service run by Amazon for consumers. service run by Amazon for consumers.
@ -260,7 +259,7 @@ Files >= this size will be downloaded via their tempLink.
Files this size or more will be downloaded via their "tempLink". This Files this size or more will be downloaded via their "tempLink". This
is to work around a problem with Amazon Drive which blocks downloads is to work around a problem with Amazon Drive which blocks downloads
of files bigger than about 10 GiB. The default for this is 9 GiB which of files bigger than about 10 GiB. The default for this is 9 GiB which
shouldn't need to be changed. shouldn't need to be changed.
To download files above this threshold, rclone requests a "tempLink" To download files above this threshold, rclone requests a "tempLink"
@ -270,7 +269,7 @@ underlying S3 storage.
- Config: templink_threshold - Config: templink_threshold
- Env Var: RCLONE_ACD_TEMPLINK_THRESHOLD - Env Var: RCLONE_ACD_TEMPLINK_THRESHOLD
- Type: SizeSuffix - Type: SizeSuffix
- Default: 9G - Default: 9Gi
#### --acd-encoding #### --acd-encoding

View File

@ -431,7 +431,7 @@ put them back in again.` >}}
* Laurens Janssen <BD69BM@insim.biz> * Laurens Janssen <BD69BM@insim.biz>
* Bob Bagwill <bobbagwill@gmail.com> * Bob Bagwill <bobbagwill@gmail.com>
* Nathan Collins <colli372@msu.edu> * Nathan Collins <colli372@msu.edu>
* lostheli <unknown> * lostheli
* kelv <kelvin@acks.org> * kelv <kelvin@acks.org>
* Milly <milly.ca@gmail.com> * Milly <milly.ca@gmail.com>
* gtorelly <gtorelly@gmail.com> * gtorelly <gtorelly@gmail.com>

View File

@ -3,8 +3,7 @@ title: "Microsoft Azure Blob Storage"
description: "Rclone docs for Microsoft Azure Blob Storage" description: "Rclone docs for Microsoft Azure Blob Storage"
--- ---
{{< icon "fab fa-windows" >}} Microsoft Azure Blob Storage # {{< icon "fab fa-windows" >}} Microsoft Azure Blob Storage
-----------------------------------------
Paths are specified as `remote:container` (or `remote:` for the `lsd` Paths are specified as `remote:container` (or `remote:` for the `lsd`
command.) You may put subdirectories in too, e.g. command.) You may put subdirectories in too, e.g.
@ -285,7 +284,7 @@ Note that this is stored in memory and there may be up to
- Config: chunk_size - Config: chunk_size
- Env Var: RCLONE_AZUREBLOB_CHUNK_SIZE - Env Var: RCLONE_AZUREBLOB_CHUNK_SIZE
- Type: SizeSuffix - Type: SizeSuffix
- Default: 4M - Default: 4Mi
#### --azureblob-list-chunk #### --azureblob-list-chunk

View File

@ -3,8 +3,7 @@ title: "B2"
description: "Backblaze B2" description: "Backblaze B2"
--- ---
{{< icon "fa fa-fire" >}} Backblaze B2 # {{< icon "fa fa-fire" >}} Backblaze B2
----------------------------------------
B2 is [Backblaze's cloud storage system](https://www.backblaze.com/b2/). B2 is [Backblaze's cloud storage system](https://www.backblaze.com/b2/).
@ -406,7 +405,7 @@ This value should be set no larger than 4.657 GiB (== 5 GB).
- Config: upload_cutoff - Config: upload_cutoff
- Env Var: RCLONE_B2_UPLOAD_CUTOFF - Env Var: RCLONE_B2_UPLOAD_CUTOFF
- Type: SizeSuffix - Type: SizeSuffix
- Default: 200M - Default: 200Mi
#### --b2-copy-cutoff #### --b2-copy-cutoff
@ -420,7 +419,7 @@ The minimum is 0 and the maximum is 4.6 GiB.
- Config: copy_cutoff - Config: copy_cutoff
- Env Var: RCLONE_B2_COPY_CUTOFF - Env Var: RCLONE_B2_COPY_CUTOFF
- Type: SizeSuffix - Type: SizeSuffix
- Default: 4G - Default: 4Gi
#### --b2-chunk-size #### --b2-chunk-size
@ -434,7 +433,7 @@ minimum size.
- Config: chunk_size - Config: chunk_size
- Env Var: RCLONE_B2_CHUNK_SIZE - Env Var: RCLONE_B2_CHUNK_SIZE
- Type: SizeSuffix - Type: SizeSuffix
- Default: 96M - Default: 96Mi
#### --b2-disable-checksum #### --b2-disable-checksum

View File

@ -3,8 +3,7 @@ title: "Box"
description: "Rclone docs for Box" description: "Rclone docs for Box"
--- ---
{{< icon "fa fa-archive" >}} Box # {{< icon "fa fa-archive" >}} Box
-----------------------------------------
Paths are specified as `remote:path` Paths are specified as `remote:path`
@ -374,7 +373,7 @@ Cutoff for switching to multipart upload (>= 50 MiB).
- Config: upload_cutoff - Config: upload_cutoff
- Env Var: RCLONE_BOX_UPLOAD_CUTOFF - Env Var: RCLONE_BOX_UPLOAD_CUTOFF
- Type: SizeSuffix - Type: SizeSuffix
- Default: 50M - Default: 50Mi
#### --box-commit-retries #### --box-commit-retries

View File

@ -3,8 +3,7 @@ title: "Cache"
description: "Rclone docs for cache remote" description: "Rclone docs for cache remote"
--- ---
{{< icon "fa fa-archive" >}} Cache (DEPRECATED) # {{< icon "fa fa-archive" >}} Cache (DEPRECATED)
-----------------------------------------
The `cache` remote wraps another existing remote and stores file structure The `cache` remote wraps another existing remote and stores file structure
and its data for long running tasks like `rclone mount`. and its data for long running tasks like `rclone mount`.
@ -361,9 +360,9 @@ will need to be cleared or unexpected EOF errors will occur.
- Config: chunk_size - Config: chunk_size
- Env Var: RCLONE_CACHE_CHUNK_SIZE - Env Var: RCLONE_CACHE_CHUNK_SIZE
- Type: SizeSuffix - Type: SizeSuffix
- Default: 5M - Default: 5Mi
- Examples: - Examples:
- "1m" - "1M"
- 1 MiB - 1 MiB
- "5M" - "5M"
- 5 MiB - 5 MiB
@ -398,7 +397,7 @@ oldest chunks until it goes under this value.
- Config: chunk_total_size - Config: chunk_total_size
- Env Var: RCLONE_CACHE_CHUNK_TOTAL_SIZE - Env Var: RCLONE_CACHE_CHUNK_TOTAL_SIZE
- Type: SizeSuffix - Type: SizeSuffix
- Default: 10G - Default: 10Gi
- Examples: - Examples:
- "500M" - "500M"
- 500 MiB - 500 MiB

View File

@ -5,6 +5,149 @@ description: "Rclone Changelog"
# Changelog # Changelog
## v1.56.0 - 2021-07-20
[See commits](https://github.com/rclone/rclone/compare/v1.55.0...v1.56.0)
* New backends
* [Uptobox](/uptobox/) (buengese)
* New commands
* [serve docker](/commands/rclone_serve_docker/) (Antoine GIRARD) (Ivan Andreev)
* and accompanying [docker volume plugin](/docker/)
* [checksum](/commands/rclone_checksum/) to check files against a file of checksums (Ivan Andreev)
* this is also available as `rclone md5sum -C` etc
* [config touch](/commands/rclone_config_touch/): ensure config exists at configured location (albertony)
* [test changenotify](/commands/rclone_test_changenotify/): command to help debugging changenotify (Nick Craig-Wood)
* Deprecations
* `dbhashsum`: Remove command deprecated a year ago (Ivan Andreev)
* `cache`: Deprecate cache backend (Ivan Andreev)
* New Features
* rework config system so it can be used non-interactively via cli and rc API.
* See docs in [config create](/commands/rclone_config_create/)
* This is a very big change to all the backends so may cause breakages - please file bugs!
* librclone - export the rclone RC as a C library (lewisxy) (Nick Craig-Wood)
* Link a C-API rclone shared object into your project
* Use the RC as an in memory interface
* Python example supplied
* Also supports Android and gomobile
* fs
* Add `--disable-http2` for global http2 disable (Nick Craig-Wood)
* Make `--dump` imply `-vv` (Alex Chen)
* Use binary prefixes for size and rate units (albertony)
* Use decimal prefixes for counts (albertony)
* Add google search widget to rclone.org (Ivan Andreev)
* accounting: Calculate rolling average speed (Haochen Tong)
* atexit: Terminate with non-zero status after receiving signal (Michael Hanselmann)
* build
* Only run event-based workflow scripts under rclone repo with manual override (Mathieu Carbou)
* Add Android build with gomobile (x0b)
* check: Log the hash in use like cryptcheck does (Nick Craig-Wood)
* version: Print os/version, kernel and bitness (Ivan Andreev)
* config
* Prevent use of Windows reserved names in config file name (albertony)
* Create config file in windows appdata directory by default (albertony)
* Treat any config file paths with filename notfound as memory-only config (albertony)
* Delay load config file (albertony)
* Replace defaultConfig with a thread-safe in-memory implementation (Chris Macklin)
* Allow `config create` and friends to take `key=value` parameters (Nick Craig-Wood)
* Fixed issues with flags/options set by environment vars. (Ole Frost)
* fshttp: Implement graceful DSCP error handling (Tyson Moore)
* lib/http - provides an abstraction for a central http server that services can bind routes to (Nolan Woods)
* Add `--template` config and flags to serve/data (Nolan Woods)
* Add default 404 handler (Nolan Woods)
* link: Use "off" value for unset expiry (Nick Craig-Wood)
* oauthutil: Raise fatal error if token expired without refresh token (Alex Chen)
* rcat: Add `--size` flag for more efficient uploads of known size (Nazar Mishturak)
* serve sftp: Add `--stdio` flag to serve via stdio (Tom)
* sync: Don't warn about `--no-traverse` when `--files-from` is set (Nick Gaya)
* `test makefiles`
* Add `--seed` flag and make data generated repeatable (Nick Craig-Wood)
* Add log levels and speed summary (Nick Craig-Wood)
* Bug Fixes
* accounting: Fix startTime of statsGroups.sum (Haochen Tong)
* cmd/ncdu: Fix out of range panic in delete (buengese)
* config
* Fix issues with memory-only config file paths (albertony)
* Fix in memory config not saving on the fly backend config (Nick Craig-Wood)
* fshttp: Fix address parsing for DSCP (Tyson Moore)
* ncdu: Update termbox-go library to fix crash (Nick Craig-Wood)
* oauthutil: Fix old authorize result not recognised (Cnly)
* operations: Don't update timestamps of files in `--compare-dest` (Nick Gaya)
* selfupdate: fix archive name on macos (Ivan Andreev)
* Mount
* Refactor before adding serve docker (Antoine GIRARD)
* VFS
* Add cache reset for `--vfs-cache-max-size` handling at cache poll interval (Leo Luan)
* Fix modtime changing when reading file into cache (Nick Craig-Wood)
* Avoid unnecessary subdir in cache path (albertony)
* Fix that umask option cannot be set as environment variable (albertony)
* Do not print notice about missing poll-interval support when set to 0 (albertony)
* Local
* Always use readlink to read symlink size for better compatibility (Nick Craig-Wood)
* Add `--local-unicode-normalization` (and remove `--local-no-unicode-normalization`) (Nick Craig-Wood)
* Skip entries removed concurrently with List() (Ivan Andreev)
* Crypt
* Support timestamped filenames from `--b2-versions` (Dominik Mydlil)
* B2
* Don't include the bucket name in public link file prefixes (Jeffrey Tolar)
* Fix versions and .files with no extension (Nick Craig-Wood)
* Factor version handling into lib/version (Dominik Mydlil)
* Box
* Use upload preflight check to avoid listings in file uploads (Nick Craig-Wood)
* Return errors instead of calling log.Fatal with them (Nick Craig-Wood)
* Drive
* Switch to the Drives API for looking up shared drives (Nick Craig-Wood)
* Fix some google docs being treated as files (Nick Craig-Wood)
* Dropbox
* Add `--dropbox-batch-mode` flag to speed up uploading (Nick Craig-Wood)
* Read the [batch mode](/dropbox/#batch-mode) docs for more info
* Set visibility in link sharing when `--expire` is set (Nick Craig-Wood)
* Simplify chunked uploads (Alexey Ivanov)
* Improve "own App IP" instructions (Ivan Andreev)
* Fichier
* Check if more than one upload link is returned (Nick Craig-Wood)
* Support downloading password protected files and folders (Florian Penzkofer)
* Make error messages report text from the API (Nick Craig-Wood)
* Fix move of files in the same directory (Nick Craig-Wood)
* Check that we actually got a download token and retry if we didn't (buengese)
* Filefabric
* Fix listing after change of from field from "int" to int. (Nick Craig-Wood)
* FTP
* Make upload error 250 indicate success (Nick Craig-Wood)
* GCS
* Make compatible with gsutil's mtime metadata (database64128)
* Clean up time format constants (database64128)
* Google Photos
* Fix read only scope not being used properly (Nick Craig-Wood)
* HTTP
* Replace httplib with lib/http (Nolan Woods)
* Clean up Bind to better use middleware (Nolan Woods)
* Jottacloud
* Fix legacy auth with state based config system (buengese)
* Fix invalid url in output from link command (albertony)
* Add no versions option (buengese)
* Onedrive
* Add `list_chunk option` (Nick Gaya)
* Also report root error if unable to cancel multipart upload (Cnly)
* Fix failed to configure: empty token found error (Nick Craig-Wood)
* Make link return direct download link (Xuanchen Wu)
* S3
* Add `--s3-no-head-object` (Tatsuya Noyori)
* Remove WebIdentityRoleProvider to fix crash on auth (Nick Craig-Wood)
* Don't check to see if remote is object if it ends with / (Nick Craig-Wood)
* Add SeaweedFS (Chris Lu)
* Update Alibaba OSS endpoints (Chuan Zh)
* SFTP
* Fix performance regression by re-enabling concurrent writes (Nick Craig-Wood)
* Expand tilde and environment variables in configured `known_hosts_file` (albertony)
* Tardigrade
* Upgrade to uplink v1.4.6 (Caleb Case)
* Use negative offset (Caleb Case)
* Add warning about `too many open files` (acsfer)
* WebDAV
* Fix sharepoint auth over http (Nick Craig-Wood)
* Add headers option (Antoon Prins)
## v1.55.1 - 2021-04-26 ## v1.55.1 - 2021-04-26
[See commits](https://github.com/rclone/rclone/compare/v1.55.0...v1.55.1) [See commits](https://github.com/rclone/rclone/compare/v1.55.0...v1.55.1)

View File

@ -3,8 +3,7 @@ title: "Chunker"
description: "Split-chunking overlay remote" description: "Split-chunking overlay remote"
--- ---
{{< icon "fa fa-cut" >}}Chunker (BETA) # {{< icon "fa fa-cut" >}}Chunker (BETA)
----------------------------------------
The `chunker` overlay transparently splits large files into smaller chunks The `chunker` overlay transparently splits large files into smaller chunks
during upload to wrapped remote and transparently assembles them back during upload to wrapped remote and transparently assembles them back
@ -332,7 +331,7 @@ Files larger than chunk size will be split in chunks.
- Config: chunk_size - Config: chunk_size
- Env Var: RCLONE_CHUNKER_CHUNK_SIZE - Env Var: RCLONE_CHUNKER_CHUNK_SIZE
- Type: SizeSuffix - Type: SizeSuffix
- Default: 2G - Default: 2Gi
#### --chunker-hash-type #### --chunker-hash-type

View File

@ -24,7 +24,7 @@ you would do:
If the remote uses OAuth the token will be updated, if you don't If the remote uses OAuth the token will be updated, if you don't
require this add an extra parameter thus: require this add an extra parameter thus:
rclone config update myremote swift env_auth=true config_refresh_token=false rclone config update myremote env_auth=true config_refresh_token=false
Note that if the config process would normally ask a question the Note that if the config process would normally ask a question the
default is taken (unless `--non-interactive` is used). Each time default is taken (unless `--non-interactive` is used). Each time

View File

@ -18,7 +18,7 @@ FUSE.
First set up your remote using `rclone config`. Check it works with `rclone ls` etc. First set up your remote using `rclone config`. Check it works with `rclone ls` etc.
On Linux and macOS, you can either run mount in foreground mode or background (daemon) mode. On Linux and OSX, you can either run mount in foreground mode or background (daemon) mode.
Mount runs in foreground mode by default, use the `--daemon` flag to specify background mode. Mount runs in foreground mode by default, use the `--daemon` flag to specify background mode.
You can only run mount in foreground mode on Windows. You can only run mount in foreground mode on Windows.
@ -608,7 +608,7 @@ rclone mount remote:path /path/to/mountpoint [flags]
--poll-interval duration Time to wait between polling for changes. Must be smaller than dir-cache-time. Only on supported remotes. Set to 0 to disable. (default 1m0s) --poll-interval duration Time to wait between polling for changes. Must be smaller than dir-cache-time. Only on supported remotes. Set to 0 to disable. (default 1m0s)
--read-only Mount read-only. --read-only Mount read-only.
--uid uint32 Override the uid field set by the filesystem. Not supported on Windows. (default 1000) --uid uint32 Override the uid field set by the filesystem. Not supported on Windows. (default 1000)
--umask int Override the permission bits set by the filesystem. Not supported on Windows. (default 18) --umask int Override the permission bits set by the filesystem. Not supported on Windows. (default 2)
--vfs-cache-max-age duration Max age of objects in the cache. (default 1h0m0s) --vfs-cache-max-age duration Max age of objects in the cache. (default 1h0m0s)
--vfs-cache-max-size SizeSuffix Max total size of objects in the cache. (default off) --vfs-cache-max-size SizeSuffix Max total size of objects in the cache. (default off)
--vfs-cache-mode CacheMode Cache mode off|minimal|writes|full (default off) --vfs-cache-mode CacheMode Cache mode off|minimal|writes|full (default off)

View File

@ -33,6 +33,7 @@ Here are the keys - press '?' to toggle the help on and off
a toggle average size in directory a toggle average size in directory
n,s,C,A sort by name,size,count,average size n,s,C,A sort by name,size,count,average size
d delete file/directory d delete file/directory
y copy current path to clipboard
Y display current path Y display current path
^L refresh screen ^L refresh screen
? to toggle help on and off ? to toggle help on and off

View File

@ -319,7 +319,7 @@ rclone serve dlna remote:path [flags]
--poll-interval duration Time to wait between polling for changes. Must be smaller than dir-cache-time. Only on supported remotes. Set to 0 to disable. (default 1m0s) --poll-interval duration Time to wait between polling for changes. Must be smaller than dir-cache-time. Only on supported remotes. Set to 0 to disable. (default 1m0s)
--read-only Mount read-only. --read-only Mount read-only.
--uid uint32 Override the uid field set by the filesystem. Not supported on Windows. (default 1000) --uid uint32 Override the uid field set by the filesystem. Not supported on Windows. (default 1000)
--umask int Override the permission bits set by the filesystem. Not supported on Windows. (default 18) --umask int Override the permission bits set by the filesystem. Not supported on Windows. (default 2)
--vfs-cache-max-age duration Max age of objects in the cache. (default 1h0m0s) --vfs-cache-max-age duration Max age of objects in the cache. (default 1h0m0s)
--vfs-cache-max-size SizeSuffix Max total size of objects in the cache. (default off) --vfs-cache-max-size SizeSuffix Max total size of objects in the cache. (default off)
--vfs-cache-mode CacheMode Cache mode off|minimal|writes|full (default off) --vfs-cache-mode CacheMode Cache mode off|minimal|writes|full (default off)

View File

@ -354,7 +354,7 @@ rclone serve docker [flags]
--socket-addr string <host:port> or absolute path (default: /run/docker/plugins/rclone.sock) --socket-addr string <host:port> or absolute path (default: /run/docker/plugins/rclone.sock)
--socket-gid int GID for unix socket (default: current process GID) (default 1000) --socket-gid int GID for unix socket (default: current process GID) (default 1000)
--uid uint32 Override the uid field set by the filesystem. Not supported on Windows. (default 1000) --uid uint32 Override the uid field set by the filesystem. Not supported on Windows. (default 1000)
--umask int Override the permission bits set by the filesystem. Not supported on Windows. (default 18) --umask int Override the permission bits set by the filesystem. Not supported on Windows. (default 2)
--vfs-cache-max-age duration Max age of objects in the cache. (default 1h0m0s) --vfs-cache-max-age duration Max age of objects in the cache. (default 1h0m0s)
--vfs-cache-max-size SizeSuffix Max total size of objects in the cache. (default off) --vfs-cache-max-size SizeSuffix Max total size of objects in the cache. (default off)
--vfs-cache-mode CacheMode Cache mode off|minimal|writes|full (default off) --vfs-cache-mode CacheMode Cache mode off|minimal|writes|full (default off)

View File

@ -403,7 +403,7 @@ rclone serve ftp remote:path [flags]
--public-ip string Public IP address to advertise for passive connections. --public-ip string Public IP address to advertise for passive connections.
--read-only Mount read-only. --read-only Mount read-only.
--uid uint32 Override the uid field set by the filesystem. Not supported on Windows. (default 1000) --uid uint32 Override the uid field set by the filesystem. Not supported on Windows. (default 1000)
--umask int Override the permission bits set by the filesystem. Not supported on Windows. (default 18) --umask int Override the permission bits set by the filesystem. Not supported on Windows. (default 2)
--user string User name for authentication. (default "anonymous") --user string User name for authentication. (default "anonymous")
--vfs-cache-max-age duration Max age of objects in the cache. (default 1h0m0s) --vfs-cache-max-age duration Max age of objects in the cache. (default 1h0m0s)
--vfs-cache-max-size SizeSuffix Max total size of objects in the cache. (default off) --vfs-cache-max-size SizeSuffix Max total size of objects in the cache. (default off)

View File

@ -398,7 +398,7 @@ rclone serve http remote:path [flags]
--server-write-timeout duration Timeout for server writing data (default 1h0m0s) --server-write-timeout duration Timeout for server writing data (default 1h0m0s)
--template string User Specified Template. --template string User Specified Template.
--uid uint32 Override the uid field set by the filesystem. Not supported on Windows. (default 1000) --uid uint32 Override the uid field set by the filesystem. Not supported on Windows. (default 1000)
--umask int Override the permission bits set by the filesystem. Not supported on Windows. (default 18) --umask int Override the permission bits set by the filesystem. Not supported on Windows. (default 2)
--user string User name for authentication. --user string User name for authentication.
--vfs-cache-max-age duration Max age of objects in the cache. (default 1h0m0s) --vfs-cache-max-age duration Max age of objects in the cache. (default 1h0m0s)
--vfs-cache-max-size SizeSuffix Max total size of objects in the cache. (default off) --vfs-cache-max-size SizeSuffix Max total size of objects in the cache. (default off)

View File

@ -419,7 +419,7 @@ rclone serve sftp remote:path [flags]
--read-only Mount read-only. --read-only Mount read-only.
--stdio Run an sftp server on run stdin/stdout --stdio Run an sftp server on run stdin/stdout
--uid uint32 Override the uid field set by the filesystem. Not supported on Windows. (default 1000) --uid uint32 Override the uid field set by the filesystem. Not supported on Windows. (default 1000)
--umask int Override the permission bits set by the filesystem. Not supported on Windows. (default 18) --umask int Override the permission bits set by the filesystem. Not supported on Windows. (default 2)
--user string User name for authentication. --user string User name for authentication.
--vfs-cache-max-age duration Max age of objects in the cache. (default 1h0m0s) --vfs-cache-max-age duration Max age of objects in the cache. (default 1h0m0s)
--vfs-cache-max-size SizeSuffix Max total size of objects in the cache. (default off) --vfs-cache-max-size SizeSuffix Max total size of objects in the cache. (default off)

View File

@ -491,7 +491,7 @@ rclone serve webdav remote:path [flags]
--server-write-timeout duration Timeout for server writing data (default 1h0m0s) --server-write-timeout duration Timeout for server writing data (default 1h0m0s)
--template string User Specified Template. --template string User Specified Template.
--uid uint32 Override the uid field set by the filesystem. Not supported on Windows. (default 1000) --uid uint32 Override the uid field set by the filesystem. Not supported on Windows. (default 1000)
--umask int Override the permission bits set by the filesystem. Not supported on Windows. (default 18) --umask int Override the permission bits set by the filesystem. Not supported on Windows. (default 2)
--user string User name for authentication. --user string User name for authentication.
--vfs-cache-max-age duration Max age of objects in the cache. (default 1h0m0s) --vfs-cache-max-age duration Max age of objects in the cache. (default 1h0m0s)
--vfs-cache-max-size SizeSuffix Max total size of objects in the cache. (default off) --vfs-cache-max-size SizeSuffix Max total size of objects in the cache. (default off)

View File

@ -37,6 +37,6 @@ See the [global flags page](/flags/) for global options not listed here.
* [rclone test changenotify](/commands/rclone_test_changenotify/) - Log any change notify requests for the remote passed in. * [rclone test changenotify](/commands/rclone_test_changenotify/) - Log any change notify requests for the remote passed in.
* [rclone test histogram](/commands/rclone_test_histogram/) - Makes a histogram of file name characters. * [rclone test histogram](/commands/rclone_test_histogram/) - Makes a histogram of file name characters.
* [rclone test info](/commands/rclone_test_info/) - Discovers file name or other limitations for paths. * [rclone test info](/commands/rclone_test_info/) - Discovers file name or other limitations for paths.
* [rclone test makefiles](/commands/rclone_test_makefiles/) - Make a random file hierarchy in <dir> * [rclone test makefiles](/commands/rclone_test_makefiles/) - Make a random file hierarchy in a directory
* [rclone test memory](/commands/rclone_test_memory/) - Load all the objects at remote:path into memory and report memory stats. * [rclone test memory](/commands/rclone_test_memory/) - Load all the objects at remote:path into memory and report memory stats.

View File

@ -1,13 +1,13 @@
--- ---
title: "rclone test makefiles" title: "rclone test makefiles"
description: "Make a random file hierarchy in <dir>" description: "Make a random file hierarchy in a directory"
slug: rclone_test_makefiles slug: rclone_test_makefiles
url: /commands/rclone_test_makefiles/ url: /commands/rclone_test_makefiles/
# autogenerated - DO NOT EDIT, instead edit the source code in cmd/test/makefiles/ and as part of making a release run "make commanddocs" # autogenerated - DO NOT EDIT, instead edit the source code in cmd/test/makefiles/ and as part of making a release run "make commanddocs"
--- ---
# rclone test makefiles # rclone test makefiles
Make a random file hierarchy in <dir> Make a random file hierarchy in a directory
``` ```
rclone test makefiles <dir> [flags] rclone test makefiles <dir> [flags]

View File

@ -3,8 +3,7 @@ title: "Compress"
description: "Compression Remote" description: "Compression Remote"
--- ---
{{< icon "fas fa-compress" >}}Compress (Experimental) # {{< icon "fas fa-compress" >}}Compress (Experimental)
-----------------------------------------
### Warning ### Warning
This remote is currently **experimental**. Things may break and data may be lost. Anything you do with this remote is This remote is currently **experimental**. Things may break and data may be lost. Anything you do with this remote is
@ -142,6 +141,6 @@ Some remotes don't allow the upload of files with unknown size.
- Config: ram_cache_limit - Config: ram_cache_limit
- Env Var: RCLONE_COMPRESS_RAM_CACHE_LIMIT - Env Var: RCLONE_COMPRESS_RAM_CACHE_LIMIT
- Type: SizeSuffix - Type: SizeSuffix
- Default: 20M - Default: 20Mi
{{< rem autogenerated options stop >}} {{< rem autogenerated options stop >}}

View File

@ -3,8 +3,7 @@ title: "Crypt"
description: "Encryption overlay remote" description: "Encryption overlay remote"
--- ---
{{< icon "fa fa-lock" >}}Crypt # {{< icon "fa fa-lock" >}}Crypt
----------------------------------------
Rclone `crypt` remotes encrypt and decrypt other remotes. Rclone `crypt` remotes encrypt and decrypt other remotes.

View File

@ -3,8 +3,7 @@ title: "Google drive"
description: "Rclone docs for Google drive" description: "Rclone docs for Google drive"
--- ---
{{< icon "fab fa-google" >}} Google Drive # {{< icon "fab fa-google" >}} Google Drive
-----------------------------------------
Paths are specified as `drive:path` Paths are specified as `drive:path`
@ -868,7 +867,7 @@ Cutoff for switching to chunked upload
- Config: upload_cutoff - Config: upload_cutoff
- Env Var: RCLONE_DRIVE_UPLOAD_CUTOFF - Env Var: RCLONE_DRIVE_UPLOAD_CUTOFF
- Type: SizeSuffix - Type: SizeSuffix
- Default: 8M - Default: 8Mi
#### --drive-chunk-size #### --drive-chunk-size
@ -882,7 +881,7 @@ Reducing this will reduce memory usage but decrease performance.
- Config: chunk_size - Config: chunk_size
- Env Var: RCLONE_DRIVE_CHUNK_SIZE - Env Var: RCLONE_DRIVE_CHUNK_SIZE
- Type: SizeSuffix - Type: SizeSuffix
- Default: 8M - Default: 8Mi
#### --drive-acknowledge-abuse #### --drive-acknowledge-abuse

View File

@ -3,8 +3,7 @@ title: "Dropbox"
description: "Rclone docs for Dropbox" description: "Rclone docs for Dropbox"
--- ---
{{< icon "fab fa-dropbox" >}} Dropbox # {{< icon "fab fa-dropbox" >}} Dropbox
---------------------------------
Paths are specified as `remote:path` Paths are specified as `remote:path`
@ -238,7 +237,7 @@ Leave blank to use the provider defaults.
#### --dropbox-chunk-size #### --dropbox-chunk-size
Upload chunk size. (< 150M). Upload chunk size. (< 150Mi).
Any files larger than this will be uploaded in chunks of this size. Any files larger than this will be uploaded in chunks of this size.
@ -250,7 +249,7 @@ memory. It can be set smaller if you are tight on memory.
- Config: chunk_size - Config: chunk_size
- Env Var: RCLONE_DROPBOX_CHUNK_SIZE - Env Var: RCLONE_DROPBOX_CHUNK_SIZE
- Type: SizeSuffix - Type: SizeSuffix
- Default: 48M - Default: 48Mi
#### --dropbox-impersonate #### --dropbox-impersonate
@ -309,6 +308,75 @@ shared folder.
- Type: bool - Type: bool
- Default: false - Default: false
#### --dropbox-batch-mode
Upload file batching sync|async|off.
This sets the batch mode used by rclone.
For full info see [the main docs](https://rclone.org/dropbox/#batch-mode)
This has 3 possible values
- off - no batching
- sync - batch uploads and check completion (default)
- async - batch upload and don't check completion
Rclone will close any outstanding batches when it exits which may make
a delay on quit.
- Config: batch_mode
- Env Var: RCLONE_DROPBOX_BATCH_MODE
- Type: string
- Default: "sync"
#### --dropbox-batch-size
Max number of files in upload batch.
This sets the batch size of files to upload. It has to be less than 1000.
By default this is 0 which means rclone which calculate the batch size
depending on the setting of batch_mode.
- batch_mode: async - default batch_size is 100
- batch_mode: sync - default batch_size is the same as --transfers
- batch_mode: off - not in use
Rclone will close any outstanding batches when it exits which may make
a delay on quit.
Setting this is a great idea if you are uploading lots of small files
as it will make them a lot quicker. You can use --transfers 32 to
maximise throughput.
- Config: batch_size
- Env Var: RCLONE_DROPBOX_BATCH_SIZE
- Type: int
- Default: 0
#### --dropbox-batch-timeout
Max time to allow an idle upload batch before uploading
If an upload batch is idle for more than this long then it will be
uploaded.
The default for this is 0 which means rclone will choose a sensible
default based on the batch_mode in use.
- batch_mode: async - default batch_timeout is 500ms
- batch_mode: sync - default batch_timeout is 10s
- batch_mode: off - not in use
- Config: batch_timeout
- Env Var: RCLONE_DROPBOX_BATCH_TIMEOUT
- Type: Duration
- Default: 0s
#### --dropbox-encoding #### --dropbox-encoding
This sets the encoding for the backend. This sets the encoding for the backend.

View File

@ -3,8 +3,7 @@ title: "1Fichier"
description: "Rclone docs for 1Fichier" description: "Rclone docs for 1Fichier"
--- ---
{{< icon "fa fa-archive" >}} 1Fichier # {{< icon "fa fa-archive" >}} 1Fichier
-----------------------------------------
This is a backend for the [1fichier](https://1fichier.com) cloud This is a backend for the [1fichier](https://1fichier.com) cloud
storage service. Note that a Premium subscription is required to use storage service. Note that a Premium subscription is required to use
@ -139,6 +138,28 @@ If you want to download a shared folder, add this parameter
- Type: string - Type: string
- Default: "" - Default: ""
#### --fichier-file-password
If you want to download a shared file that is password protected, add this parameter
**NB** Input to this must be obscured - see [rclone obscure](/commands/rclone_obscure/).
- Config: file_password
- Env Var: RCLONE_FICHIER_FILE_PASSWORD
- Type: string
- Default: ""
#### --fichier-folder-password
If you want to list the files in a shared folder that is password protected, add this parameter
**NB** Input to this must be obscured - see [rclone obscure](/commands/rclone_obscure/).
- Config: folder_password
- Env Var: RCLONE_FICHIER_FOLDER_PASSWORD
- Type: string
- Default: ""
#### --fichier-encoding #### --fichier-encoding
This sets the encoding for the backend. This sets the encoding for the backend.

View File

@ -3,8 +3,7 @@ title: "Enterprise File Fabric"
description: "Rclone docs for the Enterprise File Fabric backend" description: "Rclone docs for the Enterprise File Fabric backend"
--- ---
{{< icon "fa fa-cloud" >}} Enterprise File Fabric # {{< icon "fa fa-cloud" >}} Enterprise File Fabric
-----------------------------------------
This backend supports [Storage Made Easy's Enterprise File This backend supports [Storage Made Easy's Enterprise File
Fabric™](https://storagemadeeasy.com/about/) which provides a software Fabric™](https://storagemadeeasy.com/about/) which provides a software

View File

@ -154,7 +154,7 @@ These flags are available for every command.
--use-json-log Use json log format. --use-json-log Use json log format.
--use-mmap Use mmap allocator (see docs). --use-mmap Use mmap allocator (see docs).
--use-server-modtime Use server modified time instead of object metadata --use-server-modtime Use server modified time instead of object metadata
--user-agent string Set the user-agent to a specified string. The default is rclone/ version (default "rclone/v1.56.0-beta.5531.41f561bf2.pr-commanddocs") --user-agent string Set the user-agent to a specified string. The default is rclone/ version (default "rclone/v1.56.0")
-v, --verbose count Print lots more stuff (repeat for more) -v, --verbose count Print lots more stuff (repeat for more)
``` ```
@ -311,6 +311,8 @@ and may be set in the config file.
--dropbox-token-url string Token server url. --dropbox-token-url string Token server url.
--fichier-api-key string Your API Key, get it from https://1fichier.com/console/params.pl --fichier-api-key string Your API Key, get it from https://1fichier.com/console/params.pl
--fichier-encoding MultiEncoder This sets the encoding for the backend. (default Slash,LtGt,DoubleQuote,SingleQuote,BackQuote,Dollar,BackSlash,Del,Ctl,LeftSpace,RightSpace,InvalidUtf8,Dot) --fichier-encoding MultiEncoder This sets the encoding for the backend. (default Slash,LtGt,DoubleQuote,SingleQuote,BackQuote,Dollar,BackSlash,Del,Ctl,LeftSpace,RightSpace,InvalidUtf8,Dot)
--fichier-file-password string If you want to download a shared file that is password protected, add this parameter (obscured)
--fichier-folder-password string If you want to list the files in a shared folder that is password protected, add this parameter (obscured)
--fichier-shared-folder string If you want to download a shared folder, add this parameter --fichier-shared-folder string If you want to download a shared folder, add this parameter
--filefabric-encoding MultiEncoder This sets the encoding for the backend. (default Slash,Del,Ctl,InvalidUtf8,Dot) --filefabric-encoding MultiEncoder This sets the encoding for the backend. (default Slash,Del,Ctl,InvalidUtf8,Dot)
--filefabric-permanent-token string Permanent Authentication Token --filefabric-permanent-token string Permanent Authentication Token
@ -375,6 +377,7 @@ and may be set in the config file.
--jottacloud-encoding MultiEncoder This sets the encoding for the backend. (default Slash,LtGt,DoubleQuote,Colon,Question,Asterisk,Pipe,Del,Ctl,InvalidUtf8,Dot) --jottacloud-encoding MultiEncoder This sets the encoding for the backend. (default Slash,LtGt,DoubleQuote,Colon,Question,Asterisk,Pipe,Del,Ctl,InvalidUtf8,Dot)
--jottacloud-hard-delete Delete files permanently rather than putting them into the trash. --jottacloud-hard-delete Delete files permanently rather than putting them into the trash.
--jottacloud-md5-memory-limit SizeSuffix Files bigger than this will be cached on disk to calculate the MD5 if required. (default 10Mi) --jottacloud-md5-memory-limit SizeSuffix Files bigger than this will be cached on disk to calculate the MD5 if required. (default 10Mi)
--jottacloud-no-versions Avoid server side versioning by deleting files and recreating files instead of overwriting them.
--jottacloud-trashed-only Only show files that are in the trash. --jottacloud-trashed-only Only show files that are in the trash.
--jottacloud-upload-resume-limit SizeSuffix Files bigger than this can be resumed if the upload fail's. (default 10Mi) --jottacloud-upload-resume-limit SizeSuffix Files bigger than this can be resumed if the upload fail's. (default 10Mi)
--koofr-encoding MultiEncoder This sets the encoding for the backend. (default Slash,BackSlash,Del,Ctl,InvalidUtf8,Dot) --koofr-encoding MultiEncoder This sets the encoding for the backend. (default Slash,BackSlash,Del,Ctl,InvalidUtf8,Dot)
@ -587,7 +590,7 @@ and may be set in the config file.
--zoho-client-id string OAuth Client Id --zoho-client-id string OAuth Client Id
--zoho-client-secret string OAuth Client Secret --zoho-client-secret string OAuth Client Secret
--zoho-encoding MultiEncoder This sets the encoding for the backend. (default Del,Ctl,InvalidUtf8) --zoho-encoding MultiEncoder This sets the encoding for the backend. (default Del,Ctl,InvalidUtf8)
--zoho-region string Zoho region to connect to. You'll have to use the region you organization is registered in. --zoho-region string Zoho region to connect to.
--zoho-token string OAuth Access Token as a JSON blob. --zoho-token string OAuth Access Token as a JSON blob.
--zoho-token-url string Token server url. --zoho-token-url string Token server url.
``` ```

View File

@ -3,8 +3,7 @@ title: "FTP"
description: "Rclone docs for FTP backend" description: "Rclone docs for FTP backend"
--- ---
{{< icon "fa fa-file" >}} FTP # {{< icon "fa fa-file" >}} FTP
------------------------------
FTP is the File Transfer Protocol. Rclone FTP support is provided using the FTP is the File Transfer Protocol. Rclone FTP support is provided using the
[github.com/jlaffaye/ftp](https://godoc.org/github.com/jlaffaye/ftp) [github.com/jlaffaye/ftp](https://godoc.org/github.com/jlaffaye/ftp)

View File

@ -3,8 +3,7 @@ title: "Google Cloud Storage"
description: "Rclone docs for Google Cloud Storage" description: "Rclone docs for Google Cloud Storage"
--- ---
{{< icon "fab fa-google" >}} Google Cloud Storage # {{< icon "fab fa-google" >}} Google Cloud Storage
-------------------------------------------------
Paths are specified as `remote:bucket` (or `remote:` for the `lsd` Paths are specified as `remote:bucket` (or `remote:` for the `lsd`
command.) You may put subdirectories in too, e.g. `remote:bucket/path/to/dir`. command.) You may put subdirectories in too, e.g. `remote:bucket/path/to/dir`.

View File

@ -3,8 +3,7 @@ title: "Google Photos"
description: "Rclone docs for Google Photos" description: "Rclone docs for Google Photos"
--- ---
{{< icon "fa fa-images" >}} Google Photos # {{< icon "fa fa-images" >}} Google Photos
-------------------------------------------------
The rclone backend for [Google Photos](https://www.google.com/photos/about/) is The rclone backend for [Google Photos](https://www.google.com/photos/about/) is
a specialized backend for transferring photos and videos to and from a specialized backend for transferring photos and videos to and from

View File

@ -3,8 +3,7 @@ title: "HDFS Remote"
description: "Remote for Hadoop Distributed Filesystem" description: "Remote for Hadoop Distributed Filesystem"
--- ---
{{< icon "fa fa-globe" >}} HDFS # {{< icon "fa fa-globe" >}} HDFS
-------------------------------------------------
[HDFS](https://hadoop.apache.org/docs/stable/hadoop-project-dist/hadoop-hdfs/HdfsDesign.html) is a [HDFS](https://hadoop.apache.org/docs/stable/hadoop-project-dist/hadoop-hdfs/HdfsDesign.html) is a
distributed file-system, part of the [Apache Hadoop](https://hadoop.apache.org/) framework. distributed file-system, part of the [Apache Hadoop](https://hadoop.apache.org/) framework.
@ -190,7 +189,7 @@ Here are the advanced options specific to hdfs (Hadoop distributed file system).
Kerberos service principal name for the namenode Kerberos service principal name for the namenode
Enables KERBEROS authentication. Specifies the Service Principal Name Enables KERBEROS authentication. Specifies the Service Principal Name
(<SERVICE>/<FQDN>) for the namenode. (SERVICE/FQDN) for the namenode.
- Config: service_principal_name - Config: service_principal_name
- Env Var: RCLONE_HDFS_SERVICE_PRINCIPAL_NAME - Env Var: RCLONE_HDFS_SERVICE_PRINCIPAL_NAME

View File

@ -3,8 +3,7 @@ title: "HTTP Remote"
description: "Read only remote for HTTP servers" description: "Read only remote for HTTP servers"
--- ---
{{< icon "fa fa-globe" >}} HTTP # {{< icon "fa fa-globe" >}} HTTP
-------------------------------------------------
The HTTP remote is a read only remote for reading files of a The HTTP remote is a read only remote for reading files of a
webserver. The webserver should provide file listings which rclone webserver. The webserver should provide file listings which rclone

View File

@ -3,8 +3,7 @@ title: "Hubic"
description: "Rclone docs for Hubic" description: "Rclone docs for Hubic"
--- ---
{{< icon "fa fa-space-shuttle" >}} Hubic # {{< icon "fa fa-space-shuttle" >}} Hubic
-----------------------------------------
Paths are specified as `remote:path` Paths are specified as `remote:path`
@ -173,7 +172,7 @@ default for this is 5 GiB which is its maximum value.
- Config: chunk_size - Config: chunk_size
- Env Var: RCLONE_HUBIC_CHUNK_SIZE - Env Var: RCLONE_HUBIC_CHUNK_SIZE
- Type: SizeSuffix - Type: SizeSuffix
- Default: 5G - Default: 5Gi
#### --hubic-no-chunk #### --hubic-no-chunk

View File

@ -3,8 +3,7 @@ title: "Jottacloud"
description: "Rclone docs for Jottacloud" description: "Rclone docs for Jottacloud"
--- ---
{{< icon "fa fa-cloud" >}} Jottacloud # {{< icon "fa fa-cloud" >}} Jottacloud
-----------------------------------------
Jottacloud is a cloud storage service provider from a Norwegian company, using its own datacenters in Norway. Jottacloud is a cloud storage service provider from a Norwegian company, using its own datacenters in Norway.

View File

@ -3,8 +3,7 @@ title: "Koofr"
description: "Rclone docs for Koofr" description: "Rclone docs for Koofr"
--- ---
{{< icon "fa fa-suitcase" >}} Koofr # {{< icon "fa fa-suitcase" >}} Koofr
-----------------------------------------
Paths are specified as `remote:path` Paths are specified as `remote:path`

View File

@ -3,8 +3,7 @@ title: "Local Filesystem"
description: "Rclone docs for the local filesystem" description: "Rclone docs for the local filesystem"
--- ---
{{< icon "fas fa-hdd" >}} Local Filesystem # {{< icon "fas fa-hdd" >}} Local Filesystem
-------------------------------------------
Local paths are specified as normal filesystem paths, e.g. `/path/to/wherever`, so Local paths are specified as normal filesystem paths, e.g. `/path/to/wherever`, so
@ -367,32 +366,41 @@ points, as you explicitly acknowledge that they should be skipped.
#### --local-zero-size-links #### --local-zero-size-links
Assume the Stat size of links is zero (and read them instead) Assume the Stat size of links is zero (and read them instead) (Deprecated)
On some virtual filesystems (such ash LucidLink), reading a link size via a Stat call always returns 0. Rclone used to use the Stat size of links as the link size, but this fails in quite a few places
However, on unix it reads as the length of the text in the link. This may cause errors like this when
syncing:
Failed to copy: corrupted on transfer: sizes differ 0 vs 13 - Windows
- On some virtual filesystems (such ash LucidLink)
- Android
So rclone now always reads the link
Setting this flag causes rclone to read the link and use that as the size of the link
instead of 0 which in most cases fixes the problem.
- Config: zero_size_links - Config: zero_size_links
- Env Var: RCLONE_LOCAL_ZERO_SIZE_LINKS - Env Var: RCLONE_LOCAL_ZERO_SIZE_LINKS
- Type: bool - Type: bool
- Default: false - Default: false
#### --local-no-unicode-normalization #### --local-unicode-normalization
Don't apply unicode normalization to paths and filenames (Deprecated) Apply unicode NFC normalization to paths and filenames
This flag is deprecated now. Rclone no longer normalizes unicode file This flag can be used to normalize file names into unicode NFC form
names, but it compares them with unicode normalization in the sync that are read from the local filesystem.
routine instead.
- Config: no_unicode_normalization Rclone does not normally touch the encoding of file names it reads from
- Env Var: RCLONE_LOCAL_NO_UNICODE_NORMALIZATION the file system.
This can be useful when using macOS as it normally provides decomposed (NFD)
unicode which in some language (eg Korean) doesn't display properly on
some OSes.
Note that rclone compares filenames with unicode normalization in the sync
routine so this flag shouldn't normally be used.
- Config: unicode_normalization
- Env Var: RCLONE_LOCAL_UNICODE_NORMALIZATION
- Type: bool - Type: bool
- Default: false - Default: false

View File

@ -3,8 +3,7 @@ title: "Mailru"
description: "Mail.ru Cloud" description: "Mail.ru Cloud"
--- ---
{{< icon "fas fa-at" >}} Mail.ru Cloud # {{< icon "fas fa-at" >}} Mail.ru Cloud
----------------------------------------
[Mail.ru Cloud](https://cloud.mail.ru/) is a cloud storage provided by a Russian internet company [Mail.Ru Group](https://mail.ru). The official desktop client is [Disk-O:](https://disk-o.cloud/en), available on Windows and Mac OS. [Mail.ru Cloud](https://cloud.mail.ru/) is a cloud storage provided by a Russian internet company [Mail.Ru Group](https://mail.ru). The official desktop client is [Disk-O:](https://disk-o.cloud/en), available on Windows and Mac OS.
@ -241,7 +240,7 @@ This option allows you to disable speedup (put by hash) for large files
- Config: speedup_max_disk - Config: speedup_max_disk
- Env Var: RCLONE_MAILRU_SPEEDUP_MAX_DISK - Env Var: RCLONE_MAILRU_SPEEDUP_MAX_DISK
- Type: SizeSuffix - Type: SizeSuffix
- Default: 3G - Default: 3Gi
- Examples: - Examples:
- "0" - "0"
- Completely disable speedup (put by hash). - Completely disable speedup (put by hash).
@ -257,7 +256,7 @@ Files larger than the size given below will always be hashed on disk.
- Config: speedup_max_memory - Config: speedup_max_memory
- Env Var: RCLONE_MAILRU_SPEEDUP_MAX_MEMORY - Env Var: RCLONE_MAILRU_SPEEDUP_MAX_MEMORY
- Type: SizeSuffix - Type: SizeSuffix
- Default: 32M - Default: 32Mi
- Examples: - Examples:
- "0" - "0"
- Preliminary hashing will always be done in a temporary disk location. - Preliminary hashing will always be done in a temporary disk location.

View File

@ -3,8 +3,7 @@ title: "Mega"
description: "Rclone docs for Mega" description: "Rclone docs for Mega"
--- ---
{{< icon "fa fa-archive" >}} Mega # {{< icon "fa fa-archive" >}} Mega
-----------------------------------------
[Mega](https://mega.nz/) is a cloud storage and file hosting service [Mega](https://mega.nz/) is a cloud storage and file hosting service
known for its security feature where all files are encrypted locally known for its security feature where all files are encrypted locally

View File

@ -3,8 +3,7 @@ title: "Memory"
description: "Rclone docs for Memory backend" description: "Rclone docs for Memory backend"
--- ---
{{< icon "fas fa-memory" >}} Memory # {{< icon "fas fa-memory" >}} Memory
-----------------------------------------
The memory backend is an in RAM backend. It does not persist its The memory backend is an in RAM backend. It does not persist its
data - use the local backend for that. data - use the local backend for that.

View File

@ -3,8 +3,7 @@ title: "Microsoft OneDrive"
description: "Rclone docs for Microsoft OneDrive" description: "Rclone docs for Microsoft OneDrive"
--- ---
{{< icon "fab fa-windows" >}} Microsoft OneDrive # {{< icon "fab fa-windows" >}} Microsoft OneDrive
-----------------------------------------
Paths are specified as `remote:path` Paths are specified as `remote:path`
@ -277,7 +276,7 @@ Note that the chunks will be buffered into memory.
- Config: chunk_size - Config: chunk_size
- Env Var: RCLONE_ONEDRIVE_CHUNK_SIZE - Env Var: RCLONE_ONEDRIVE_CHUNK_SIZE
- Type: SizeSuffix - Type: SizeSuffix
- Default: 10M - Default: 10Mi
#### --onedrive-drive-id #### --onedrive-drive-id

View File

@ -3,8 +3,7 @@ title: "OpenDrive"
description: "Rclone docs for OpenDrive" description: "Rclone docs for OpenDrive"
--- ---
{{< icon "fa fa-file" >}} OpenDrive # {{< icon "fa fa-file" >}} OpenDrive
------------------------------------
Paths are specified as `remote:path` Paths are specified as `remote:path`
@ -148,7 +147,7 @@ increase memory use.
- Config: chunk_size - Config: chunk_size
- Env Var: RCLONE_OPENDRIVE_CHUNK_SIZE - Env Var: RCLONE_OPENDRIVE_CHUNK_SIZE
- Type: SizeSuffix - Type: SizeSuffix
- Default: 10M - Default: 10Mi
{{< rem autogenerated options stop >}} {{< rem autogenerated options stop >}}

View File

@ -3,8 +3,7 @@ title: "pCloud"
description: "Rclone docs for pCloud" description: "Rclone docs for pCloud"
--- ---
{{< icon "fa fa-cloud" >}} pCloud # {{< icon "fa fa-cloud" >}} pCloud
-----------------------------------------
Paths are specified as `remote:path` Paths are specified as `remote:path`

View File

@ -3,8 +3,7 @@ title: "premiumize.me"
description: "Rclone docs for premiumize.me" description: "Rclone docs for premiumize.me"
--- ---
{{< icon "fa fa-user" >}} premiumize.me # {{< icon "fa fa-user" >}} premiumize.me
-----------------------------------------
Paths are specified as `remote:path` Paths are specified as `remote:path`

View File

@ -3,8 +3,7 @@ title: "put.io"
description: "Rclone docs for put.io" description: "Rclone docs for put.io"
--- ---
{{< icon "fas fa-parking" >}} put.io # {{< icon "fas fa-parking" >}} put.io
---------------------------------
Paths are specified as `remote:path` Paths are specified as `remote:path`

View File

@ -3,8 +3,7 @@ title: "QingStor"
description: "Rclone docs for QingStor Object Storage" description: "Rclone docs for QingStor Object Storage"
--- ---
{{< icon "fas fa-hdd" >}} QingStor # {{< icon "fas fa-hdd" >}} QingStor
---------------------------------------
Paths are specified as `remote:bucket` (or `remote:` for the `lsd` Paths are specified as `remote:bucket` (or `remote:` for the `lsd`
command.) You may put subdirectories in too, e.g. `remote:bucket/path/to/dir`. command.) You may put subdirectories in too, e.g. `remote:bucket/path/to/dir`.
@ -232,7 +231,7 @@ The minimum is 0 and the maximum is 5 GiB.
- Config: upload_cutoff - Config: upload_cutoff
- Env Var: RCLONE_QINGSTOR_UPLOAD_CUTOFF - Env Var: RCLONE_QINGSTOR_UPLOAD_CUTOFF
- Type: SizeSuffix - Type: SizeSuffix
- Default: 200M - Default: 200Mi
#### --qingstor-chunk-size #### --qingstor-chunk-size
@ -250,7 +249,7 @@ enough memory, then increasing this will speed up the transfers.
- Config: chunk_size - Config: chunk_size
- Env Var: RCLONE_QINGSTOR_CHUNK_SIZE - Env Var: RCLONE_QINGSTOR_CHUNK_SIZE
- Type: SizeSuffix - Type: SizeSuffix
- Default: 4M - Default: 4Mi
#### --qingstor-upload-concurrency #### --qingstor-upload-concurrency

View File

@ -525,8 +525,14 @@ This takes the following parameters
- name - name of remote - name - name of remote
- parameters - a map of \{ "key": "value" \} pairs - parameters - a map of \{ "key": "value" \} pairs
- type - type of the new remote - type - type of the new remote
- obscure - optional bool - forces obscuring of passwords - opt - a dictionary of options to control the configuration
- noObscure - optional bool - forces passwords not to be obscured - obscure - declare passwords are plain and need obscuring
- noObscure - declare passwords are already obscured and don't need obscuring
- nonInteractive - don't interact with a user, return questions
- continue - continue the config process with an answer
- all - ask all the config questions not just the post config ones
- state - state to restart with - used with continue
- result - result to restart with - used with continue
See the [config create command](/commands/rclone_config_create/) command for more information on the above. See the [config create command](/commands/rclone_config_create/) command for more information on the above.
@ -600,8 +606,14 @@ This takes the following parameters
- name - name of remote - name - name of remote
- parameters - a map of \{ "key": "value" \} pairs - parameters - a map of \{ "key": "value" \} pairs
- obscure - optional bool - forces obscuring of passwords - opt - a dictionary of options to control the configuration
- noObscure - optional bool - forces passwords not to be obscured - obscure - declare passwords are plain and need obscuring
- noObscure - declare passwords are already obscured and don't need obscuring
- nonInteractive - don't interact with a user, return questions
- continue - continue the config process with an answer
- all - ask all the config questions not just the post config ones
- state - state to restart with - used with continue
- result - result to restart with - used with continue
See the [config update command](/commands/rclone_config_update/) command for more information on the above. See the [config update command](/commands/rclone_config_update/) command for more information on the above.
@ -775,7 +787,7 @@ Returns the following values:
"lastError": last error string, "lastError": last error string,
"renames" : number of files renamed, "renames" : number of files renamed,
"retryError": boolean showing whether there has been at least one non-NoRetryError, "retryError": boolean showing whether there has been at least one non-NoRetryError,
"speed": average speed in bytes/sec since start of the group, "speed": average speed in bytes per second since start of the group,
"totalBytes": total number of bytes in the group, "totalBytes": total number of bytes in the group,
"totalChecks": total number of checks in the group, "totalChecks": total number of checks in the group,
"totalTransfers": total number of transfers in the group, "totalTransfers": total number of transfers in the group,

View File

@ -3,8 +3,7 @@ title: "Amazon S3"
description: "Rclone docs for Amazon S3" description: "Rclone docs for Amazon S3"
--- ---
{{< icon "fab fa-amazon" >}} Amazon S3 Storage Providers # {{< icon "fab fa-amazon" >}} Amazon S3 Storage Providers
--------------------------------------------------------
The S3 backend can be used with a number of different providers: The S3 backend can be used with a number of different providers:
@ -894,6 +893,10 @@ Endpoint for OSS API.
- Type: string - Type: string
- Default: "" - Default: ""
- Examples: - Examples:
- "oss-accelerate.aliyuncs.com"
- Global Accelerate
- "oss-accelerate-overseas.aliyuncs.com"
- Global Accelerate (outside mainland China)
- "oss-cn-hangzhou.aliyuncs.com" - "oss-cn-hangzhou.aliyuncs.com"
- East China 1 (Hangzhou) - East China 1 (Hangzhou)
- "oss-cn-shanghai.aliyuncs.com" - "oss-cn-shanghai.aliyuncs.com"
@ -905,9 +908,17 @@ Endpoint for OSS API.
- "oss-cn-zhangjiakou.aliyuncs.com" - "oss-cn-zhangjiakou.aliyuncs.com"
- North China 3 (Zhangjiakou) - North China 3 (Zhangjiakou)
- "oss-cn-huhehaote.aliyuncs.com" - "oss-cn-huhehaote.aliyuncs.com"
- North China 5 (Huhehaote) - North China 5 (Hohhot)
- "oss-cn-wulanchabu.aliyuncs.com"
- North China 6 (Ulanqab)
- "oss-cn-shenzhen.aliyuncs.com" - "oss-cn-shenzhen.aliyuncs.com"
- South China 1 (Shenzhen) - South China 1 (Shenzhen)
- "oss-cn-heyuan.aliyuncs.com"
- South China 2 (Heyuan)
- "oss-cn-guangzhou.aliyuncs.com"
- South China 3 (Guangzhou)
- "oss-cn-chengdu.aliyuncs.com"
- West China 1 (Chengdu)
- "oss-cn-hongkong.aliyuncs.com" - "oss-cn-hongkong.aliyuncs.com"
- Hong Kong (Hong Kong) - Hong Kong (Hong Kong)
- "oss-us-west-1.aliyuncs.com" - "oss-us-west-1.aliyuncs.com"
@ -1029,6 +1040,8 @@ Required when using an S3 clone.
- Digital Ocean Spaces Amsterdam 3 - Digital Ocean Spaces Amsterdam 3
- "sgp1.digitaloceanspaces.com" - "sgp1.digitaloceanspaces.com"
- Digital Ocean Spaces Singapore 1 - Digital Ocean Spaces Singapore 1
- "localhost:8333"
- SeaweedFS S3 localhost
- "s3.wasabisys.com" - "s3.wasabisys.com"
- Wasabi US East endpoint - Wasabi US East endpoint
- "s3.us-west-1.wasabisys.com" - "s3.us-west-1.wasabisys.com"
@ -1334,7 +1347,7 @@ The storage class to use when storing new objects in S3.
### Advanced Options ### Advanced Options
Here are the advanced options specific to s3 (Amazon S3 Compliant Storage Providers including AWS, Alibaba, Ceph, Digital Ocean, Dreamhost, IBM COS, Minio, and Tencent COS). Here are the advanced options specific to s3 (Amazon S3 Compliant Storage Providers including AWS, Alibaba, Ceph, Digital Ocean, Dreamhost, IBM COS, Minio, SeaweedFS, and Tencent COS).
#### --s3-bucket-acl #### --s3-bucket-acl
@ -1420,7 +1433,7 @@ The minimum is 0 and the maximum is 5 GiB.
- Config: upload_cutoff - Config: upload_cutoff
- Env Var: RCLONE_S3_UPLOAD_CUTOFF - Env Var: RCLONE_S3_UPLOAD_CUTOFF
- Type: SizeSuffix - Type: SizeSuffix
- Default: 200M - Default: 200Mi
#### --s3-chunk-size #### --s3-chunk-size
@ -1449,7 +1462,7 @@ larger files then you will need to increase chunk_size.
- Config: chunk_size - Config: chunk_size
- Env Var: RCLONE_S3_CHUNK_SIZE - Env Var: RCLONE_S3_CHUNK_SIZE
- Type: SizeSuffix - Type: SizeSuffix
- Default: 5M - Default: 5Mi
#### --s3-max-upload-parts #### --s3-max-upload-parts
@ -1482,7 +1495,7 @@ The minimum is 0 and the maximum is 5 GiB.
- Config: copy_cutoff - Config: copy_cutoff
- Env Var: RCLONE_S3_COPY_CUTOFF - Env Var: RCLONE_S3_COPY_CUTOFF
- Type: SizeSuffix - Type: SizeSuffix
- Default: 4.656G - Default: 4.656Gi
#### --s3-disable-checksum #### --s3-disable-checksum
@ -1684,6 +1697,15 @@ very small even with this flag.
- Type: bool - Type: bool
- Default: false - Default: false
#### --s3-no-head-object
If set, don't HEAD objects
- Config: no_head_object
- Env Var: RCLONE_S3_NO_HEAD_OBJECT
- Type: bool
- Default: false
#### --s3-encoding #### --s3-encoding
This sets the encoding for the backend. This sets the encoding for the backend.
@ -1884,7 +1906,7 @@ Then use it as normal with the name of the public bucket, e.g.
You will be able to list and copy data but not upload it. You will be able to list and copy data but not upload it.
### Ceph ### ## Ceph
[Ceph](https://ceph.com/) is an open source unified, distributed [Ceph](https://ceph.com/) is an open source unified, distributed
storage system designed for excellent performance, reliability and storage system designed for excellent performance, reliability and
@ -1940,7 +1962,7 @@ removed).
Because this is a json dump, it is encoding the `/` as `\/`, so if you Because this is a json dump, it is encoding the `/` as `\/`, so if you
use the secret key as `xxxxxx/xxxx` it will work fine. use the secret key as `xxxxxx/xxxx` it will work fine.
### Dreamhost ### ## Dreamhost
Dreamhost [DreamObjects](https://www.dreamhost.com/cloud/storage/) is Dreamhost [DreamObjects](https://www.dreamhost.com/cloud/storage/) is
an object storage system based on CEPH. an object storage system based on CEPH.
@ -1964,7 +1986,7 @@ server_side_encryption =
storage_class = storage_class =
``` ```
### DigitalOcean Spaces ### ## DigitalOcean Spaces
[Spaces](https://www.digitalocean.com/products/object-storage/) is an [S3-interoperable](https://developers.digitalocean.com/documentation/spaces/) object storage service from cloud provider DigitalOcean. [Spaces](https://www.digitalocean.com/products/object-storage/) is an [S3-interoperable](https://developers.digitalocean.com/documentation/spaces/) object storage service from cloud provider DigitalOcean.
@ -2010,7 +2032,7 @@ rclone mkdir spaces:my-new-space
rclone copy /path/to/files spaces:my-new-space rclone copy /path/to/files spaces:my-new-space
``` ```
### IBM COS (S3) ### ## IBM COS (S3)
Information stored with IBM Cloud Object Storage is encrypted and dispersed across multiple geographic locations, and accessed through an implementation of the S3 API. This service makes use of the distributed storage technologies provided by IBMs Cloud Object Storage System (formerly Cleversafe). For more information visit: (http://www.ibm.com/cloud/object-storage) Information stored with IBM Cloud Object Storage is encrypted and dispersed across multiple geographic locations, and accessed through an implementation of the S3 API. This service makes use of the distributed storage technologies provided by IBMs Cloud Object Storage System (formerly Cleversafe). For more information visit: (http://www.ibm.com/cloud/object-storage)
@ -2182,7 +2204,7 @@ acl> 1
rclone delete IBM-COS-XREGION:newbucket/file.txt rclone delete IBM-COS-XREGION:newbucket/file.txt
``` ```
### Minio ### ## Minio
[Minio](https://minio.io/) is an object storage server built for cloud application developers and devops. [Minio](https://minio.io/) is an object storage server built for cloud application developers and devops.
@ -2249,7 +2271,7 @@ So once set up, for example to copy files into a bucket
rclone copy /path/to/files minio:bucket rclone copy /path/to/files minio:bucket
``` ```
### Scaleway {#scaleway} ## Scaleway
[Scaleway](https://www.scaleway.com/object-storage/) The Object Storage platform allows you to store anything from backups, logs and web assets to documents and photos. [Scaleway](https://www.scaleway.com/object-storage/) The Object Storage platform allows you to store anything from backups, logs and web assets to documents and photos.
Files can be dropped from the Scaleway console or transferred through our API and CLI or using any S3-compatible tool. Files can be dropped from the Scaleway console or transferred through our API and CLI or using any S3-compatible tool.
@ -2271,7 +2293,7 @@ server_side_encryption =
storage_class = storage_class =
``` ```
### SeaweedFS ### ## SeaweedFS
[SeaweedFS](https://github.com/chrislusf/seaweedfs/) is a distributed storage system for [SeaweedFS](https://github.com/chrislusf/seaweedfs/) is a distributed storage system for
blobs, objects, files, and data lake, with O(1) disk seek and a scalable file metadata store. blobs, objects, files, and data lake, with O(1) disk seek and a scalable file metadata store.
@ -2321,7 +2343,7 @@ So once set up, for example to copy files into a bucket
rclone copy /path/to/files seaweedfs_s3:foo rclone copy /path/to/files seaweedfs_s3:foo
``` ```
### Wasabi ### ## Wasabi
[Wasabi](https://wasabi.com) is a cloud-based object storage service for a [Wasabi](https://wasabi.com) is a cloud-based object storage service for a
broad range of applications and use cases. Wasabi is designed for broad range of applications and use cases. Wasabi is designed for
@ -2434,7 +2456,7 @@ server_side_encryption =
storage_class = storage_class =
``` ```
### Alibaba OSS {#alibaba-oss} ## Alibaba OSS {#alibaba-oss}
Here is an example of making an [Alibaba Cloud (Aliyun) OSS](https://www.alibabacloud.com/product/oss/) Here is an example of making an [Alibaba Cloud (Aliyun) OSS](https://www.alibabacloud.com/product/oss/)
configuration. First run: configuration. First run:
@ -2544,7 +2566,7 @@ d) Delete this remote
y/e/d> y y/e/d> y
``` ```
### Tencent COS {#tencent-cos} ## Tencent COS {#tencent-cos}
[Tencent Cloud Object Storage (COS)](https://intl.cloud.tencent.com/product/cos) is a distributed storage service offered by Tencent Cloud for unstructured data. It is secure, stable, massive, convenient, low-delay and low-cost. [Tencent Cloud Object Storage (COS)](https://intl.cloud.tencent.com/product/cos) is a distributed storage service offered by Tencent Cloud for unstructured data. It is secure, stable, massive, convenient, low-delay and low-cost.
@ -2676,13 +2698,13 @@ Name Type
cos s3 cos s3
``` ```
### Netease NOS ## Netease NOS
For Netease NOS configure as per the configurator `rclone config` For Netease NOS configure as per the configurator `rclone config`
setting the provider `Netease`. This will automatically set setting the provider `Netease`. This will automatically set
`force_path_style = false` which is necessary for it to run properly. `force_path_style = false` which is necessary for it to run properly.
### Limitations ## Limitations
`rclone about` is not supported by the S3 backend. Backends without `rclone about` is not supported by the S3 backend. Backends without
this capability cannot determine free space for an rclone mount or this capability cannot determine free space for an rclone mount or

View File

@ -3,8 +3,7 @@ title: "Seafile"
description: "Seafile" description: "Seafile"
--- ---
{{< icon "fa fa-server" >}}Seafile # {{< icon "fa fa-server" >}}Seafile
----------------------------------------
This is a backend for the [Seafile](https://www.seafile.com/) storage service: This is a backend for the [Seafile](https://www.seafile.com/) storage service:
- It works with both the free community edition or the professional edition. - It works with both the free community edition or the professional edition.

View File

@ -3,8 +3,7 @@ title: "SFTP"
description: "SFTP" description: "SFTP"
--- ---
{{< icon "fa fa-server" >}} SFTP # {{< icon "fa fa-server" >}} SFTP
----------------------------------------
SFTP is the [Secure (or SSH) File Transfer SFTP is the [Secure (or SSH) File Transfer
Protocol](https://en.wikipedia.org/wiki/SSH_File_Transfer_Protocol). Protocol](https://en.wikipedia.org/wiki/SSH_File_Transfer_Protocol).
@ -531,6 +530,21 @@ If concurrent reads are disabled, the use_fstat option is ignored.
- Type: bool - Type: bool
- Default: false - Default: false
#### --sftp-disable-concurrent-writes
If set don't use concurrent writes
Normally rclone uses concurrent writes to upload files. This improves
the performance greatly, especially for distant servers.
This option disables concurrent writes should that be necessary.
- Config: disable_concurrent_writes
- Env Var: RCLONE_SFTP_DISABLE_CONCURRENT_WRITES
- Type: bool
- Default: false
#### --sftp-idle-timeout #### --sftp-idle-timeout
Max time before closing idle connections Max time before closing idle connections

View File

@ -3,7 +3,7 @@ title: "Citrix ShareFile"
description: "Rclone docs for Citrix ShareFile" description: "Rclone docs for Citrix ShareFile"
--- ---
## {{< icon "fas fa-share-square" >}} Citrix ShareFile # {{< icon "fas fa-share-square" >}} Citrix ShareFile
[Citrix ShareFile](https://sharefile.com) is a secure file sharing and transfer service aimed as business. [Citrix ShareFile](https://sharefile.com) is a secure file sharing and transfer service aimed as business.
@ -191,7 +191,7 @@ Cutoff for switching to multipart upload.
- Config: upload_cutoff - Config: upload_cutoff
- Env Var: RCLONE_SHAREFILE_UPLOAD_CUTOFF - Env Var: RCLONE_SHAREFILE_UPLOAD_CUTOFF
- Type: SizeSuffix - Type: SizeSuffix
- Default: 128M - Default: 128Mi
#### --sharefile-chunk-size #### --sharefile-chunk-size
@ -205,7 +205,7 @@ Reducing this will reduce memory usage but decrease performance.
- Config: chunk_size - Config: chunk_size
- Env Var: RCLONE_SHAREFILE_CHUNK_SIZE - Env Var: RCLONE_SHAREFILE_CHUNK_SIZE
- Type: SizeSuffix - Type: SizeSuffix
- Default: 64M - Default: 64Mi
#### --sharefile-endpoint #### --sharefile-endpoint

View File

@ -3,8 +3,7 @@ title: "SugarSync"
description: "Rclone docs for SugarSync" description: "Rclone docs for SugarSync"
--- ---
{{< icon "fas fa-dove" >}} SugarSync # {{< icon "fas fa-dove" >}} SugarSync
-----------------------------------------
[SugarSync](https://sugarsync.com) is a cloud service that enables [SugarSync](https://sugarsync.com) is a cloud service that enables
active synchronization of files across computers and other devices for active synchronization of files across computers and other devices for

View File

@ -3,8 +3,7 @@ title: "Swift"
description: "Swift" description: "Swift"
--- ---
{{< icon "fa fa-space-shuttle" >}}Swift # {{< icon "fa fa-space-shuttle" >}}Swift
----------------------------------------
Swift refers to [OpenStack Object Storage](https://docs.openstack.org/swift/latest/). Swift refers to [OpenStack Object Storage](https://docs.openstack.org/swift/latest/).
Commercial implementations of that being: Commercial implementations of that being:
@ -449,7 +448,7 @@ default for this is 5 GiB which is its maximum value.
- Config: chunk_size - Config: chunk_size
- Env Var: RCLONE_SWIFT_CHUNK_SIZE - Env Var: RCLONE_SWIFT_CHUNK_SIZE
- Type: SizeSuffix - Type: SizeSuffix
- Default: 5G - Default: 5Gi
#### --swift-no-chunk #### --swift-no-chunk

View File

@ -3,8 +3,7 @@ title: "Tardigrade"
description: "Rclone docs for Tardigrade" description: "Rclone docs for Tardigrade"
--- ---
{{< icon "fas fa-dove" >}} Tardigrade # {{< icon "fas fa-dove" >}} Tardigrade
-----------------------------------------
[Tardigrade](https://tardigrade.io) is an encrypted, secure, and [Tardigrade](https://tardigrade.io) is an encrypted, secure, and
cost-effective object storage service that enables you to store, back up, and cost-effective object storage service that enables you to store, back up, and

View File

@ -3,8 +3,7 @@ title: "Union"
description: "Remote Unification" description: "Remote Unification"
--- ---
{{< icon "fa fa-link" >}} Union # {{< icon "fa fa-link" >}} Union
-----------------------------------------
The `union` remote provides a unification similar to UnionFS using other remotes. The `union` remote provides a unification similar to UnionFS using other remotes.

View File

@ -3,8 +3,7 @@ title: "Uptobox"
description: "Rclone docs for Uptobox" description: "Rclone docs for Uptobox"
--- ---
{{< icon "fa fa-archive" >}} Uptobox # {{< icon "fa fa-archive" >}} Uptobox
-----------------------------------------
This is a Backend for Uptobox file storage service. Uptobox is closer to a one-click hoster than a traditional This is a Backend for Uptobox file storage service. Uptobox is closer to a one-click hoster than a traditional
cloud storage provider and therefore not suitable for long term storage. cloud storage provider and therefore not suitable for long term storage.
@ -13,9 +12,9 @@ Paths are specified as `remote:path`
Paths may be as deep as required, e.g. `remote:directory/subdirectory`. Paths may be as deep as required, e.g. `remote:directory/subdirectory`.
## Setup ### Setup
To configure an Uptobox backend you'll need your personal api token. You'll find it in you To configure an Uptobox backend you'll need your personal api token. You'll find it in your
[account settings](https://uptobox.com/my_account) [account settings](https://uptobox.com/my_account)
@ -107,12 +106,12 @@ as they can't be used in XML strings.
Here are the standard options specific to uptobox (Uptobox). Here are the standard options specific to uptobox (Uptobox).
#### --uptobox-api-key #### --uptobox-access-token
Your API Key, get it from https://uptobox.com/my_account Your access Token, get it from https://uptobox.com/my_account
- Config: api_key - Config: access_token
- Env Var: RCLONE_UPTOBOX_API_KEY - Env Var: RCLONE_UPTOBOX_ACCESS_TOKEN
- Type: string - Type: string
- Default: "" - Default: ""
@ -129,7 +128,7 @@ See: the [encoding section in the overview](/overview/#encoding) for more info.
- Config: encoding - Config: encoding
- Env Var: RCLONE_UPTOBOX_ENCODING - Env Var: RCLONE_UPTOBOX_ENCODING
- Type: MultiEncoder - Type: MultiEncoder
- Default: Slash,LtGt,DoubleQuote,SingleQuote,BackQuote,Dollar,BackSlash,Del,Ctl,LeftSpace,RightSpace,InvalidUtf8,Dot - Default: Slash,LtGt,DoubleQuote,BackQuote,Del,Ctl,LeftSpace,InvalidUtf8,Dot
{{< rem autogenerated options stop >}} {{< rem autogenerated options stop >}}
@ -138,4 +137,4 @@ See: the [encoding section in the overview](/overview/#encoding) for more info.
Uptobox will delete inactive files that have not been accessed in 60 days. Uptobox will delete inactive files that have not been accessed in 60 days.
`rclone about` is not supported by this backend an overview of used space can however `rclone about` is not supported by this backend an overview of used space can however
been seen in the uptobox web interface. been seen in the uptobox web interface.

View File

@ -3,8 +3,7 @@ title: "WebDAV"
description: "Rclone docs for WebDAV" description: "Rclone docs for WebDAV"
--- ---
{{< icon "fa fa-globe" >}} WebDAV # {{< icon "fa fa-globe" >}} WebDAV
-----------------------------------------
Paths are specified as `remote:path` Paths are specified as `remote:path`

View File

@ -3,8 +3,7 @@ title: "Yandex"
description: "Yandex Disk" description: "Yandex Disk"
--- ---
{{< icon "fa fa-space-shuttle" >}}Yandex Disk # {{< icon "fa fa-space-shuttle" >}}Yandex Disk
----------------------------------------
[Yandex Disk](https://disk.yandex.com) is a cloud storage solution created by [Yandex](https://yandex.com). [Yandex Disk](https://disk.yandex.com) is a cloud storage solution created by [Yandex](https://yandex.com).

View File

@ -3,8 +3,7 @@ title: "Zoho"
description: "Zoho WorkDrive" description: "Zoho WorkDrive"
--- ---
{{< icon "fas fa-folder" >}}Zoho Workdrive # {{< icon "fas fa-folder" >}}Zoho Workdrive
----------------------------------------
[Zoho WorkDrive](https://www.zoho.com/workdrive/) is a cloud storage solution created by [Zoho](https://zoho.com). [Zoho WorkDrive](https://www.zoho.com/workdrive/) is a cloud storage solution created by [Zoho](https://zoho.com).
@ -150,7 +149,11 @@ Leave blank normally.
#### --zoho-region #### --zoho-region
Zoho region to connect to. You'll have to use the region you organization is registered in. Zoho region to connect to.
You'll have to use the region your organization is registered in. If
not sure use the same top level domain as you connect to in your
browser.
- Config: region - Config: region
- Env Var: RCLONE_ZOHO_REGION - Env Var: RCLONE_ZOHO_REGION

3946
rclone.1 generated

File diff suppressed because it is too large Load Diff