docs: improve grammar and fix typos (#5361)

This alters some comments in source files, but is interested mainly in documentation files and help messages.
This commit is contained in:
Atílio Antônio
2021-11-04 08:50:43 -03:00
committed by GitHub
parent 454574e2cc
commit c08d48a50d
59 changed files with 179 additions and 179 deletions

View File

@@ -19,8 +19,8 @@ notoc: true
## About rclone {#about}
Rclone is a command line program to manage files on cloud storage. It
is a feature rich alternative to cloud vendors' web storage
Rclone is a command-line program to manage files on cloud storage. It
is a feature-rich alternative to cloud vendors' web storage
interfaces. [Over 40 cloud storage products](#providers) support
rclone including S3 object stores, business & consumer file storage
services, as well as standard transfer protocols.
@@ -43,7 +43,7 @@ bandwidth use and transfers from one provider to another without
using local disk.
Virtual backends wrap local and cloud file systems to apply
[encryption](/crypt/),
[encryption](/crypt/),
[compression](/compress/),
[chunking](/chunker/),
[hashing](/hasher/) and
@@ -58,13 +58,13 @@ macOS, linux and FreeBSD, and also serves these over
[FTP](/commands/rclone_serve_ftp/) and
[DLNA](/commands/rclone_serve_dlna/).
Rclone is mature, open source software originally inspired by rsync
Rclone is mature, open-source software originally inspired by rsync
and written in [Go](https://golang.org). The friendly support
community are familiar with varied use cases. Official Ubuntu, Debian,
community is familiar with varied use cases. Official Ubuntu, Debian,
Fedora, Brew and Chocolatey repos. include rclone. For the latest
version [downloading from rclone.org](/downloads/) is recommended.
Rclone is widely used on Linux, Windows and Mac. Third party
Rclone is widely used on Linux, Windows and Mac. Third-party
developers create innovative backup, restore, GUI and business
process solutions using the rclone command line or API.
@@ -77,7 +77,7 @@ Rclone helps you:
- Backup (and encrypt) files to cloud storage
- Restore (and decrypt) files from cloud storage
- Mirror cloud data to other cloud services or locally
- Migrate data to cloud, or between cloud storage vendors
- Migrate data to the cloud, or between cloud storage vendors
- Mount multiple, encrypted, cached or diverse cloud storage as a disk
- Analyse and account for data held on cloud storage using [lsf](/commands/rclone_lsf/), [ljson](/commands/rclone_lsjson/), [size](/commands/rclone_size/), [ncdu](/commands/rclone_ncdu/)
- [Union](/union/) file systems together to present multiple local and/or cloud file systems as one

View File

@@ -36,7 +36,7 @@ which pass through it.
Since rclone doesn't currently have its own Amazon Drive credentials
so you will either need to have your own `client_id` and
`client_secret` with Amazon Drive, or use a third party oauth proxy
`client_secret` with Amazon Drive, or use a third-party oauth proxy
in which case you will need to enter `client_id`, `client_secret`,
`auth_url` and `token_url`.
@@ -148,7 +148,7 @@ as they can't be used in JSON strings.
Any files you delete with rclone will end up in the trash. Amazon
don't provide an API to permanently delete files, nor to empty the
trash, so you will have to do that with one of Amazon's apps or via
the Amazon Drive website. As of November 17, 2016, files are
the Amazon Drive website. As of November 17, 2016, files are
automatically deleted by Amazon from the trash after 30 days.
### Using with non `.com` Amazon accounts

View File

@@ -22,11 +22,11 @@ Millions of files in a directory tends to occur on bucket-based remotes
(e.g. S3 buckets) since those remotes do not segregate subdirectories within
the bucket.
### Bucket based remotes and folders
### Bucket-based remotes and folders
Bucket based remotes (e.g. S3/GCS/Swift/B2) do not have a concept of
Bucket-based remotes (e.g. S3/GCS/Swift/B2) do not have a concept of
directories. Rclone therefore cannot create directories in them which
means that empty directories on a bucket based remote will tend to
means that empty directories on a bucket-based remote will tend to
disappear.
Some software creates empty keys ending in `/` as directory markers.

View File

@@ -376,7 +376,7 @@ description: "Rclone Changelog"
* New Features
* [Connection strings](/docs/#connection-strings)
* Config parameters can now be passed as part of the remote name as a connection string.
* For example to do the equivalent of `--drive-shared-with-me` use `drive,shared_with_me:`
* For example, to do the equivalent of `--drive-shared-with-me` use `drive,shared_with_me:`
* Make sure we don't save on the fly remote config to the config file (Nick Craig-Wood)
* Make sure backends with additional config have a different name for caching (Nick Craig-Wood)
* This work was sponsored by CERN, through the [CS3MESH4EOSC Project](https://cs3mesh4eosc.eu/).
@@ -629,7 +629,7 @@ description: "Rclone Changelog"
* And thanks to these people for many doc fixes too numerous to list
* Ameer Dawood, Antoine GIRARD, Bob Bagwill, Christopher Stewart
* CokeMine, David, Dov Murik, Durval Menezes, Evan Harris, gtorelly
* Ilyess Bachiri, Janne Johansson, Kerry Su, Marcin Zelent,
* Ilyess Bachiri, Janne Johansson, Kerry Su, Marcin Zelent,
* Martin Michlmayr, Milly, Sơn Trần-Nguyễn
* Mount
* Update systemd status with cache stats (Hekmon)
@@ -1174,7 +1174,7 @@ all the docs and Edward Barker for helping re-write the front page.
* [Union](/union/) re-write to have multiple writable remotes (Max Sum)
* [Seafile](/seafile) for Seafile server (Fred @creativeprojects)
* New commands
* backend: command for backend specific commands (see backends) (Nick Craig-Wood)
* backend: command for backend-specific commands (see backends) (Nick Craig-Wood)
* cachestats: Deprecate in favour of `rclone backend stats cache:` (Nick Craig-Wood)
* dbhashsum: Deprecate in favour of `rclone hashsum DropboxHash` (Nick Craig-Wood)
* New Features
@@ -1211,7 +1211,7 @@ all the docs and Edward Barker for helping re-write the front page.
* lsjson: Add `--hash-type` parameter and use it in lsf to speed up hashing (Nick Craig-Wood)
* rc
* Add `-o`/`--opt` and `-a`/`--arg` for more structured input (Nick Craig-Wood)
* Implement `backend/command` for running backend specific commands remotely (Nick Craig-Wood)
* Implement `backend/command` for running backend-specific commands remotely (Nick Craig-Wood)
* Add `mount/mount` command for starting `rclone mount` via the API (Chaitanya)
* rcd: Add Prometheus metrics support (Gary Kim)
* serve http
@@ -1638,7 +1638,7 @@ all the docs and Edward Barker for helping re-write the front page.
* Add flag `--vfs-case-insensitive` for windows/macOS mounts (Ivan Andreev)
* Make objects of unknown size readable through the VFS (Nick Craig-Wood)
* Move writeback of dirty data out of close() method into its own method (FlushWrites) and remove close() call from Flush() (Brett Dutro)
* Stop empty dirs disappearing when renamed on bucket based remotes (Nick Craig-Wood)
* Stop empty dirs disappearing when renamed on bucket-based remotes (Nick Craig-Wood)
* Stop change notify polling clearing so much of the directory cache (Nick Craig-Wood)
* Azure Blob
* Disable logging to the Windows event log (Nick Craig-Wood)
@@ -1791,7 +1791,7 @@ all the docs and Edward Barker for helping re-write the front page.
* rcd: Fix permissions problems on cache directory with web gui download (Nick Craig-Wood)
* Mount
* Default `--daemon-timout` to 15 minutes on macOS and FreeBSD (Nick Craig-Wood)
* Update docs to show mounting from root OK for bucket based (Nick Craig-Wood)
* Update docs to show mounting from root OK for bucket-based (Nick Craig-Wood)
* Remove nonseekable flag from write files (Nick Craig-Wood)
* VFS
* Make write without cache more efficient (Nick Craig-Wood)
@@ -1858,7 +1858,7 @@ all the docs and Edward Barker for helping re-write the front page.
* controlled with `--multi-thread-cutoff` and `--multi-thread-streams`
* Use rclone.conf from rclone executable directory to enable portable use (albertony)
* Allow sync of a file and a directory with the same name (forgems)
* this is common on bucket based remotes, e.g. s3, gcs
* this is common on bucket-based remotes, e.g. s3, gcs
* Add `--ignore-case-sync` for forced case insensitivity (garry415)
* Implement `--stats-one-line-date` and `--stats-one-line-date-format` (Peter Berbec)
* Log an ERROR for all commands which exit with non-zero status (Nick Craig-Wood)
@@ -1872,7 +1872,7 @@ all the docs and Edward Barker for helping re-write the front page.
* lsjson
* Added EncryptedPath to output (calisro)
* Support showing the Tier of the object (Nick Craig-Wood)
* Add IsBucket field for bucket based remote listing of the root (Nick Craig-Wood)
* Add IsBucket field for bucket-based remote listing of the root (Nick Craig-Wood)
* rc
* Add `--loopback` flag to run commands directly without a server (Nick Craig-Wood)
* Add operations/fsinfo: Return information about the remote (Nick Craig-Wood)
@@ -1888,7 +1888,7 @@ all the docs and Edward Barker for helping re-write the front page.
* Make move and copy individual files obey `--backup-dir` (Nick Craig-Wood)
* If `--ignore-checksum` is in effect, don't calculate checksum (Nick Craig-Wood)
* moveto: Fix case-insensitive same remote move (Gary Kim)
* rc: Fix serving bucket based objects with `--rc-serve` (Nick Craig-Wood)
* rc: Fix serving bucket-based objects with `--rc-serve` (Nick Craig-Wood)
* serve webdav: Fix serveDir not being updated with changes from webdav (Gary Kim)
* Mount
* Fix poll interval documentation (Animosity022)
@@ -2573,7 +2573,7 @@ Point release to fix hubic and azureblob backends.
* Always forget parent dir for notifications
* Integrate with Plex websocket
* Add rc cache/stats (seuffert)
* Add info log on notification
* Add info log on notification
* Box
* Fix failure reading large directories - parse file/directory size as float
* Dropbox
@@ -2754,7 +2754,7 @@ Point release to fix hubic and azureblob backends.
* Fix following of symlinks
* Fix reading config file outside of Fs setup
* Fix reading $USER in username fallback not $HOME
* Fix running under crontab - Use correct OS way of reading username
* Fix running under crontab - Use correct OS way of reading username
* Swift
* Fix refresh of authentication token
* in v1.39 a bug was introduced which ignored new tokens - this fixes it
@@ -2917,7 +2917,7 @@ Point release to fix hubic and azureblob backends.
* HTTP - thanks to Vasiliy Tolstov
* New commands
* rclone ncdu - for exploring a remote with a text based user interface.
* rclone lsjson - for listing with a machine readable output
* rclone lsjson - for listing with a machine-readable output
* rclone dbhashsum - to show Dropbox style hashes of files (local or Dropbox)
* New Features
* Implement --fast-list flag
@@ -3181,7 +3181,7 @@ Point release to fix hubic and azureblob backends.
* Unix: implement `-x`/`--one-file-system` to stay on a single file system
* thanks Durval Menezes and Luiz Carlos Rumbelsperger Viana
* Windows: ignore the symlink bit on files
* Windows: Ignore directory based junction points
* Windows: Ignore directory-based junction points
* B2
* Make sure each upload has at least one upload slot - fixes strange upload stats
* Fix uploads when using crypt
@@ -3284,7 +3284,7 @@ Point release to fix hubic and azureblob backends.
* Retry more errors
* Add --ignore-size flag - for uploading images to onedrive
* Log -v output to stdout by default
* Display the transfer stats in more human readable form
* Display the transfer stats in more human-readable form
* Make 0 size files specifiable with `--max-size 0b`
* Add `b` suffix so we can specify bytes in --bwlimit, --min-size, etc.
* Use "password:" instead of "password>" prompt - thanks Klaus Post and Leigh Klotz

View File

@@ -18,7 +18,7 @@ a remote.
First check your chosen remote is working - we'll call it `remote:path` here.
Note that anything inside `remote:path` will be chunked and anything outside
won't. This means that if you are using a bucket based remote (e.g. S3, B2, swift)
won't. This means that if you are using a bucket-based remote (e.g. S3, B2, swift)
then you should probably put the bucket in the remote `s3:bucket`.
Now configure `chunker` using `rclone config`. We will call this one `overlay`

View File

@@ -224,7 +224,7 @@ it when needed.
If you intend to use the wrapped remote both directly for keeping
unencrypted content, as well as through a crypt remote for encrypted
content, it is recommended to point the crypt remote to a separate
directory within the wrapped remote. If you use a bucket based storage
directory within the wrapped remote. If you use a bucket-based storage
system (e.g. Swift, S3, Google Compute Storage, B2, Hubic) it is generally
advisable to wrap the crypt remote around a specific bucket (`s3:bucket`).
If wrapping around the entire root of the storage (`s3:`), and use the

View File

@@ -278,7 +278,7 @@ This will make `parameter` be `with"quote` and `parameter2` be
`with'quote`.
If you leave off the `=parameter` then rclone will substitute `=true`
which works very well with flags. For example to use s3 configured in
which works very well with flags. For example, to use s3 configured in
the environment you could use:
rclone lsd :s3,env_auth:
@@ -485,7 +485,7 @@ it will give an error.
This option controls the bandwidth limit. For example
--bwlimit 10M
would mean limit the upload and download bandwidth to 10 MiB/s.
**NB** this is **bytes** per second not **bits** per second. To use a
single limit, specify the desired bandwidth in KiB/s, or use a
@@ -664,12 +664,12 @@ they are incorrect as it would normally.
### --compare-dest=DIR ###
When using `sync`, `copy` or `move` DIR is checked in addition to the
destination for files. If a file identical to the source is found that
file is NOT copied from source. This is useful to copy just files that
When using `sync`, `copy` or `move` DIR is checked in addition to the
destination for files. If a file identical to the source is found that
file is NOT copied from source. This is useful to copy just files that
have changed since the last backup.
You must use the same remote as the destination of the sync. The
You must use the same remote as the destination of the sync. The
compare directory must not overlap the destination directory.
See `--copy-dest` and `--backup-dir`.
@@ -772,9 +772,9 @@ connection to go through to a remote object storage system. It is
### --copy-dest=DIR ###
When using `sync`, `copy` or `move` DIR is checked in addition to the
destination for files. If a file identical to the source is found that
file is server-side copied from DIR to the destination. This is useful
When using `sync`, `copy` or `move` DIR is checked in addition to the
destination for files. If a file identical to the source is found that
file is server-side copied from DIR to the destination. This is useful
for incremental backup.
The remote in use must support server-side copy and you must
@@ -951,7 +951,7 @@ default, and responds to key `u` for toggling human-readable format.
### --ignore-case-sync ###
Using this option will cause rclone to ignore the case of the files
Using this option will cause rclone to ignore the case of the files
when synchronizing so files will not be copied/synced when the
existing filenames are the same, even if the casing is different.
@@ -1097,7 +1097,7 @@ warnings and significant events.
### --use-json-log ###
This switches the log format to JSON for rclone. The fields of json log
This switches the log format to JSON for rclone. The fields of json log
are level, msg, source, time.
### --low-level-retries NUMBER ###
@@ -1479,7 +1479,7 @@ Disable retries with `--retries 1`.
### --retries-sleep=TIME ###
This sets the interval between each retry specified by `--retries`
This sets the interval between each retry specified by `--retries`
The default is `0`. Use `0` to disable.
@@ -1516,9 +1516,9 @@ Note that on macOS you can send a SIGINFO (which is normally ctrl-T in
the terminal) to make the stats print immediately.
### --stats-file-name-length integer ###
By default, the `--stats` output will truncate file names and paths longer
than 40 characters. This is equivalent to providing
`--stats-file-name-length 40`. Use `--stats-file-name-length 0` to disable
By default, the `--stats` output will truncate file names and paths longer
than 40 characters. This is equivalent to providing
`--stats-file-name-length 40`. Use `--stats-file-name-length 0` to disable
any truncation of file names printed by stats.
### --stats-log-level string ###
@@ -1562,14 +1562,14 @@ The default is `bytes`.
### --suffix=SUFFIX ###
When using `sync`, `copy` or `move` any files which would have been
overwritten or deleted will have the suffix added to them. If there
is a file with the same path (after the suffix has been added), then
overwritten or deleted will have the suffix added to them. If there
is a file with the same path (after the suffix has been added), then
it will be overwritten.
The remote in use must support server-side move or copy and you must
use the same remote as the destination of the sync.
This is for use with files to add the suffix in the current directory
This is for use with files to add the suffix in the current directory
or with `--backup-dir`. See `--backup-dir` for more info.
For example
@@ -1633,7 +1633,7 @@ will depend on the backend. For HTTP based backends it is an HTTP
PUT/GET/POST/etc and its response. For FTP/SFTP it is a round trip
transaction over TCP.
For example to limit rclone to 10 transactions per second use
For example, to limit rclone to 10 transactions per second use
`--tpslimit 10`, or to 1 transaction every 2 seconds use `--tpslimit
0.5`.
@@ -1749,7 +1749,7 @@ quickly using the least amount of memory.
However, some remotes have a way of listing all files beneath a
directory in one (or a small number) of transactions. These tend to
be the bucket based remotes (e.g. S3, B2, GCS, Swift, Hubic).
be the bucket-based remotes (e.g. S3, B2, GCS, Swift, Hubic).
If you use the `--fast-list` flag then rclone will use this method for
listing directories. This will have the following consequences for
@@ -1898,8 +1898,8 @@ This option defaults to `false`.
Configuration Encryption
------------------------
Your configuration file contains information for logging in to
your cloud services. This means that you should keep your
Your configuration file contains information for logging in to
your cloud services. This means that you should keep your
`rclone.conf` file in a secure location.
If you are in an environment where that isn't possible, you can
@@ -1947,8 +1947,8 @@ encryption from your configuration.
There is no way to recover the configuration if you lose your password.
rclone uses [nacl secretbox](https://godoc.org/golang.org/x/crypto/nacl/secretbox)
which in turn uses XSalsa20 and Poly1305 to encrypt and authenticate
rclone uses [nacl secretbox](https://godoc.org/golang.org/x/crypto/nacl/secretbox)
which in turn uses XSalsa20 and Poly1305 to encrypt and authenticate
your configuration with secret-key cryptography.
The password is SHA-256 hashed, which produces the key for secretbox.
The hashed password is not stored.
@@ -2000,8 +2000,8 @@ script method of supplying the password enhances the security of
the config password considerably.
If you are running rclone inside a script, unless you are using the
`--password-command` method, you might want to disable
password prompts. To do that, pass the parameter
`--password-command` method, you might want to disable
password prompts. To do that, pass the parameter
`--ask-password=false` to rclone. This will make rclone fail instead
of asking for a password if `RCLONE_CONFIG_PASS` doesn't contain
a valid password, and `--password-command` has not been supplied.
@@ -2039,9 +2039,9 @@ Write CPU profile to file. This can be analysed with `go tool pprof`.
The `--dump` flag takes a comma separated list of flags to dump info
about.
Note that some headers including `Accept-Encoding` as shown may not
Note that some headers including `Accept-Encoding` as shown may not
be correct in the request and the response may not show `Content-Encoding`
if the go standard libraries auto gzip encoding was in effect. In this case
if the go standard libraries auto gzip encoding was in effect. In this case
the body of the request will be gunzipped before showing it.
The available flags are:
@@ -2279,7 +2279,7 @@ this order and the first one with a value is used.
- Parameters in connection strings, e.g. `myRemote,skip_links:`
- Flag values as supplied on the command line, e.g. `--skip-links`
- Remote specific environment vars, e.g. `RCLONE_CONFIG_MYREMOTE_SKIP_LINKS` (see above).
- Backend specific environment vars, e.g. `RCLONE_LOCAL_SKIP_LINKS`.
- Backend-specific environment vars, e.g. `RCLONE_LOCAL_SKIP_LINKS`.
- Backend generic environment vars, e.g. `RCLONE_SKIP_LINKS`.
- Config file, e.g. `skip_links = true`.
- Default values, e.g. `false` - these can't be changed.

View File

@@ -6,7 +6,7 @@ type: page
# {{< icon "fa fa-heart heart" >}} Donations to the rclone project
Rclone is a free open source project with thousands of contributions
Rclone is a free open-source project with thousands of contributions
from volunteers all round the world and I would like to thank all of
you for donating your time to the project.

View File

@@ -190,7 +190,7 @@ issues with DNS resolution. See the [name resolution section in the go docs](htt
### The total size reported in the stats for a sync is wrong and keeps changing
It is likely you have more than 10,000 files that need to be
synced. By default rclone only gets 10,000 files ahead in a sync so as
synced. By default, rclone only gets 10,000 files ahead in a sync so as
not to use up too much memory. You can change this default with the
[--max-backlog](/docs/#max-backlog-n) flag.

View File

@@ -386,7 +386,7 @@ statement. For more flexibility use the `--filter-from` flag.
### `--filter` - Add a file-filtering rule
Specifies path/file names to an rclone command, based on a single
include or exclude rule, in `+` or `-` format.
include or exclude rule, in `+` or `-` format.
This flag can be repeated. See above for the order filter flags are
processed in.
@@ -555,7 +555,7 @@ input to `--files-from-raw`.
### `--ignore-case` - make searches case insensitive
By default rclone filter patterns are case sensitive. The `--ignore-case`
By default, rclone filter patterns are case sensitive. The `--ignore-case`
flag makes all of the filters patterns on the command line case
insensitive.

View File

@@ -17,7 +17,7 @@ rclone rcd --rc-web-gui
```
This will produce logs like this and rclone needs to continue to run to serve the GUI:
```
2019/08/25 11:40:14 NOTICE: A new release for gui is present at https://github.com/rclone/rclone-webui-react/releases/download/v0.0.6/currentbuild.zip
2019/08/25 11:40:14 NOTICE: Downloading webgui binary. Please wait. [Size: 3813937, Path : /home/USER/.cache/rclone/webgui/v0.0.6.zip]
@@ -28,12 +28,12 @@ This will produce logs like this and rclone needs to continue to run to serve th
This assumes you are running rclone locally on your machine. It is
possible to separate the rclone and the GUI - see below for details.
If you wish to check for updates then you can add `--rc-web-gui-update`
If you wish to check for updates then you can add `--rc-web-gui-update`
to the command line.
If you find your GUI broken, you may force it to update by add `--rc-web-gui-force-update`.
By default, rclone will open your browser. Add `--rc-web-gui-no-open-browser`
By default, rclone will open your browser. Add `--rc-web-gui-no-open-browser`
to disable this feature.
## Using the GUI
@@ -55,7 +55,7 @@ On the left hand side you will see a series of view buttons you can click on:
When you run the `rclone rcd --rc-web-gui` this is what happens
- Rclone starts but only runs the remote control API ("rc").
- The API is bound to localhost with an auto generated username and password.
- The API is bound to localhost with an auto-generated username and password.
- If the API bundle is missing then rclone will download it.
- rclone will start serving the files from the API bundle over the same port as the API
- rclone will open the browser with a `login_token` so it can log straight in.

View File

@@ -48,12 +48,12 @@ Copy binary file
sudo cp rclone /usr/bin/
sudo chown root:root /usr/bin/rclone
sudo chmod 755 /usr/bin/rclone
Install manpage
sudo mkdir -p /usr/local/share/man/man1
sudo cp rclone.1 /usr/local/share/man/man1/
sudo mandb
sudo mandb
Run `rclone config` to setup. See [rclone config docs](/docs/) for more details.
@@ -229,7 +229,7 @@ Instructions
1. `git clone https://github.com/stefangweichinger/ansible-rclone.git` into your local roles-directory
2. add the role to the hosts you want rclone installed to:
```
- hosts: rclone-hosts
roles:
@@ -346,7 +346,7 @@ your rclone command, as an alternative to scheduled task configured to run at st
##### Mount command built-in service integration ####
For mount commands, Rclone has a built-in Windows service integration via the third party
For mount commands, Rclone has a built-in Windows service integration via the third-party
WinFsp library it uses. Registering as a regular Windows service easy, as you just have to
execute the built-in PowerShell command `New-Service` (requires administrative privileges).
@@ -366,9 +366,9 @@ Windows standard methods for managing network drives. This is currently not
officially supported by Rclone, but with WinFsp version 2019.3 B2 / v1.5B2 or later
it should be possible through path rewriting as described [here](https://github.com/rclone/rclone/issues/3340).
##### Third party service integration ####
##### Third-party service integration #####
To Windows service running any rclone command, the excellent third party utility
To Windows service running any rclone command, the excellent third-party utility
[NSSM](http://nssm.cc), the "Non-Sucking Service Manager", can be used.
It includes some advanced features such as adjusting process periority, defining
process environment variables, redirect to file anything written to stdout, and

View File

@@ -107,7 +107,7 @@ Choose a number from below, or type in an existing value
1 > Archive
2 > Links
3 > Sync
Mountpoints> 1
--------------------
[jotta]
@@ -200,7 +200,7 @@ as they can't be used in XML strings.
### Deleting files
By default rclone will send all files to the trash when deleting files. They will be permanently
By default, rclone will send all files to the trash when deleting files. They will be permanently
deleted automatically after 30 days. You may bypass the trash and permanently delete files immediately
by using the [--jottacloud-hard-delete](#jottacloud-hard-delete) flag, or set the equivalent environment variable.
Emptying the trash is supported by the [cleanup](/commands/rclone_cleanup/) command.

View File

@@ -8,7 +8,7 @@ description: "Rclone docs for Memory backend"
The memory backend is an in RAM backend. It does not persist its
data - use the local backend for that.
The memory backend behaves like a bucket based remote (e.g. like
The memory backend behaves like a bucket-based remote (e.g. like
s3). Because it has no parameters you can just use it with the
`:memory:` remote name.

View File

@@ -406,7 +406,7 @@ remote itself may assign the MIME type.
## Optional Features ##
All rclone remotes support a base command set. Other features depend
upon backend specific capabilities.
upon backend-specific capabilities.
| Name | Purge | Copy | Move | DirMove | CleanUp | ListR | StreamUpload | LinkSharing | About | EmptyDir |
| ---------------------------- |:-----:|:----:|:----:|:-------:|:-------:|:-----:|:------------:|:------------:|:-----:|:--------:|
@@ -428,7 +428,7 @@ upon backend specific capabilities.
| Jottacloud | Yes | Yes | Yes | Yes | Yes | Yes | No | Yes | Yes | Yes |
| Mail.ru Cloud | Yes | Yes | Yes | Yes | Yes | No | No | Yes | Yes | Yes |
| Mega | Yes | No | Yes | Yes | Yes | No | No | Yes | Yes | Yes |
| Memory | No | Yes | No | No | No | Yes | Yes | No | No | No |
| Memory | No | Yes | No | No | No | Yes | Yes | No | No | No |
| Microsoft Azure Blob Storage | Yes | Yes | No | No | No | Yes | Yes | No | No | No |
| Microsoft OneDrive | Yes | Yes | Yes | Yes | Yes | No | No | Yes | Yes | Yes |
| OpenDrive | Yes | Yes | Yes | Yes | No | No | No | No | No | Yes |
@@ -529,4 +529,4 @@ See [rclone about command](https://rclone.org/commands/rclone_about/)
### EmptyDir ###
The remote supports empty directories. See [Limitations](/bugs/#limitations)
for details. Most Object/Bucket based remotes do not support this.
for details. Most Object/Bucket-based remotes do not support this.

View File

@@ -55,7 +55,7 @@ This website may use social sharing buttons which help share web content directl
## Use of Cloud API User Data ##
Rclone is a command line program to manage files on cloud storage. Its sole purpose is to access and manipulate user content in the [supported](/overview/) cloud storage systems from a local machine of the end user. For accessing the user content via the cloud provider API, Rclone uses authentication mechanisms, such as OAuth or HTTP Cookies, depending on the particular cloud provider offerings. Use of these authentication mechanisms and user data is governed by the privacy policies mentioned in the [Resources & Further Information](/privacy/#resources-further-information) section and followed by the privacy policy of Rclone.
Rclone is a command-line program to manage files on cloud storage. Its sole purpose is to access and manipulate user content in the [supported](/overview/) cloud storage systems from a local machine of the end user. For accessing the user content via the cloud provider API, Rclone uses authentication mechanisms, such as OAuth or HTTP Cookies, depending on the particular cloud provider offerings. Use of these authentication mechanisms and user data is governed by the privacy policies mentioned in the [Resources & Further Information](/privacy/#resources-further-information) section and followed by the privacy policy of Rclone.
* Rclone provides the end user with access to their files available in a storage system associated by the authentication credentials via the publicly exposed API of the storage system.
* Rclone allows storing the authentication credentials on the user machine in the local configuration file.

View File

@@ -1632,7 +1632,7 @@ parameters or by supplying "Content-Type: application/json" and a JSON
blob in the body. There are examples of these below using `curl`.
The response will be a JSON blob in the body of the response. This is
formatted to be reasonably human readable.
formatted to be reasonably human-readable.
### Error returns

View File

@@ -151,7 +151,7 @@ Choose a number from below, or type in your own value
region> 1
Endpoint for S3 API.
Leave blank if using AWS to use the default endpoint for the region.
endpoint>
endpoint>
Location constraint - must be set to match the Region. Used when creating buckets only.
Choose a number from below, or type in your own value
1 / Empty for US Region, Northern Virginia, or Pacific Northwest.
@@ -239,16 +239,16 @@ env_auth = false
access_key_id = XXX
secret_access_key = YYY
region = us-east-1
endpoint =
location_constraint =
endpoint =
location_constraint =
acl = private
server_side_encryption =
storage_class =
server_side_encryption =
storage_class =
--------------------
y) Yes this is OK
e) Edit this remote
d) Delete this remote
y/e/d>
y/e/d>
```
### Modified time
@@ -268,7 +268,7 @@ request as the metadata isn't returned in object listings.
#### Avoiding HEAD requests to read the modification time
By default rclone will use the modification time of objects stored in
By default, rclone will use the modification time of objects stored in
S3 for syncing. This is stored in object metadata which unfortunately
takes an extra HEAD request to read which can be expensive (in time
and money).
@@ -347,7 +347,7 @@ Note that `--fast-list` isn't required in the top-up sync.
#### Avoiding HEAD requests after PUT
By default rclone will HEAD every object it uploads. It does this to
By default, rclone will HEAD every object it uploads. It does this to
check the object got uploaded correctly.
You can disable this with the [--s3-no-head](#s3-no-head) option - see
@@ -513,7 +513,7 @@ Example policy:
"Effect": "Allow",
"Action": "s3:ListAllMyBuckets",
"Resource": "arn:aws:s3:::*"
}
}
]
}
```
@@ -1940,14 +1940,14 @@ up looking like this:
type = s3
provider = AWS
env_auth = false
access_key_id =
secret_access_key =
access_key_id =
secret_access_key =
region = us-east-1
endpoint =
location_constraint =
endpoint =
location_constraint =
acl = private
server_side_encryption =
storage_class =
server_side_encryption =
storage_class =
```
Then use it as normal with the name of the public bucket, e.g.
@@ -1983,7 +1983,7 @@ upload_cutoff = 0
### Ceph
[Ceph](https://ceph.com/) is an open source unified, distributed
[Ceph](https://ceph.com/) is an open-source, unified, distributed
storage system designed for excellent performance, reliability and
scalability. It has an S3 compatible object storage interface.
@@ -2340,7 +2340,7 @@ location_constraint =
server_side_encryption =
```
So once set up, for example to copy files into a bucket
So once set up, for example, to copy files into a bucket
```
rclone copy /path/to/files minio:bucket