Skip to content

Commit 33ee208

Browse files
authored
Raise Acrolinx score above minimum
1 parent f79c8ce commit 33ee208

File tree

1 file changed

+5
-5
lines changed

1 file changed

+5
-5
lines changed

articles/storage/common/storage-ref-azcopy-copy.md

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -58,7 +58,7 @@ azcopy copy [source] [destination] [flags]
5858

5959
## Examples
6060

61-
Upload a single file by using OAuth authentication. If you haven't yet logged into AzCopy, please run the azcopy login command before you run the following command.
61+
Upload a single file by using OAuth authentication. If you haven't yet logged into AzCopy, run the azcopy login command before you run the following command.
6262

6363
`azcopy cp "/path/to/file.txt" "https://[account].blob.core.windows.net/[container]/[path/to/blob]"`
6464

@@ -101,7 +101,7 @@ Upload files and directories to Azure Storage account and set the query-string e
101101
- Keys and values are URL encoded and the key-value pairs are separated by an ampersand('&')
102102
- While setting tags on the blobs, there are additional permissions('t' for tags) in SAS without which the service will give authorization error back.
103103

104-
Download a single file by using OAuth authentication. If you haven't yet logged into AzCopy, please run the azcopy login command before you run the following command.
104+
Download a single file by using OAuth authentication. If you haven't yet logged into AzCopy, run the azcopy login command before you run the following command.
105105

106106
`azcopy cp "https://[account].blob.core.windows.net/[container]/[path/to/blob]" "/path/to/file.txt"`
107107

@@ -179,7 +179,7 @@ Copy all buckets to Blob Storage from an Amazon Web Services (AWS) region by usi
179179

180180
`azcopy cp "https://s3-[region].amazonaws.com/" "https://[destaccount].blob.core.windows.net?[SAS]" --recursive=true`
181181

182-
Copy a subset of buckets by using a wildcard symbol (*) in the bucket name. Like the previous examples, you'll need an access key and a SAS token. Make sure to set the environment variable AWS_ACCESS_KEY_ID and AWS_SECRET_ACCESS_KEY for AWS S3 source.
182+
Copy a subset of buckets by using a wildcard symbol (*) in the bucket name. Like the previous examples, you need an access key and a SAS token. Make sure to set the environment variable AWS_ACCESS_KEY_ID and AWS_SECRET_ACCESS_KEY for AWS S3 source.
183183

184184
`azcopy cp "https://s3.amazonaws.com/[bucket*name]/" "https://[destaccount].blob.core.windows.net?[SAS]" --recursive=true`
185185

@@ -247,7 +247,7 @@ Copy a subset of buckets by using a wildcard symbol (*) in the bucket name from
247247

248248
`--cpk-by-name` (string) Client provided key by name that gives clients making requests against Azure Blob storage an option to provide an encryption key on a per-request basis. Provided key name will be fetched from Azure Key Vault and will be used to encrypt the data
249249

250-
`--cpk-by-value` Client provided key by name that let clients making requests against Azure Blob storage an option to provide an encryption key on a per-request basis. Provided key and its hash will be fetched from environment variables
250+
`--cpk-by-value` Client provided key by name that lets clients making requests against Azure Blob storage an option to provide an encryption key on a per-request basis. Provided key and its hash will be fetched from environment variables
251251

252252
`--decompress` Automatically decompress files when downloading, if their content-encoding indicates that they're compressed. The supported content-encoding values are 'gzip' and 'deflate'. File extensions of '.gz'/'.gzip' or '.zz' aren't necessary, but will be removed if present.
253253

@@ -313,7 +313,7 @@ Copy a subset of buckets by using a wildcard symbol (*) in the bucket name from
313313

314314
`--recursive` Look into subdirectories recursively when uploading from local file system.
315315

316-
`--s2s-detect-source-changed` Detect if the source file/blob changes while it is being read. (This parameter only applies to service-to-service copies, because the corresponding check is permanently enabled for uploads and downloads.)
316+
`--s2s-detect-source-changed` Detect if the source file/blob changes while it's being read. (This parameter only applies to service-to-service copies, because the corresponding check is permanently enabled for uploads and downloads.)
317317

318318
`--s2s-handle-invalid-metadata` (string) Specifies how invalid metadata keys are handled. Available options: ExcludeIfInvalid, FailIfInvalid, RenameIfInvalid. (default 'ExcludeIfInvalid'). (default "ExcludeIfInvalid")
319319

0 commit comments

Comments
 (0)