404

[ Avaa Bypassed ]




Upload:

Command:

botdev@13.59.219.20: ~ $
**Copying a local file to S3**

The following ``cp`` command copies a single file to a specified
bucket and key::

    aws s3 cp test.txt s3://mybucket/test2.txt

Output::

    upload: test.txt to s3://mybucket/test2.txt

**Copying a local file to S3 with an expiration date**

The following ``cp`` command copies a single file to a specified
bucket and key that expires at the specified ISO 8601 timestamp::

    aws s3 cp test.txt s3://mybucket/test2.txt --expires 2014-10-01T20:30:00Z

Output::

    upload: test.txt to s3://mybucket/test2.txt


**Copying a file from S3 to S3**

The following ``cp`` command copies a single s3 object to a specified bucket and key::

    aws s3 cp s3://mybucket/test.txt s3://mybucket/test2.txt

Output::

    copy: s3://mybucket/test.txt to s3://mybucket/test2.txt


**Copying an S3 object to a local file**

The following ``cp`` command copies a single object to a specified file locally::

    aws s3 cp s3://mybucket/test.txt test2.txt

Output::

    download: s3://mybucket/test.txt to test2.txt


**Copying an S3 object from one bucket to another**

The following ``cp`` command copies a single object to a specified bucket while retaining its original name::

    aws s3 cp s3://mybucket/test.txt s3://mybucket2/

Output::

    copy: s3://mybucket/test.txt to s3://mybucket2/test.txt

**Recursively copying S3 objects to a local directory**

When passed with the parameter ``--recursive``, the following ``cp`` command recursively copies all objects under a
specified prefix and bucket to a specified directory.  In this example, the bucket ``mybucket`` has the objects
``test1.txt`` and ``test2.txt``::

    aws s3 cp s3://mybucket . --recursive

Output::

    download: s3://mybucket/test1.txt to test1.txt
    download: s3://mybucket/test2.txt to test2.txt

**Recursively copying local files to S3**

When passed with the parameter ``--recursive``, the following ``cp`` command recursively copies all files under a
specified directory to a specified bucket and prefix while excluding some files by using an ``--exclude`` parameter.  In
this example, the directory ``myDir`` has the files ``test1.txt`` and ``test2.jpg``::

    aws s3 cp myDir s3://mybucket/ --recursive --exclude "*.jpg"

Output::

    upload: myDir/test1.txt to s3://mybucket/test1.txt

**Recursively copying S3 objects to another bucket**

When passed with the parameter ``--recursive``, the following ``cp`` command recursively copies all objects under a
specified bucket to another bucket while excluding some objects by using an ``--exclude`` parameter.  In this example,
the bucket ``mybucket`` has the objects ``test1.txt`` and ``another/test1.txt``::

    aws s3 cp s3://mybucket/ s3://mybucket2/ --recursive --exclude "another/*"

Output::

    copy: s3://mybucket/test1.txt to s3://mybucket2/test1.txt

You can combine ``--exclude`` and ``--include`` options to copy only objects that match a pattern, excluding all others::

    aws s3 cp s3://mybucket/logs/ s3://mybucket2/logs/ --recursive --exclude "*" --include "*.log"

Output::

    copy: s3://mybucket/logs/test/test.log to s3://mybucket2/logs/test/test.log
    copy: s3://mybucket/logs/test3.log to s3://mybucket2/logs/test3.log

**Setting the Access Control List (ACL) while copying an S3 object**

The following ``cp`` command copies a single object to a specified bucket and key while setting the ACL to
``public-read-write``::

    aws s3 cp s3://mybucket/test.txt s3://mybucket/test2.txt --acl public-read-write

Output::

    copy: s3://mybucket/test.txt to s3://mybucket/test2.txt

Note that if you're using the ``--acl`` option, ensure that any associated IAM
policies include the ``"s3:PutObjectAcl"`` action::

    aws iam get-user-policy --user-name myuser --policy-name mypolicy

Output::

    {
        "UserName": "myuser",
        "PolicyName": "mypolicy",
        "PolicyDocument": {
            "Version": "2012-10-17",
            "Statement": [
                {
                    "Action": [
                        "s3:PutObject",
                        "s3:PutObjectAcl"
                    ],
                    "Resource": [
                        "arn:aws:s3:::mybucket/*"
                    ],
                    "Effect": "Allow",
                    "Sid": "Stmt1234567891234"
                }
            ]
        }
    }

**Granting permissions for an S3 object**

The following ``cp`` command illustrates the use of the ``--grants`` option to grant read access to all users identified
by URI and full control to a specific user identified by their Canonical ID::

  aws s3 cp file.txt s3://mybucket/ --grants read=uri=http://acs.amazonaws.com/groups/global/AllUsers full=id=79a59df900b949e55d96a1e698fbacedfd6e09d98eacf8f8d5218e7cd47ef2be

Output::

    upload: file.txt to s3://mybucket/file.txt

**Uploading a local file stream to S3**

.. WARNING:: PowerShell may alter the encoding of or add a CRLF to piped input.

The following ``cp`` command uploads a local file stream from standard input to a specified bucket and key::

    aws s3 cp - s3://mybucket/stream.txt

**Uploading a local file stream that is larger than 50GB to S3**

The following ``cp`` command uploads a 51GB local file stream from standard input to a specified bucket and key.  The ``--expected-size`` option must be provided, or the upload may fail when it reaches the default part limit of 10,000::

    aws s3 cp - s3://mybucket/stream.txt --expected-size 54760833024

**Downloading an S3 object as a local file stream**

.. WARNING:: PowerShell may alter the encoding of or add a CRLF to piped or redirected output.

The following ``cp`` command downloads an S3 object locally as a stream to standard output. Downloading as a stream is not currently compatible with the ``--recursive`` parameter::

    aws s3 cp s3://mybucket/stream.txt -

**Uploading to an S3 access point**

The following ``cp`` command uploads a single file (``mydoc.txt``) to the access point (``myaccesspoint``) at the key (``mykey``)::

    aws s3 cp mydoc.txt s3://arn:aws:s3:us-west-2:123456789012:accesspoint/myaccesspoint/mykey

Output::

    upload: mydoc.txt to s3://arn:aws:s3:us-west-2:123456789012:accesspoint/myaccesspoint/mykey


**Downloading from an S3 access point**

The following ``cp`` command downloads a single object (``mykey``) from the access point (``myaccesspoint``) to the local file (``mydoc.txt``)::

    aws s3 cp s3://arn:aws:s3:us-west-2:123456789012:accesspoint/myaccesspoint/mykey mydoc.txt

Output::

    download: s3://arn:aws:s3:us-west-2:123456789012:accesspoint/myaccesspoint/mykey to mydoc.txt

Filemanager

Name Type Size Permission Actions
_concepts.rst File 7.65 KB 0644
cp.rst File 6.45 KB 0644
ls.rst File 3.53 KB 0644
mb.rst File 535 B 0644
mv.rst File 3.13 KB 0644
presign.rst File 1.42 KB 0644
rb.rst File 641 B 0644
rm.rst File 1.67 KB 0644
sync.rst File 5.67 KB 0644
website.rst File 883 B 0644