Transfer directories to and from Amazon S3 using the AWS SDK for PHP Version 3
You use the Transfer class in AWS SDK for PHP Version 3 to upload entire directories to an
Amazon S3 bucket and download entire buckets to a local directory.
Uploading a local directory to Amazon S3
The Aws\S3\Transfer object performs the transfers. The following example
shows how to recursively upload a local directory of files to an Amazon S3 bucket.
// Create an S3 client. $client = new \Aws\S3\S3Client([ 'region' => 'us-west-2', 'version' => '2006-03-01', ]); // Where the files will be sourced from. $source = '/path/to/source/files'; // Where the files will be transferred to. $dest = 's3://bucket'; // Create a transfer object. $directoryTransfer = new \Aws\S3\Transfer($client, $source, $dest); // Perform the transfer synchronously. $directoryTransfer->transfer();
In this example, we created an Amazon S3 client, created a Transfer object, and
performed transfer synchronously.
The previous example demonstrates the minimum amount of code needed to perform a transfer.
A Transfer object can also perform transfers asynchronously and has various
configuration options that you can use to customize the transfers.
You can upload the local files to a “subfolder” of an Amazon S3 bucket by providing a key
prefix in the s3:// URI. The following example uploads the local files on disk to
the bucket bucket and stores the files under the foo key
prefix.
$source = '/path/to/source/files'; $dest = 's3://bucket/foo'; $directoryTransfer = new \Aws\S3\Transfer($client, $source, $dest); $directoryTransfer->transfer();
Downloading an Amazon S3 bucket
You can recursively download an Amazon S3 bucket to a local directory on disk by specifying the
$source argument as an Amazon S3 URI (e.g., s3://bucket) and the
$dest argument as the path to a local directory.
// Where the files will be sourced from. $source = 's3://bucket'; // Where the files will be transferred to. $dest = '/path/to/destination/dir'; $directoryTransfer = new \Aws\S3\Transfer($client, $source, $dest); $directoryTransfer->transfer();
Note
The SDK automatically creates any necessary directories when downloading the objects in the bucket.
You can include a key prefix in the Amazon S3 URI after the bucket to download only objects stored under a “pseudo-folder”. The following example downloads only files stored under the “/foo” key prefix of the given bucket.
$source = 's3://bucket/foo'; $dest = '/path/to/destination/dir'; $directoryTransfer = new \Aws\S3\Transfer($client, $source, $dest); $directoryTransfer->transfer();
Configuration
The Transfer object constructor accepts the following arguments.
-
$client -
The
Aws\ClientInterfaceobject to use to perform the transfers. -
$source(string |Iterator) -
The source data being transferred. This can point to a local path on disk (e.g.,
/path/to/files) or an Amazon S3 bucket (e.g.,s3://bucket). Thes3://URI may also contain a key prefix that can be used to only transfer objects under a common prefix.If the
$sourceargument is an Amazon S3 URI, the$destargument must be a local directory (and vice versa).In addition to providing a string value, you can also provide an
\Iteratorobject that yields absolute file names. If you provide an\Iteratorobject, you must provide abase_diroption in the$optionsassociative array. -
$dest -
The destination where the files will be transferred. If the
$sourceargument is a local path on disk,$destmust be an Amazon S3 bucket URI (e.g.,s3://bucket). If the$sourceargument is an Amazon S3 bucket URI, the$destargument must be a local path on disk. -
$options -
An associative array of transfer options. The following transfer options are valid:
add_content_md5(bool)-
Set to
trueto calculate the MD5 checksum for uploads. -
base_dir(string) -
Base directory of the source, if
$sourceis an iterator. If the$sourceoption is not an array, then this option is ignored. -
before(callable) -
A callback to invoke before each transfer. The callback should have a function signature like
function (Aws\Command $command) {...}. The provided command will be aGetObject,PutObject,CreateMultipartUpload,UploadPart, orCompleteMultipartUploadcommand. -
mup_threshold(int) -
Size in bytes in which a multipart upload should be used instead of
PutObject. Defaults to16777216(16 MB). -
concurrency(int, default=5) -
Number of files to upload concurrently. The ideal concurrency value will vary based on the number of files being uploaded and the average size of each file. Generally, smaller files benefit from a higher concurrency while larger files do not.
-
debug(bool) -
Set to
trueto print out debug information for transfers. Set to anfopen()resource to write to a specific stream instead of writing to STDOUT.
Async transfers
The Transfer object is an instance of
GuzzleHttp\Promise\PromisorInterface. This means that the transfer can occur
asynchronously and is initiated by calling the promise method of the
object.
$source = '/path/to/source/files'; $dest = 's3://bucket'; $directoryTransfer = new \Aws\S3\Transfer($client, $source, $dest); // Initiate the transfer and get a promise. $promise = $directoryTransfer->promise(); // Do something when the transfer is complete using the then() method. $promise->then(function () { echo 'Done!'; });
The promise is rejected if any file fails to transfer. You can handle the failed transfer
asynchronously using the otherwise method of the promise. The
otherwise function accepts a callback to invoke when an error occurs. The
callback accepts the $reason for the rejection, which typically is an instance of
Aws\Exception\AwsException (although a value of any type can be delivered to the callback).
$promise->otherwise(function ($reason) { echo 'Transfer failed: '; var_dump($reason); });
Because the Transfer object returns a promise, these transfers can occur
concurrently with other asynchronous promises.
Customizing the directory transfer
By adding a callback to the constructor, you can customize the options that the
Transfer executes.
$uploader = new Transfer($s3Client, $source, $dest, [ 'before' => function (\Aws\Command $command) { // Commands can vary for multipart uploads, so check which command // is being processed. if (in_array($command->getName(), ['PutObject', 'CreateMultipartUpload'])) { // Set custom cache-control metadata. $command['CacheControl'] = 'max-age=3600'; // Apply a canned ACL. $command['ACL'] = strpos($command['Key'], 'CONFIDENTIAL') ### false ? 'public-read' : 'private'; } }, ]);