Versions objects in an S3 bucket, by pattern-matching filenames to identify version numbers.
-
bucket
: Required. The name of the bucket. -
access_key_id
: Optional. The AWS access key to use when accessing the bucket. -
secret_access_key
: Optional. The AWS secret key to use when accessing the bucket. -
session_token
: Optional. The AWS STS session token to use when accessing the bucket. -
region_name
: Optional. The region the bucket is in. Defaults tous-east-1
. -
private
: Optional. Indicates that the bucket is private, so that any URLs provided are signed. -
cloudfront_url
: Optional. The URL (scheme and domain) of your CloudFront distribution that is fronting this bucket (e.ghttps://d5yxxxxx.cloudfront.net
). This will affectin
but notcheck
andput
.in
will ignore thebucket
name setting, exclusively using thecloudfront_url
. When configuring CloudFront with versioned buckets, setQuery String Forwarding and Caching
toForward all, cache based on all
to ensure S3 calls succeed. -
endpoint
: Optional. Custom endpoint for using S3 compatible provider. -
disable_ssl
: Optional. Disable SSL for the endpoint, useful for S3 compatible providers without SSL. -
skip_ssl_verification
: Optional. Skip SSL verification for S3 endpoint. Useful for S3 compatible providers using self-signed SSL certificates. -
skip_download
: Optional. Skip downloading object from S3. Useful only trigger the pipeline without using the object. -
server_side_encryption
: Optional. An encryption algorithm to use when storing objects in S3. -
sse_kms_key_id
: Optional. The ID of the AWS KMS master encryption key used for the object. -
use_v2_signing
: Optional. Use signature v2 signing, useful for S3 compatible providers that do not support v4.
One of the following two options must be specified:
-
regexp
: Optional. The pattern to match filenames against within S3. The first grouped match is used to extract the version, or if a group is explicitly namedversion
, that group is used. At least one capture group must be specified, with parentheses.The version extracted from this pattern is used to version the resource. Semantic versions, or just numbers, are supported. Accordingly, full regular expressions are supported, to specify the capture groups.
-
versioned_file
: Optional If you enable versioning for your S3 bucket then you can keep the file name the same and upload new versions of your file without resorting to version numbers. This property is the path to the file in your S3 bucket.
If no resource versions exist you can set up this resource to emit an initial version with a specified content. This won't create a real resource in S3 but only create an initial version for Concourse. The resource file will be created as usual when you get
a resource with an initial version.
You can define one of the following two options:
-
initial_path
: Optional. Must be used with theregexp
option. You should set this to the file path containing the initial version which would match the given regexp. E.g. ifregexp
isfile/build-(.*).zip
, theninitial_path
might befile/build-0.0.0.zip
. The resource version will be0.0.0
in this case. -
initial_version
: Optional. Must be used with theversioned_file
option. This will be the resource version.
By default the resource file will be created with no content when get
runs. You can set the content by using one of the following options:
-
initial_content_text
: Optional. Initial content as a string. -
initial_content_binary
: Optional. You can pass binary content as a base64 encoded string.
Objects will be found via the pattern configured by regexp
. The versions
will be used to order them (using semver). Each
object's filename is the resulting version.
Places the following files in the destination:
-
(filename)
: The file fetched from the bucket. -
url
: A file containing the URL of the object. Ifprivate
is true, this URL will be signed. -
version
: The version identified in the file name.
unpack
: Optional. If true and the file is an archive (tar, gzipped tar, other gzipped file, or zip), unpack the file. Gzipped tarballs will be both ungzipped and untarred. It is ignored whenget
is running on the initial version.
Given a file specified by file
, upload it to the S3 bucket. If regexp
is
specified, the new file will be uploaded to the directory that the regex
searches in. If versioned_file
is specified, the new file will be uploaded as
a new version of that file.
-
file
: Required. Path to the file to upload, provided by an output of a task. If multiple files are matched by the glob, an error is raised. The file which matches will be placed into the directory structure on S3 as defined inregexp
in the resource definition. The matching syntax is bash glob expansion, so no capture groups, etc. -
acl
: Optional. Canned Acl for the uploaded object. -
content_type
: Optional. MIME Content-Type describing the contents of the uploaded object
When the file has the version name in the filename
- name: release
type: s3
source:
bucket: releases
regexp: directory_on_s3/release-(.*).tgz
access_key_id: ACCESS-KEY
secret_access_key: SECRET
or
When the file is being versioned by s3
- name: release
type: s3
source:
bucket: releases
versioned_file: directory_on_s3/release.tgz
access_key_id: ACCESS-KEY
secret_access_key: SECRET
- get: release
- put: release
params:
file: path/to/release-*.tgz
acl: public-read
The bucket itself (e.g. "arn:aws:s3:::your-bucket"
):
s3:ListBucket
The objects in the bucket (e.g. "arn:aws:s3:::your-bucket/*"
):
s3:PutObject
s3:PutObjectAcl
s3:GetObject
Everything above and...
The bucket itself (e.g. "arn:aws:s3:::your-bucket"
):
s3:ListBucketVersions
s3:GetBucketVersioning
The objects in the bucket (e.g. "arn:aws:s3:::your-bucket/*"
):
s3:GetObjectVersion
s3:PutObjectVersionAcl
First get the resource via:
go get github.com/concourse/s3-resource
- golang is required - version 1.9.x is tested; earlier versions may also work.
- docker is required - version 17.06.x is tested; earlier versions may also work.
- godep is used for dependency management of the golang packages.
The tests have been embedded with the Dockerfile
; ensuring that the testing
environment is consistent across any docker
enabled platform. When the docker
image builds, the test are run inside the docker container, on failure they
will stop the build.
Run the tests with the following command:
docker build -t s3-resource .
The integration requires two AWS S3 buckets, one without versioning and another
with. The docker build
step requires setting --build-args
so the
integration will run.
Run the tests with the following command:
docker build . -t s3-resource --build-arg S3_TESTING_ACCESS_KEY_ID="access-key" --build-arg S3_TESTING_SECRET_ACCESS_KEY="some-secret" --build-arg S3_TESTING_BUCKET="bucket-non-versioned" --build-arg S3_VERSIONED_TESTING_BUCKET="bucket-versioned" --build-arg S3_TESTING_REGION="us-east-1" --build-arg S3_ENDPOINT="https://s3.amazonaws.com"
Please make all pull requests to the master
branch and ensure tests pass
locally.