Compliant Kubernetes Service documentation has moved

Please note: You are not reading Kubernetes documentation. If you're looking for Compliant Kubernetes Service documentation, it has moved. Read more here.

How can I access my Datica S3 Bucket?

When Datica creates S3 Buckets for The Platform environments, we place three environment variables in your environment:


Datica also sets a number of bucket policies that secure the bucket contents.

Using the AWS CLI

For moving large amounts of content into your bucket, the AWS CLI can be a handy way for interacting with the environment.

Setup your S3 Credentials

Use the below command and add in your S3_ACCESS_KEY_ID and S3_SECRET_ACCESS_KEY

aws configure

Use the AWS S3 Commands

After setting up credentials, you can use the aws s3 sub command to interact with the container.

Example Commands:

  • List Bucket
    • aws s3 ls s3://your_bucket_name/
  • Copy Single File to Bucket
    • aws s3 cp local_file s3://your_bucket_name/ --sse
  • Copy Folder to Bucket
    • aws s3 cp local_folder s3://your_bucket_name/ --recursive --sse

Note that copying data to the bucket requires the server side encryption flag (–sse) due to bucket policies.

See here for more details.

For using the S3 API routes, see here.

Using Your Application

Most development frameworks contain libraries for interacting with the S3 API. Those libraries can make the task of interacting with S3 much simpler.

Amazon SDKs

Bucket Region

All S3 buckets setup by Datica are put in the US Standard region. When specification of the region is required for an SDK that you are using, you should specify the us-east-1 region.