How can I access my Datica S3 Bucket?
When Datica creates S3 Buckets for The Platform environments, we place three environment variables in your environment:
Datica also sets a number of bucket policies that secure the bucket contents.
Using the AWS CLI
For moving large amounts of content into your bucket, the AWS CLI can be a handy way for interacting with the environment.
Setup your S3 Credentials
Use the below command and add in your
Use the AWS S3 Commands
After setting up credentials, you can use the
aws s3 sub command to interact with the container.
- List Bucket
aws s3 ls s3://your_bucket_name/
- Copy Single File to Bucket
aws s3 cp local_file s3://your_bucket_name/ --sse
- Copy Folder to Bucket
aws s3 cp local_folder s3://your_bucket_name/ --recursive --sse
Note that copying data to the bucket requires the server side encryption flag (–sse) due to bucket policies.
See here for more details.
For using the S3 API routes, see here.
Using Your Application
Most development frameworks contain libraries for interacting with the S3 API. Those libraries can make the task of interacting with S3 much simpler.
All S3 buckets setup by Datica are put in the
US Standard region. When specification of the region is required for an SDK that you are using, you should specify the