Upload to S3
Use DuploCloud to upload to S3 from Github Actions
The goal of this section is to show how you can upload to an S3 bucket from Github Actions.
This process is done in two basic steps:
- Logs in to AWS ECR using Just-In-Time (JIT) AWS credentials from DuploCloud
- Upload your website content to S3
To use any of the below examples you will need to change:
- The local path to upload from, if it is not
build
duplo_host
env varTENANT_NAME
env varBUCKET_NAME
env var
The following example uploads a single directory to S3.
It does not show more advanced things like the following items:
- Setting cache control directives
- Making the uploaded content public
- Making AWS delete older content
name: Build and Deploy
on:
# Triggers the workflow on push to matching branches
push:
branches:
- master
env:
duplo_host: https://mysystem.duplocloud.net # CHANGE ME!
duplo_token: "${{ secrets.DUPLO_TOKEN }}"
TENANT_NAME: mytenant # CHANGE ME!
BUCKET_NAME: duploservices-mytenant-website-1234 # CHANGE ME!
# A workflow run is made up of one or more jobs that can run sequentially or in parallel
jobs:
build_and_deploy:
runs-on: ubuntu-latest
needs:
- build
steps:
- name: Checkout
uses: actions/[email protected]
# Upload to S3
- name: Get AWS credentials
uses: duplocloud/ghactions-aws-[email protected]
with:
tenant: "${{ env.TENANT_NAME }}"
# How to sync an entire folder
- name: Sync build folder to S3
run: aws s3 sync build/ "s3://$BUCKET_NAME/"
# How to copy an individual file
- name: Copy individual file to S3
run: aws s3 cp my-archive.zip "s3://$BUCKET_NAME/"
This example uploads multiple directories to S3.
It assumes that the following is true:
- There is a subfolder named
static
, whose contents can be cached for one year. - The rest of the contents can change at any time, so the cache uses
must-revalidate
.
name: Build and Deploy
on:
# Triggers the workflow on push to matching branches
push:
branches:
- master
env:
duplo_host: https://mysystem.duplocloud.net # CHANGE ME!
duplo_token: "${{ secrets.DUPLO_TOKEN }}"
TENANT_NAME: mytenant # CHANGE ME!
BUCKET_NAME: duploservices-mytenant-website-1234 # CHANGE ME!
# A workflow run is made up of one or more jobs that can run sequentially or in parallel
jobs:
build_and_deploy:
runs-on: ubuntu-latest
needs:
- build
steps:
- name: Checkout
uses: actions/[email protected]
# Upload to S3
- name: Get AWS credentials
uses: duplocloud/ghactions-aws-[email protected]
with:
tenant: "${{ env.TENANT_NAME }}"
- name: Sync files to S3
run: |-
# First, upload the "static/" subdirectory - it can be cached for one year
aws s3 sync build/static/ "s3://$BUCKET_NAME/static/" --cache-control 31536000,public
# Then, upload the rest. It can change at any time - so it uses "must-revalidate"
aws s3 sync build/ "s3://$BUCKET_NAME/" --exclude static --cache-control max-age=120,must-revalidate
Last modified 7mo ago