Initial code commit
This commit is contained in:
parent
5b71cecab6
commit
b3382d880b
37 changed files with 1778 additions and 1 deletions
358
.github/workflows/deploy.yml
vendored
Normal file
358
.github/workflows/deploy.yml
vendored
Normal file
|
|
@ -0,0 +1,358 @@
|
|||
name: Deploy the Helm Chart
|
||||
|
||||
on:
|
||||
push:
|
||||
branches:
|
||||
- main
|
||||
workflow_dispatch:
|
||||
|
||||
jobs:
|
||||
# "Build" from the source code
|
||||
build:
|
||||
runs-on: self-hosted
|
||||
outputs:
|
||||
chart-name: ${{ steps.update-helm-repo.outputs.CHART_NAME }}
|
||||
steps:
|
||||
# Get the source code from the repository
|
||||
- name: Checkout repository
|
||||
uses: actions/checkout@v4
|
||||
|
||||
# "Build"/package the source code int the appropriate format (Helm chart etc...)
|
||||
- name: Update Helm Repository (if needed)
|
||||
id: update-helm-repo
|
||||
run: |
|
||||
# Parse the chart name from the Chart.yaml
|
||||
CHART_NAME=$(yq '.name' Chart.yaml)
|
||||
CHART_NAME=${CHART_NAME#\"} # Remove leading quote
|
||||
CHART_NAME=${CHART_NAME%\"} # Remove trailing quote
|
||||
echo "CHART_NAME=$CHART_NAME" >> $GITHUB_OUTPUT
|
||||
|
||||
echo "Chart Name: $CHART_NAME"
|
||||
|
||||
# Note, this depends on the [Harbor Helm Index](https://github.com/AlanBridgeman/harbor-helm-index) pretty heavily
|
||||
# In particular, that tool allows us to treat what is an OCI registry as a Helm repository (which includes using `helm search repo`)
|
||||
helm repo add BridgemanAccessible https://helm.bridgemanaccessible.ca
|
||||
|
||||
# Check if the chart is already in the repository or not
|
||||
REMOTE_CHART_WORK_OUTPUT=$(helm search repo BridgemanAccessible/$CHART_NAME 2>/dev/null || echo "")
|
||||
if [ -n "$REMOTE_CHART_WORK_OUTPUT" ]; then
|
||||
# The chart is already in the repository, so we need to check if the version is the same or not
|
||||
echo "Chart already exists in the repository. Checking version..."
|
||||
|
||||
IFS=' ' read -r -a REMOTE_VERSIONS <<< $(helm search repo BridgemanAccessible/$CHART_NAME --output json | jq '.[].version | @sh')
|
||||
|
||||
echo "Remote Chart Versions: ${REMOTE_VERSIONS[@]}"
|
||||
else
|
||||
# The chart is not in the repository, so we'll need to add it
|
||||
echo "Chart not found in the repository. Adding it..."
|
||||
|
||||
# Set a blank value so that it WON'T match the local version
|
||||
IFS=' ' read -r -a REMOTE_VERSIONS <<< ""
|
||||
fi
|
||||
|
||||
# Just to keep things clean/safe/etc... remove the repo
|
||||
helm repo remove BridgemanAccessible
|
||||
|
||||
# Get the local version from the Chart.yaml file
|
||||
LOCAL_VERSION=$(yq '.version' Chart.yaml)
|
||||
LOCAL_VERSION=${LOCAL_VERSION#\"} # Remove leading quote
|
||||
LOCAL_VERSION=${LOCAL_VERSION%\"} # Remove trailing quote
|
||||
echo "Local Chart Version: $LOCAL_VERSION"
|
||||
|
||||
has_match='false'
|
||||
if [ ${#REMOTE_VERSIONS[@]} -gt 0 ]; then
|
||||
# Loop through the remote tags and check if any of them match the local version
|
||||
for REMOTE_VERSION in ${REMOTE_VERSIONS[@]}; do
|
||||
REMOTE_VERSION=${REMOTE_VERSION#\'} # Remove leading quote
|
||||
REMOTE_VERSION=${REMOTE_VERSION%\'} # Remove trailing quote
|
||||
|
||||
# Check if the remote tag is the same as the local tag
|
||||
if [ "$REMOTE_VERSION" == "$LOCAL_VERSION" ]; then
|
||||
echo "Remote version matches local version!"
|
||||
has_match='true'
|
||||
break
|
||||
fi
|
||||
done
|
||||
fi
|
||||
|
||||
# If the versions match, we want to increment the chart's patch version
|
||||
if [ "has_match" ]; then
|
||||
echo "Versions match!"
|
||||
|
||||
# Increment the patch version of the local version (Ex. 1.0.0 -> 1.0.1)
|
||||
IFS='.' read -r major minor patch <<< "$LOCAL_VERSION"
|
||||
patch=$((patch + 1))
|
||||
NEW_LOCAL_VERSION="$major.$minor.$patch"
|
||||
|
||||
echo "New Local Version: $NEW_LOCAL_VERSION"
|
||||
echo "Committing new chart version change..."
|
||||
|
||||
sed -i "s|version: \"$LOCAL_VERSION\"|version: \"$NEW_LOCAL_VERSION\"|g" Chart.yaml
|
||||
|
||||
LOCAL_VERSION=$NEW_LOCAL_VERSION
|
||||
|
||||
# Update remote URL to use the GITHUB_TOKEN for authentication
|
||||
git remote set-url origin https://x-access-token:${{ secrets.GITHUB_TOKEN }}@github.com/${{ github.repository }}.git
|
||||
|
||||
# Setup git user details for committing the version change and tag
|
||||
git config user.name "GitHub Actions"
|
||||
git config user.email "actions@github.com"
|
||||
|
||||
# Commit the version change to the `package.json` file
|
||||
git add Chart.yaml
|
||||
git commit -m "[Github Actions] Update Helm chart version to $(yq -r '.version' Chart.yaml)"
|
||||
|
||||
# Push the changes to the repository
|
||||
git push origin HEAD:main
|
||||
else
|
||||
echo "Versions do not match!"
|
||||
fi
|
||||
|
||||
helm package .
|
||||
helm push ./$CHART_NAME-$LOCAL_VERSION.tgz oci://${{ secrets.REPOSITORY_HOSTNAME }}/helm
|
||||
|
||||
# Deploy to our environment (stagging cluster)
|
||||
deploy:
|
||||
runs-on: self-hosted
|
||||
needs: build
|
||||
outputs:
|
||||
backup-created: ${{ steps.backup.outputs.created }}
|
||||
restore-time: ${{ steps.deploy-helm-chart.outputs.RESTORE_TIME }}
|
||||
env:
|
||||
NAMESPACE: ciam-service-dashboard
|
||||
RELEASE_NAME: services
|
||||
CHART_NAME: ${{ needs.build.outputs.chart-name }}
|
||||
steps:
|
||||
# Check if it's already deployed
|
||||
# This determines if we do the backup and takedown before deploying the new version
|
||||
- name: Already Deployed Check
|
||||
id: already-deployed-check
|
||||
run: |
|
||||
# Check if the Helm chart is already deployed
|
||||
if helm status -n ${{ env.NAMESPACE }} ${{ env.RELEASE_NAME }}; then
|
||||
echo "Already deployed"
|
||||
echo "already-deployed=true" >> $GITHUB_OUTPUT
|
||||
else
|
||||
echo "Not deployed"
|
||||
echo "already-deployed=false" >> $GITHUB_OUTPUT
|
||||
fi
|
||||
|
||||
# Tale a backup of the current state of the resources
|
||||
- name: Backup Data
|
||||
id: backup
|
||||
if: steps.already-deployed-check.outputs.already-deployed == 'true'
|
||||
run: |
|
||||
# -- Setup Backup Sidecar metadata variables --
|
||||
|
||||
# Variable for the image name (so that there can't be typos etc...)
|
||||
BACKUP_SIDECAR_IMAGE_NAME="backup-sidecar"
|
||||
|
||||
# Get the available tags from the image registry
|
||||
IFS=' ' read -r -a BACKUP_SIDECAR_TAGS <<< $(skopeo list-tags docker://${{ secrets.REPOSITORY_HOSTNAME }}/k8s/$BACKUP_SIDECAR_IMAGE_NAME | jq -r '.Tags | @sh')
|
||||
|
||||
# Get the latest tag from the list of tags
|
||||
LATEST_BACKUP_SIDECAR_TAG=$(echo ${BACKUP_SIDECAR_TAGS[@]} | tr ' ' '\n' | sort -V | tail -n 1)
|
||||
LATEST_BACKUP_SIDECAR_TAG=${LATEST_BACKUP_SIDECAR_TAG#\'} # Remove leading quote
|
||||
LATEST_BACKUP_SIDECAR_TAG=${LATEST_BACKUP_SIDECAR_TAG%\'} # Remove trailing quote
|
||||
echo "Latest Backup Sidecar Tag: $LATEST_BACKUP_SIDECAR_TAG"
|
||||
|
||||
# Get/Generate the "full" image name (including the tag) for the backup sidecar
|
||||
FULL_BACKUP_SIDECAR_IMAGE_NAME="${{ secrets.REPOSITORY_HOSTNAME }}/k8s/$BACKUP_SIDECAR_IMAGE_NAME:$LATEST_BACKUP_SIDECAR_TAG"
|
||||
echo "Looking for backup sidecar: $FULL_BACKUP_SIDECAR_IMAGE_NAME"
|
||||
|
||||
# -- END: Setup Backup Sidecar metadata variables --
|
||||
|
||||
# -- Create a backup --
|
||||
|
||||
# Get the name of the running main pod (the one that has the backup sidecar container in it)
|
||||
RUNNING_MAIN_POD=$(kubectl -n ${{ env.NAMESPACE }} get pods -o json | jq -r --arg img "$FULL_BACKUP_SIDECAR_IMAGE_NAME" '.items[] | select(.spec.containers[].image == $img) | .metadata.name')
|
||||
echo "Running Main Pod: $RUNNING_MAIN_POD"
|
||||
|
||||
# Make the request to create the backups
|
||||
CREATE_BACKUP_OUTPUT=$(kubectl -n ${{ env.NAMESPACE }} exec $RUNNING_MAIN_POD -c ${{ env.RELEASE_NAME }} -- /bin/bash -c "curl -X PUT -sSL http://localhost:4000/backup -o backup-output.txt && cat backup-output.txt")
|
||||
echo "Create Backup Output: $CREATE_BACKUP_OUTPUT"
|
||||
|
||||
# Parse the output to check if the backup was created successfully
|
||||
BACKUP_OUTPUT_MESSAGE=$(echo $CREATE_BACKUP_OUTPUT | jq -r '.message')
|
||||
if [ "$BACKUP_OUTPUT_MESSAGE" == "Backup created successfully" ]; then
|
||||
echo "Backup created successfully."
|
||||
echo "created=true" >> $GITHUB_OUTPUT
|
||||
else
|
||||
echo "Error creating backup: $BACKUP_OUTPUT_MESSAGE"
|
||||
exit 1
|
||||
fi
|
||||
|
||||
# -- END: Create a backup --
|
||||
|
||||
# Because we've taken a backup of it's current state and it's easier to start fresh etc...
|
||||
# We delete all the existing resources (including the Helm chart) before (re-)deploying the new one
|
||||
- name: Remove old resources
|
||||
if: steps.backup.outputs.created == 'true'
|
||||
run: |
|
||||
# Use a custom script (found in [Helpful Scripts]()) to uninstall the Helm chart and delete any associated resources
|
||||
k8s-delete \
|
||||
--namespace ${{ env.NAMESPACE }} \
|
||||
--release-name ${{ env.RELEASE_NAME }} \
|
||||
--rwx-volumes vault-role-vars
|
||||
|
||||
# Deploy the resources to the cluster
|
||||
- name: Deploy Helm Chart
|
||||
id: deploy-helm-chart
|
||||
run: |
|
||||
FILLED_VALUES_FILE="values.filled.yaml"
|
||||
|
||||
# Download a filled version of the `values.yaml` file from a secure location
|
||||
DOWNLOAD_FILE=$(curl -sSL https://secure-storage.bridgemanaccessible.ca/services-dashboard/filled-values?token=${{ secrets.SECURE_STORAGE_TOKEN }} -o $FILLED_VALUES_FILE || echo "Failed to download filled values file.")
|
||||
if [ "$DOWNLOAD_FILE" == "Failed to download filled values file." ]; then
|
||||
echo "Error: $DOWNLOAD_FILE"
|
||||
exit 1
|
||||
fi
|
||||
|
||||
if [ -n "${{ steps.backup.outputs.created }}" ]; then
|
||||
CURR_DATETIME=$(date -u +"%Y-%m-%dT%H:%M:%SZ")
|
||||
|
||||
echo "Setting restoreFromBackup to: $CURR_DATETIME"
|
||||
|
||||
# If the backup was created successfully, we want to set the `restoreFromBackup` variable to true in the filled values file
|
||||
sed -i "s|restoreFromBackup: \"\"|restoreFromBackup: \"$CURR_DATETIME\"|g" $FILLED_VALUES_FILE
|
||||
|
||||
echo "RESTORE_TIME=$CURR_DATETIME" >> $GITHUB_OUTPUT
|
||||
fi
|
||||
|
||||
# Parse the chart name from the Chart.yaml
|
||||
CHART_NAME=${{ env.CHART_NAME }}
|
||||
echo "Chart Name: $CHART_NAME"
|
||||
|
||||
# Can run `k8s-deploy --help` if you want to see all the options available
|
||||
k8s-deploy \
|
||||
--namespace ${{ env.NAMESPACE }} \
|
||||
--release-name ${{ env.RELEASE_NAME }} \
|
||||
--filled-values-file $FILLED_VALUES_FILE \
|
||||
--chart-name $CHART_NAME \
|
||||
--rwx-volumes vault-role-vars
|
||||
|
||||
# Once deployed, we want to restore it to it's previous state (if applicable)
|
||||
restore:
|
||||
runs-on: self-hosted
|
||||
needs: deploy
|
||||
if: needs.deploy.outputs.backup-created == 'true'
|
||||
env:
|
||||
NAMESPACE: ciam-services-dashboard
|
||||
CONTAINER_NAME: services
|
||||
RESTORE_TIME: ${{ needs.deploy.outputs.restore-time }}
|
||||
SITE_HOSTNAME: services.bridgemanaccessible.ca
|
||||
steps:
|
||||
# Restore the data from the backup
|
||||
- name: Restore data
|
||||
run: |
|
||||
# -- Setup Backup Sidecar metadata variables --
|
||||
|
||||
# Variable for the image name (so that there can't be typos etc...)
|
||||
BACKUP_SIDECAR_IMAGE_NAME="backup-sidecar"
|
||||
|
||||
# Get the available tags from the image registry
|
||||
IFS=' ' read -r -a BACKUP_SIDECAR_TAGS <<< $(skopeo list-tags docker://${{ secrets.REPOSITORY_HOSTNAME }}/k8s/$BACKUP_SIDECAR_IMAGE_NAME | jq -r '.Tags | @sh')
|
||||
|
||||
# Get the latest tag from the list of tags
|
||||
LATEST_BACKUP_SIDECAR_TAG=$(echo ${BACKUP_SIDECAR_TAGS[@]} | tr ' ' '\n' | sort -V | tail -n 1)
|
||||
LATEST_BACKUP_SIDECAR_TAG=${LATEST_BACKUP_SIDECAR_TAG#\'} # Remove leading quote
|
||||
LATEST_BACKUP_SIDECAR_TAG=${LATEST_BACKUP_SIDECAR_TAG%\'} # Remove trailing quote
|
||||
echo "Latest Backup Sidecar Tag: $LATEST_BACKUP_SIDECAR_TAG"
|
||||
|
||||
# Get/Generate the "full" image name (including the tag) for the backup sidecar
|
||||
FULL_BACKUP_SIDECAR_IMAGE_NAME="${{ secrets.REPOSITORY_HOSTNAME }}/k8s/$BACKUP_SIDECAR_IMAGE_NAME:$LATEST_BACKUP_SIDECAR_TAG"
|
||||
echo "Looking for backup sidecar: $FULL_BACKUP_SIDECAR_IMAGE_NAME"
|
||||
|
||||
# -- END: Setup Backup Sidecar metadata variables --
|
||||
|
||||
# Get the name of the running main pod (the one that has the backup sidecar container in it)
|
||||
RUNNING_MAIN_POD=$(kubectl -n ${{ env.NAMESPACE }} get pods -o json | jq -r --arg img "$FULL_BACKUP_SIDECAR_IMAGE_NAME" '.items[] | select(.spec.containers[].image == $img) | .metadata.name')
|
||||
echo "Running Main Pod: $RUNNING_MAIN_POD"
|
||||
|
||||
# Variables for retry logic
|
||||
MAX_RETRIES=5
|
||||
RETRY_INTERVAL=30
|
||||
RETRY_COUNT=0
|
||||
|
||||
# Because Kubernetes (or at least our stagging cluster) can be a bit tempramental,
|
||||
# We allow for "resetting" a few times.
|
||||
# By resetting, we mean re-detecting the main running pod etc...
|
||||
MAX_RESETS=3
|
||||
RESET_COUNT=0
|
||||
|
||||
POD_STATUS=$(kubectl -n ${{ env.NAMESPACE }} get pod $RUNNING_MAIN_POD -o json | jq -r '.status.phase')
|
||||
echo "Pod Status: $POD_STATUS"
|
||||
|
||||
# Retry (wait) a few times if the pod isn't running yet
|
||||
while [ "$POD_STATUS" != "Running" ] && [ $RETRY_COUNT -lt $MAX_RETRIES ]; do
|
||||
echo "Pod is not running yet (Current Status: $POD_STATUS). Waiting for $RETRY_INTERVAL seconds (attempt $((RETRY_COUNT + 1))/$MAX_RETRIES)..."
|
||||
sleep $RETRY_INTERVAL
|
||||
|
||||
# Get the current status of the pod
|
||||
POD_STATUS=$(kubectl -n ciam-account-dashboard get pod $RUNNING_MAIN_POD -o json | jq -r '.status.phase')
|
||||
|
||||
# Increment the retry count
|
||||
RETRY_COUNT=$((RETRY_COUNT + 1))
|
||||
|
||||
# If the pod is in a failed state, we want to "reset"
|
||||
# Though, we only want to reset a few times (to avoid infinite loops if something else is wrong etc...)
|
||||
if [[ "$POD_STATUS" == "Failed" ]] && [ $RESET_COUNT -lt $MAX_RESETS ]; then
|
||||
echo "Pod is in a failed state. Resetting..."
|
||||
|
||||
# Reset the pod and increment the reset count
|
||||
RUNNING_MAIN_POD=$(kubectl -n ${{ env.NAMESPACE }} get pods -o json | jq -r --arg img "$FULL_BACKUP_SIDECAR_IMAGE_NAME" --arg prev "$RUNNING_MAIN_POD" '.items[] | select(.spec.containers[].image == $img and .metadata.name != $prev) | .metadata.name')
|
||||
echo "Running Main Pod Reset: $RUNNING_MAIN_POD"
|
||||
|
||||
# Reset the retry count
|
||||
RETRY_COUNT=0
|
||||
echo "Retry Count Reset"
|
||||
|
||||
RESET_COUNT=$((RESET_COUNT + 1))
|
||||
fi
|
||||
done
|
||||
|
||||
if [ "$POD_STATUS" != "Running" ]; then
|
||||
echo "Error: Pod did not start running after $((RETRY_COUNT * RETRY_INTERVAL)) seconds."
|
||||
exit 1
|
||||
fi
|
||||
|
||||
LATEST_BACKUPS=$(kubectl -n ${{ env.NAMESPACE }} exec $RUNNING_MAIN_POD -c ${{ env.CONTAINER_NAME }} -- /bin/bash -c "curl -sSL http://localhost:4000/backups/latest -o latest-backup.json && cat latest-backup.json")
|
||||
echo "Latest Backups: $LATEST_BACKUPS"
|
||||
|
||||
LATEST_DB_BACKUP=$(echo $LATEST_BACKUPS | jq -r '.backups.db')
|
||||
echo "Latest DB Backup: $LATEST_DB_BACKUP"
|
||||
|
||||
LATEST_VAULT_BACKUP=$(echo $LATEST_BACKUPS | jq -r '.backups.vault')
|
||||
echo "Latest Vault Backup: $LATEST_VAULT_BACKUP"
|
||||
|
||||
echo "Restoring Vault Backup: $LATEST_VAULT_BACKUP at restore time: ${{ env.RESTORE_TIME }}"
|
||||
restore_output=$(kubectl -n ${{ env.NAMESPACE }} exec $RUNNING_MAIN_POD -c ${{ env.CONTAINER_NAME }} -- /bin/bash -c "curl -s -X POST -H 'Content-Type: application/json' -d '{ \"restoreTime\": \"${{ env.RESTORE_TIME }}\" }' http://localhost:4000/restore/latest -o restore-output.txt && cat restore-output.txt")
|
||||
echo "Restore Output: $restore_output"
|
||||
|
||||
# Verify that the site starts up after the restore
|
||||
- name: Verify Restore
|
||||
run: |
|
||||
SITE_REQUEST_STATUS_CODE=$(curl -sSL https://$SITE_HOSTNAME -w '%{http_code}' -o /dev/null)
|
||||
|
||||
MAX_RETRIES=5
|
||||
RETRY_INTERVAL=30
|
||||
RETRY_COUNT=0
|
||||
|
||||
# Retry (wait) a few times if the site isn't up yet
|
||||
while [ "$SITE_REQUEST_STATUS_CODE" != "200" ] && [ $RETRY_COUNT -lt $MAX_RETRIES ]; do
|
||||
echo "Site is not up yet (Current Status Code: $SITE_REQUEST_STATUS_CODE). Waiting for $RETRY_INTERVAL seconds (attempt $((RETRY_COUNT + 1))/$MAX_RETRIES)..."
|
||||
sleep $RETRY_INTERVAL
|
||||
|
||||
# Get the current status of the site
|
||||
SITE_REQUEST_STATUS_CODE=$(curl -sSL http://account.bridgemanaccessible.ca -w '%{http_code}' -o /dev/null)
|
||||
|
||||
# Increment the retry count
|
||||
RETRY_COUNT=$((RETRY_COUNT + 1))
|
||||
done
|
||||
|
||||
if [ "$SITE_REQUEST_STATUS_CODE" != "200" ]; then
|
||||
echo "Error: Site did not start up after $((RETRY_COUNT * RETRY_INTERVAL)) seconds."
|
||||
exit 1
|
||||
fi
|
||||
|
||||
echo "Site is up and running (Status Code: $SITE_REQUEST_STATUS_CODE)."
|
||||
9
.gitignore
vendored
Normal file
9
.gitignore
vendored
Normal file
|
|
@ -0,0 +1,9 @@
|
|||
# Exclude the filled values.yaml file
|
||||
values.filled.yaml
|
||||
|
||||
# Exclude any packaged charts
|
||||
services-dashboard-*.tgz
|
||||
|
||||
# Random other stuff
|
||||
templates.old
|
||||
input.json
|
||||
21
.helmignore
Normal file
21
.helmignore
Normal file
|
|
@ -0,0 +1,21 @@
|
|||
# Ignore the ignore file
|
||||
.helmignore
|
||||
|
||||
# Ignore the Helm chart's packaged tarball
|
||||
*.tgz
|
||||
|
||||
# Ignore git related files
|
||||
.git
|
||||
.gitignore
|
||||
|
||||
# Ignore github automation files
|
||||
.github
|
||||
|
||||
# Ignore inputs file
|
||||
input.json
|
||||
|
||||
# Ignore the filled in values file
|
||||
values.filled.yaml
|
||||
|
||||
# Ignore .old (backup) files
|
||||
*.old*
|
||||
11
Chart.yaml
Normal file
11
Chart.yaml
Normal file
|
|
@ -0,0 +1,11 @@
|
|||
apiVersion: v1
|
||||
appVersion: "1.0.298"
|
||||
description: A Helm chart for deploying Bridgeman Accessible's Services Dashboard
|
||||
home: https://github.com/Bridgeman-Accessible/services.bridgemanaccessible.ca-helm
|
||||
maintainers:
|
||||
- email: info@bridgemanaccessible.ca
|
||||
name: Bridgeman Accessible
|
||||
name: services-dashboard
|
||||
sources:
|
||||
- https://github.com/Bridgeman-Accessible/services.bridgemanaccessible.ca
|
||||
version: "1.5.1"
|
||||
20
README.md
20
README.md
|
|
@ -1,2 +1,20 @@
|
|||
# services.bridgemanaccessible.ca-helm
|
||||
# Services Dashboard Helm Chart
|
||||
A Helm chart for deploying Bridgeman Accessible's Services Dashboard
|
||||
|
||||
## Automation (Githu Actions)
|
||||
This repository is setup with CI/CD automation that allows for easy deployment when new code is pushed to the repository.
|
||||
|
||||
This workflow can be broken down into a few key actions:
|
||||
1. Build a new version of the Helm Chart
|
||||
2. If already deployed, backup old data
|
||||
3. Redeploy the Helm Chart (by deleting and then re-creating)
|
||||
4. Restore the data using the created backup, if applicable
|
||||
|
||||
### New Version
|
||||
If the Chart version matches a version already in the remote registry/repository the patch number will be automatically incremented.
|
||||
|
||||
### Creating backup of current status
|
||||
Note, there is an inherent assumption that the pod is currently up and running (stable)
|
||||
|
||||
### Restoring data from backup
|
||||
Can fail if redeployment doesn't go smoothly though tries to accommodate for common issues.
|
||||
|
|
|
|||
9
templates/acs-email-secret.yaml
Normal file
9
templates/acs-email-secret.yaml
Normal file
|
|
@ -0,0 +1,9 @@
|
|||
{{- if and (.Values.comms.email.enabled) (eq .Values.comms.email.type "acs") -}}
|
||||
apiVersion: v1
|
||||
kind: Secret
|
||||
metadata:
|
||||
name: {{ .Release.Name }}-email-secret
|
||||
type: opaque
|
||||
data:
|
||||
connection-string: {{ .Values.comms.email.connectionString | b64enc }}
|
||||
{{- end -}}
|
||||
12
templates/cache-configmap.yaml
Normal file
12
templates/cache-configmap.yaml
Normal file
|
|
@ -0,0 +1,12 @@
|
|||
apiVersion: v1
|
||||
kind: ConfigMap
|
||||
metadata:
|
||||
name: {{ .Release.Name }}-cache-configmap
|
||||
namespace: {{ .Release.Namespace }}
|
||||
data:
|
||||
{{- if and (eq .Values.cache.type "redis") (.Values.cache.create) }}
|
||||
hostname: {{ .Release.Name }}-redis
|
||||
{{- else }}
|
||||
hostname: {{ .Values.cache.hostname }}
|
||||
{{- end }}
|
||||
port: {{ .Values.cache.port | quote }}
|
||||
7
templates/cache-credentials-secret.yaml
Normal file
7
templates/cache-credentials-secret.yaml
Normal file
|
|
@ -0,0 +1,7 @@
|
|||
apiVersion: v1
|
||||
kind: Secret
|
||||
metadata:
|
||||
name: {{ .Release.Name }}-cache-credentials
|
||||
type: Opaque
|
||||
data:
|
||||
password: {{ .Values.cache.password | b64enc }}
|
||||
37
templates/database.yaml
Normal file
37
templates/database.yaml
Normal file
|
|
@ -0,0 +1,37 @@
|
|||
{{- if and (eq .Values.database.type "postgres") (.Values.database.create) -}}
|
||||
apiVersion: postgresql.org/v1
|
||||
kind: PostgresDatabase
|
||||
metadata:
|
||||
name: {{ .Release.Name }}-db
|
||||
namespace: {{ .Release.Namespace }}
|
||||
spec:
|
||||
dbName:
|
||||
envFrom:
|
||||
configMapKeyRef:
|
||||
- name: {{ .Release.Name }}-db-credentials
|
||||
namespace: postgres-controller
|
||||
key: db-name
|
||||
dbRoleName:
|
||||
envFrom:
|
||||
configMapKeyRef:
|
||||
- name: {{ .Release.Name }}-db-credentials
|
||||
namespace: postgres-controller
|
||||
key: db-user
|
||||
dbRolePassword:
|
||||
envFrom:
|
||||
secretKeyRef:
|
||||
- name: {{ .Release.Name }}-db-password
|
||||
namespace: postgres-controller
|
||||
key: password
|
||||
# Because we've adopted a "throw away"/"ephemeral"/"container-esk" approach to our database, we want it to be dropped/deleted when everything else is deleted.
|
||||
# This is because we re-create it and restore from a backup on every deploy.
|
||||
# Which helps keep the data current and reinforces the utility of the backup and restore systems.
|
||||
onDeletion:
|
||||
# Whether to drop the database when the resource is deleted
|
||||
dropDB: true
|
||||
# Whether to drop the role when the resource is deleted
|
||||
dropRole: true
|
||||
{{- if .Values.database.instance_id }}
|
||||
dbInstanceId: {{ .Values.database.instance_id }}
|
||||
{{- end }}
|
||||
{{- end -}}
|
||||
16
templates/db-credentials-config-map-postgres-controller.yaml
Normal file
16
templates/db-credentials-config-map-postgres-controller.yaml
Normal file
|
|
@ -0,0 +1,16 @@
|
|||
{{- if and (eq .Values.database.type "postgres") (.Values.database.create) -}}
|
||||
apiVersion: v1
|
||||
kind: ConfigMap
|
||||
metadata:
|
||||
name: {{ .Release.Name }}-db-credentials
|
||||
namespace: postgres-controller
|
||||
data:
|
||||
db-host: {{ .Values.database.host }}
|
||||
db-name: {{ .Values.database.name }}
|
||||
db-user: {{ .Values.database.user }}
|
||||
{{- if .Values.database.port }}
|
||||
db-port: {{ .Values.database.port | quote }}
|
||||
{{- else }}
|
||||
db-port: "5432"
|
||||
{{- end }}
|
||||
{{- end -}}
|
||||
13
templates/db-credentials-config-map.yaml
Normal file
13
templates/db-credentials-config-map.yaml
Normal file
|
|
@ -0,0 +1,13 @@
|
|||
apiVersion: v1
|
||||
kind: ConfigMap
|
||||
metadata:
|
||||
name: {{ .Release.Name }}-db-credentials
|
||||
data:
|
||||
db-host: {{ .Values.database.host }}
|
||||
db-name: {{ .Values.database.name }}
|
||||
db-user: {{ .Values.database.user }}
|
||||
{{- if .Values.database.port }}
|
||||
db-port: {{ .Values.database.port | quote }}
|
||||
{{- else }}
|
||||
db-port: "5432"
|
||||
{{- end }}
|
||||
10
templates/db-password-secret-postgres-controller.yaml
Normal file
10
templates/db-password-secret-postgres-controller.yaml
Normal file
|
|
@ -0,0 +1,10 @@
|
|||
{{- if and (eq .Values.database.type "postgres") (.Values.database.create) -}}
|
||||
apiVersion: v1
|
||||
kind: Secret
|
||||
metadata:
|
||||
name: {{ .Release.Name }}-db-password
|
||||
namespace: postgres-controller
|
||||
type: Opaque
|
||||
data:
|
||||
password: {{ .Values.database.password | b64enc }}
|
||||
{{- end -}}
|
||||
7
templates/db-password-secret.yaml
Normal file
7
templates/db-password-secret.yaml
Normal file
|
|
@ -0,0 +1,7 @@
|
|||
apiVersion: v1
|
||||
kind: Secret
|
||||
metadata:
|
||||
name: {{ .Release.Name }}-db-password
|
||||
type: Opaque
|
||||
data:
|
||||
password: {{ .Values.database.password | b64enc }}
|
||||
418
templates/deployment.yaml
Normal file
418
templates/deployment.yaml
Normal file
|
|
@ -0,0 +1,418 @@
|
|||
apiVersion: apps/v1
|
||||
kind: Deployment
|
||||
metadata:
|
||||
name: {{ .Release.Name }}
|
||||
labels:
|
||||
app: {{ .Release.Name }}
|
||||
spec:
|
||||
replicas: {{ .Values.replicaCount }}
|
||||
selector:
|
||||
matchLabels:
|
||||
app: {{ .Release.Name }}
|
||||
template:
|
||||
metadata:
|
||||
labels:
|
||||
app: {{ .Release.Name }}
|
||||
spec:
|
||||
containers:
|
||||
- name: {{ .Release.Name }}
|
||||
image: {{ .Values.image.repository }}:{{ .Values.image.tag }}
|
||||
imagePullPolicy: {{ .Values.image.pullPolicy }}
|
||||
ports:
|
||||
- containerPort: {{ .Values.container.port }}
|
||||
env:
|
||||
- name: NODE_ENV
|
||||
value: {{ .Values.container.env }}
|
||||
- name: PORT
|
||||
value: "{{ .Values.container.port }}"
|
||||
- name: DOMAIN
|
||||
value: "bridgemanaccessible.ca"
|
||||
- name: ACCOUNTS_DEV_PORT
|
||||
value: "9010"
|
||||
- name: ACCOUNTS_SUBDOMAIN_PREFIX
|
||||
value: "account"
|
||||
# OAuth Implementation Stuff
|
||||
- name: BASE_APP_URL
|
||||
valueFrom:
|
||||
configMapKeyRef:
|
||||
name: {{ .Release.Name }}-oauth-credentials
|
||||
key: base-app-url
|
||||
- name: APP_ABBRV
|
||||
valueFrom:
|
||||
configMapKeyRef:
|
||||
name: {{ .Release.Name }}-oauth-credentials
|
||||
key: app-abbreviation
|
||||
- name: APP_NAME
|
||||
valueFrom:
|
||||
configMapKeyRef:
|
||||
name: {{ .Release.Name }}-oauth-credentials
|
||||
key: app-name
|
||||
- name: SERVICE_NAME
|
||||
valueFrom:
|
||||
configMapKeyRef:
|
||||
name: {{ .Release.Name }}-oauth-credentials
|
||||
key: service-name
|
||||
- name: DEV_PORT
|
||||
valueFrom:
|
||||
configMapKeyRef:
|
||||
name: {{ .Release.Name }}-oauth-credentials
|
||||
key: dev-port
|
||||
# Database credentials
|
||||
- name: DB_HOST
|
||||
valueFrom:
|
||||
configMapKeyRef:
|
||||
name: {{ .Release.Name }}-db-credentials
|
||||
key: db-host
|
||||
- name: DB_NAME
|
||||
valueFrom:
|
||||
configMapKeyRef:
|
||||
name: {{ .Release.Name }}-db-credentials
|
||||
key: db-name
|
||||
- name: DB_PASSWORD
|
||||
valueFrom:
|
||||
secretKeyRef:
|
||||
name: {{ .Release.Name }}-db-password
|
||||
key: password
|
||||
- name: DB_PORT
|
||||
valueFrom:
|
||||
configMapKeyRef:
|
||||
name: {{ .Release.Name }}-db-credentials
|
||||
key: db-port
|
||||
- name: DB_USER
|
||||
valueFrom:
|
||||
configMapKeyRef:
|
||||
name: {{ .Release.Name }}-db-credentials
|
||||
key: db-user
|
||||
# NoSQL Credentials
|
||||
{{- if eq .Values.nosql.type "mongodb" }}
|
||||
- name: STORAGE_ACCOUNT_CONNECTION_STRING
|
||||
valueFrom:
|
||||
secretKeyRef:
|
||||
name: {{ .Release.Name }}-mongo-credentials
|
||||
key: connection-string
|
||||
{{- else if eq .Values.nosql.type "azure" }}
|
||||
- name: STORAGE_ACCOUNT_KEY
|
||||
valueFrom:
|
||||
secretKeyRef:
|
||||
name: {{ .Release.Name }}-azure-tables-credentials
|
||||
key: key
|
||||
- name: STORAGE_ACCOUNT_NAME
|
||||
valueFrom:
|
||||
configMapKeyRef:
|
||||
name: {{ .Release.Name }}-azure-tables-config
|
||||
key: name
|
||||
{{- end }}
|
||||
# NoSQL Grouping Names
|
||||
- name: ACCESS_PROPERTIES_STORAGE_TABLE_NAME
|
||||
valueFrom:
|
||||
configMapKeyRef:
|
||||
name: {{ .Release.Name }}-nosql-grouping
|
||||
key: access-properties
|
||||
- name: LOCALE_PROPERTIES_STORAGE_TABLE_NAME
|
||||
valueFrom:
|
||||
configMapKeyRef:
|
||||
name: {{ .Release.Name }}-nosql-grouping
|
||||
key: locales
|
||||
- name: ORDER_PROPERTIES_STORAGE_TABLE_NAME
|
||||
valueFrom:
|
||||
configMapKeyRef:
|
||||
name: {{ .Release.Name }}-nosql-grouping
|
||||
key: order-properties
|
||||
- name: PRICE_PROPERTIES_STORAGE_TABLE_NAME
|
||||
valueFrom:
|
||||
configMapKeyRef:
|
||||
name: {{ .Release.Name }}-nosql-grouping
|
||||
key: price-properties
|
||||
- name: SERVICE_CATEGORY_PROPERTIES_STORAGE_TABLE_NAME
|
||||
valueFrom:
|
||||
configMapKeyRef:
|
||||
name: {{ .Release.Name }}-nosql-grouping
|
||||
key: service-category-properties
|
||||
- name: SERVICE_PROPERTIES_STORAGE_TABLE_NAME
|
||||
valueFrom:
|
||||
configMapKeyRef:
|
||||
name: {{ .Release.Name }}-nosql-grouping
|
||||
key: service-properties
|
||||
# -- Secrets Vault (Hashicorp Vault OR Azure Key Vault) --
|
||||
{{- if .Values.vault.enabled }}
|
||||
{{- if eq .Values.vault.type "azure" }}
|
||||
- name: KEYVAULT_CLIENT_ID
|
||||
valueFrom:
|
||||
secretKeyRef:
|
||||
name: {{ .Release.Name }}-vault-secret
|
||||
key: client-id
|
||||
- name: KEYVAULT_CLIENT_SECRET
|
||||
valueFrom:
|
||||
secretKeyRef:
|
||||
name: {{ .Release.Name }}-vault-secret
|
||||
key: client-secret
|
||||
- name: KEYVAULT_NAME
|
||||
valueFrom:
|
||||
secretKeyRef:
|
||||
name: {{ .Release.Name }}-vault-secret
|
||||
key: name
|
||||
- name: KEYVAULT_TENANT_ID
|
||||
valueFrom:
|
||||
secretKeyRef:
|
||||
name: {{ .Release.Name }}-vault-secret
|
||||
key: tenant-id
|
||||
{{- else if eq .Values.vault.type "hashicorp" }}
|
||||
- name: VAULT_NAME
|
||||
valueFrom:
|
||||
secretKeyRef:
|
||||
name: {{ .Release.Name }}-vault-secret
|
||||
key: vault-name
|
||||
- name: VAULT_PORT
|
||||
valueFrom:
|
||||
secretKeyRef:
|
||||
name: {{ .Release.Name }}-vault-secret
|
||||
key: vault-port
|
||||
{{- end }}
|
||||
{{- end }}
|
||||
# Caching Server Variables
|
||||
- name: CACHE_HOSTNAME
|
||||
valueFrom:
|
||||
configMapKeyRef:
|
||||
name: {{ .Release.Name }}-cache-configmap
|
||||
key: hostname
|
||||
- name: CACHE_PORT
|
||||
valueFrom:
|
||||
configMapKeyRef:
|
||||
name: {{ .Release.Name }}-cache-configmap
|
||||
key: port
|
||||
- name: CACHE_PASSWORD
|
||||
valueFrom:
|
||||
secretKeyRef:
|
||||
name: {{ .Release.Name }}-cache-credentials
|
||||
key: password
|
||||
# Email (Azure Communication Services API credentials, etc...)
|
||||
{{- if and (.Values.comms.email.enabled) (eq .Values.comms.email.type "acs") }}
|
||||
- name: EMAIL_CONNECTION_STRING
|
||||
valueFrom:
|
||||
secretKeyRef:
|
||||
name: {{ .Release.Name }}-email-secret
|
||||
key: connection-string
|
||||
{{- end }}
|
||||
# Third-Party Integrations
|
||||
{{- if .Values.thirdParty.stripe.enabled }}
|
||||
- name: STRIPE_PUBLIC_KEY
|
||||
valueFrom:
|
||||
secretKeyRef:
|
||||
name: {{ .Release.Name }}-stripe-secret
|
||||
key: public-key
|
||||
- name: STRIPE_SECRET_KEY
|
||||
valueFrom:
|
||||
secretKeyRef:
|
||||
name: {{ .Release.Name }}-stripe-secret
|
||||
key: secret-key
|
||||
- name: STRIPE_TEST_PUBLIC_KEY
|
||||
valueFrom:
|
||||
secretKeyRef:
|
||||
name: {{ .Release.Name }}-stripe-secret
|
||||
key: test-public-key
|
||||
- name: STRIPE_TEST_SECRET_KEY
|
||||
valueFrom:
|
||||
secretKeyRef:
|
||||
name: {{ .Release.Name }}-stripe-secret
|
||||
key: test-secret-key
|
||||
{{- end }}
|
||||
{{- if .Values.thirdParty.moneris.enabled }}
|
||||
- name: MONERIS_MERCHANT_ID
|
||||
valueFrom:
|
||||
secretKeyRef:
|
||||
name: {{ .Release.Name }}-moneris-secret
|
||||
key: merchant-id
|
||||
- name: MONERIS_STORE_ID
|
||||
valueFrom:
|
||||
secretKeyRef:
|
||||
name: {{ .Release.Name }}-moneris-secret
|
||||
key: store-id
|
||||
- name: MONERIS_HT_PROFILE_ID
|
||||
valueFrom:
|
||||
secretKeyRef:
|
||||
name: {{ .Release.Name }}-moneris-secret
|
||||
key: ht-profile-id
|
||||
- name: MONERIS_APP_ID
|
||||
valueFrom:
|
||||
secretKeyRef:
|
||||
name: {{ .Release.Name }}-moneris-secret
|
||||
key: app-id
|
||||
- name: MONERIS_APP_SECRET
|
||||
valueFrom:
|
||||
secretKeyRef:
|
||||
name: {{ .Release.Name }}-moneris-secret
|
||||
key: app-secret
|
||||
- name: MONERIS_TEST_MERCHANT_ID
|
||||
valueFrom:
|
||||
secretKeyRef:
|
||||
name: {{ .Release.Name }}-moneris-secret
|
||||
key: test-merchant-id
|
||||
- name: MONERIS_TEST_STORE_ID
|
||||
valueFrom:
|
||||
secretKeyRef:
|
||||
name: {{ .Release.Name }}-moneris-secret
|
||||
key: test-store-id
|
||||
- name: MONERIS_TEST_HT_PROFILE_ID
|
||||
valueFrom:
|
||||
secretKeyRef:
|
||||
name: {{ .Release.Name }}-moneris-secret
|
||||
key: test-ht-profile-id
|
||||
- name: MONERIS_TEST_APP_ID
|
||||
valueFrom:
|
||||
secretKeyRef:
|
||||
name: {{ .Release.Name }}-moneris-secret
|
||||
key: test-app-id
|
||||
- name: MONERIS_TEST_APP_SECRET
|
||||
valueFrom:
|
||||
secretKeyRef:
|
||||
name: {{ .Release.Name }}-moneris-secret
|
||||
key: test-app-secret
|
||||
{{- end }}
|
||||
# Logging Sidecar related environment variables
|
||||
{{- if .Values.loggingSidecar.enabled }}
|
||||
- name: LOGGING_SIDE_CAR_PORT
|
||||
value: {{ .Values.loggingSidecar.port | quote }}
|
||||
{{- end }}
|
||||
# Due to subtleties related to how the entrypoint scripts detects how/when to proceed
|
||||
# This environment variable indicates if the entrypoint should wait for a restore to complete
|
||||
{{- if ne .Values.container.restoreFromBackup "" }}
|
||||
- name: RESTORE_FROM_BACKUP
|
||||
value: {{ .Values.container.restoreFromBackup | quote }}
|
||||
{{- end }}
|
||||
{{- if .Values.vault.create.enabled }}
|
||||
volumeMounts:
|
||||
# Mount for a shared volume for app's vault credentials
|
||||
# Note, this is required because the credentials are created as part of the vault's setup and not known ahead of time
|
||||
# This seems a better approach than passing, as an example, a preset root token as an environment variable for a few reasons:
|
||||
# 1. This way even if the app's credentials become compromised their more limited in scope/permissions
|
||||
# 2. This creates a semi-unintentional wait lock for the app until the vault is ready (and the credentials are created)
|
||||
- name: role-vars
|
||||
mountPath: /role_vars
|
||||
readOnly: true
|
||||
{{- end }}
|
||||
# Logging sidecar for sending logs to a log aggregator
|
||||
{{- if .Values.loggingSidecar.enabled }}
|
||||
- name: {{ .Values.loggingSidecar.name }}
|
||||
image: {{ .Values.loggingSidecar.image.repository }}:{{ .Values.loggingSidecar.image.tag }}
|
||||
imagePullPolicy: {{ .Values.loggingSidecar.image.pullPolicy }}
|
||||
ports:
|
||||
- containerPort: {{ .Values.loggingSidecar.port }}
|
||||
env:
|
||||
- name: PORT
|
||||
value: {{ .Values.loggingSidecar.port | quote }}
|
||||
# Log aggregator (OpenObserve) auth variables
|
||||
- name: LOGGER_AUTH_USERNAME
|
||||
valueFrom:
|
||||
configMapKeyRef:
|
||||
name: {{ .Release.Name }}-logging-sidecar-credentials
|
||||
key: username
|
||||
- name: LOGGER_AUTH_PASSWORD
|
||||
valueFrom:
|
||||
secretKeyRef:
|
||||
name: {{ .Release.Name }}-logging-sidecar-password
|
||||
key: password
|
||||
{{- end }}
|
||||
# Backup sidecar for backing up service data
|
||||
{{- if .Values.backupSidecar.enabled }}
|
||||
- name: {{ .Values.backupSidecar.name }}
|
||||
image: {{ .Values.backupSidecar.image.repository }}:{{ .Values.backupSidecar.image.tag }}
|
||||
imagePullPolicy: {{ .Values.backupSidecar.image.pullPolicy }}
|
||||
ports:
|
||||
- containerPort: {{ .Values.backupSidecar.port }}
|
||||
env:
|
||||
# Release name (used to identify the service/release the backups came from in remote storage)
|
||||
- name: RELEASE_NAME
|
||||
value: {{ .Release.Name }}
|
||||
# Database related environment variables
|
||||
- name: DB_HOST
|
||||
valueFrom:
|
||||
configMapKeyRef:
|
||||
name: {{ .Release.Name }}-db-credentials
|
||||
key: db-host
|
||||
- name: DB_NAME
|
||||
valueFrom:
|
||||
configMapKeyRef:
|
||||
name: {{ .Release.Name }}-db-credentials
|
||||
key: db-name
|
||||
- name: DB_PASSWORD
|
||||
valueFrom:
|
||||
secretKeyRef:
|
||||
name: {{ .Release.Name }}-db-password
|
||||
key: password
|
||||
- name: DB_PORT
|
||||
valueFrom:
|
||||
configMapKeyRef:
|
||||
name: {{ .Release.Name }}-db-credentials
|
||||
key: db-port
|
||||
- name: DB_USER
|
||||
valueFrom:
|
||||
configMapKeyRef:
|
||||
name: {{ .Release.Name }}-db-credentials
|
||||
key: db-user
|
||||
{{- if .Values.vault.create.snapshotServer.enabled }}
|
||||
- name: VAULT_NAME
|
||||
valueFrom:
|
||||
secretKeyRef:
|
||||
name: {{ .Release.Name }}-pass-vault-secret
|
||||
key: vault-name
|
||||
- name: VAULT_SNAPSHOT_SERVER_PORT
|
||||
valueFrom:
|
||||
configMapKeyRef:
|
||||
name: {{ .Release.Name }}-vault-snapshot-config
|
||||
key: port
|
||||
{{- end }}
|
||||
# NoSQL storage related environment variables
|
||||
# Note, we only worry about self-hosted options as cloud-based should have their own backups etc...
|
||||
{{- if eq .Values.nosql.type "mongodb" }}
|
||||
- name: STORAGE_ACCOUNT_CONNECTION_STRING
|
||||
valueFrom:
|
||||
secretKeyRef:
|
||||
name: {{ .Release.Name }}-mongo-credentials
|
||||
key: connection-string
|
||||
{{- end }}
|
||||
# Redis related environment variables
|
||||
# Redis is used for BullMQ, which is how we schedule backups
|
||||
# We use this instead of, for instance cron jobs, as it lets us deal with failures
|
||||
- name: REDIS_HOSTNAME
|
||||
valueFrom:
|
||||
configMapKeyRef:
|
||||
name: {{ .Release.Name }}-cache-configmap
|
||||
key: hostname
|
||||
- name: REDIS_PORT
|
||||
valueFrom:
|
||||
configMapKeyRef:
|
||||
name: {{ .Release.Name }}-cache-configmap
|
||||
key: port
|
||||
- name: REDIS_PASSWORD
|
||||
valueFrom:
|
||||
secretKeyRef:
|
||||
name: {{ .Release.Name }}-cache-credentials
|
||||
key: password
|
||||
{{- if .Values.passVault.create.snapshotServer.enabled }}
|
||||
volumeMounts:
|
||||
# Mount for a shared volume for Vault credentials
|
||||
# This is separate from the app's `role vars` volume because it includes other credentials
|
||||
# In particular, the unseal keys which we require when/if we restore from the backup
|
||||
# This volume is also read-only where the `role-vars` is read-write (see description below for why)
|
||||
- name: creds
|
||||
mountPath: /vault-creds
|
||||
readOnly: true
|
||||
# Mount for a shared volume for the Vault's role variables for the app
|
||||
# This is required by the backup sidecar because if a restart of the app occurs AFTER a vault has been reset (ex. vault using a different container instance),
|
||||
# despite the vault data being restored the app would receive incorrect credentials (because this is ONLY written during setup of the vault)
|
||||
# The backup sidecar mitigates this by doing it's own write (to overwrite) once it's done a restore
|
||||
- name: role-vars
|
||||
mountPath: /role_vars
|
||||
{{- end }}
|
||||
{{- end }}
|
||||
{{- if .Values.vault.create.enabled }}
|
||||
volumes:
|
||||
- name: role-vars
|
||||
persistentVolumeClaim:
|
||||
claimName: {{ .Release.Name }}-vault-role-vars
|
||||
{{- if .Values.vault.create.snapshotServer.enabled }}
|
||||
- name: creds
|
||||
persistentVolumeClaim:
|
||||
claimName: {{ .Release.Name }}-vault-creds
|
||||
{{- end }}
|
||||
{{- end }}
|
||||
21
templates/ingress.yaml
Normal file
21
templates/ingress.yaml
Normal file
|
|
@ -0,0 +1,21 @@
|
|||
{{- if .Values.ingress.enabled -}}
|
||||
apiVersion: networking.k8s.io/v1
|
||||
kind: Ingress
|
||||
metadata:
|
||||
name: {{ .Release.Name }}
|
||||
annotations:
|
||||
nginx.ingress.kubernetes.io/rewrite-target: /
|
||||
spec:
|
||||
rules:
|
||||
- host: {{ .Values.ingress.host }}
|
||||
http:
|
||||
paths:
|
||||
- path: /
|
||||
pathType: Prefix
|
||||
backend:
|
||||
service:
|
||||
name: {{ .Release.Name }}
|
||||
port:
|
||||
number: 80
|
||||
ingressClassName: {{ .Values.ingress.class }}
|
||||
{{- end -}}
|
||||
10
templates/logging-sidecar-credentials-configmap.yaml
Normal file
10
templates/logging-sidecar-credentials-configmap.yaml
Normal file
|
|
@ -0,0 +1,10 @@
|
|||
{{- if .Values.loggingSidecar.enabled -}}
|
||||
apiVersion: v1
|
||||
kind: ConfigMap
|
||||
metadata:
|
||||
name: {{ .Release.Name }}-logging-sidecar-credentials
|
||||
labels:
|
||||
app: {{ .Values.loggingSidecar.name }}
|
||||
data:
|
||||
username: {{ .Values.loggingSidecar.auth.username | quote }}
|
||||
{{- end -}}
|
||||
11
templates/logging-sidecar-password-secret.yaml
Normal file
11
templates/logging-sidecar-password-secret.yaml
Normal file
|
|
@ -0,0 +1,11 @@
|
|||
{{- if .Values.loggingSidecar.enabled -}}
|
||||
apiVersion: v1
|
||||
kind: Secret
|
||||
metadata:
|
||||
name: {{ .Release.Name }}-logging-sidecar-password
|
||||
labels:
|
||||
app: {{ .Values.loggingSidecar.name }}
|
||||
type: opaque
|
||||
data:
|
||||
password: {{ .Values.loggingSidecar.auth.password | b64enc | quote }}
|
||||
{{- end -}}
|
||||
20
templates/moneris-secret.yaml
Normal file
20
templates/moneris-secret.yaml
Normal file
|
|
@ -0,0 +1,20 @@
|
|||
{{- if .Values.thirdParty.moneris.enabled -}}
|
||||
apiVersion: v1
|
||||
kind: Secret
|
||||
metadata:
|
||||
name: {{ .Release.Name }}-moneris-secret
|
||||
labels:
|
||||
app: {{ .Release.Name }}
|
||||
type: Opaque
|
||||
data:
|
||||
merchant-id: {{ .Values.thirdParty.moneris.merchantId | b64enc }}
|
||||
store-id: {{ .Values.thirdParty.moneris.storeId | b64enc }}
|
||||
ht-profile-id: {{ .Values.thirdParty.moneris.htProfileId | b64enc }}
|
||||
app-id: {{ .Values.thirdParty.moneris.appId | b64enc }}
|
||||
app-secret: {{ .Values.thirdParty.moneris.appSecret | b64enc }}
|
||||
test-merchant-id: {{ .Values.thirdParty.moneris.testMerchantId | b64enc }}
|
||||
test-store-id: {{ .Values.thirdParty.moneris.testStoreId | b64enc }}
|
||||
test-ht-profile-id: {{ .Values.thirdParty.moneris.testHtProfileId | b64enc }}
|
||||
test-app-id: {{ .Values.thirdParty.moneris.testAppId | b64enc }}
|
||||
test-app-secret: {{ .Values.thirdParty.moneris.testAppSecret | b64enc }}
|
||||
{{- end -}}
|
||||
15
templates/mongo-credentials-secret.yaml
Normal file
15
templates/mongo-credentials-secret.yaml
Normal file
|
|
@ -0,0 +1,15 @@
|
|||
{{- if eq .Values.nosql.type "mongodb" -}}
|
||||
apiVersion: v1
|
||||
kind: Secret
|
||||
metadata:
|
||||
name: {{ .Release.Name }}-mongo-credentials
|
||||
type: Opaque
|
||||
data:
|
||||
user: {{ .Values.nosql.user | b64enc }}
|
||||
password: {{ .Values.nosql.password | b64enc }}
|
||||
{{- if and (.Values.nosql.connectionString) (not .values.nosql.create) }}
|
||||
connection-string: {{ .Values.nosql.connectionString | b64enc }}
|
||||
{{- else if .Values.nosql.create }}
|
||||
connection-string: {{ printf "mongodb://%s:%s@%s-mongo-svc.%s.svc.cluster.local:27017/%s?replicaSet=%s-mongo" .Values.nosql.user .Values.nosql.password .Release.Name .Release.Namespace .Values.nosql.name .Release.Name | b64enc }}
|
||||
{{- end }}
|
||||
{{- end -}}
|
||||
16
templates/mongo-role-binding-database.yaml
Normal file
16
templates/mongo-role-binding-database.yaml
Normal file
|
|
@ -0,0 +1,16 @@
|
|||
{{- if and (eq .Values.nosql.type "mongodb") (.Values.nosql.create) -}}
|
||||
apiVersion: rbac.authorization.k8s.io/v1
|
||||
kind: RoleBinding
|
||||
metadata:
|
||||
name: mongodb-database
|
||||
namespace: {{ .Release.Namespace }}
|
||||
subjects:
|
||||
- kind: ServiceAccount
|
||||
name: mongodb-database
|
||||
namespace: {{ .Release.Namespace }}
|
||||
roleRef:
|
||||
kind: Role
|
||||
name: mongodb-database
|
||||
namespace: {{ .Release.Namespace }}
|
||||
apiGroup: rbac.authorization.k8s.io
|
||||
{{- end -}}
|
||||
22
templates/mongo-role-database.yaml
Normal file
22
templates/mongo-role-database.yaml
Normal file
|
|
@ -0,0 +1,22 @@
|
|||
{{- if and (eq .Values.nosql.type "mongodb") (.Values.nosql.create) -}}
|
||||
apiVersion: rbac.authorization.k8s.io/v1
|
||||
kind: Role
|
||||
metadata:
|
||||
name: mongodb-database
|
||||
namespace: {{ .Release.Namespace }}
|
||||
rules:
|
||||
- apiGroups:
|
||||
- ""
|
||||
resources:
|
||||
- secrets
|
||||
verbs:
|
||||
- get
|
||||
- apiGroups:
|
||||
- ""
|
||||
resources:
|
||||
- pods
|
||||
verbs:
|
||||
- patch
|
||||
- delete
|
||||
- get
|
||||
{{- end -}}
|
||||
7
templates/mongo-service-account-database.yaml
Normal file
7
templates/mongo-service-account-database.yaml
Normal file
|
|
@ -0,0 +1,7 @@
|
|||
{{- if and (eq .Values.nosql.type "mongodb") (.Values.nosql.create) -}}
|
||||
apiVersion: v1
|
||||
kind: ServiceAccount
|
||||
metadata:
|
||||
name: mongodb-database
|
||||
namespace: {{ .Release.Namespace }}
|
||||
{{- end -}}
|
||||
31
templates/mongo.yaml
Normal file
31
templates/mongo.yaml
Normal file
|
|
@ -0,0 +1,31 @@
|
|||
{{- if and (eq .Values.nosql.type "mongodb") (.Values.nosql.create) -}}
|
||||
apiVersion: mongodbcommunity.mongodb.com/v1
|
||||
kind: MongoDBCommunity
|
||||
metadata:
|
||||
name: {{ .Release.Name }}-mongo
|
||||
namespace: {{ .Release.Namespace }}
|
||||
spec:
|
||||
members: {{ .Values.nosql.replicaCount }}
|
||||
type: ReplicaSet
|
||||
version: 4.4.0
|
||||
security:
|
||||
authentication:
|
||||
ignoreUnknownUsers: true
|
||||
modes:
|
||||
- SCRAM
|
||||
tls:
|
||||
enabled: {{ .Values.nosql.tls.enabled }}
|
||||
readinessProbe:
|
||||
initialDelaySeconds: 30
|
||||
periodSeconds: 10
|
||||
users:
|
||||
- name: {{ .Values.nosql.user }}
|
||||
db: {{ .Values.nosql.name }}
|
||||
passwordSecretRef:
|
||||
name: {{ .Release.Name }}-mongo-credentials
|
||||
key: password
|
||||
roles:
|
||||
- name: readWrite
|
||||
db: {{ .Values.nosql.name }}
|
||||
scramCredentialsSecretName: {{ .Release.Name }}-mongo-scram
|
||||
{{- end -}}
|
||||
13
templates/nosql-grouping-config-map.yaml
Normal file
13
templates/nosql-grouping-config-map.yaml
Normal file
|
|
@ -0,0 +1,13 @@
|
|||
apiVersion: v1
|
||||
kind: ConfigMap
|
||||
metadata:
|
||||
name: {{ .Release.Name }}-nosql-grouping
|
||||
labels:
|
||||
app: {{ .Release.Name }}
|
||||
data:
|
||||
access-properties: {{ .Values.nosql.grouping.accessProperties }}
|
||||
locales: {{ .Values.nosql.grouping.locales }}
|
||||
order-properties: {{ .Values.nosql.grouping.orderProperties }}
|
||||
price-properties: {{ .Values.nosql.grouping.priceProperties }}
|
||||
service-category-properties: {{ .Values.nosql.grouping.serviceCategoryProperties }}
|
||||
service-properties: {{ .Values.nosql.grouping.serviceProperties }}
|
||||
12
templates/oauth-credentials-config-map.yaml
Normal file
12
templates/oauth-credentials-config-map.yaml
Normal file
|
|
@ -0,0 +1,12 @@
|
|||
apiVersion: v1
|
||||
kind: ConfigMap
|
||||
metadata:
|
||||
name: {{ .Release.Name }}-oauth-credentials
|
||||
labels:
|
||||
app: {{ .Release.Name }}
|
||||
data:
|
||||
base-app-url: {{ .Values.oauth.baseAppUrl }}
|
||||
app-abbreviation: {{ .Values.oauth.appAbbreviation }}
|
||||
app-name: {{ .Values.oauth.appName }}
|
||||
service-name: {{ .Values.oauth.serviceName }}
|
||||
dev-port: {{ .Values.oauth.devPort | quote }}
|
||||
49
templates/redis-deployment.yaml
Normal file
49
templates/redis-deployment.yaml
Normal file
|
|
@ -0,0 +1,49 @@
|
|||
{{- if and (eq .Values.cache.type "redis") (.Values.cache.create) -}}
|
||||
apiVersion: apps/v1
|
||||
kind: Deployment
|
||||
metadata:
|
||||
name: {{ .Release.Name }}-redis
|
||||
labels:
|
||||
app: redis
|
||||
spec:
|
||||
replicas: {{ .Values.cache.replicaCount }}
|
||||
selector:
|
||||
matchLabels:
|
||||
app: redis
|
||||
template:
|
||||
metadata:
|
||||
labels:
|
||||
app: redis
|
||||
spec:
|
||||
containers:
|
||||
- name: redis
|
||||
image: {{ .Values.cache.image.repository | default "bitnami/redis" }}:{{ .Values.cache.image.tag | default "7.0.5" }}
|
||||
ports:
|
||||
- containerPort: {{ .Values.cache.port }}
|
||||
{{- if .Values.cache.tls.enabled }}
|
||||
- containerPort: {{ .Values.cache.tls.port }}
|
||||
{{- end }}
|
||||
env:
|
||||
- name: ALLOW_EMPTY_PASSWORD
|
||||
value: "false"
|
||||
- name: REDIS_PASSWORD
|
||||
valueFrom:
|
||||
secretKeyRef:
|
||||
name: {{ .Release.Name }}-cache-credentials
|
||||
key: password
|
||||
- name: REDIS_DISABLE_COMMANDS
|
||||
value: "FLUSHDB,FLUSHALL"
|
||||
# TLS configuration
|
||||
#- name: REDIS_TLS_ENABLED
|
||||
# value: "{{ .Values.cache.tls.enabled }}"
|
||||
#- name: REDIS_TLS_AUTH_CLIENTS
|
||||
# value: "yes"
|
||||
#- name: REDIS_TLS_PORT_NUMBER
|
||||
# value: "{{ .Values.cache.tls.port }}"
|
||||
volumeMounts:
|
||||
- name: redis-data
|
||||
mountPath: /bitnami/redis
|
||||
volumes:
|
||||
- name: redis-data
|
||||
emptyDir: {}
|
||||
{{- end -}}
|
||||
15
templates/redis-service.yaml
Normal file
15
templates/redis-service.yaml
Normal file
|
|
@ -0,0 +1,15 @@
|
|||
{{- if and (eq .Values.cache.type "redis") (.Values.cache.create) -}}
|
||||
apiVersion: v1
|
||||
kind: Service
|
||||
metadata:
|
||||
name: {{ .Release.Name }}-redis
|
||||
labels:
|
||||
app: redis
|
||||
spec:
|
||||
ports:
|
||||
- port: {{ .Values.cache.port }}
|
||||
targetPort: {{ .Values.cache.port }}
|
||||
selector:
|
||||
app: redis
|
||||
type: ClusterIP
|
||||
{{- end -}}
|
||||
11
templates/service.yaml
Normal file
11
templates/service.yaml
Normal file
|
|
@ -0,0 +1,11 @@
|
|||
apiVersion: v1
|
||||
kind: Service
|
||||
metadata:
|
||||
name: {{ .Release.Name }}
|
||||
spec:
|
||||
selector:
|
||||
app: {{ .Release.Name }}
|
||||
ports:
|
||||
- protocol: TCP
|
||||
port: 80
|
||||
targetPort: {{ .Values.container.port }}
|
||||
14
templates/stripe-secret.yaml
Normal file
14
templates/stripe-secret.yaml
Normal file
|
|
@ -0,0 +1,14 @@
|
|||
{{- if .Values.thirdParty.stripe.enabled -}}
|
||||
apiVersion: v1
|
||||
kind: Secret
|
||||
metadata:
|
||||
name: {{ .Release.Name }}-stripe-secret
|
||||
labels:
|
||||
app: {{ .Release.Name }}
|
||||
type: Opaque
|
||||
data:
|
||||
public-key: {{ .Values.thirdParty.stripe.publicKey | b64enc }}
|
||||
secret-key: {{ .Values.thirdParty.stripe.secretKey | b64enc }}
|
||||
test-public-key: {{ .Values.thirdParty.stripe.testPublicKey | b64enc }}
|
||||
test-secret-key: {{ .Values.thirdParty.stripe.testSecretKey | b64enc }}
|
||||
{{- end -}}
|
||||
15
templates/vault-creds-persistant-volume-claim.yaml
Normal file
15
templates/vault-creds-persistant-volume-claim.yaml
Normal file
|
|
@ -0,0 +1,15 @@
|
|||
{{- if .Values.vault.create.enabled -}}
|
||||
apiVersion: v1
|
||||
kind: PersistentVolumeClaim
|
||||
metadata:
|
||||
name: {{ .Release.Name }}-vault-creds
|
||||
labels:
|
||||
app: {{ .Release.Name }}-vault
|
||||
spec:
|
||||
storageClassName: {{ .Values.vault.storage.storageClass }}
|
||||
accessModes:
|
||||
- ReadWriteMany
|
||||
resources:
|
||||
requests:
|
||||
storage: {{ .Values.vault.create.storage.size }}
|
||||
{{- end -}}
|
||||
66
templates/vault-deployment.yaml
Normal file
66
templates/vault-deployment.yaml
Normal file
|
|
@ -0,0 +1,66 @@
|
|||
{{- if and (.Values.vault.create.enabled) (eq .Values.vault.type "hashicorp") -}}
|
||||
apiVersion: apps/v1
|
||||
kind: Deployment
|
||||
metadata:
|
||||
name: {{ .Release.Name }}-vault
|
||||
labels:
|
||||
app: {{ .Release.Name }}-vault
|
||||
spec:
|
||||
replicas: 1
|
||||
selector:
|
||||
matchLabels:
|
||||
app: {{ .Release.Name }}-vault
|
||||
template:
|
||||
metadata:
|
||||
labels:
|
||||
app: {{ .Release.Name }}-vault
|
||||
spec:
|
||||
containers:
|
||||
- name: {{ .Release.Name }}-vault
|
||||
image: {{ .Values.vault.create.image.repository }}:{{ .Values.vault.create.image.tag }}
|
||||
ports:
|
||||
- containerPort: 8200
|
||||
- containerPort: 8201
|
||||
{{- if .Values.vault.create.snapshotServer.enabled }}
|
||||
- containerPort: {{ .Values.vault.create.snapshotServer.internalPort }}
|
||||
{{- end }}
|
||||
env:
|
||||
- name: VAULT_ADDR
|
||||
value: http://0.0.0.0:8200
|
||||
- name: POLICY_CAPABILITIES
|
||||
value: {{ .Values.vault.create.policyCapabilities | join "," }}
|
||||
- name: ROLE_ID_SECRET_NAME
|
||||
value: VAULT_ROLE_ID
|
||||
- name: SECRET_ID_SECRET_NAME
|
||||
value: VAULT_SECRET_ID
|
||||
{{- if .Values.vault.create.snapshotServer.enabled }}
|
||||
- name: SNAPSHOT_SERVER_PORT
|
||||
value: {{ .Values.vault.create.snapshotServer.internalPort | quote }}
|
||||
{{- end }}
|
||||
volumeMounts:
|
||||
- name: vault-data
|
||||
mountPath: /vault/data
|
||||
- name: vault-log
|
||||
mountPath: /vault/logs
|
||||
- name: vault-creds
|
||||
mountPath: /vault/creds
|
||||
- name: vault-role-vars
|
||||
mountPath: /role_vars
|
||||
capAdd:
|
||||
- IPC_LOCK
|
||||
volumes:
|
||||
- name: vault-data
|
||||
emptyDir: {}
|
||||
- name: vault-log
|
||||
emptyDir: {}
|
||||
- name: vault-creds
|
||||
{{- if .Values.vault.create.snapshotServer.enabled }}
|
||||
persistentVolumeClaim:
|
||||
claimName: {{ .Release.Name }}-vault-creds
|
||||
{{- else }}
|
||||
emptyDir: {}
|
||||
{{- end }}
|
||||
- name: vault-role-vars
|
||||
persistentVolumeClaim:
|
||||
claimName: {{ .Release.Name }}-vault-role-vars
|
||||
{{- end -}}
|
||||
24
templates/vault-hashicorp-secret.yaml
Normal file
24
templates/vault-hashicorp-secret.yaml
Normal file
|
|
@ -0,0 +1,24 @@
|
|||
{{- if and (.Values.vault.enabled) (eq .Values.vault.type "hashicorp") -}}
|
||||
apiVersion: v1
|
||||
kind: Secret
|
||||
metadata:
|
||||
name: {{ .Release.Name }}-vault-secret
|
||||
type: opaque
|
||||
data:
|
||||
{{- if .Values.vault.create.enabled }}
|
||||
# Because we create the Hashicorp Vault instance as part of the Helm chart,
|
||||
# we can use the name of the created resource (utilizing k8s built-in container connections)
|
||||
# to connect to the Vault instance without having to hard-code the Vault name.
|
||||
vault-name: {{ printf "%s-vault" .Release.Name | b64enc }}
|
||||
# Because we create the Hashicorp Vault instance as part of the Helm chart,
|
||||
# We know the port that the Vault instance is running on.
|
||||
vault-port: {{ printf "%d" 80 | b64enc }}
|
||||
{{- else }}
|
||||
# Because the Vault wasn't created as part of the Helm chart,
|
||||
# we need the deployer to specify the name of the Vault instance to connect to.
|
||||
vault-name: {{ .Values.vault.vaultName | b64enc }}
|
||||
# Because the Vault wasn't created as part of the Helm chart,
|
||||
# we need the deployer to specify the port that the Vault instance is running on.
|
||||
vault-port: {{ .Values.passVault.vaultPort | b64enc }}
|
||||
{{- end }}
|
||||
{{- end -}}
|
||||
21
templates/vault-ingress.yaml
Normal file
21
templates/vault-ingress.yaml
Normal file
|
|
@ -0,0 +1,21 @@
|
|||
{{- if .Values.vault.create.ingress.enabled -}}
|
||||
apiVersion: networking.k8s.io/v1
|
||||
kind: Ingress
|
||||
metadata:
|
||||
name: {{ .Release.Name }}-vault-ingress
|
||||
labels:
|
||||
app: {{ .Release.Name }}-vault
|
||||
spec:
|
||||
ingressClassName: nginx
|
||||
rules:
|
||||
- host: {{ .Values.vault.create.ingress.host }}
|
||||
http:
|
||||
paths:
|
||||
- path: /
|
||||
pathType: Prefix
|
||||
backend:
|
||||
service:
|
||||
name: {{ .Release.Name }}-vault
|
||||
port:
|
||||
number: 80
|
||||
{{- end -}}
|
||||
15
templates/vault-role-vars-persistent-volume-claim.yaml
Normal file
15
templates/vault-role-vars-persistent-volume-claim.yaml
Normal file
|
|
@ -0,0 +1,15 @@
|
|||
{{- if .Values.vault.create.enabled -}}
|
||||
apiVersion: v1
|
||||
kind: PersistentVolumeClaim
|
||||
metadata:
|
||||
name: {{ .Release.Name }}-vault-role-vars
|
||||
labels:
|
||||
app: {{ .Release.Name }}-vault
|
||||
spec:
|
||||
storageClassName: {{ .Values.vault.create.storage.storageClass }}
|
||||
accessModes:
|
||||
- ReadWriteMany
|
||||
resources:
|
||||
requests:
|
||||
storage: {{ .Values.vault.create.storage.size }}
|
||||
{{- end -}}
|
||||
22
templates/vault-service.yaml
Normal file
22
templates/vault-service.yaml
Normal file
|
|
@ -0,0 +1,22 @@
|
|||
{{- if .Values.vault.create.enabled -}}
|
||||
apiVersion: v1
|
||||
kind: Service
|
||||
metadata:
|
||||
name: {{ .Release.Name }}-vault
|
||||
labels:
|
||||
app: {{ .Release.Name }}-vault
|
||||
spec:
|
||||
selector:
|
||||
app: {{ .Release.Name }}-vault
|
||||
ports:
|
||||
{{- if .Values.vault.create.snapshotServer.enabled }}
|
||||
- name: custom-snapshot-server
|
||||
protocol: TCP
|
||||
port: {{ .Values.vault.create.snapshotServer.externalPort }}
|
||||
targetPort: {{ .Values.vault.create.snapshotServer.internalPort }}
|
||||
{{- end }}
|
||||
- name: http
|
||||
protocol: TCP
|
||||
port: 80
|
||||
targetPort: 8200
|
||||
{{- end -}}
|
||||
8
templates/vault-snapshot-server-configmap.yaml
Normal file
8
templates/vault-snapshot-server-configmap.yaml
Normal file
|
|
@ -0,0 +1,8 @@
|
|||
{{- if .Values.vault.create.snapshotServer.enabled -}}
|
||||
apiVersion: v1
|
||||
kind: ConfigMap
|
||||
metadata:
|
||||
name: {{ .Release.Name }}-vault-snapshot-config
|
||||
data:
|
||||
port: {{ .Values.vault.create.snapshotServer.externalPort | quote }}
|
||||
{{- end -}}
|
||||
393
values.yaml
Normal file
393
values.yaml
Normal file
|
|
@ -0,0 +1,393 @@
|
|||
# The number of instances (replicas) of the app to run
|
||||
replicaCount: 1
|
||||
|
||||
image:
|
||||
# The repository of the image to use for the app
|
||||
# Should be in the format `<Image Repository (Ex. containers.example.com)>/<Image Name (Ex. app)>`
|
||||
repository: "<Image Repository>/<Image Name>"
|
||||
|
||||
# The specific image tag to use. It's recommended to use some kind of versioning tag scheme as it makes updating the container without having to fully redeploy easier.
|
||||
# Ex. v1.0.0
|
||||
tag: "v1.0.0"
|
||||
|
||||
# How often the image should be pulled. The possible values are "Always", "Never", and "IfNotPresent"
|
||||
# It's recommended for production to use "IfNotPresent" to avoid pulling the image every time the pod starts
|
||||
# Though, for development, "Always" is recommended to ensure the latest changes are being tested
|
||||
pullPolicy: "IfNotPresent"
|
||||
|
||||
container:
|
||||
# The port that the container listens on (Ex. 8080)
|
||||
port: 8080
|
||||
|
||||
# The environment that the container is running in (Ex. development, production, etc...)
|
||||
# This is used for the NODE_ENV environment variable
|
||||
env: "production"
|
||||
|
||||
# The timestamp of the backup that the entrypoint script should wait for a restore to complete
|
||||
restoreFromBackup: ""
|
||||
|
||||
ingress:
|
||||
# We want an ingress resource if we are deploying to a cluster that has a ingress controller/load balancer
|
||||
# This includes most public cloud providers like EKS, GKE, and AKS
|
||||
enabled: true
|
||||
|
||||
# The DNS Name (Ex. app.example.com) where the app will be accessible
|
||||
host: "<hostname>"
|
||||
|
||||
# The class of the ingress controller that is being used (defaulted here to an NGINX ingress controller as it's popular for Kubernetes clusters)
|
||||
class: nginx
|
||||
|
||||
# Configuration for using OAuth within the app
|
||||
oauth:
|
||||
baseAppUrl: "<Base App URL>"
|
||||
appAbbreviation: "<App Abbreviation>"
|
||||
appName: "<App Name>"
|
||||
serviceName: "<Service Name>"
|
||||
devPort: "<Dev Port>"
|
||||
|
||||
# Configuration for the relational database
|
||||
database:
|
||||
# The type of the relational database that is used.
|
||||
#
|
||||
# The following table lists the possible values for this field:
|
||||
#
|
||||
# | Value | Description |
|
||||
# | ---------- | ------------------------------------------ |
|
||||
# | `postgres` | Uses PostgreSQL as the relational database |
|
||||
#
|
||||
# Note, for use of `postgres`, it uses a [`postgres-controller` CRD](https://github.com/AlanBridgeman/postgres-controller) to create the database
|
||||
#
|
||||
type: "postgres"
|
||||
|
||||
# If set to `true`, the database will be created as part of the deployment
|
||||
# This uses the [`postgres-controller` CRD](https://github.com/AlanBridgeman/postgres-controller) to create the database
|
||||
create: false
|
||||
|
||||
# The host that the database is located on
|
||||
host: "<DB Host>"
|
||||
|
||||
# The name of the database to be used
|
||||
name: "<DB Name>"
|
||||
|
||||
# The user that is used to access the database
|
||||
user: "<DB User>"
|
||||
|
||||
# The password that is used to access the database
|
||||
password: "<DB Password>"
|
||||
|
||||
# The port that the database listens on
|
||||
#port: 5432
|
||||
|
||||
# Allows for distinguishing between multiple database instances/servers
|
||||
#instance_id: ""
|
||||
|
||||
# Configurations for the secrets vault
|
||||
vault:
|
||||
# If a secrets vault should be used
|
||||
# That is, if a dedicated software for secret management should be used
|
||||
# This should virtually always be true if storing any kind of sensitive information as it's the most secure option
|
||||
enabled: true
|
||||
|
||||
# The type of secrets vault to use.
|
||||
#
|
||||
# Vaults
|
||||
# ------
|
||||
# The following table lists the supported vault types:
|
||||
#
|
||||
# | Type | Description | Current Status | Required Fields |
|
||||
# | ----------- | -------------------- | -------------- | --------------------------------------------------- |
|
||||
# | `hashicorp` | Uses Hashicorp Vault | Implemented | `vaultName` (if `create` not true) |
|
||||
# | `azure` | Uses Azure Key Vault | Implemented | `vaultName`, `clientId`, `clientSecret`, `tenantId` |
|
||||
#
|
||||
type: "<Vault Type>"
|
||||
|
||||
# Configurations to create a Hashicorp Vault instance as part of the Helm chart
|
||||
#
|
||||
# THIS IS ONLY RELEVANT IF `type` IS SET TO `hashicorp`
|
||||
create:
|
||||
# If a Hashicorp Vault instance should be created as part of the Helm chart
|
||||
enabled: <true/false>
|
||||
|
||||
# Configurations for the image to use if creating the Hashicorp Vault instance
|
||||
# as part of the Helm chart
|
||||
image:
|
||||
# The repository of the image to use
|
||||
repository: <image repository>
|
||||
|
||||
# The tag of the image to use
|
||||
tag: <image tag>
|
||||
|
||||
# Configurations for the ingress of the created Hashicorp Vault instance
|
||||
ingress:
|
||||
# If an ingress should be created for the created Hashicorp Vault instance
|
||||
enabled: <true/false>
|
||||
|
||||
# The host of the ingress for the created Hashicorp Vault instance
|
||||
host: <DNS Name for vault>
|
||||
|
||||
# As part of the custom Hashicorp Vault image it includes a "snapshot server"
|
||||
# This allows for the triggering and receiving of manual backups ("snapshots") and restoration via HTTP
|
||||
snapshotServer:
|
||||
# If the snapshot server should be enabled/running
|
||||
enabled: true
|
||||
|
||||
# The external port that is opened via the service resource
|
||||
externalPort: 81
|
||||
|
||||
# The internal port that the snapshot server listens on
|
||||
internalPort: 8300
|
||||
|
||||
# Configurations for the storage of the created Hashicorp Vault instance
|
||||
storage:
|
||||
# The storage class to use for the created Hashicorp Vault instance's Persistent Volume Claim
|
||||
storageClass: <Storage Class>
|
||||
|
||||
# The size of the created Hashicorp Vault instance's Persistent Volume Claim
|
||||
size: <Storage Size>
|
||||
|
||||
# The capabilities of the policy to create for the app
|
||||
policyCapabilities:
|
||||
- read
|
||||
- create
|
||||
- update
|
||||
|
||||
# The name of the vault instance to connect to
|
||||
#
|
||||
# This is relevant if type is set to `hashicorp` or `azure`
|
||||
# Note, if `create` is true this is ignored
|
||||
#
|
||||
# For `hashicorp`, this is generally the hostname of the Hashicorp Vault instance to connect to
|
||||
# For `azure`, this is the name of the Azure Key Vault instance to connect to
|
||||
#vaultName: "<Vault Name>"
|
||||
|
||||
# The port of the vault instance to connect to
|
||||
#
|
||||
# ONLY RELEVANT iF `type` IS SET TO `hashicorp` AND `create` IS NOT TRUE
|
||||
#vaultPort: <Vault Port>
|
||||
|
||||
# The client ID of the Azure Key Vault instance
|
||||
#
|
||||
# ONLY RELEVANT IF `type` IS SET TO `azure`
|
||||
#client-id: <Azure Key Vault Client ID>
|
||||
|
||||
# The client secret of the Azure Key Vault instance
|
||||
#
|
||||
# ONLY RELEVANT IF `type` IS SET TO `azure`
|
||||
#client-secret: <Azure Key Vault Client Secret>
|
||||
|
||||
# The tenant ID of the Azure Key Vault instance
|
||||
#
|
||||
# ONLY RELEVANT IF `type` IS SET TO `azure`
|
||||
#tenant-id: <Azure Key Vault Tenant ID>
|
||||
|
||||
# Configuration the NoSQL database
|
||||
# Within the parlance of the system these are often called "properties" databases (and store less structured data)
|
||||
nosql:
|
||||
# Determines the type of NoSQL storage that is used
|
||||
#
|
||||
# The following table lists the possible values for this field:
|
||||
#
|
||||
# | Value | Description |
|
||||
# | --------- | ------------------------------------------------------------------------------------------ |
|
||||
# | `mongodb` | Uses MongoDB as the NoSQL database for the default account properties database |
|
||||
# | `azure` | Uses Azure Table Storage as the NoSQL database for the default account properties database |
|
||||
#
|
||||
type: <NoSQL Storage Type>
|
||||
|
||||
# If to create a resource as part of the deployment process
|
||||
# ONLY relevant if `type` is set to `mongodb`
|
||||
# This uses the [MongoDBCommunity CRD](https://github.com/mongodb/mongodb-kubernetes-operator) to create the resource
|
||||
create: false
|
||||
|
||||
# The number of replicas/members as part of the Mongo deployment
|
||||
# See the `member` parameter of the [MongoDBCommunity CRD](https://github.com/mongodb/mongodb-kubernetes-operator) for more information
|
||||
# ONLY relevant if `type` is set to `mongodb` and `create` is set to `true`
|
||||
#replicaCount: <Number of replicas>
|
||||
|
||||
# The TLS configuration for the connection to the NoSQL database
|
||||
# ONLY relevant if `type` is set to `mongodb` and `create` is set to `true`
|
||||
tls:
|
||||
# If to use TLS for the connection to the NoSQL database
|
||||
enabled: <true/false>
|
||||
|
||||
# The connection string used to access the NoSQL database
|
||||
# ONLY relevant if `type` is set to `mongodb` and `create` is set to `false`
|
||||
# Should be in the following format: `mongodb://<hostname>:<port>`
|
||||
#connectionString: "mongodb://mongo.example.com:27017"
|
||||
|
||||
# The key used to access the NoSQL database
|
||||
# ONLY relevant if `type` is set to `azure`
|
||||
#key: ""
|
||||
|
||||
# The name of the NoSQL database
|
||||
name: "<NoSQL DB Name>"
|
||||
|
||||
# The username used to access the NoSQL database
|
||||
# ONLY relevant if `type` is set to `mongodb`
|
||||
user: "<MongoDB User>"
|
||||
|
||||
# The password used to access the NoSQL database
|
||||
# ONLY relevant if `type` is set to `mongodb`
|
||||
password: "<MongoDB Password>"
|
||||
|
||||
# Configurable NoSQL information groupings
|
||||
# For Azure Table Storage these are table names
|
||||
# For MongoDB these are collection names
|
||||
grouping:
|
||||
<Table Intermediate Name>: "<NoSQL Table Name>"
|
||||
|
||||
# Configuration for cache server
|
||||
cache:
|
||||
type: "redis"
|
||||
|
||||
# If to create a Redis instance/resource as part of the deployment process
|
||||
create: false
|
||||
|
||||
# The image to use for the Redis instance
|
||||
# ONLY relevant if `create` is set to `true`
|
||||
image: {}
|
||||
|
||||
# The number of replicas of the Redis instance
|
||||
# ONLY relevant if `create` is set to `true`
|
||||
#replicaCount: <Number of replicas (Ex. 1)>
|
||||
|
||||
# Hostname of the Redis server
|
||||
# ONLY relevant if `create` is set to `false`
|
||||
#hostName: "<Redis Host Name>"
|
||||
|
||||
# The password to use for the Redis server
|
||||
password: "<Redis Password>"
|
||||
|
||||
# The port of the Redis server
|
||||
port: "<Redis Port>"
|
||||
|
||||
# Redis TLS Configurations
|
||||
tls:
|
||||
# If TLS is enabled for the Redis instance
|
||||
enabled: false
|
||||
|
||||
# The port of the Redis instance for TLS
|
||||
# ONLY relevant if `tls.enabled` is set to `true`
|
||||
#port: "<TLS Port (Ex. 6380)>"
|
||||
|
||||
# Configurations for communication services/relays
|
||||
comms:
|
||||
# Configurations for Email
|
||||
email:
|
||||
# If emailing is enabled
|
||||
#
|
||||
# Note, if `true` then the `type` must be set (along with any required fields for that type)
|
||||
#
|
||||
# Please see the notes on the `type` property below for more information.
|
||||
# Including required fields for each type
|
||||
enabled: true
|
||||
|
||||
# The type/provider to use for sending emails
|
||||
#
|
||||
# The following types are supported:
|
||||
#
|
||||
# | Type | Description | Current Status | Required Fields |
|
||||
# | --------- | --------------------------------------- | -------------- | --------------------------------------- |
|
||||
# | `acs` | Uses Azure Communication Services (ACS) | Implemented | `connectionString` |
|
||||
# | `sendgrid`| Uses SendGrid | To-Do | `apiKey` |
|
||||
# | `smtp` | Uses SMTP | To-Do | `host`, `port`, `username`, `password` |
|
||||
#
|
||||
type: 'acs'
|
||||
|
||||
# The connection string for the Azure Communication Services (ACS) Email service
|
||||
#
|
||||
# THIS IS REQUIRED IF `type` IS SET TO `acs`. Ignored otherwise.
|
||||
connectionString: "<ACS Connection String>"
|
||||
|
||||
# Configurations for integration with third-party services
|
||||
thirdParty:
|
||||
# Configurations for the Stripe integration
|
||||
stripe:
|
||||
# If the integration is enabled
|
||||
enabled: true
|
||||
|
||||
# The public key (live) of the Stripe integration
|
||||
publicKey: <Stripe Live Public Key>
|
||||
|
||||
# The secret key (live) of the Stripe integration
|
||||
secretKey: <Stripe Live Secret Key>
|
||||
|
||||
# The public key (test) of the Stripe integration
|
||||
testPublicKey: <Stripe Test Public Key>
|
||||
|
||||
# The secret key (test) of the Stripe integration
|
||||
testSecretKey: <Stripe Test Secret Key>
|
||||
|
||||
# Configurations for the Moneris integration
|
||||
moneris:
|
||||
# If the integration is enabled
|
||||
enabled: true
|
||||
|
||||
# The Merchant ID (production) for the Moneris integration
|
||||
merchantId: <Moneris Merchant ID>
|
||||
|
||||
# The Store ID (production) for the Moneris integration
|
||||
storeId: <Moneris Store ID>
|
||||
|
||||
# The Hosted Tokenization Profile ID (production) for the Moneris iframe
|
||||
htProfileId: <Moneris Hosted Tokenization Profile ID>
|
||||
|
||||
# The Application/client ID (production) for the Moneris API
|
||||
appId: <Moneris Application ID>
|
||||
|
||||
# The Application/client Secret (production) for the Moneris API
|
||||
appSecret: <Moneris App Secret>
|
||||
|
||||
# The Merchant ID (test) for the Moneris integration
|
||||
testMerchantId: <Moneris Test Merchant ID>
|
||||
|
||||
# The Store ID (test) for the Moneris integration
|
||||
testStoreId: <Moneris Test/QA Store ID>
|
||||
|
||||
# The Hosted Tokenization Profile ID (test) for the Moneris iframe
|
||||
testHtProfileId: <Moneris Test/QA Hosted Tokenization Profile ID>
|
||||
|
||||
# The Application/client ID (test) for the Moneris API
|
||||
testAppId: <Moneris Sandbox/Tes Application Id>
|
||||
|
||||
# The Application/client Secret (test) for the Moneris API
|
||||
testAppSecret: <Moneris Sandbox/Test App Secret>
|
||||
|
||||
# Configuration for the logging sidecar
|
||||
loggingSidecar:
|
||||
# If the logging sidecar should be used (enabled)
|
||||
enabled: true
|
||||
|
||||
# The image used for the logging sidecar
|
||||
image:
|
||||
repository: "<repository>/<Logging Sidecar Image Name>"
|
||||
tag: v1.0.0
|
||||
pullPolicy: IfNotPresent
|
||||
|
||||
# The name of the logging sidecar container
|
||||
name: logging-sidecar
|
||||
|
||||
# The port that the logging sidecar listens on
|
||||
port: 3000
|
||||
|
||||
# Log aggretator authentication details
|
||||
auth:
|
||||
username: "<Log Aggregator (OpenObserve) Username>"
|
||||
password: "<Log Aggregator (OpenObserve) Password>"
|
||||
|
||||
# Configuration for the backup sidecar
|
||||
backupSidecar:
|
||||
# If the backup sidecar should be used (enabled)
|
||||
enabled: true
|
||||
|
||||
# The image used for the backup sidecar
|
||||
image:
|
||||
repository: "<repository>/<Backup Sidecar Image Name>"
|
||||
tag: v1.0.0
|
||||
pullPolicy: IfNotPresent
|
||||
|
||||
# The name of the backup sidecar container
|
||||
name: backup-sidecar
|
||||
|
||||
# The port that the backup sidecar listens on
|
||||
port: 3001
|
||||
Loading…
Add table
Add a link
Reference in a new issue