Skip to content

Helm chart to deploy the AAE infrastructure

Notifications You must be signed in to change notification settings

Alfresco/alfresco-process-infrastructure-deployment

Repository files navigation

alfresco-process-infrastructure

Build Status pre-commit

Helm chart to install the Alfresco Activiti Enterprise (AAE) infrastructure to model and deploy your process applications:

  • Alfresco Identity Service
  • Modeling Service
  • Modeling App
  • Deployment Service
  • Admin App
  • Transformation (Tika) Service

Once installed, you can deploy new AAE applications:

For all the available values, see the chart README.md.

Prerequisites

setup cluster

Setup a Kubernetes cluster following your preferred procedure.

install helm

Install the latest version of helm.

ingress

An ingress-nginx should be installed and bound to an external DNS address, for example:

helm upgrade -i ingress-nginx ingress-nginx \
  --repo https://kubernetes.github.io/ingress-nginx \
  -n ingress-nginx --create-namespace

helm tips

For any command on helm, please verify the output with --dry-run option, then execute without.

To install from the development chart repo, use alfresco-incubator rather than alfresco as CHART_REPO variable.

kubectl tips

Check deployment progress with kubectl get pods -w -A until all containers are running. If anything is stuck, check events with kubectl get events -w -A.

configure installation namespace

export DESIRED_NAMESPACE=${DESIRED_NAMESPACE:-aae}
kubectl create ns $DESIRED_NAMESPACE

add quay-registry-secret

Configure access to pull images from quay.io in the installation namespace:

kubectl create secret \
  -n $DESIRED_NAMESPACE \
  docker-registry quay-registry-secret \
    --docker-server=quay.io \
    --docker-username=$QUAY_USERNAME \
    --docker-password=$QUAY_PASSWORD

where:

  • QUAY_USERNAME is your username on Quay
  • QUAY_PASSWORD is your password on Quay

set main helm env variables

export RELEASE_NAME=aae
export CHART_NAME=alfresco-process-infrastructure
export HELM_OPTS="-n $DESIRED_NAMESPACE"

set environment specific variables

for localhost

A custom extra values file to add settings for localhost is provided:

export DOMAIN=host.docker.internal
HELM_OPTS+=" -f values-localhost.yaml"

Make sure your local cluster has at least 16GB of memory and 8 CPUs. The startup might take as much as 10 minutes, use kubectl get pods -A -w to check the status.

NB if not already present in your /etc/hosts file, please add a DNS mapping from host.docker.internal to 127.0.0.1.

If the hostname host.docker.internal is not resolved correctly on some deployments, patch them after calling helm via:

kubectl patch deployment -n $DESIRED_NAMESPACE ${RELEASE_NAME}-alfresco-modeling-service -p "$(cat deployment-localhost-patch.yaml)"

for a cloud environment

export CLUSTER=aaedev
export DOMAIN=$CLUSTER.envalfresco.com

set helm env variables

HELM_OPTS+=" \
  --set global.gateway.domain=$DOMAIN"

disable alfresco-deployment-service

To disable alfresco-deployment-service in the infrastructure:

HELM_OPTS+="
  --set alfresco-deployment-service.enabled=false
"

Multi-AZ K8S cloud StorageClass for project release

A StorageClass that can work across multiple availability zones need to be available to store project release files per each application:

  • for EKS always use EFS
  • for AKS only if Multi-AZ is configured, use AFS

Add the helm values to use it:

HELM_OPTS+="
  --set alfresco-deployment-service.projectReleaseVolume.storageClass=${STORAGE_CLASS_NAME} \
  --set alfresco-deployment-service.projectReleaseVolume.permission=ReadWriteMany
"

Email Connector

NB In order to set email connector all the variables need to be set. If these variables are set then deployment service will use these configs as default for any applications deployed. Once these variables are configured at the deployment of chart via Helm customer won’t have the possibility to override these values from the admin app. In case you want to configure email connector variable from admin-app please dont not configure email connector during helm deployment.

Add the helm properties to configure email connector:

HELM_OPTS+="
  --set alfresco-deployment-service.applications.connectors.emailConnector.username=${email_connecor_username}
  --set alfresco-deployment-service.applications.connectors.emailConnector.password=${email_connector_password}
  --set alfresco-deployment-service.applications.connectors.emailConnector.host=${email_connector_host}
  --set alfresco-deployment-service.applications.connectors.emailConnector.port=${email_connector_port}
"

debug and dry run

To verify the k8s yaml output:

HELM_OPTS+=" --debug --dry-run"

If all good then launch again without --dry-run.

launch helm

Install from the stable repo using a released chart version:

helm upgrade -i --wait \
  --repo https://kubernetes-charts.alfresco.com/stable \
  $HELM_OPTS $RELEASE_NAME $CHART_NAME

or from the incubator repo for a development chart version:

helm upgrade -i --wait \
  --repo https://kubernetes-charts.alfresco.com/incubator \
  $HELM_OPTS $RELEASE_NAME $CHART_NAME

or from the current repository directory:

helm repo update
helm dependency update helm/$CHART_NAME
helm upgrade -i --wait \
  $HELM_OPTS $RELEASE_NAME helm/$CHART_NAME

Testing

Access IDS

Open browser and login to IDS:

open $SSO_URL

Verify Realm

To read back the realm from the secret, use:

kubectl get secret \
  -n $DESIRED_NAMESPACE \
  realm-secret -o jsonpath="{['data']['alfresco-realm\.json']}" | base64 -D > alfresco-realm.json

Use an internal Docker Registry

In an air gapped environment where the Kubernetes cluster has no direct access to external image repositories, use a tool like helm-image-mirror to tag and push images to your internal registry and modify helm charts with the new image locations.

Use an external PostgreSQL database

Modify the file values-external-postgresql.yaml providing values for your external database per each service, then run:

HELM_OPTS+=" -f values-external-postgresql.yaml"

CI/CD

Running on GH Actions.

For Dependabot PRs to be validated by CI, the label "CI" should be added to the PR.

Requires the following secrets to be set:

Name Description
BOT_GITHUB_TOKEN Token to launch other builds on GH
BOT_GITHUB_USERNAME Username to issue propagation PRs
RANCHER2_URL Rancher URL to perform helm tests
RANCHER2_ACCESS_KEY Rancher access key
RANCHER2_SECRET_KEY Rancher secret key
TEAMS_NOTIFICATION_AUTOMATE_BACKEND_WEBHOOK Token to notify Teams on failure