# goto directory for this lab
cd ../module05
Helm is template engine (deployment engine) for kubernetes.
Please change IP address of your ingress controller and name of your Azure Container Registry in helm upgrade
command below.
# create namespace
kubectl create namespace myapp
# create secrets
POSTGRESQL_URL="jdbc:postgresql://${POSTGRESQL_NAME}.postgres.database.azure.com:5432/todo?user=${POSTGRESQL_USER}@${POSTGRESQL_NAME}&password=${POSTGRESQL_PASSWORD}&ssl=true"
kubectl create secret generic myrelease-myapp \
--from-literal=postgresqlurl="$POSTGRESQL_URL" \
--namespace myapp
# add helm charts from local to ACR repo
az configure --defaults acr=${ACR_NAME}
# get access token for helm (similar to docker login)
az acr helm repo add
# pack helm repo
helm package helm/myapp
# push repo to ACR(helm)
az acr helm push myapp-0.1.0.tgz
# list repos (two examples)
az acr helm list
helm update
helm search ${ACR_NAME}
# Get ingress public IP
export INGRESS_IP=$(kubectl get svc ingress-nginx-ingress-controller -o=custom-columns=EXTERNAL-IP:.status.loadBalancer.ingress[*].ip | grep -v "EXTERNAL-IP")
echo "You will be able to access application on this URL: http://${INGRESS_IP}.xip.io"
# deploy from ACR helm repository
helm upgrade --install myrelease ${ACR_NAME}/myapp --namespace='myapp' --set-string appspa.image.repository="${ACR_NAME}.azurecr.io/myappspa",appspa.image.tag='v1',apptodo.image.repository="${ACR_NAME}.azurecr.io/myapptodo",apptodo.image.tag='v1',apphost="${INGRESS_IP}.xip.io"
# clean-up deployment
helm delete --purge myrelease
# delete namespace
kubectl delete namespace myapp
This CI/CD demo contains CI simple pipeline in Azure Container Registry and CD pipeline in FLUX (git based delivery system - https://github.com/weaveworks/flux ).
There we will define two build tasks - for building SPA web GUI and TODO microservice.
# set default ACR name
az configure --defaults acr=${ACR_NAME}
# build manualy / last parameter of command is your forked github repo
az acr run -f module05/acr-flux/myapp-ci.yaml https://github.com/valda-z/java-k8s-workshop.git
Now let's prepare task triggered by github commit. We will need access token for this task, there is description: https://help.github.com/articles/creating-a-personal-access-token-for-the-command-line/ .
# now prepare task triggered by github commit
az acr task create -n myapp -f module05/acr-flux/myapp-ci.yaml \
-c https://github.com/valda-z/java-k8s-workshop.git \
--pull-request-trigger-enabled true \
--git-access-token 0000000000000000000000000000000000000000
# list tasks
az acr task list -o table
# list run history
az acr task list-runs -o table
Now we can create small commit in our repo and check via az acr task list-runs -o table
if build was triggered and status of build.
There is space for some homework and creative solution. You can use Azure LogicApp or any other technique to handle webhooks from ACR to trigger CD tasks (by commit changes to github repo with flux configuration).
Fork our base FLUX repo: https://github.com/azurecz/java-k8s-workshop-flux and setup configuration files in repo.
Yaml files for deployment are stored in folder wokloads
.
In bash change your working directory now to java-k8s-workshop-flux
.
# Change yaml files to your ACR name
sed -i 's/YOURACRNAME/'$ACR_NAME'/g' workloads/*.yaml
# Change yaml files to your ingress public IP
export INGRESS_IP=$(kubectl get svc ingress-nginx-ingress-controller -o=custom-columns=EXTERNAL-IP:.status.loadBalancer.ingress[*].ip | grep -v "EXTERNAL-IP")
echo "You will be able to access application on this URL: http://${INGRESS_IP}.xip.io"
# Change YAML files for ingress
sed -i 's/YOURINGRESSIP/'$INGRESS_IP'/g' workloads/*.yaml
# and finally you have to commit changes
git add *
git commit -m "configuration changed"
git push
Now let's install flux to AKS cluster. https://github.com/weaveworks/flux/blob/master/site/helm-get-started.md#install-weave-flux
# add flux repo
helm repo add weaveworks https://weaveworks.github.io/flux
# apply CRD
kubectl apply -f https://raw.githubusercontent.com/weaveworks/flux/master/deploy-helm/flux-helm-release-crd.yaml
# install and configure FLUX
helm upgrade -i flux \
--set helmOperator.create=true \
--set helmOperator.createCRD=false \
--set git.url=git@github.com:valda-z/java-k8s-workshop-flux.git \
--namespace flux \
weaveworks/flux
See description there: https://github.com/weaveworks/flux/blob/master/site/helm-get-started.md#giving-write-access
# get flux public key
fluxctl identity --k8s-fwd-ns flux
In order to sync your cluster state with git you need to copy the public key and create a deploy key with write access on your GitHub repository.
Open GitHub, navigate to your fork, go to Setting > Deploy keys, click on Add deploy key, give it a name, check Allow write access, paste the Flux public key and click Add key.
Once Flux has confirmed access to the repository, it will start deploying the workloads of java-k8s-workshop-flux.
# check pods
kubectl get pods --all-namespaces
Now we can see that myapptodo POD is failing because we have no secrets with connection string to DB.
# create secret for myapptodo
POSTGRESQL_URL="jdbc:postgresql://${POSTGRESQL_NAME}.postgres.database.azure.com:5432/todo?user=${POSTGRESQL_USER}@${POSTGRESQL_NAME}&password=${POSTGRESQL_PASSWORD}&ssl=true"
kubectl create secret generic myapptodo-secret \
--from-literal=postgresqlurl="$POSTGRESQL_URL" \
--namespace myapp
Now you can access application on this URL:
# Get ingress Public IP
export INGRESS_IP=$(kubectl get svc ingress-nginx-ingress-controller -o=custom-columns=EXTERNAL-IP:.status.loadBalancer.ingress[*].ip | grep -v "EXTERNAL-IP")
echo "You will be able to access application on this URL: http://${INGRESS_IP}.xip.io"
Finally you can change some configuration in flux configuration github repo and observe how flux will deploy changes.
# check flux logs
kubectl -n flux logs deployment/flux -f
How it works?
Jenkis is deployed directly to AKS cluster, master jenkins container exposes user interface on public IP address on port 8080. Jenkins build agents are deployed ad-hoc when build job needs agent and is destroyed just after build.
Build pipeline is in Jenkinsfile which is part of source code tree in github. Build pipeline consists these steps:
- clone source codes from git
- build docker images (and applications)
- push images to Azure Container Registry
- deploy new version via helm to AKS cluster
# install jenkins (from repository java-k8s-workshop, directory module05)
cd module05
# Get ingress Public IP
export INGRESS_IP=$(kubectl get svc ingress-nginx-ingress-controller -o=custom-columns=EXTERNAL-IP:.status.loadBalancer.ingress[*].ip | grep -v "EXTERNAL-IP")
# postgres URL
POSTGRESQL_URL="jdbc:postgresql://${POSTGRESQL_NAME}.postgres.database.azure.com:5432/todo?user=${POSTGRESQL_USER}@${POSTGRESQL_NAME}&password=${POSTGRESQL_PASSWORD}&ssl=true"
# install jenkins - please put your github repository name to --giturl parameter
./jenkins/deploy.sh \
--ingressdns "${INGRESS_IP}.xip.io" \
--postgresjdbcurl "${POSTGRESQL_URL}" \
--acrname "${ACR_NAME}" \
--acrkey "${ACR_KEY}" \
--giturl "https://github.com/azurecz/java-k8s-workshop.git"
You have to have microsoft account or MSDN license
Open browser on https://dev.azure.com
1. Create a project
2 Import a repo from github for simplicity we use only master branch
3 Create service connection for ACR and AKS
4 Create variable group (Library) and needed variables
5 You have to link variables to concrete pipeline --> edit --> variables --> variable groups
6 Create a build pipeline
trigger:
- master
#runinng on MS hosted images
pool:
vmImage: 'Ubuntu-16.04'
# available variables from Global
# KeyVault secrets must be linked
variables:
- group: BaseVariables
steps:
- bash: |
echo $(acr-name-demo)
# myapptodo
cd module01/src/myapptodo
docker build -t $(acr-name-demo)/myapptodo:latest . # should be tagged $(Build.BuildId) or ReleaseId
docker login -u $(acr-name-demo-user) -p $(acr-psw) $(acr-name-demo)
docker push $(acr-name-demo)/myapptodo:latest
# myappspa
cd module01/src/myappspa
docker build -t $(acr-name-demo)/myappspa:latest . # $(Build.BuildId)
docker push $(acr-name-demo)/myappspa:latest
displayName: 'Build, tag and push image'
- task: CopyFiles@2
inputs:
sourceFolder: '$(Build.SourcesDirectory)/module05/helm'
#contents: '**'
targetFolder: '$(Build.ArtifactStagingDirectory)'
#cleanTargetFolder: false # Optional
overWrite: true
#flattenFolders: false # Optional
- task: PublishBuildArtifacts@1
inputs:
pathtoPublish: '$(Build.ArtifactStagingDirectory)'
artifactName: 'drop'
1 Create a stage (currently only Deploy)
2 Choose Deploy app to AKS via its Helm chart
3 Install helm (2.12.2)
4 helm init with arg --service-account tiller --upgrade
5 helm deploy (set chart path and --install param)
- task: Docker@1
displayName: Build image
inputs:
command: build
azureSubscriptionEndpoint: $(azureSubscriptionEndpoint)
azureContainerRegistry: $(azureContainerRegistry)
dockerFile: Dockerfile
imageName: $(Build.Repository.Name)
- task: Docker@1
displayName: Tag image
inputs:
command: tag
azureSubscriptionEndpoint: $(azureSubscriptionEndpoint)
azureContainerRegistry: $(azureContainerRegistry)
imageName: $(azureContainerRegistry)/$(Build.Repository.Name):latest
arguments: $(azureContainerRegistry)/$(Build.Repository.Name):$(Build.BuildId)
- task: Docker@1
displayName: Push image
inputs:
command: push
azureSubscriptionEndpoint: $(azureSubscriptionEndpoint)
azureContainerRegistry: $(azureContainerRegistry)
imageName: $(Build.Repository.Name):$(Build.BuildId)