Skip to content

g3rzi/CKA

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

45 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

CKA - Summary for the exam

General commands

kubectl run --dry-run=false --restart=Never -o yaml --image nginx nginx -> will print the object without sending it and only pod will be created (not deployment).
--restart=Never => will create Pod, not Deployment.

Creating role and rolebindings for a user:

kubectl create role role-john -n development  --verb=create,list,get,update,delete --resource=pods 
kubectl create rolebinding binding-john -n development --role=role-john --user=john

Creating user in kubeconfig:

kubectl config --kubeconfig=config-demo set-credentials john --client-certificate=/root/john.crt --client-key=/root/john.key
kubectl config --kubeconfig=config-demo set-context john-ctx --cluster=kubernetes --namespace=development

Create object without writing it to file:

kubectl apply -f - <<EOF
apiVersion: v1
kind: Pod
metadata:
  name: super-user-pod
spec:
  containers:
  - name: redis
    image: busybox:1.28
    securityContext:
      capabilities:
        add: ["SYS_TIME"]
EOF

Sections

Section 3: Scheduling

47. Taints and Tolerations

kubectl taint nodes node-name key=value:taint-effect

The taint-effect is what happens to PODs that DO NOT TOLERATE this taint. The options are: NoSchedule -> Pod won't schedule PreferNoSchedule -> The system will try to avoid placing a pod on the node but that is not guaranteed. NoExecute -> New pods will not be scheduled on the node and existing pods on the node if any will be evicted if they do not tolearate the taint.

Example: kubectl taint nodes node1 app=blue:NoSchedule

apiVersion:
kind: Pod
metadata:
  name: myapp-pod
spec:
  containers:
  - name: nginx-container
    image: nginx
  tolerations:
  - key: app
    operator: "Equal"
    value: blue
    effect: NoSchedule

To remove taint from the master run: kubectl taint nodes master node-role.kubernetes.io/master:NoSchedule-.
Where "node-role.kubernetes.io/master" is the key and "NoSchedule" is the effect.

48. Node Selectors

Two ways to assign pod to nodes:

  1. Node Selectors
    Assign lable to node:
    kubectl label nodse <node-name> <label-key>=<label-value>
    Example: kubectl lable nodes node-1 size=Large

    Create the pod with nodeSelector field like that:

    apiVersion: v1
    kind: Pod
    metadata:
      name: myapp-pod
    spec:
      nodeSelector:
        kubernetes.io/hostname: node03
      containers:
      - image: data-processor
        name: data-processor
    

    Good method but sometimes you need more complexed assignments like: Large OR Medium, Not Small, etc.
    For this, you will need to use Node Affinity.

  2. Node Affinity Node affinity allows pod placement on specific nodes.
    This configuration allows pods to run on nodes with the labels "size=Large" or "size=Medium". The In operator make sure the pod will be places on a node whose label size has any value in the list of values specified.

apiVersion: v1
kind: Pod
metadata:
  name: myapp-pod
spec:
  containers:
  - image: data-processor
    name: data-processor
affinity:
  nodeAffinity:
    requireDuringSchedulingIngnoredDuringExecution:
      nodeSelectorTerms:
      - matchExpressions:
        - key: size
          operator: In
          values:
          - Large
          - Medium

You can also use the NotIn operator like that:

        - key: size
          operator: NotIn
          values:
          - Small

Section 5: Application Lifecycle Management

77. Rolling Updates and Rollbacks

To states of your rollout run:
kubectl rollout status deployment/myapp-deployment

To see the history of the rollout run:
kubectl rollout history deployment/myapp-deployment
The problem with that for some period the application is down and no one can access it.

Two rollout of deployment startegy:

  1. Recreate startegy (not default)
    Destroy all the deployments and then create them.

  2. RollingUpdate
    Take down older version and create new one one by one.

To update image you can use the kubectl apply -f file.yaml command or use kubectl set image deployment/myapp-deployment nginx=nginx:1.9.1. The latter will result change in the deployment definition file.

When running kubectl describe deployment myapp you will notice a difference between the two startegies. In the events you will see how the "recreate" take all the applications down and then up.
You have the field named StartegyType where you will the name of the strategy.

When you upgrage your application, Kubernetes create new replica set with the same amount of pods you had and create them with the new version and take down the ones with the old image.

If you need to rollback to the previous revious you can run kubectl rollout undo deployment/myapp-deployment.

Summarize commands:

Section 6: Cluster Maintenance

100. Practice Test - OS Upgrades

To mark node as unschedulable and remove any pods from it run:
kubectl drain node01 --ignore-daemonsets

One reason if it failed is because there is an unmanaged Pod, then you need to use the --force flag.
To make the node scheduled again run:
kubectl uncordon node01

To just make the node as unschedulable without removing any workload run:
kubectl cordon node01

105. Backup and Restore Methods

When configure ETCD, it is possible to configure where all the data will be store.
It appears in the etcd.service (which can be seen by cat /etc/kubernetes/manifests/etcd.yaml) under the flag --data-dir=/var/lib/etcd.

We can take a snapshot of the database by running:
ETCDCTL_API=3 etcdctl snapshot save snapshot.db

Remember to use it will all the details:

ETCDCTL_API=3 etcdctl snapshot save snapshot.db \
              --endpoints=https://127.0.0.1:2379
	      --cacert=/etc/etcd/ca.crt \
	      --cert=/etc/etcd/etcd-server.crt \
	      --key=/etc/etcd/etcd-server.key

To restore the backup:
Stop the service:
service kube-apiserver stop
Then run:

ETCDCTL_API=3 etcdctl snapshot restore snapshot.db \
--data-dir /var/lib/etcd-from-backup \
--initial-cluster master-1=https://192.168.5.11:2380,master-2=https://192.168.5.12:2380 \
--initial-advertise-peer-urls https://${INTERNAL_IP}:2380

Then we will update etcd.service file with:

--initial-cluster-token etcd-cluster-1 \
...
--data-dir=/var/lib/etcd-from-backup

Then restart the daemon:

systemctl daemon-reload
service etcd restart
service kube-apiserver start

Section 7: Security

openssl x509 -req -in /etc/kubernetes/pki/apiserver-etcd-client.csr \
-CA /etc/kubernetes/pki/etcd/ca.crt \
-CAkey /etc/kubernetes/pki/etcd/ca.key \
-CAcreateserial \
-out /etc/kubernetes/pki/apiserver-etcd-client.crt \

view certificate

openssl x509 -in <certificate path> -text -noout

116. Certificates API

openssl genrsa -out jane.key 2048
openssl req -new -key jane.key -subj "/CN=jane" -out jane.csr
Creating jane-csr.yaml:
kind: CertificateSigningRequest
...

kubectl get csr
kubectl certificate approve jane

117. TLS in Kubernetes - Certificate Creation

Certificate Authority (CA):

# Generate private key ca.key
openssl genrsa -out ca.key 2048
# Generate CertificateSigningRequest ca.csr
openssl req -new -key ca.key -subj "/CN=Kubernetes-CA" -out ca.csr
# Sign certificate  
openssl x509 -req -in ca.csr -signkey ca.key -out ca.crt

Admin user:

# Generate private key admin.key
openssl genrsa -out ca.key 2048
# Generate CertificateSigningRequest admin.csr
openssl req -new -key admin.key -subj "/CN=kube-admin/O=system:masters" -out admin.csr
# Sign certificate  
# Notice that we are signing with -CAkey
openssl x509 -req -in admin.csr -CA ca.crt -CAkey ca.key -out admin.crt

121. Certificates API

Instead of the user sign the certificate by himself, he can use Kubernetes API to create is and reviewed by the administrators:

  1. Create CertificateSigningRequest Object
  2. Review Requests
  3. Approve Requests
  4. Share Certs to Users

Create private key with CSR:

openssl genrsa -out jane.key 2048
openssl req -new -key jane.key -subj "/CN=jane" -out jane.csr

Create CertificateSigningRequest Object:

apiVersion: certificates.k8s.io/v1beta1
kind: CertificateSigningRequest
metadata:
  name: jane
spec:
  groups:
  - system:authenticated
  usages:
  - digital signature
  - key encipherment
  -server auth
  request: 
    # cat jane.csr | base64
    <base64_of_jane.csr_content_WITHOUT_'\n'>

Another option is like that:

kubectl apply -f - <<EOF
apiVersion: certificates.k8s.io/v1beta1
kind: CertificateSigningRequest
metadata:
  name: jane
spec:
  groups:
  - system:authenticated
  usages:
  - digital signature
  - key encipherment
  - server auth
  request: $(cat jane.csr | base64 | tr -d '\n')
EOF

Administrators can see the pending certificates and approve it:

# View request
kubectl get csr
# Approve request
kubectl certificate approve jane

All the operations of the certificates are done by the Controller Manager. It as controllers in it called "CSR-APPROVING" and "CSR-SIGNING".

Network:

159 & 160. CNI weave

cat /etc/cni/net.d/ -> check the config file for the network provider
ip addr show weave -> show ip addresses of the provider

161. Practice Test Deploy Network Solution

kubectl apply -f "https://cloud.weave.works/k8s/net?k8s-version=$(kubectl version | base64 | tr -d '\n')" -> Install weave-network

163. Service Networking

kubectl get pods -o wide
kubectl get service
# Showing the rules from the service to the pod's port.
iptables -L -t net | grep db-service
# Showing what proxy is being used
cat /var/log/kube-proxy.log

164. Practice Test - Service Networking

Check the range of IP addresses configured for PODs on the cluster:
The network is configured with weave. Check the weave pods logs using command kubectl logs -n kube-system <weave-pod-name> weave and look for ipalloc-range

Check the IP range configured for the services within the cluster:
Inspect the setting on kube-api server by running on command ps -aux | grep kube-api You should find something like that:
--service-cluster-ip-range=10.96.0.0/12

Checl what type of proxy the kube-proxy configured to use:
Check the logs of the kube-proxy pods. Command: kubectl logs -n kube-system <kube-proxy pod name> Most of the time it will be iptables.

166. CoreDNS in Kubernetes

Configuration file of the DNS:
cat /etc/coredns/Corefile
This file can be seen using:
kubectl get configmap -n kube-system

The DNS also creates a service named: kube-dns to be available through the cluster. We can see the DNS configuration in the kubelet configuration file:
/var/lib/kubelet/config.yaml

167. Practice Test - Explore DNS

Check what dns domain/zone configured on the cluster run: kubectl describe configmaps coredns -n kube-system

168. Ingress

Ingress Controller

#Deployment of the nginx-ingress image
apiVersion: apps/v1
kind: Deployment
metadata:
  name: nginx-ingress-controller
spec:
  replicas: 1
  selector:
    matchLabels:
      name: nginx-ingress
    template:
      metadata:
        labels:
          name: nginx-ingress
    spec:
      containers:
        - name: nginx-ingress-controller
          image: quay.io/kubernetes-ingress-controller/nginx-ingress-controller:0.21.0
        args:
          - /nginx-ingress-controller
          - --configmap=$(POD_NAMESPACE)/nginx-coinfguration
        env:
          - name: POD_NAME
            valueFrom:
              fieldRef:
                fieldPath: metadata.name
          - name: POD_NAMESPACE
            valueFrom:
              fieldRef:
                fieldPath: metadata.namespace
        ports:
          - name: http
            containerPort: 80
          - name: https
            containerPort: 443
# ConfiguMap to feed nginx configuration data
apiVersion: v1
kind: ConfigMap
metadata:
  name: nginx-coinfguration
# Service to expose the deployment
apiVersion: v1
kind: Service
metadata:
  name: nginx-ingress
spec:
  type: NodePort
  ports:
  - port: 80
    targetPort: 80
    protocol: TCP
    name: http
  - port: 443
    targetPort: 443
    protocol: TCP
    name: https
  selector:
    name: nginx-ingress
# The service account should have Roles, ClusterRoles and RoleBindings to access the above objects
apiVersion: v1
kind: ServiceAccount
metadata:
  name: nginx-ingress-serviceaccount

Ingress Resource

Ingress-wear.yaml

apiVersion: extensions/v1beta1
kind: Ingress
metdata:
  name: ingress-wear
spec:
  backend:
    serviceName: wear-service
	servicePort: 80

Rules

First method: Split traffic by URL (one rule)
apiVersion: extensions/v1beta1
kind: Ingress
metdata:
  name: ingress-wear-watch
spec:
  rules:
  - http:
      paths:
      - path: /wear
        backend:
          serviceName: wear-service
          servicePort: 80
      - path: /watch
        backend:
          serviceName: watch-service
          servicePort: 80  	
Second method: Using domain names (two rules)
apiVersion: extensions/v1beta1
kind: Ingress
metdata:
  name: ingress-wear-watch
spec:
  rules:
  - host: wear.my-online-store.com
    http:
      paths:
      - backend:
          serviceName: wear-service
          servicePort: 80
    - host: watch.my-online-store.com
      http:
        paths:
        - backend: 
            serviceName: watch-service
            servicePort: 80   

kubectl describe ingress ingress-wear-watch The Default backend is for the case where a user tries to access a URL that does not match any of the rules. The user is directed to the service specified as the default backend. You must deploy the specified service.

217. Practice Test - Worker Node Failure

Check kubelet status on the nodes:
journalctl -u kubelet
service kubelet status

Related files to kubelet:

/etc/kubernetes/kubelet.conf  
/etc/systemd/system/kubelet.service.d/10-kubeadm.conf  
/var/lib/kubelet/config.yaml -> IMPORTANT, contains kubelet settings  

222. Practice Test - Advanced Kubetl Commands

Find user aws-user inside a context:
kubectl config view --kubeconfig=my-kube-config -o jsonpath="{.contexts[?(@.context.user=='aws-user')].name}" > /opt/outputs/aws-context-name

Section 15: Mock Exams

225. Mock Exam - 2

  1. Create a new deployment called nginx-deploy, with image nginx:1.16 and 1 replica. Record the version. Next upgrade the deployment to version 1.17 using rolling update. Make sure that the version upgrade is recorded in the resource annotation. To solve it we will need to use --record option when create the deployment. To update version we can do it with kubectl edit or kubectl set image.

View the history and update the image:

kubectl rollout history deployment nginx-deploy
kubectl set image deployment/nginx-deploy nginx-deploy=nginx:1.17 --record

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published