In this course we will manage some basic concepts about container orchestation tool.
Minikube is local Kubernetes, focusing on making it easy to learn and develop for Kubernetes. All you need is Docker (or similarly compatible) container or a Virtual Machine environment
To install the latest minikube stable release on x86-64 Linux using binary download. Check the following link on case of diferent machine architecture https://minikube.sigs.k8s.io/docs/start/
curl -LO https://storage.googleapis.com/minikube/releases/latest/minikube-linux-amd64
sudo install minikube-linux-amd64 /usr/local/bin/minikube
- 2 CPUs or more
- 2GB of free memory
- 20GB of free disk space
- Internet connection
- Container or virtual machine manager, such as: Docker, QEMU, Hyperkit, Hyper-V, KVM, Parallels, Podman, VirtualBox, or VMware Fusion/Workstation
minikube start
kubectl delete all --all
## From an especific namespace
kubectl delete all --all -n namespace
Prometheus server is responsible for send alerts of monitoring to the alert manager and this one is in charge to alert to our different integrations suach send email or slack etc.
The first file is our namespace that are a way to organize clusters into virtual sub-clusters. (1-monitoring/namespaces.yaml)
kubectl appliy -f namespace.yaml
# namespace/monitoring created β
Then we must create some role or permissions in that monitoring subcluster that we have already created. This with the objective that prometheus can access to system files such as extract metrics.
kubectl apply -f prometheus-cluster-role.yaml
#clusterrole.rbac.authorization.k8s.io/prometheus created β
#clusterrolebinding.rbac.authorization.k8s.io/prometheus created β
Also, we need to deploy our prometheus pod server that it will create our container instance base on the following image prom/prometheus:v2.1.0 in monitoring
namespace and it will attach some config and storage volume in the root cluster.
kubectl apply -f prometheus-cluster-role.yaml
# deployment.apps/prometheus-deployment created β
Check what we have at this moment:
kubectl -n monitoring get all
NAME | READY | STATUS | RESTARTS | AGE |
---|---|---|---|---|
pod/prometheus-deployment-5875b79765-nzpjp | 0/1 | ContainerCreating | 0 | 2m31s |
deployment.apps/prometheus-deployment | 0/1 | 1 | 0 | 2m31s |
replicaset.apps/prometheus-deployment-5875b79765 | 1 | 1 | 0 | 2m31s |
Config maps are manifiest that you inject to an specific namespace and then whatever container running on this ns can read those configuration maps and generated its configuration base on that template (Global Enviroment Namespaces)
Basic Check of Memory Use
rules:
- alert: High Pod Memory
expr: sum(container_memory_usage_bytes) > 1
for: 1m
labels:
severity: critical
annotations:
summary: High Memory Usage
Applying Map Configuration:
kubectl apply -f prometheus-configmap.yaml
# configmap/prometheus-server-conf created β
Creation of a service to expose our prometheus pods using a node port that basically open a port in the workers nodes that redirects the traffic to the pod with specific label or selector.
kubectl apply -f prometheus-service.yaml
# service/prometheus-service created β
-
Service creation (Cluster Ip because we want that prometheus can access to that container assigning an specific ip in the cluster)
kubectl apply -f alert-manager-service.yaml #service/alertmanager created β
-
Deployment in this case we have a pod that will have two containers, the firts one is the alert manager and the second one a side car that will be watching for changes in global config map. Also, we have a Persistent Volume Claim that is a request for storage by a user. It is similar to a Pod. Pods consume node resources and PVCs consume PV resources, claims can request specific size and access modes.
kubectl apply -f alert-manager-configmap.yaml # service/alertmanager created β kubectl apply -f alert-manager-pvc.yaml # persistentvolumeclaim/alertmanager created β kubectl apply -f alert-manager-deployment.yaml # deployment.apps/alertmanager created β
- Pods
NAME | READY | STATUS | RESTARTS | AGE |
---|---|---|---|---|
pod/alertmanager-85bc4f4759-s25c5 | 2/2 | Running | 0 | 23m |
pod/prometheus-deployment-5875b79765-nzpjp | 1/1 | Running | 0 | 91m |
- Services
NAME | TYPE | CLUSTER-IP | EXTERNAL-IP | PORT(S) | AGE |
---|---|---|---|---|---|
service/alertmanager | ClusterIP | 10.110.82.83 | 9093/TCP | 59m | |
service/prometheus-service | NodePort | 10.100.0.75 | 8080:30000/TCP | 79m |
- Deployments
NAME | READY | UP-TO-DATE | AVAILABLE | AGE |
---|---|---|---|---|
deployment.apps/alertmanager | 1/1 | 1 | 1 | 23m |
deployment.apps/prometheus-deployment | 1/1 | 1 | 1 | 91m |
Due to we are using minikube for our local kubernetes cluster we just have one node and for accesing externally we must use the following command:
minikube -n monitoring service prometheus-service --url
#Β http://192.168.49.2:30000 β οΈ
However i have my minikube configuration inside an ec2 vm hosted in aws cloud provider π . That is why i must port forward my vm local entry set in 31000 port and the service port 8080 of the prometheus node port:
kubectl port-forward --address 0.0.0.0 service/prometheus-service --namespace monitoring 31000:8080
# Forwarding from 0.0.0.0:31000 -> 9090
# Handling connection for 31000
It's a package manager that makes it possible to download charts, which are pre-packaged collections of all the necessary versioned, pre-configured resources required to deploy a container. Helm charts are written in YAML and contain everything your developers need to deploy a container to a Kubernetes cluster.
sudo snap install helm --classic
Helm chart creation : chart_name
helm create varelachart
Debugging the manifiests that helm will install in our cluster
helm install --dry-run debug .
Some linter guardians or constraints in our yamls files
helm lint
Creating a package of this helm chart due to we need other repositorios can use it.
helm package varelachart/
mv ./varelachart-0.1.0.tgz charts/
Adding your custom configurations in values.yaml file.
healthcheck:
readinessProbe:
path: /
port: http
livenessProbe:
path: /
port: http
path: {{.Values.healthcheck.readinessProbe.path}}
port: {{.Values.healthcheck.readinessProbe.path}}
Creation of the index file use for helm repository to know what charts are inside this package. It will look for all tgz inside the current folder and map ot into this index entrypoint.
helm repo index .
Push your changes to your github or distributed version management and for using repository:
helm repo add varela https://alejandro945.github.io/kubernetes/2-helm
Validation of repository creation
helm search repo varela
Installation on your kubernetes cluster
helm install test varela/varelachart
Delete a helm chart or list them
helm list
helm delete test
Loki is a log database developed by Grafana Labs. In this repo we will deploy Loki on a Kubernetes cluster, and we will use it to monitor the log of our pods.
kubectl apply -f app.yaml
- Port forward service from 30000:80
kubectl port-forward --address 0.0.0.0 service/hello-kubernetes 30000:80
Grafana is an open source interactive data-visualization platform, developed by Grafana Labs, which allows users to see their data via charts and graphs that are unified into one dashboard (or multiple dashboards!) for easier interpretation and understanding.
- Apply yaml files using kustomize
kubectl apply -k .
- Port forward service from 31000:80
kubectl port-forward --address 0.0.0.0 service/grafana-service 31000:80
- Authentication credentials
Username: admin Password: admin
Definition: Loki is a log aggregation system designed to store and query logs from all your applications and infrastructure.
Another concept to have in mind is a fully qualified domain name (FQDN) that is the complete address of an internet host or computer. It provides its exact location within the domain name system (DNS) by specifying the hostname, domain name and top-level domain (TLD).
Contextualization: Loki is a great database that is easy to create and the only thing we will do is describe that it will need a volume so that it is persistent between service drops (pod)
- Apply all YAML files in order and wait
kubectl apply -k .
- Remember the FQDN
http://loki-service.kube-logging.svc.cluster.local:3100
Service capable of monitoring the system logs and processing it in the way that we tell it in this case for loki, we will use daemon set to deploy a copy of fluentd in each of the nodes to centralize this information in the loki database to view them in grafana.
Using configmap to grab the logs and insert them into loki, for this a fluentd configuration is injected into its container, here we tell you where those logs are, how we parse that information (automation core)
- Apply all YAML files in order and wait
kubectl apply -k .