8.2 KiB
title | description | lead | date | draft | images | menu | weight | toc |
---|---|---|---|---|---|---|---|---|
Deploy to kubeadm | Deploy kube-prometheus to Kubernets kubeadm. | Deploy kube-prometheus to Kubernets kubeadm. | 2021-03-08T23:04:32+01:00 | false | [] | [{docs [{parent kube}]}] | 500 | true |
The kubeadm tool is linked by Kubernetes as the offical way to deploy and manage self-hosted clusters. kubeadm does a lot of heavy lifting by automatically configuring your Kubernetes cluster with some common options. This guide is intended to show you how to deploy Prometheus, Prometheus Operator and Kube Prometheus to get you started monitoring your cluster that was deployed with kubeadm.
This guide assumes you have a basic understanding of how to use the functionality the Prometheus Operator implements. If you haven't yet, we recommend reading through the getting started guide as well as the alerting guide.
kubeadm Pre-requisites
This guide assumes you have some familiarity with kubeadm
or at least have deployed a cluster using kubeadm
. By default, kubeadm
does not expose two of the services that we will be monitoring. Therefore, in order to get the most out of the kube-prometheus
package, we need to make some quick tweaks to the Kubernetes cluster. Since we will be monitoring the kube-controller-manager
and kube-scheduler
, we must expose them to the cluster.
By default, kubeadm
runs these pods on your master and bound to 127.0.0.1
. There are a couple of ways to change this. The recommended way to change these features is to use the kubeadm config file. An example configuration file can be used:
apiVersion: kubeadm.k8s.io/v1alpha1
kind: MasterConfiguration
api:
advertiseAddress: 192.168.1.173
bindPort: 6443
authorizationModes:
- Node
- RBAC
certificatesDir: /etc/kubernetes/pki
cloudProvider:
etcd:
dataDir: /var/lib/etcd
endpoints: null
imageRepository: gcr.io/google_containers
kubernetesVersion: v1.8.3
networking:
dnsDomain: cluster.local
serviceSubnet: 10.96.0.0/12
nodeName: your-dev
tokenTTL: 24h0m0s
controllerManagerExtraArgs:
address: 0.0.0.0
schedulerExtraArgs:
address: 0.0.0.0
Notice the schedulerExtraArgs
and controllerManagerExtraArgs
. This exposes the kube-controller-manager
and kube-scheduler
services to the rest of the cluster. If you have kubernetes core components as pods in the kube-system namespace, ensure that the kube-prometheus-exporter-kube-scheduler
and kube-prometheus-exporter-kube-controller-manager
services' spec.selector
values match those of pods.
In addition, we will be using node-exporter
to monitor the cAdvisor
service on all the nodes. This, however requires a change to the kubelet
service on the master as well as all the nodes. According to the Kubernetes documentation
The kubeadm deb package ships with configuration for how the kubelet should be run. Note that the
kubeadm
CLI command will never touch this drop-in file. This drop-in file belongs to the kubeadm deb/rpm package.
Again, we need to expose the cadvisor
that is installed and managed by the kubelet
daemon and allow webhook token authentication. To do so, we do the following on all the masters and nodes:
KUBEADM_SYSTEMD_CONF=/etc/systemd/system/kubelet.service.d/10-kubeadm.conf
sed -e "/cadvisor-port=0/d" -i "$KUBEADM_SYSTEMD_CONF"
if ! grep -q "authentication-token-webhook=true" "$KUBEADM_SYSTEMD_CONF"; then
sed -e "s/--authorization-mode=Webhook/--authentication-token-webhook=true --authorization-mode=Webhook/" -i "$KUBEADM_SYSTEMD_CONF"
fi
systemctl daemon-reload
systemctl restart kubelet
In case you already have a Kubernetes deployed with kubeadm, change the address kube-controller-manager and kube-scheduler listens in addition to previous kubelet change:
sed -e "s/- --address=127.0.0.1/- --address=0.0.0.0/" -i /etc/kubernetes/manifests/kube-controller-manager.yaml
sed -e "s/- --address=127.0.0.1/- --address=0.0.0.0/" -i /etc/kubernetes/manifests/kube-scheduler.yaml
With these changes, your Kubernetes cluster is ready.
Metric Sources
Monitoring a Kubernetes cluster with Prometheus is a natural choice as Kubernetes components themselves are instrumented with Prometheus metrics, therefore those components simply have to be discovered by Prometheus and most of the cluster is monitored.
Metrics that are rather about cluster state than a single component's metrics is exposed by the add-on component kube-state-metrics.
Additionally, to have an overview of cluster nodes' resources the Prometheus node_exporter is used. The node_exporter allows monitoring a node's resources: CPU, memory and disk utilization and more.
Once you complete this guide you will monitor the following:
- cluster state via kube-state-metrics
- nodes via the node_exporter
- kubelets
- apiserver
- kube-scheduler
- kube-controller-manager
Getting Up and Running Fast with Kube-Prometheus
To help get started more quickly with monitoring Kubernetes clusters, kube-prometheus was created. It is a collection of manifests including dashboards and alerting rules that can easily be deployed. It utilizes the Prometheus Operator and all the manifests demonstrated in this guide.
This section represent a quick installation and is not intended to teach you about all the components. The easiest way to get started is to clone this repository and use the kube-prometheus
section of the code.
git clone https://github.com/coreos/kube-prometheus
cd kube-prometheus/
First, create the namespace in which you want the monitoring tool suite to be running.
export NAMESPACE='monitoring'
kubectl create namespace "$NAMESPACE"
Now we will create the components for the Prometheus operator
kubectl --namespace="$NAMESPACE" apply -f manifests/prometheus-operator
This will create all the Prometheus Operator components. You might need to wait a short amount of time before the Custom Resource Definitions are available in the cluster. You can wait for them:
until kubectl --namespace="$NAMESPACE" get alertmanagers.monitoring.coreos.com > /dev/null 2>&1; do sleep 1; printf "."; done
Next, we will install the node exporter and then kube-state-metrics:
kubectl --namespace="$NAMESPACE" apply -f manifests/node-exporter
kubectl --namespace="$NAMESPACE" apply -f manifests/kube-state-metrics
Then, we can deploy the grafana credentials. By default, the username/password will be admin/admin
, you should change these for your production clusters.
kubectl --namespace="$NAMESPACE" apply -f manifests/grafana/grafana-credentials.yaml
Then install grafana itself:
kubectl --namespace="$NAMESPACE" apply -f manifests/grafana
Next up is the Prometheus
object itself. We will deploy the application, and then the roles/role-bindings.
find manifests/prometheus -type f ! -name prometheus-k8s-roles.yaml ! -name prometheus-k8s-role-bindings.yaml -exec kubectl --namespace "$NAMESPACE" apply -f {} \;
kubectl apply -f manifests/prometheus/prometheus-k8s-roles.yaml
kubectl apply -f manifests/prometheus/prometheus-k8s-role-bindings.yaml
Finally, install the Alertmanager
kubectl --namespace="$NAMESPACE" apply -f manifests/alertmanager
Now you should have a working cluster. After all the pods are ready, you should be able to reach:
- Prometheus UI on node port
30900
- Alertmanager UI on node port
30903
- Grafana on node port
30902
These can of course be changed via the Service definitions. It is recommended to look at the Exposing Prometheus and Alert Manager documentation for more detailed information on how to expose these services.