×
Community Blog Kubernetes Pod Disruption Budgets (PDB)

Kubernetes Pod Disruption Budgets (PDB)

This tutorial demonstrates Pod Disruption Budgets (PDB) using deployments in Kubernetes.

By Alwyn Botha, Alibaba Cloud Community Blog author.

Pod Disruption Budgets specify the number of concurrent disruptions that you allow for your Pods. Disruptions may be caused by deliberate or accidental Pod deletion.

When a Kubernetes node run out of RAM or disk space it will evict Pods in an attempt to keep the node running properly. This eviction also causes disruptions.

You can specify Pod Disruption Budgets for Pods managed by these built-in Kubernetes controllers:

  • Deployment
  • ReplicationController
  • ReplicaSet
  • StatefulSet

This tutorial demonstrates Pod Disruption Budgets using deployments. Before proceeding with this tutorial, you should experiment with Kubernetes Eviction Policies since it demonstrates how Pod Disruption Budgets handle evictions.

As in the Kubernetes Eviction Policies tutorial we start with eviction-hard="memory.available<480M

We will then deploy some Pods that exceed this limit to see how Pod Disruption Budgets work.

As detailed in Kubernetes Eviction Policies tutorials you need to run these exercises on a node with 2200 MB RAM.

Pod Disruption Budgets Basics

Start minikube with hard threshold.

minikube start --extra-config=kubelet.eviction-hard="memory.available<480M" --extra-config=kubelet.feature-gates="ExperimentalCriticalPodAnnotation=true"  --extra-config=kubelet.eviction-pressure-transition-period="30s" 

Check available RAM:

source ./mem  | tail -n1
memory.available_in_mb 1211

Check if node has MemoryPressure:

kubectl describe node minikube | grep MemoryPressure
  MemoryPressure   False   Tue, 05 Feb 2019 07:45:32 +0200   Mon, 04 Feb 2019 12:06:54 +0200   KubeletHasSufficientMemory   kubelet has sufficient memory available

No MemoryPressure: 700 MB away from that.

We will now create 3 Pods using 50 MB each.

kubectl create -f my50mb-ram-pod-a.yaml
kubectl create -f my50mb-ram-pod-b.yaml
kubectl create -f my50mb-ram-pod-c.yaml

Check available RAM:

source ./mem  | tail -n1
memory.available_in_mb 1042

500 MB away from MemoryPressure

Now we are going to create 10 more Pods that will surely push the node into a MemoryPressure condition.

To learn more we will create 2 deployments with 5 replicas each.

The my5mbDeployment Pods uses 5 MB each - via stress benchmark application.

The my10mbDeployment Pods uses 10 MB each - via stress benchmark application.

We also have those three 50 MB Pods already running.

Edit / create YAML files for deployments:

nano my5mbDeployment.yaml

apiVersion: apps/v1
kind: Deployment
metadata:
  name: 5mb-deployment
  labels:
    app: 5mbpods
spec:
  replicas: 5
  strategy: 
    type: RollingUpdate
  selector:
    matchLabels:
      app: 5mbpods
  template:
    metadata:
      labels:
        app: 5mbpods
    spec:
      containers:
      - name: myram-container-1
        image: mytutorials/centos:bench
        imagePullPolicy: IfNotPresent
    
        command: ['sh', '-c', 'stress --vm 1 --vm-bytes 5M --vm-hang 3000 -t 3600']
    
        resources:
          requests:
            memory: "1k"
    
      terminationGracePeriodSeconds: 0
nano my10mbDeployment.yaml

apiVersion: apps/v1
kind: Deployment
metadata:
  name: 10mb-deployment
  labels:
    app: 10mbpods
spec:
  replicas: 5
  strategy: 
    type: RollingUpdate
  selector:
    matchLabels:
      app: 10mbpods
  template:
    metadata:
      labels:
        app: 10mbpods
    spec:
      containers:
      - name: myram-container-1
        image: mytutorials/centos:bench
        imagePullPolicy: IfNotPresent
    
        command: ['sh', '-c', 'stress --vm 1 --vm-bytes 10M --vm-hang 3000 -t 3600']
    
        resources:
          requests:
            memory: "1k"
    
      terminationGracePeriodSeconds: 0

Create deployments.

kubectl create -f my5mbDeployment.yaml
deployment.apps/5mb-deployment created

kubectl create -f my10mbDeployment.yaml
deployment.apps/10mb-deployment created

Note that in the deployment we specified replicas: 5 each. We need 5 identical Pods running for each of those 2 deployments.

Now we specify a Pod Disruption Budget:

  • for 5m Pods we need a minimum of 2 Pods to be always running minAvailable: 2
  • for 10m Pods we need a minimum of 1 Pod to be always running minAvailable: 1

Important: the app: 5mbpods in the Pod Disruption Budget must match the same line in the deployment.

ONLY this line links the Pod Disruption Budget to an application - the Pods in the deployment.

nano mypdb5.yaml

apiVersion: policy/v1beta1
kind: PodDisruptionBudget
metadata:
  name: 5mbpods-pdb
spec:
  minAvailable: 2
  selector:
    matchLabels:
      app: 5mbpods
nano mypdb10.yaml

apiVersion: policy/v1beta1
kind: PodDisruptionBudget
metadata:
  name: 5mbpods-pdb
spec:
  minAvailable: 1
  selector:
    matchLabels:
      app: 10mbpods

Create Pod Disruption Budgets :

kubectl create -f mypdb5.yaml
kubectl create -f mypdb10.yaml

Check available RAM:

memory.available_in_mb 478

MemoryPressure? YES

  MemoryPressure   True    Tue, 05 Feb 2019 07:49:23 +0200   Tue, 05 Feb 2019 07:48:43 +0200   KubeletHasInsufficientMemory   kubelet has insufficient memory available

Mere moments later:

kubectl get deploy

NAME              DESIRED   CURRENT   UP-TO-DATE   AVAILABLE   AGE
10mb-deployment   5         5         5            5           29s
5mb-deployment    5         5         5            5           33s

Our deployments are running perfectly: 5 DESIRED replicas, 5 AVAILABLE.

List all running Pods.

kubectl get pods

NAME                               READY   STATUS    RESTARTS   AGE
10mb-deployment-558b949886-f6dvj   1/1     Running   0          34s
10mb-deployment-558b949886-lx8kp   1/1     Running   0          34s
10mb-deployment-558b949886-n2bct   1/1     Running   0          34s
10mb-deployment-558b949886-t2652   1/1     Running   0          34s
10mb-deployment-558b949886-w52hz   1/1     Running   0          34s

5mb-deployment-5964c58c48-2qwvf    1/1     Running   0          38s
5mb-deployment-5964c58c48-4b7z6    1/1     Running   0          38s
5mb-deployment-5964c58c48-9shg9    1/1     Running   0          38s
5mb-deployment-5964c58c48-crsp6    1/1     Running   0          38s
5mb-deployment-5964c58c48-x848b    1/1     Running   0          38s

my50mb-ram-pod-a                   1/1     Running   0          84s
my50mb-ram-pod-b                   0/1     Evicted   0          85s
my50mb-ram-pod-c                   0/1     Evicted   0          85s

We see 2 of our 50 MB Pods got evicted.

We specified minAvailable: 1 for our 10mb-deployment : meaning we allow 4 of the 5 replicas to be disrupted.

Kubernetes choose to evict the 50 MB Pods. Why?

As detailed in the Kubernetes Eviction Policies tutorial, Pods are ranked for eviction.

pods are ranked by Priority, and then usage above request.

Our 50 MB Pods rank much higher than our 10 MB Pods.

Here is an edited extract of the kubelet eviction manager logs.

my50mb-ram-pod-c and my50mb-ram-pod-b listed first - got evicted. System never got close to our 2 deployments.

Feb 05 07:48:40 pods ranked for eviction:
  kube-apiserver-minikube_kube-system,
  etcd-minikube_kube-system,
  kube-controller-manager-minikube_kube-system,
  kube-scheduler-minikube_kube-system,
  my50mb-ram-pod-c_default,   < ----------
  my50mb-ram-pod-b_default,   < ----------
  my50mb-ram-pod-a_default,
  metrics-server-6486d4db88-rlsdd_kube-system,
  kubernetes-dashboard-5bff5f8fb8-27mh9_kube-system,
  10mb-deployment-558b949886-n2bct_default,
  10mb-deployment-558b949886-w52hz_default,
  10mb-deployment-558b949886-f6dvj_default,
  10mb-deployment-558b949886-lx8kp_default,
  10mb-deployment-558b949886-t2652_default,
  5mb-deployment-5964c58c48-9shg9_default,
  5mb-deployment-5964c58c48-x848b_default,
  5mb-deployment-5964c58c48-2qwvf_default,
  5mb-deployment-5964c58c48-4b7z6_default,
  5mb-deployment-5964c58c48-crsp6_default,
  
Feb 05 07:48:40 pod my50mb-ram-pod-c_default is evicted successfully

Feb 05 07:49:03 pods ranked for eviction:
  kube-apiserver-minikube_kube-system,
  kube-controller-manager-minikube_kube-system,
  etcd-minikube_kube-system,
  kube-scheduler-minikube_kube-system,
  my50mb-ram-pod-b_default,   < ----------
  my50mb-ram-pod-a_default,   < ----------
  metrics-server-6486d4db88-rlsdd_kube-system,
  kubernetes-dashboard-5bff5f8fb8-27mh9_kube-system,
  10mb-deployment-558b949886-n2bct_default,
  10mb-deployment-558b949886-w52hz_default,
  10mb-deployment-558b949886-lx8kp_default,
  10mb-deployment-558b949886-f6dvj_default,
  10mb-deployment-558b949886-t2652_default,
  5mb-deployment-5964c58c48-crsp6_default,
  5mb-deployment-5964c58c48-x848b_default,
  5mb-deployment-5964c58c48-9shg9_default,
  5mb-deployment-5964c58c48-2qwvf_default,
  5mb-deployment-5964c58c48-4b7z6_default,
  
Feb 05 07:49:03 pod my50mb-ram-pod-b_default is evicted successfully

Kubernetes cannot evict critical static pods :

  • kube-apiserver-minikube_kube-system,
  • etcd-minikube_kube-system,
  • kube-controller-manager-minikube_kube-system,
  • kube-scheduler-minikube_kube-system,

So those Pods are not evicted. I deleted those lines from output since it adds no value to this discussion.

Check if node still has MemoryPressure:

kubectl describe node minikube | grep MemoryPressure
  MemoryPressure   False   Tue, 05 Feb 2019 07:50:23 +0200   Tue, 05 Feb 2019 07:49:43 +0200   KubeletHasSufficientMemory   kubelet has sufficient memory available

No, those evictions fixed the problem.

source ./mem  | tail -n1
memory.available_in_mb 497

17 MB above threshold limit.

Let us use another 20 MB RAM by scaling our 10 MB deployment to 7 replicas.

kubectl scale -f my10mbDeployment.yaml --replicas=7

A second later.

MemoryPressure   False   Tue, 05 Feb 2019 07:52:13 +0200   Tue, 05 Feb 2019 07:49:43 +0200   KubeletHasSufficientMemory   kubelet has sufficient memory available

Too fast, system did not realize its MemoryPressure should be TRUE .

We can see RAM below 480 MB limit.

source ./mem  | tail -n1
memory.available_in_mb 454

Let's list the pods

kubectl get pods

NAME                               READY   STATUS    RESTARTS   AGE
10mb-deployment-558b949886-8j2rb   1/1     Running   0          30s
10mb-deployment-558b949886-f6dvj   1/1     Running   0          4m9s
10mb-deployment-558b949886-lx8kp   1/1     Running   0          4m9s
10mb-deployment-558b949886-n2bct   1/1     Running   0          4m9s
10mb-deployment-558b949886-t2652   1/1     Running   0          4m9s
10mb-deployment-558b949886-w52hz   1/1     Running   0          4m9s
10mb-deployment-558b949886-zbkkv   1/1     Running   0          30s

5mb-deployment-5964c58c48-2qwvf    1/1     Running   0          4m13s
5mb-deployment-5964c58c48-4b7z6    1/1     Running   0          4m13s
5mb-deployment-5964c58c48-9shg9    1/1     Running   0          4m13s
5mb-deployment-5964c58c48-crsp6    1/1     Running   0          4m13s
5mb-deployment-5964c58c48-x848b    1/1     Running   0          4m13s

my50mb-ram-pod-a                   0/1     Evicted   0          4m59s
my50mb-ram-pod-b                   0/1     Evicted   0          5m
my50mb-ram-pod-c                   0/1     Evicted   0          5m

Another 50 MB Pod evicted.

Lesson learned: Pod Disruption Budgets do not mean such Pods will be disrupted first.

pods are ranked by usage above request.

Not ranked by if they have a budget or not.

Pod Disruption Budgets are considered WHEN a Pod qualifies for eviction.

The eviction manager logs shows my50mb-ram-pod-a rank higher: it got evicted.

Feb 05 07:52:28 pods ranked for eviction:
  kube-apiserver-minikube_kube-system,
  etcd-minikube_kube-system,
  kube-controller-manager-minikube_kube-system,
  kube-scheduler-minikube_kube-system,
  my50mb-ram-pod-a_default,         < - - - - - - - -
  metrics-server-6486d4db88-rlsdd_kube-system,
  kubernetes-dashboard-5bff5f8fb8-27mh9_kube-system,
  10mb-deployment-558b949886-8j2rb_default,
  10mb-deployment-558b949886-n2bct_default,
  10mb-deployment-558b949886-w52hz_default,
  10mb-deployment-558b949886-lx8kp_default,
  10mb-deployment-558b949886-f6dvj_default,
  10mb-deployment-558b949886-t2652_default,
  5mb-deployment-5964c58c48-crsp6_default,
  5mb-deployment-5964c58c48-x848b_default,
  5mb-deployment-5964c58c48-2qwvf_default,
  5mb-deployment-5964c58c48-9shg9_default,
  5mb-deployment-5964c58c48-4b7z6_default,
  10mb-deployment-558b949886-zbkkv_default,
  
Feb 05 07:52:28 pod my50mb-ram-pod-a_default is evicted successfully

Check available RAM:

source ./mem  | tail -n1
memory.available_in_mb 498

Check if node has MemoryPressure:

  MemoryPressure   False   Tue, 05 Feb 2019 07:54:03 +0200   Tue, 05 Feb 2019 07:53:03 +0200   KubeletHasSufficientMemory   kubelet has sufficient memory available

Problem solved.

source ./mem  | tail -n1
memory.available_in_mb 495

15 MB above eviction threshold.

Let's use more RAM to cause Pod Disruption Budgets to get used.

kubectl scale -f my5mbDeployment.yaml --replicas=10

List deployments - we can see on first line one Pod got evicted. ( AVAILABLE = 6 but DESIRED = 7 )

kubectl get deploy

NAME              DESIRED   CURRENT   UP-TO-DATE   AVAILABLE   AGE
10mb-deployment   7         7         7            6           8m53s
5mb-deployment    10        10        10           10          8m57s

MemoryPressure True

kubectl describe node minikube | grep MemoryPressure
  MemoryPressure   True    Tue, 05 Feb 2019 07:57:34 +0200   Tue, 05 Feb 2019 07:57:24 +0200   KubeletHasInsufficientMemory   kubelet has insufficient memory available

Way below threshold.

source ./mem  | tail -n1
memory.available_in_mb 456

List all Pods.

kubectl get pods

NAME                               READY   STATUS    RESTARTS   AGE
10mb-deployment-558b949886-8j2rb   1/1     Running   0          5m25s
10mb-deployment-558b949886-f6dvj   1/1     Running   0          9m4s
10mb-deployment-558b949886-lx8kp   1/1     Running   0          9m4s
10mb-deployment-558b949886-n2bct   1/1     Running   0          9m4s

10mb-deployment-558b949886-r97pz   0/1     Pending   0          15s

10mb-deployment-558b949886-t2652   1/1     Running   0          9m4s
10mb-deployment-558b949886-w52hz   1/1     Running   0          9m4s

10mb-deployment-558b949886-zbkkv   0/1     Evicted   0          5m25s


5mb-deployment-5964c58c48-2qwvf    1/1     Running   0          9m8s
5mb-deployment-5964c58c48-4b7z6    1/1     Running   0          9m8s
5mb-deployment-5964c58c48-8jf4r    1/1     Running   0          25s
5mb-deployment-5964c58c48-9shg9    1/1     Running   0          9m8s
5mb-deployment-5964c58c48-cjh2m    1/1     Running   0          25s
5mb-deployment-5964c58c48-crsp6    1/1     Running   0          9m8s
5mb-deployment-5964c58c48-dzx9x    1/1     Running   0          25s
5mb-deployment-5964c58c48-gwbwn    1/1     Running   0          25s
5mb-deployment-5964c58c48-jvkbk    1/1     Running   0          25s
5mb-deployment-5964c58c48-x848b    1/1     Running   0          9m8s

my50mb-ram-pod-a                   0/1     Evicted   0          9m54s
my50mb-ram-pod-b                   0/1     Evicted   0          9m55s
my50mb-ram-pod-c                   0/1     Evicted   0          9m55s

Based on experience this node now has a permanent problem.

One 10mb-deployment evicted, seconds later a replacement 10mb-deployment Pod is Pending

Node does not allow new Pods to start running while in MemoryPressure condition ( Pods are in Pending state ).

source ./mem  | tail -n1
memory.available_in_mb 465

Still not enough RAM available. More Pods will be evicted.

The 10 MB deployment notices its evicted Pods and creates replacement Pods - which stay pending.

This very second 3 Pods are pending ... waiting for RAM to become available.

The moment RAM becomes available some of those Pods will become 'running'. Even just one 10 MB Pod will push node into below eviction threshold condition ... and so the cycle continues.

The problem is the system does not have standalone Pods running for only a few minutes and then completing successfully. All we have is these 2 massive deployments - both running for an hour.

If you have a node running like this: slightly over capacity with NO Pods constantly completing successfully then evictions cannot save you.

Pod Disruption Budgets cannot save a permanently over-capacity server.

Pod Disruption Budgets DO HELP if some Pods complete successfully over time. Then replacement Pods will not push the node into low RAM again.

kubectl get pods

NAME                               READY   STATUS    RESTARTS   AGE
10mb-deployment-558b949886-66tml   0/1     Pending   0          22s
10mb-deployment-558b949886-8j2rb   0/1     Evicted   0          6m59s
10mb-deployment-558b949886-9mfx4   0/1     Pending   0          24s
10mb-deployment-558b949886-f6dvj   1/1     Running   0          10m
10mb-deployment-558b949886-hrp2s   0/1     Evicted   0          34s
10mb-deployment-558b949886-lx8kp   1/1     Running   0          10m
10mb-deployment-558b949886-n2bct   1/1     Running   0          10m
10mb-deployment-558b949886-npvjf   0/1     Pending   0          29s
10mb-deployment-558b949886-r97pz   0/1     Evicted   0          109s
10mb-deployment-558b949886-t2652   1/1     Running   0          10m
10mb-deployment-558b949886-w52hz   0/1     Evicted   0          10m
10mb-deployment-558b949886-zbkkv   0/1     Evicted   0          6m59s

5mb-deployment-5964c58c48-2qwvf    1/1     Running   0          10m
5mb-deployment-5964c58c48-4b7z6    1/1     Running   0          10m
5mb-deployment-5964c58c48-8jf4r    1/1     Running   0          119s
5mb-deployment-5964c58c48-9shg9    1/1     Running   0          10m
5mb-deployment-5964c58c48-cjh2m    1/1     Running   0          119s
5mb-deployment-5964c58c48-crsp6    1/1     Running   0          10m
5mb-deployment-5964c58c48-dzx9x    1/1     Running   0          119s
5mb-deployment-5964c58c48-gwbwn    1/1     Running   0          119s
5mb-deployment-5964c58c48-jvkbk    1/1     Running   0          119s
5mb-deployment-5964c58c48-x848b    1/1     Running   0          10m

my50mb-ram-pod-a                   0/1     Evicted   0          11m
my50mb-ram-pod-b                   0/1     Evicted   0          11m
my50mb-ram-pod-c                   0/1     Evicted   0          11m

Note that our 5mb-deployment never lost a Pod due to eviction.

pods are ranked by Priority, and then usage above request.

10mb-deployment Pods rank higher and get evicted all the time.

Kubernetes does not spread evictions evenly across Pod Disruption Budgets. The above simple rule still holds.

We specified minAvailable: 1 for our 10mb-deployment : meaning we allow 4 of the 5 replicas to be disrupted.

The output above shows 4 Pods still running - kubelet focus all its effort on this deployment since it still has scope to delete 3 more Pods as allowed in the Pod Disruption Budget.

The obvious solution in this case is to reduce number of replicas for 10mb-deployment and / or 5mb-deployment . There are no other Pods to consider.

Exercise complete. Delete ...

kubectl delete -f my10mbDeployment.yaml
deployment.apps "10mb-deployment" deleted

kubectl delete -f my5mbDeployment.yaml
deployment.apps "5mb-deployment" deleted

kubectl delete -f my50mb-ram-pod-a.yaml
pod "my50mb-ram-pod-a" deleted

kubectl delete -f my50mb-ram-pod-b.yaml
pod "my50mb-ram-pod-b" deleted

kubectl delete -f my50mb-ram-pod-c.yaml
pod "my50mb-ram-pod-c" deleted

Delete Pod Disruption Budgets .

kubectl delete -f mypdb5.yaml
kubectl delete -f mypdb10.yaml
kubectl delete -f mypdb5-10.yaml  
minikube stop

3 Identical Deployments

I want to see Kubernetes handle 3 identical simultaneous deployments with identical Pod Disruption Budgets.

Same minikube startup configurations as before.

minikube start --extra-config=kubelet.eviction-hard="memory.available<480M" --extra-config=kubelet.feature-gates="ExperimentalCriticalPodAnnotation=true"  --extra-config=kubelet.eviction-pressure-transition-period="30s" 

Here are the 3 new Pod Disruption Budgets. Create it on your node.

nano mypdb10-a.yaml

apiVersion: policy/v1beta1
kind: PodDisruptionBudget
metadata:
  name: 10mbpods-pdb-a
spec:
  minAvailable: 1
  selector:
    matchLabels:
      app: 10mbpods-a
nano mypdb10-b.yaml

apiVersion: policy/v1beta1
kind: PodDisruptionBudget
metadata:
  name: 10mbpods-pdb-b
spec:
  minAvailable: 1
  selector:
    matchLabels:
      app: 10mbpods-b
nano mypdb10-v.yaml

apiVersion: policy/v1beta1
kind: PodDisruptionBudget
metadata:
  name: 10mbpods-pdb-v
spec:
  minAvailable: 1
  selector:
    matchLabels:
      app: 10mbpods-c
kubectl create -f mypdb10-a.yaml
kubectl create -f mypdb10-b.yaml
kubectl create -f mypdb10-c.yaml

get poddisruptionbudgets does not show useful information at this point.

kubectl get poddisruptionbudgets

NAME             MIN AVAILABLE   MAX UNAVAILABLE   ALLOWED DISRUPTIONS   AGE
10mbpods-pdb-a   1               N/A               0                     4m19s
10mbpods-pdb-b   1               N/A               0                     4m19s
10mbpods-pdb-c   1               N/A               0                     4m19s
source ./mem  | tail -n1
memory.available_in_mb 1193

YAML manifests for the deployments :

nano my10mbDeployment-a.yaml

apiVersion: apps/v1
kind: Deployment
metadata:
  name: 10mb-deployment-a
  labels:
    app: 10mbpods-a
spec:
  replicas: 4
  strategy: 
    type: RollingUpdate
  selector:
    matchLabels:
      app: 10mbpods-a
  template:
    metadata:
      labels:
        app: 10mbpods-a
    spec:
      containers:
      - name: myram-container-1
        image: mytutorials/centos:bench
        imagePullPolicy: IfNotPresent
    
        command: ['sh', '-c', 'stress --vm 1 --vm-bytes 10M --vm-hang 3000 -t 3600']
    
        resources:
          requests:
            memory: "1k"
    
      terminationGracePeriodSeconds: 0
nano my10mbDeployment-b.yaml

apiVersion: apps/v1
kind: Deployment
metadata:
  name: 10mb-deployment-b
  labels:
    app: 10mbpods-b
spec:
  replicas: 4
  strategy: 
    type: RollingUpdate
  selector:
    matchLabels:
      app: 10mbpods-b
  template:
    metadata:
      labels:
        app: 10mbpods-b
    spec:
      containers:
      - name: myram-container-1
        image: mytutorials/centos:bench
        imagePullPolicy: IfNotPresent
    
        command: ['sh', '-c', 'stress --vm 1 --vm-bytes 10M --vm-hang 3000 -t 3600']
    
        resources:
          requests:
            memory: "1k"
    
      terminationGracePeriodSeconds: 0
nano my10mbDeployment-c.yaml

apiVersion: apps/v1
kind: Deployment
metadata:
  name: 10mb-deployment-c
  labels:
    app: 10mbpods-c
spec:
  replicas: 4
  strategy: 
    type: RollingUpdate
  selector:
    matchLabels:
      app: 10mbpods-c
  template:
    metadata:
      labels:
        app: 10mbpods-c
    spec:
      containers:
      - name: myram-container-1
        image: mytutorials/centos:bench
        imagePullPolicy: IfNotPresent
    
        command: ['sh', '-c', 'stress --vm 1 --vm-bytes 10M --vm-hang 3000 -t 3600']
    
        resources:
          requests:
            memory: "1k"
    
      terminationGracePeriodSeconds: 0

Create deployments.

kubectl create -f my10mbDeployment-a.yaml
kubectl create -f my10mbDeployment-b.yaml
kubectl create -f my10mbDeployment-c.yaml

Check RAM

memory.available_in_mb 676
  MemoryPressure   False   Tue, 05 Feb 2019 09:01:05 +0200   Tue, 05 Feb 2019 08:50:34 +0200   KubeletHasSufficientMemory   kubelet has sufficient memory available

No problem so far.

kubectl get poddisruptionbudgets

NAME             MIN AVAILABLE   MAX UNAVAILABLE   ALLOWED DISRUPTIONS   AGE
10mbpods-pdb-a   1               N/A               3                     8m41s
10mbpods-pdb-b   1               N/A               3                     8m41s
10mbpods-pdb-c   1               N/A               3                     8m41s
memory.available_in_mb 522

One more replica per deployment should push node into RAM problem.

kubectl scale -f my10mbDeployment-a.yaml --replicas=5
kubectl scale -f my10mbDeployment-b.yaml --replicas=5
kubectl scale -f my10mbDeployment-c.yaml --replicas=5
kubectl get poddisruptionbudgets

NAME             MIN AVAILABLE   MAX UNAVAILABLE   ALLOWED DISRUPTIONS   AGE
10mbpods-pdb-a   1               N/A               4                     11m
10mbpods-pdb-b   1               N/A               4                     11m
10mbpods-pdb-c   1               N/A               4                     11m

No problem seconds later.

Everything seems OK.

kubectl get pods

NAME                                 READY   STATUS    RESTARTS   AGE
10mb-deployment-a-86589488d-8kxsk    1/1     Running   0          3m36s
10mb-deployment-a-86589488d-9sxbd    1/1     Running   0          3m36s
10mb-deployment-a-86589488d-gdbx9    1/1     Running   0          3m36s
10mb-deployment-a-86589488d-r8z56    1/1     Running   0          3m36s
10mb-deployment-a-86589488d-wm4zr    1/1     Running   0          24s

10mb-deployment-b-6776ffb68c-2zz84   1/1     Running   0          3m36s
10mb-deployment-b-6776ffb68c-4z5dz   1/1     Running   0          24s
10mb-deployment-b-6776ffb68c-q4mmk   1/1     Running   0          3m36s
10mb-deployment-b-6776ffb68c-qrw5z   1/1     Running   0          3m36s
10mb-deployment-b-6776ffb68c-rxknz   1/1     Running   0          3m36s

10mb-deployment-c-b5b56fd58-2thf2    1/1     Running   0          3m36s
10mb-deployment-c-b5b56fd58-2zd78    1/1     Running   0          3m36s
10mb-deployment-c-b5b56fd58-66fjj    1/1     Running   0          24s
10mb-deployment-c-b5b56fd58-m95mn    1/1     Running   0          3m36s
10mb-deployment-c-b5b56fd58-rmn4t    1/1     Running   0          3m36s

A minute later. Kubernetes + Docker + Linux kernel system overhead uses around 30 MB RAM to be able to manage the influx of new replicas.

MemoryPressure True

memory.available_in_mb 460
  MemoryPressure   True    Tue, 05 Feb 2019 09:05:05 +0200   Tue, 05 Feb 2019 09:05:05 +0200   KubeletHasInsufficientMemory   kubelet has insufficient memory available

List Pods:

kubectl get pods

NAME                                 READY   STATUS    RESTARTS   AGE
10mb-deployment-a-86589488d-8kxsk    1/1     Running   0          5m3s
10mb-deployment-a-86589488d-9sxbd    1/1     Running   0          5m3s
10mb-deployment-a-86589488d-gdbx9    1/1     Running   0          5m3s
10mb-deployment-a-86589488d-r8z56    1/1     Running   0          5m3s
10mb-deployment-a-86589488d-wm4zr    1/1     Running   0          111s

10mb-deployment-b-6776ffb68c-2zz84   1/1     Running   0          5m3s
10mb-deployment-b-6776ffb68c-4z5dz   1/1     Running   0          111s
10mb-deployment-b-6776ffb68c-q4mmk   1/1     Running   0          5m3s
10mb-deployment-b-6776ffb68c-qrw5z   1/1     Running   0          5m3s
10mb-deployment-b-6776ffb68c-rxknz   1/1     Running   0          5m3s

10mb-deployment-c-b5b56fd58-2r4sp    0/1     Pending   0          26s
10mb-deployment-c-b5b56fd58-2thf2    1/1     Running   0          5m3s
10mb-deployment-c-b5b56fd58-2zd78    0/1     Evicted   0          5m3s
10mb-deployment-c-b5b56fd58-66fjj    0/1     Evicted   0          111s
10mb-deployment-c-b5b56fd58-fff2g    0/1     Pending   0          28s
10mb-deployment-c-b5b56fd58-m95mn    1/1     Running   0          5m3s
10mb-deployment-c-b5b56fd58-rbkbx    0/1     Evicted   0          34s
10mb-deployment-c-b5b56fd58-rmn4t    1/1     Running   0          5m3s

To fix MemoryPressure 3 Pods get evicted ( all from 10mb-deployment-c )

Perfect fairness would have demanded one Pod evicted per different deployment.

Somehow three 10mb-deployment-c Pods got ranked first.

We can see in the eviction manager logs below that the remaining Pods are ranked roughly randomly evenly for our a to c series deployments.

07:05:05 pods ranked for eviction: 

10mb-deployment-c-b5b56fd58-2zd78_default,
metrics-server-6486d4db88-k249j_kube-system,
10mb-deployment-c-b5b56fd58-66fjj_default,
10mb-deployment-b-6776ffb68c-4z5dz_default,
10mb-deployment-a-86589488d-wm4zr_default,
kubernetes-dashboard-5bff5f8fb8-rq727_kube-system,
10mb-deployment-a-86589488d-9sxbd_default,
10mb-deployment-a-86589488d-gdbx9_default,
10mb-deployment-b-6776ffb68c-qrw5z_default,
10mb-deployment-b-6776ffb68c-q4mmk_default,
10mb-deployment-c-b5b56fd58-m95mn_default,
10mb-deployment-b-6776ffb68c-2zz84_default,
10mb-deployment-b-6776ffb68c-rxknz_default,
10mb-deployment-a-86589488d-r8z56_default,
10mb-deployment-c-b5b56fd58-rmn4t_default,
10mb-deployment-a-86589488d-8kxsk_default,
10mb-deployment-c-b5b56fd58-2thf2_default,
kube-proxy-tgcqv_kube-system,

07:05:05 pod 10mb-deployment-c-b5b56fd58-2zd78_default is evicted successfully

07:05:08 pods ranked for eviction: 

metrics-server-6486d4db88-k249j_kube-system,
10mb-deployment-c-b5b56fd58-66fjj_default,
10mb-deployment-c-b5b56fd58-rbkbx_default,
10mb-deployment-b-6776ffb68c-4z5dz_default,
10mb-deployment-a-86589488d-wm4zr_default,
kubernetes-dashboard-5bff5f8fb8-rq727_kube-system,
10mb-deployment-a-86589488d-9sxbd_default,
10mb-deployment-a-86589488d-gdbx9_default,
10mb-deployment-b-6776ffb68c-qrw5z_default,
10mb-deployment-b-6776ffb68c-q4mmk_default,
10mb-deployment-c-b5b56fd58-m95mn_default,
10mb-deployment-b-6776ffb68c-2zz84_default,
10mb-deployment-b-6776ffb68c-rxknz_default,
10mb-deployment-a-86589488d-r8z56_default,
10mb-deployment-c-b5b56fd58-rmn4t_default,
10mb-deployment-a-86589488d-8kxsk_default,
10mb-deployment-c-b5b56fd58-2thf2_default,
kube-proxy-tgcqv_kube-system,

07:05:09 pod metrics-server-6486d4db88-k249j_kube-system is evicted successfully

07:05:11 pods ranked for eviction: 

10mb-deployment-c-b5b56fd58-66fjj_default,
10mb-deployment-c-b5b56fd58-rbkbx_default,
10mb-deployment-b-6776ffb68c-4z5dz_default,
10mb-deployment-a-86589488d-wm4zr_default,
kubernetes-dashboard-5bff5f8fb8-rq727_kube-system,
10mb-deployment-a-86589488d-9sxbd_default,
10mb-deployment-a-86589488d-gdbx9_default,
10mb-deployment-b-6776ffb68c-qrw5z_default,
10mb-deployment-b-6776ffb68c-q4mmk_default,
10mb-deployment-c-b5b56fd58-m95mn_default,
10mb-deployment-b-6776ffb68c-2zz84_default,
10mb-deployment-b-6776ffb68c-rxknz_default,
10mb-deployment-a-86589488d-r8z56_default,
10mb-deployment-c-b5b56fd58-rmn4t_default,
10mb-deployment-a-86589488d-8kxsk_default,
10mb-deployment-c-b5b56fd58-2thf2_default,
kube-proxy-tgcqv_kube-system,

07:05:11 pod 10mb-deployment-c-b5b56fd58-66fjj_default is evicted successfully

07:05:13 pods ranked for eviction: 

10mb-deployment-c-b5b56fd58-rbkbx_default,
10mb-deployment-b-6776ffb68c-4z5dz_default,
10mb-deployment-a-86589488d-wm4zr_default,
kubernetes-dashboard-5bff5f8fb8-rq727_kube-system,
10mb-deployment-a-86589488d-9sxbd_default,
10mb-deployment-a-86589488d-gdbx9_default,
10mb-deployment-b-6776ffb68c-qrw5z_default,
10mb-deployment-b-6776ffb68c-q4mmk_default,
10mb-deployment-c-b5b56fd58-m95mn_default,
10mb-deployment-b-6776ffb68c-2zz84_default,
10mb-deployment-b-6776ffb68c-rxknz_default,
10mb-deployment-a-86589488d-r8z56_default,
10mb-deployment-c-b5b56fd58-rmn4t_default,
10mb-deployment-a-86589488d-8kxsk_default,
10mb-deployment-c-b5b56fd58-2thf2_default,
kube-proxy-tgcqv_kube-system,

07:05:13 pod 10mb-deployment-c-b5b56fd58-rbkbx_default is evicted successfully

A minute later we start to see a Pod each from the 3 different deployments get evicted.

kubectl get poddisruptionbudgets

NAME             MIN AVAILABLE   MAX UNAVAILABLE   ALLOWED DISRUPTIONS   AGE
10mbpods-pdb-a   1               N/A               4                     13m
10mbpods-pdb-b   1               N/A               4                     13m
10mbpods-pdb-c   1               N/A               3                     13m

A minute later.

kubectl get poddisruptionbudgets

NAME             MIN AVAILABLE   MAX UNAVAILABLE   ALLOWED DISRUPTIONS   AGE
10mbpods-pdb-a   1               N/A               3                     14m
10mbpods-pdb-b   1               N/A               3                     14m
10mbpods-pdb-c   1               N/A               3                     14m
kubectl get pods

NAME                                 READY   STATUS              RESTARTS   AGE
10mb-deployment-a-86589488d-8kxsk    1/1     Running             0          6m12s
10mb-deployment-a-86589488d-8wdk8    0/1     ContainerCreating   0          43s
10mb-deployment-a-86589488d-9sxbd    1/1     Running             0          6m12s
10mb-deployment-a-86589488d-gdbx9    1/1     Running             0          6m12s
10mb-deployment-a-86589488d-r8z56    1/1     Running             0          6m12s
10mb-deployment-a-86589488d-wm4zr    0/1     Evicted             0          3m

10mb-deployment-b-6776ffb68c-2zz84   1/1     Running             0          6m12s
10mb-deployment-b-6776ffb68c-4z5dz   0/1     Evicted             0          3m
10mb-deployment-b-6776ffb68c-9kx67   0/1     ContainerCreating   0          46s
10mb-deployment-b-6776ffb68c-q4mmk   1/1     Running             0          6m12s
10mb-deployment-b-6776ffb68c-qrw5z   1/1     Running             0          6m12s
10mb-deployment-b-6776ffb68c-rxknz   1/1     Running             0          6m12s

10mb-deployment-c-b5b56fd58-2r4sp    0/1     Evicted             0          95s
10mb-deployment-c-b5b56fd58-2thf2    1/1     Running             0          6m12s
10mb-deployment-c-b5b56fd58-2zd78    0/1     Evicted             0          6m12s
10mb-deployment-c-b5b56fd58-66fjj    0/1     Evicted             0          3m
10mb-deployment-c-b5b56fd58-8wzgv    0/1     ContainerCreating   0          49s
10mb-deployment-c-b5b56fd58-fff2g    0/1     Evicted             0          97s
10mb-deployment-c-b5b56fd58-m95mn    1/1     Running             0          6m12s
10mb-deployment-c-b5b56fd58-rbkbx    0/1     Evicted             0          103s
10mb-deployment-c-b5b56fd58-rmn4t    1/1     Running             0          6m12s
10mb-deployment-c-b5b56fd58-xz8nw    1/1     Running             0          53s

At the top of the ranks we can see a,b, and c deployments Pods.

07:05:55 pods ranked for eviction: 

10mb-deployment-c-b5b56fd58-fff2g_default,
10mb-deployment-b-6776ffb68c-4z5dz_default,
10mb-deployment-a-86589488d-wm4zr_default,
kubernetes-dashboard-5bff5f8fb8-rq727_kube-system,
10mb-deployment-a-86589488d-9sxbd_default,
10mb-deployment-a-86589488d-gdbx9_default,
10mb-deployment-b-6776ffb68c-qrw5z_default,
10mb-deployment-b-6776ffb68c-q4mmk_default,
10mb-deployment-c-b5b56fd58-m95mn_default,
10mb-deployment-b-6776ffb68c-2zz84_default,
10mb-deployment-b-6776ffb68c-rxknz_default,
10mb-deployment-a-86589488d-r8z56_default,
10mb-deployment-c-b5b56fd58-rmn4t_default,
10mb-deployment-a-86589488d-8kxsk_default,
10mb-deployment-c-b5b56fd58-2thf2_default,
metrics-server-6486d4db88-rdqmw_kube-system,
kube-proxy-tgcqv_kube-system,

07:05:55 pod 10mb-deployment-c-b5b56fd58-fff2g_default is evicted successfully

07:05:57 pods ranked for eviction: 

10mb-deployment-c-b5b56fd58-2r4sp_default,
10mb-deployment-b-6776ffb68c-4z5dz_default,
10mb-deployment-a-86589488d-wm4zr_default,
10mb-deployment-c-b5b56fd58-xz8nw_default,
kubernetes-dashboard-5bff5f8fb8-rq727_kube-system,
10mb-deployment-a-86589488d-9sxbd_default,
10mb-deployment-a-86589488d-gdbx9_default,
10mb-deployment-b-6776ffb68c-qrw5z_default,
10mb-deployment-b-6776ffb68c-q4mmk_default,
10mb-deployment-c-b5b56fd58-m95mn_default,
10mb-deployment-b-6776ffb68c-2zz84_default,
10mb-deployment-b-6776ffb68c-rxknz_default,
10mb-deployment-a-86589488d-r8z56_default,
10mb-deployment-c-b5b56fd58-rmn4t_default,
10mb-deployment-a-86589488d-8kxsk_default,
10mb-deployment-c-b5b56fd58-2thf2_default,
metrics-server-6486d4db88-rdqmw_kube-system,
kube-proxy-tgcqv_kube-system,

07:05:59 pod 10mb-deployment-c-b5b56fd58-2r4sp_default is evicted successfully

07:06:02 pods ranked for eviction: 

10mb-deployment-b-6776ffb68c-4z5dz_default,
10mb-deployment-a-86589488d-wm4zr_default,
10mb-deployment-c-b5b56fd58-xz8nw_default,
kubernetes-dashboard-5bff5f8fb8-rq727_kube-system,
10mb-deployment-a-86589488d-9sxbd_default,
10mb-deployment-a-86589488d-gdbx9_default,
10mb-deployment-b-6776ffb68c-qrw5z_default,
10mb-deployment-c-b5b56fd58-m95mn_default,
10mb-deployment-b-6776ffb68c-2zz84_default,
10mb-deployment-b-6776ffb68c-rxknz_default,
10mb-deployment-a-86589488d-r8z56_default,
10mb-deployment-a-86589488d-8kxsk_default,
10mb-deployment-b-6776ffb68c-q4mmk_default,
10mb-deployment-c-b5b56fd58-rmn4t_default,
10mb-deployment-c-b5b56fd58-2thf2_default,
metrics-server-6486d4db88-rdqmw_kube-system,
kube-proxy-tgcqv_kube-system,

07:06:03 pod 10mb-deployment-b-6776ffb68c-4z5dz_default is evicted successfully

07:06:05 pods ranked for eviction: 

10mb-deployment-a-86589488d-wm4zr_default,
10mb-deployment-c-b5b56fd58-xz8nw_default,
kubernetes-dashboard-5bff5f8fb8-rq727_kube-system,
10mb-deployment-a-86589488d-9sxbd_default,
10mb-deployment-a-86589488d-gdbx9_default,
10mb-deployment-b-6776ffb68c-qrw5z_default,
10mb-deployment-c-b5b56fd58-m95mn_default,
10mb-deployment-b-6776ffb68c-2zz84_default,
10mb-deployment-b-6776ffb68c-rxknz_default,
10mb-deployment-a-86589488d-r8z56_default,
10mb-deployment-b-6776ffb68c-q4mmk_default,
10mb-deployment-c-b5b56fd58-rmn4t_default,
10mb-deployment-a-86589488d-8kxsk_default,
10mb-deployment-c-b5b56fd58-2thf2_default,
metrics-server-6486d4db88-rdqmw_kube-system,
kube-proxy-tgcqv_kube-system,

07:06:05 pod 10mb-deployment-a-86589488d-wm4zr_default is evicted successfully

When we started the node easily handled 4 replicas per deployment.

Now, due to increasing system RAM overheads, it is unable to do so.

memory.available_in_mb 464

Reduce replicas to 4.

kubectl scale -f my10mbDeployment-a.yaml --replicas=4
kubectl scale -f my10mbDeployment-b.yaml --replicas=4
kubectl scale -f my10mbDeployment-c.yaml --replicas=4

RAM still below needed minimum of 480.

memory.available_in_mb 458

More Pods get evicted.

kubectl get poddisruptionbudgets

NAME             MIN AVAILABLE   MAX UNAVAILABLE   ALLOWED DISRUPTIONS   AGE
10mbpods-pdb-a   1               N/A               2                     17m
10mbpods-pdb-b   1               N/A               3                     17m
10mbpods-pdb-c   1               N/A               2                     17m

A minute later - more evictions.

kubectl get poddisruptionbudgets

NAME             MIN AVAILABLE   MAX UNAVAILABLE   ALLOWED DISRUPTIONS   AGE
10mbpods-pdb-a   1               N/A               1                     18m
10mbpods-pdb-b   1               N/A               3                     18m
10mbpods-pdb-c   1               N/A               2                     18m

This cycle of evict, recreate and more evict, recreate uses even more RAM.

memory.available_in_mb 405

Scale replicas to 3.

kubectl scale -f my10mbDeployment-a.yaml --replicas=3
kubectl scale -f my10mbDeployment-b.yaml --replicas=3
kubectl scale -f my10mbDeployment-c.yaml --replicas=3

Still does not fix RAM problem.

memory.available_in_mb 473
  MemoryPressure   True    Tue, 05 Feb 2019 09:12:16 +0200   Tue, 05 Feb 2019 09:12:16 +0200   KubeletHasInsufficientMemory   kubelet has insufficient memory available
kubectl get poddisruptionbudgets
NAME             MIN AVAILABLE   MAX UNAVAILABLE   ALLOWED DISRUPTIONS   AGE
10mbpods-pdb-a   1               N/A               1                     20m
10mbpods-pdb-b   1               N/A               0                     20m
10mbpods-pdb-c   1               N/A               1                     20m

Scale replicas to 3.

kubectl scale -f my10mbDeployment-a.yaml --replicas=2
kubectl scale -f my10mbDeployment-b.yaml --replicas=2
kubectl scale -f my10mbDeployment-c.yaml --replicas=2

Still below threshold of 480.

memory.available_in_mb 463

memory.available_in_mb 457

Delete all 3 deployments completely.

kubectl delete -f my10mbDeployment-a.yaml
kubectl delete -f my10mbDeployment-b.yaml
kubectl delete -f my10mbDeployment-c.yaml

MemoryPressure?

  MemoryPressure   False   Tue, 05 Feb 2019 09:15:07 +0200   Tue, 05 Feb 2019 09:14:57 +0200   KubeletHasSufficientMemory   kubelet has sufficient memory available

No. 5 MB above threshold.

Ongoing growing system RAM overhead now makes it impossible to run these 3 deployments with 4 replicas each.

memory.available_in_mb 485

The kubelet eviction manager logs for last 8 minutes.

There is no one large Pod to fix problem forever.

No large standalone Pod completes to give back RAM to rest of node.

Do not try to run Kubernetes nodes permanently at the RAM eviction threshold.

07:06:47 pod metrics-server-6486d4db88-rdqmw_kube-system is evicted successfully
07:06:51 pod 10mb-deployment-c-b5b56fd58-8wzgv_default is evicted successfully
07:06:54 pod 10mb-deployment-c-b5b56fd58-bpf5g_default is evicted successfully
07:06:58 pod 10mb-deployment-c-b5b56fd58-p652t_default is evicted successfully
07:07:03 pod 10mb-deployment-a-86589488d-8wdk8_default is evicted successfully
07:07:07 pod 10mb-deployment-c-b5b56fd58-xz8nw_default is evicted successfully
07:07:11 pod 10mb-deployment-b-6776ffb68c-9kx67_default is evicted successfully
07:07:56 pod 10mb-deployment-b-6776ffb68c-bgc62_default is evicted successfully
07:08:00 pod 10mb-deployment-b-6776ffb68c-s4ngr_default is evicted successfully
07:08:05 pod 10mb-deployment-a-86589488d-qk74d_default is evicted successfully
07:08:09 pod metrics-server-6486d4db88-577j9_kube-system is evicted successfully
07:08:13 pod 10mb-deployment-c-b5b56fd58-28zpq_default is evicted successfully
07:08:16 pod 10mb-deployment-c-b5b56fd58-cdnm7_default is evicted successfully
07:08:39 pod kubernetes-dashboard-5bff5f8fb8-rq727_kube-system is evicted successfully
07:09:23 pod 10mb-deployment-c-b5b56fd58-m6g95_default is evicted successfully
07:09:27 pod 10mb-deployment-c-b5b56fd58-2rp98_default is evicted successfully
07:09:31 pod 10mb-deployment-a-86589488d-9sxbd_default is evicted successfully
07:10:03 pod 10mb-deployment-a-86589488d-gdbx9_default is evicted successfully
07:10:49 pod 10mb-deployment-a-86589488d-rphsh_default is evicted successfully
07:11:20 pod metrics-server-6486d4db88-66mcg_kube-system is evicted successfully
07:11:23 pod 10mb-deployment-c-b5b56fd58-45nhg_default is evicted successfully
07:11:25 pod 10mb-deployment-a-86589488d-rphsh_default is evicted successfully
07:11:27 pod 10mb-deployment-a-86589488d-jkh2s_default is evicted successfully
07:11:30 pod kubernetes-dashboard-5bff5f8fb8-56tgx_kube-system is evicted successfully
07:11:32 pod 10mb-deployment-b-6776ffb68c-qrw5z_default is evicted successfully
07:12:15 pod 10mb-deployment-a-86589488d-klsk5_default is evicted successfully
07:12:19 pod 10mb-deployment-a-86589488d-8kg8q_default is evicted successfully
07:12:21 pod 10mb-deployment-b-6776ffb68c-zdpmk_default is evicted successfully
07:12:23 pod metrics-server-6486d4db88-xm8z9_kube-system is evicted successfully
07:12:24 pod 10mb-deployment-b-6776ffb68c-q4mmk_default is evicted successfully
07:12:27 pod 10mb-deployment-c-b5b56fd58-m95mn_default is evicted successfully
07:13:10 pod 10mb-deployment-b-6776ffb68c-4j9dr_default is evicted successfully
07:13:13 pod 10mb-deployment-b-6776ffb68c-4qqtx_default is evicted successfully
07:13:15 pod 10mb-deployment-a-86589488d-lvgxg_default is evicted successfully
07:13:18 pod 10mb-deployment-b-6776ffb68c-b6clc_default is evicted successfully
07:13:20 pod 10mb-deployment-b-6776ffb68c-rxknz_default is evicted successfully
07:13:22 pod 10mb-deployment-a-86589488d-r8z56_default is evicted successfully
07:14:05 pod 10mb-deployment-b-6776ffb68c-tgpck_default is evicted successfully
07:14:09 pod 10mb-deployment-a-86589488d-fd2bt_default is evicted successfully
07:14:13 pod 10mb-deployment-b-6776ffb68c-9rs8g_default is evicted successfully
07:14:15 pod metrics-server-6486d4db88-sqqtp_kube-system is evicted successfully

Syntax: Specifying a PodDisruptionBudget

You can find comprehensive detail on the allowed syntax:

https://kubernetes.io/docs/tasks/run-application/configure-pdb/#specifying-a-poddisruptionbudget

Pod Disruption Budget Documentation

This tutorial focused on practical exercises. The official Kubernetes documentation has more details on the concepts discussed here:

Run Your Own Experiments

As a follow up, consider this experiment.

Define 4 Pod Disruption Budgets for 4 different deployments. Pods should have same RAM usage as in previous example.

Pod Disruption Budgets should only allow 1 or 2 maxUnavailable

This way you can see Kubernetes act on Pod Disruption Budgets for 4 deployments.

0 0 0
Share on

Alibaba Clouder

1,408 posts | 218 followers

You may also like

Comments

Alibaba Clouder

1,408 posts | 218 followers

Related Products