在专用节点池上调度kube-dns

Fau*_*aun 5 kubernetes google-kubernetes-engine kube-dns

我有一个在GCP上运行的群集,该群集目前完全由可兑换节点组成。我们遇到了kube-dns不可用的问题(大概是因为节点已被抢占)。我们希望通过将kube-dnsPod移至更稳定的节点来提高DNS的弹性。

是否可以kube-dnskube-system仅不可抢占节点的节点池上调度系统群集关键Pod (或命名空间中的所有Pod )?我对使用亲和力或反亲和力或污点持谨慎态度,因为这些Pod是在群集引导时自动创建的,所做的任何更改都可能会被Kubernetes版本升级所破坏。有没有办法做到这一点在升级中仍然存在?

Fau*_*aun 5

解决方案是将污点和容忍与节点亲和力结合使用。我们创建了第二个节点池,并向可抢占池添加了污点。

地形配置:

resource "google_container_node_pool" "preemptible_worker_pool" {
  node_config {
    ...
    preemptible     = true

    labels {
      preemptible = "true"
      dedicated   = "preemptible-worker-pool"
    }

    taint {
      key    = "dedicated"
      value  = "preemptible-worker-pool"
      effect = "NO_SCHEDULE"
    }
  }
}
Run Code Online (Sandbox Code Playgroud)

然后,我们使用tolerationnodeAffinity来允许现有工作负载在受污染的节点池上运行,从而有效地强制集群关键 Pod 在未受污染(不可抢占)的节点池上运行。

库伯内特斯配置:

spec:
  template:
    spec:
      # The affinity + tolerations sections together allow and enforce that the workers are
      # run on dedicated nodes tainted with "dedicated=preemptible-worker-pool:NoSchedule".
      affinity:
        nodeAffinity:
          requiredDuringSchedulingIgnoredDuringExecution:
            nodeSelectorTerms:
            - matchExpressions:
              - key: dedicated
                operator: In
                values:
                - preemptible-worker-pool
      tolerations:
      - key: dedicated
        operator: "Equal"
        value: preemptible-worker-pool
        effect: "NoSchedule"
Run Code Online (Sandbox Code Playgroud)