重置Kubernetes集群 [英] Reset Kubernetes cluster

查看:529
本文介绍了重置Kubernetes集群的处理方法,对大家解决问题具有一定的参考价值,需要的朋友们下面随着小编来一起学习吧!

问题描述

我的网络中有六台台式机,我想构建两个Kubernetes集群.每台机器都安装了Ubuntu 16.04 LTS.最初,所有机器都是单个集群的一部分.但是,我删除了三台计算机以设置另一个集群,并在每台计算机上执行以下命令:

I have six desktop machines in my network and I want to build two Kubernetes clusters. Each machine has Ubuntu 16.04 LTS installed. Initially, all the machines were part of a single cluster. However, I removed three of the machines to setup another cluster, and executed the following command on each of these machine:

RESET COMMAND:
sudo kubeadm reset -f && 
 sudo systemctl stop kubelet && 
 sudo systemctl stop docker && 
 sudo rm -rf /var/lib/cni/ && 
 sudo rm -rf /var/lib/kubelet/* && 
 sudo rm -rf /etc/cni/ && 
 sudo ifconfig cni0 down && 
 sudo ifconfig flannel.1 down && 
 sudo ifconfig docker0 down && 
 sudo ip link delete cni0 && 
 sudo ip link delete flannel.1

此后,我重新启动了每台计算机,并通过设置主节点来进行新集群的设置:

After this I rebooted each machine, and proceeded with the setup of a new cluster, by setting up the master node:

INSTALL COMMAND:
sudo kubeadm init phase certs all && 
 sudo kubeadm init phase kubeconfig all && 
 sudo kubeadm init phase control-plane all --pod-network-cidr 10.244.0.0/16 &&
 sudo sed -i 's/initialDelaySeconds: [0-9][0-9]/initialDelaySeconds: 240/g' /etc/kubernetes/manifests/kube-apiserver.yaml &&
 sudo sed -i 's/failureThreshold: [0-9]/failureThreshold: 18/g' /etc/kubernetes/manifests/kube-apiserver.yaml &&
 sudo sed -i 's/timeoutSeconds: [0-9][0-9]/timeoutSeconds: 20/g' /etc/kubernetes/manifests/kube-apiserver.yaml &&
 sudo kubeadm init \
   --v=1 \
   --skip-phases=certs,kubeconfig,control-plane \
   --ignore-preflight-errors=all \
   --pod-network-cidr 10.244.0.0/16  

在此之后,我还安装了法兰绒.成功安装主服务器后,我继续进行kubeadm连接以添加其他两台计算机.添加这些机器后,我在主节点上安装了NGINX-Ingress.

After this I also installed flannel. After the master was successfully installed, I proceeded with the kubeadm join to add the other two machines. After these machines were added, I installed the NGINX-Ingress on the master node.

现在,我想重置集群并再次重新进行此设置.我使用"RESET COMMAND"(重置命令)重置每台计算机,然后在主节点上执行"INSTALL"命令.但是,在运行INSTALL命令并运行kubectl get pods --all-namespaces之后,我仍然可以看到以前安装的吊舱:

Now, I wanted to reset the cluster and to re-do this setup again. I reset each machine using the RESET COMMAND and proceeded with the INSTALL command on the master node. However, after I ran the INSTALL command and ran kubectl get pods --all-namespaces I can still see the pods from the previous installation:

NAMESPACE       NAME                              READY   STATUS              RESTARTS   AGE
kube-system     coredns-fb8b8dccf-h5hhk           0/1     ContainerCreating   1          20h
kube-system     coredns-fb8b8dccf-jblmv           0/1     ContainerCreating   1          20h
kube-system     etcd-ubuntu6                      1/1     Running             0          19h
kube-system     kube-apiserver-ubuntu6            1/1     Running             0          76m
kube-system     kube-controller-manager-ubuntu6   0/1     CrashLoopBackOff    7          75m
kube-system     kube-flannel-ds-amd64-4pqq6       1/1     Running             0          20h
kube-system     kube-flannel-ds-amd64-dvfmp       0/1     CrashLoopBackOff    7          20h
kube-system     kube-flannel-ds-amd64-dz9st       1/1     Terminating         0          20h
kube-system     kube-proxy-9vfjx                  1/1     Running             0          20h
kube-system     kube-proxy-q5c86                  1/1     Running             0          20h
kube-system     kube-proxy-zlw4v                  1/1     Running             0          20h
kube-system     kube-scheduler-ubuntu6            1/1     Running             0          76m
nginx-ingress   nginx-ingress-6957586bf6-fg2tt    0/1     Terminating         22         19h

为什么我看到以前安装的吊舱?

Why am I seeing the pods from the previous installation?

推荐答案

是的,基本上,当您使用

So yes, basically when you create a single control-plane cluster using kubeadm - you are installing cluster that has single control-plane node, with a single etcd database running on it.

在控制平面节点上,kubeadm使用的默认etcd目录为/var/lib/etcd.您应该清理它,以避免恢复以前的群集配置.

The default etcd directory used by kubeadm is /var/lib/etcd on the control-plane node. You should clean it up to avoid restoring previous cluster configuration.

顺便说一句,有同样的问题对于k8s 1.15.并且应该固定为1.15.1 https: //github.com/kubernetes/sig-release/blob/3a3c9f92ef484656f0cb4867f32491777d629952/releases/patch-releases.md#115

BTW, there is the same issue for k8s 1.15. And it should be fixed 1.15.1 https://github.com/kubernetes/sig-release/blob/3a3c9f92ef484656f0cb4867f32491777d629952/releases/patch-releases.md#115

这篇关于重置Kubernetes集群的文章就介绍到这了,希望我们推荐的答案对大家有所帮助,也希望大家多多支持IT屋!

查看全文
登录 关闭
扫码关注1秒登录
发送“验证码”获取 | 15天全站免登陆