1 2 3 4 5 作者:李晓辉 微信联系:lxh_chat 联系邮箱: 939958092@qq.com
角色 IP 主机名 操作系统平台 软件版本 硬件配置 k8s-master etcd haproxy keepalived 主要IP:192.168.8.128 keepalived vip: 192.168.8.200 k8s-master01.xiaohui.cn Rocky 9.4 kubernetes 1.31.1 haproxy 3.0.5 Keepalived 2.3.1 CPU: 2核心以上 内存:4G以上 硬盘:100G以上 k8s-master etcd haproxy keepalived 主要IP:192.168.8.129 keepalived vip: 192.168.8.200 k8s-master02.xiaohui.cn Rocky 9.4 kubernetes 1.31.1 haproxy 3.0.5 Keepalived 2.3.1 CPU: 2核心以上 内存:4G以上 硬盘:100G以上 k8s-master etcd haproxy keepalived 主要IP:192.168.8.130 keepalived vip: 192.168.8.200 k8s-master03.xiaohui.cn Rocky 9.4 kubernetes 1.31.1 haproxy 3.0.5 Keepalived 2.3.1 CPU: 2核心以上 内存:4G以上 硬盘:100G以上 k8s-master etcd haproxy keepalived 主要IP:192.168.8.131 keepalived vip: 192.168.8.200 k8s-master04.xiaohui.cn Rocky 9.4 kubernetes 1.31.1 haproxy 3.0.5 Keepalived 2.3.1 CPU: 2核心以上 内存:4G以上 硬盘:100G以上 k8s-worker 主要IP:192.168.8.132 k8s-worker01.xiaohui.cn Rocky 9.4 kubernetes 1.31.1 CPU: 2核心以上 内存:2G以上 硬盘:100G以上 仅供参考,以负载为准
这里是k8s-master04只是用于演示在token过期后,还怎么将master加入到集群,所以你可以认为本文档提供的是3节点的控制平面部署指南
文档拓扑描述 流量走向:Client—>Keepalived vip—>Haproxy—>—>K8s-master
添加域名解析 我的k8s-master04只是用于演示token过期后,如何加入更多节点而已,你集群的master最好是奇数个机器
1 2 3 4 5 6 7 cat << EOF >> /etc/hosts 192.168.8.128 k8s-master01.xiaohui.cn k8s-master01 192.168.8.129 k8s-master02.xiaohui.cn k8s-master02 192.168.8.130 k8s-master03.xiaohui.cn k8s-master03 192.168.8.131 k8s-master04.xiaohui.cn k8s-master04 192.168.8.132 k8s-worker01.xiaohui.cn k8s-worker01 EOF
分别给每个主机设置主机名,我这里以第一个master举例
1 hostnamectl hostname k8s-master01.xiaohui.cn
部署 Haproxy服务 Haproxy 在这里承担了K8S Master节点之间的负载均衡角色,需要在所有的控制平面机器上都安装
安装haproxy最新版 这里可以找到最新版的haproxy
1 https://github.com/haproxy/haproxy/tags
我写文章的时候,最新版是3.0.5,我用我自己的仓库做了加速,你用的时候在上面的链接找最新版
1 2 wget https://www.haproxy.org/download/3.0/src/haproxy-3.0.5.tar.gz tar xf haproxy-3.0.5.tar.gz
编译安装的时候还需要lua支持,最新版可以在下面的网址找到最新版
1 2 wget https://www.lua.org/ftp/lua-5.4.7.tar.gz tar xf lua-5.4.7.tar.gz
编译安装需要有make和gcc支持
1 2 3 yum clean all yum makecache yum install make gcc -y
编译安装lua
1 2 3 4 cd lua-5.4.7/make make install cd
为haproxy准备编译先决条件
执行这个apt install有可能会跳出界面让你选择,你可以用上下键选择,用空格勾选所有,然后tab键选中ok并回车
1 2 cd haproxy-3.0.5/yum install pcre2-devel openssl-devel -y
编译安装haproxy
1 2 3 make -j $(nproc ) TARGET=linux-glibc \ USE_OPENSSL=1 USE_QUIC=1 USE_QUIC_OPENSSL_COMPAT=1 \ USE_LUA=1 USE_PCRE2=1
准备haproxy配置文件 这里要注意,你在文本最后将k8s-master04这台机器加入到k8s集群确认没问题后,将其他所有master机器上的haproxy配置文件都更新一下,包含新的控制节点
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 mkdir /etc/haproxy /var/lib/haproxycat << EOF > /etc/haproxy/haproxy.cfg global log /dev/log local0 log /dev/log local1 notice daemon defaults mode http log global option httplog option dontlognull option http-server-close option forwardfor except 127.0.0.0/8 option redispatch retries 1 timeout http-request 10s timeout queue 20s timeout connect 5s timeout client 20s timeout server 20s timeout http-keep-alive 10s timeout check 10s frontend k8s-api bind *:64433 #非并置的情况下,这里写6443和k8s更一致 mode tcp option tcplog default_backend apiserver backend apiserver option httpchk GET /healthz http-check expect status 200 mode tcp option ssl-hello-chk balance roundrobin server k8s-master01 k8s-master01.xiaohui.cn:6443 check server k8s-master02 k8s-master02.xiaohui.cn:6443 check server k8s-master03 k8s-master03.xiaohui.cn:6443 check server k8s-master04 k8s-master04.xiaohui.cn:6443 check EOF
1 2 3 4 5 6 7 8 9 10 11 cat << EOF > /etc/systemd/system/haproxy.service [Unit] Description=HAProxy Load Balancer After=syslog.target network.target [Service] ExecStartPre=/usr/local/sbin/haproxy -f /etc/haproxy/haproxy.cfg -c -q ExecStart=/usr/local/sbin/haproxy -Ws -f /etc/haproxy/haproxy.cfg -p /var/lib/haproxy/haproxy.pid ExecReload=/bin/kill -USR2 $MAINPID [Install] WantedBy=multi-user.target EOF
启动haproxy服务 1 2 systemctl daemon-reload systemctl enable haproxy --now
部署 Keepalived服务 keepalived提供了一个虚拟IP,需要在所有的控制平面都安装
安装依赖包 1 2 3 yum install epel-release -y crb enable yum makecache
1 2 3 4 yum install -y make autoconf automake openssl-devel libnl3-devel \ iptables-devel ipset-devel file-devel net-snmp-devel \ glib2-devel pcre2-devel libnftnl-devel libmnl-devel \ systemd-devel kmod-devel NetworkManager-libnm-devel python-sphinx
这里应该会更新systemd软件,建议重启一下所有机器
编译安装keepalived 1 2 3 4 yum install git -y git clone https://github.com/acassen/keepalived.git cd keepalived./autogen.sh
准备keepalived配置文件 第一台 这是keepalived的检测脚本,主要是判断haproxy是否工作正常,这里需要注意你的网卡名称是否为ens160
1 2 3 4 5 6 7 8 9 10 11 cat <<- 'EOF' > /etc/keepalived/check_haproxy.shif systemctl is-active --quiet haproxy; then exit 0 else exit 1 fi EOF
以下为第一台keepalived配置文件
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 cat > /etc/keepalived/keepalived.conf <<- 'EOF' global_defs { router_id K8S_VIP } vrrp_script check_haproxy { script "/etc/keepalived/check_haproxy.sh" interval 2 weight 20 } vrrp_instance VI_1 { state MASTER interface ens160 virtual_router_id 51 priority 100 advert_int 1 authentication { auth_type PASS auth_pass 1111 } virtual_ipaddress { 192.168.8.200 } track_script { check_haproxy } accept } EOF
1 2 3 chmod +x /etc/keepalived/check_haproxy.shsystemctl daemon-reload systemctl enable keepalived --now
开通vrrp协议的放获取
1 2 firewall-cmd --add-protocol=vrrp --permanent firewall-cmd --reload
第二台 这是keepalived的检测脚本,主要是判断haproxy是否工作正常,这里需要注意你的网卡名称是否为ens160
1 2 3 4 5 6 7 8 9 10 11 cat <<- 'EOF' > /etc/keepalived/check_haproxy.shif systemctl is-active --quiet haproxy; then exit 0 else exit 1 fi EOF
以下为第二台keepalived配置文件
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 cat > /etc/keepalived/keepalived.conf <<- 'EOF' global_defs { router_id K8S_VIP } vrrp_script check_haproxy { script "/etc/keepalived/check_haproxy.sh" interval 2 weight 20 } vrrp_instance VI_1 { state BACKUP interface ens160 virtual_router_id 51 priority 99 advert_int 1 authentication { auth_type PASS auth_pass 1111 } virtual_ipaddress { 192.168.8.200 } track_script { check_haproxy } accept } EOF
1 2 3 chmod +x /etc/keepalived/check_haproxy.shsystemctl daemon-reload systemctl enable keepalived --now
开通vrrp协议的放获取
1 2 firewall-cmd --add-protocol=vrrp --permanent firewall-cmd --reload
第三台 这是keepalived的检测脚本,主要是判断haproxy是否工作正常,这里需要注意你的网卡名称是否为ens160
1 2 3 4 5 6 7 8 9 10 11 cat <<- 'EOF' > /etc/keepalived/check_haproxy.shif systemctl is-active --quiet haproxy; then exit 0 else exit 1 fi EOF
以下为第三台keepalived配置文件
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 cat > /etc/keepalived/keepalived.conf <<- 'EOF' global_defs { router_id K8S_VIP } vrrp_script check_haproxy { script "/etc/keepalived/check_haproxy.sh" interval 2 weight 20 } vrrp_instance VI_1 { state BACKUP interface ens160 virtual_router_id 51 priority 98 advert_int 1 authentication { auth_type PASS auth_pass 1111 } virtual_ipaddress { 192.168.8.200 } track_script { check_haproxy } accept } EOF
1 2 3 chmod +x /etc/keepalived/check_haproxy.shsystemctl daemon-reload systemctl enable keepalived --now
开通vrrp协议的防火墙
1 2 firewall-cmd --add-protocol=vrrp --permanent firewall-cmd --reload
部署高可用Kubernetes 先决条件 先决条件的部分,在所有k8s节点都要完成
禁用交换分区 1 2 sed -ri 's/.*swap.*/#&/' /etc/fstab swapoff -a
部署Docker和CRI-Docker 部署docker
1 2 3 4 5 6 7 wget -O /etc/yum.repos.d/docker-ce.repo https://mirrors.nju.edu.cn/docker-ce/linux/rhel/docker-ce.repo sed -i 's|download.docker.com|mirrors.nju.edu.cn/docker-ce|g' /etc/yum.repos.d/docker-ce.repo yum install docker-ce docker-ce-cli containerd.io docker-buildx-plugin docker-compose-plugin -y systemctl daemon-reload systemctl restart docker.service systemctl enable docker.service
部署CRI-Docker
1 2 3 4 5 6 7 8 9 10 cat > /etc/yum.repos.d/fedora36.repo <<-'EOF' [fedora36] name=fedora36 repo baseurl=https://mirrors.nju.edu.cn/fedora-archive/fedora/linux/releases/36/Everything/x86_64/os/ enabled=1 gpgcheck=0 EOF wget https://github.com/Mirantis/cri-dockerd/releases/download/v0.3.15/cri-dockerd-0.3.15-3.fc36.x86_64.rpm yum localinstall cri-dockerd-0.3.15-3.fc36.x86_64.rpm -y
将镜像指引到国内
1 2 3 4 5 sed -i 's|ExecStart=.*|ExecStart=/usr/bin/cri-dockerd --container-runtime-endpoint fd:// --network-plugin=cni --pod-infra-container-image=registry.cn-hangzhou.aliyuncs.com/google_containers/pause:3.10|' /lib/systemd/system/cri-docker.service systemctl daemon-reload systemctl restart cri-docker.service systemctl enable cri-docker.service
安装kubeadm工具 1 2 3 4 5 6 7 8 9 10 cat > /etc/yum.repos.d/k8s.repo <<EOF [k8s] name=k8s repo baseurl=https://mirrors.aliyun.com/kubernetes-new/core/stable/v1.31/rpm enabled=1 gpgcheck=0 EOF yum install -y kubelet kubeadm kubectl socat systemctl enable kubelet --now
添加命令自动补齐功能
1 2 3 4 kubectl completion bash > /etc/bash_completion.d/kubectl kubeadm completion bash > /etc/bash_completion.d/kubeadm source /etc/bash_completion.d/kubectlsource /etc/bash_completion.d/kubeadm
集成CRI-Docker
1 2 crictl config runtime-endpoint unix:///run/cri-dockerd.sock crictl images
允许 iptables 检查桥接流量 1 2 3 4 5 6 7 8 9 10 cat <<EOF | sudo tee /etc/modules-load.d/k8s.conf br_netfilter EOF modprobe br_netfilter cat <<EOF | sudo tee /etc/sysctl.d/k8s.conf net.bridge.bridge-nf-call-ip6tables = 1 net.bridge.bridge-nf-call-iptables = 1 net.ipv4.ip_forward = 1 EOF sudo sysctl --system
开通Master节点防火墙 这里的防火墙是master所需要的
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 firewall-cmd --zone=public --add-port=6443/tcp --permanent firewall-cmd --zone=public --add-port=2379-2380/tcp --permanent firewall-cmd --zone=public --add-port=10250/tcp --permanent firewall-cmd --zone=public --add-port=10259/tcp --permanent firewall-cmd --zone=public --add-port=10257/tcp --permanent firewall-cmd --zone=public --add-port=64433/tcp --permanent firewall-cmd --add-port=179/tcp --permanent firewall-cmd --add-port=4789/udp --permanent firewall-cmd --add-port=5473/tcp --permanent firewall-cmd --add-port=9099/tcp --permanent firewall-cmd --reload
集群部署 这个部署只在第一台k8s-master01完成,其他的master稍后以join的方式加入到集群中
1 2 3 4 5 6 7 8 9 10 kubeadm init \ --apiserver-advertise-address=192.168.8.128 \ --apiserver-bind-port=6443 \ --control-plane-endpoint=192.168.8.200:64433 \ --image-repository=registry.cn-hangzhou.aliyuncs.com/google_containers \ --kubernetes-version=v1.31.1 \ --service-cidr=10.96.0.0/12 \ --service-dns-domain=lixiaohui.cn \ --cri-socket unix:///var/run/cri-dockerd.sock \ --upload-certs
部分输出内容:
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 Your Kubernetes control-plane has initialized successfully! To start using your cluster, you need to run the following as a regular user: mkdir -p $HOME/.kube sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config sudo chown $(id -u):$(id -g) $HOME/.kube/config ... You can now join any number of the control-plane node running the following command on each as root: kubeadm join 192.168.8.200:64433 --token b8iv94.qg3nhbioxz6e0lzr \ --discovery-token-ca-cert-hash sha256:2fd9a0461559cbb7a018ccb47350531d4d7ac4419458f4766694d9d05ba4af9c \ --control-plane --certificate-key 7d99ab3cec720c9ac0c4fe1b929942acc0bad3bfbf5ad3a8bab068422ac7ea5c ... Then you can join any number of worker nodes by running the following on each as root: kubeadm join 192.168.8.200:64433 --token b8iv94.qg3nhbioxz6e0lzr \ --discovery-token-ca-cert-hash sha256:2fd9a0461559cbb7a018ccb47350531d4d7ac4419458f4766694d9d05ba4af9c
加入更多Master 所有的Master节点都必须得完成本地hosts添加、Haproxy、Keepalived部署,以及k8s部署的所有先决条件,请注意在haproxy的后端添加包含新节点的所有后端,在keepalived配置文件中的优先级继续减1
第二台和第三台都输入下面的命令即可,下面的命令也是从第一台部署好之后的输出中复制粘贴的,注意加了–cri-socket参数
1 2 3 4 kubeadm join 192.168.8.200:64433 --token b8iv94.qg3nhbioxz6e0lzr \ --discovery-token-ca-cert-hash sha256:2fd9a0461559cbb7a018ccb47350531d4d7ac4419458f4766694d9d05ba4af9c \ --control-plane --certificate-key 7d99ab3cec720c9ac0c4fe1b929942acc0bad3bfbf5ad3a8bab068422ac7ea5c \ --cri-socket unix:///var/run/cri-dockerd.sock
输出内容:
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 This node has joined the cluster and a new control plane instance was created: * Certificate signing request was sent to apiserver and approval was received. * The Kubelet was informed of the new secure connection details. * Control plane label and taint were applied to the new node. * The Kubernetes control plane instances scaled up. * A new etcd member was added to the local/stacked etcd cluster. To start administering your cluster from this node, you need to run the following as a regular user: mkdir -p $HOME/.kube sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config sudo chown $(id -u):$(id -g) $HOME/.kube/config Run 'kubectl get nodes' to see this node join the cluster.
这个加入过程可能会有一些警告,不过只要在授予管理权限以及部署完calico网络后,获取节点状态和kube-system下的pod状态没问题就可以忽略警告
在每台上,都授予自己管理权限
1 2 3 mkdir -p $HOME /.kubesudo cp -i /etc/kubernetes/admin.conf $HOME /.kube/configsudo chown $(id -u):$(id -g) $HOME /.kube/config
部署Calico网络插件 这个在任何一台master上做都行,只需要做一次,这里会从国外下载容器镜像,需要自己解决镜像下载问题
1 kubectl create -f https://docs.projectcalico.org/manifests/calico.yaml
在这里稍等一会儿,等待容器镜像下载完毕,在此期间,你可以用下面的方式查看系统pod,比如我的,可以看到还在努力初始化中,继续等待即可
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 [root@k8s-master01 ~]# kubectl get pod -A NAMESPACE NAME READY STATUS RESTARTS AGE kube-system calico-kube-controllers-564b9d64dd-f9mnm 0/1 ContainerCreating 0 23s kube-system calico-node-4vwv8 0/1 Init:2/3 0 23s kube-system calico-node-gl7wr 0/1 Init:0/3 0 23s kube-system calico-node-pz9tb 0/1 Init:2/3 0 23s kube-system coredns-fcd6c9c4-7gvqd 0/1 ContainerCreating 0 2m49s kube-system coredns-fcd6c9c4-9925q 0/1 ContainerCreating 0 2m49s kube-system etcd-k8s-master01.xiaohui.cn 1/1 Running 0 2m54s kube-system etcd-k8s-master02.xiaohui.cn 1/1 Running 0 75s kube-system etcd-k8s-master03.xiaohui.cn 1/1 Running 0 63s kube-system kube-apiserver-k8s-master01.xiaohui.cn 1/1 Running 0 2m54s kube-system kube-apiserver-k8s-master02.xiaohui.cn 1/1 Running 0 75s kube-system kube-apiserver-k8s-master03.xiaohui.cn 1/1 Running 0 63s kube-system kube-controller-manager-k8s-master01.xiaohui.cn 1/1 Running 0 2m54s kube-system kube-controller-manager-k8s-master02.xiaohui.cn 1/1 Running 0 75s kube-system kube-controller-manager-k8s-master03.xiaohui.cn 1/1 Running 0 63s kube-system kube-proxy-9dj5v 1/1 Running 0 70s kube-system kube-proxy-b675s 1/1 Running 0 76s kube-system kube-proxy-hsskk 1/1 Running 0 2m50s kube-system kube-scheduler-k8s-master01.xiaohui.cn 1/1 Running 0 2m54s kube-system kube-scheduler-k8s-master02.xiaohui.cn 1/1 Running 0 75s kube-system kube-scheduler-k8s-master03.xiaohui.cn 1/1 Running 0 63s
初始化好之后:
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 root@k8s-master01:~# kubectl get nodes NAME STATUS ROLES AGE VERSION k8s-master01.xiaohui.cn Ready control-plane 7m11s v1.31.1 k8s-master02.xiaohui.cn Ready control-plane 5m32s v1.31.1 k8s-master03.xiaohui.cn Ready control-plane 5m26s v1.31.1 root@k8s-master01:~# kubectl get pod -A NAMESPACE NAME READY STATUS RESTARTS AGE kube-system calico-kube-controllers-564b9d64dd-f9mnm 1/1 Running 0 4m3s kube-system calico-node-4vwv8 1/1 Running 0 4m3s kube-system calico-node-gl7wr 1/1 Running 0 4m3s kube-system calico-node-pz9tb 1/1 Running 0 4m3s kube-system coredns-fcd6c9c4-7gvqd 1/1 Running 0 6m29s kube-system coredns-fcd6c9c4-9925q 1/1 Running 0 6m29s kube-system etcd-k8s-master01.xiaohui.cn 1/1 Running 0 6m34s kube-system etcd-k8s-master02.xiaohui.cn 1/1 Running 0 4m55s kube-system etcd-k8s-master03.xiaohui.cn 1/1 Running 0 4m43s kube-system kube-apiserver-k8s-master01.xiaohui.cn 1/1 Running 0 6m34s kube-system kube-apiserver-k8s-master02.xiaohui.cn 1/1 Running 0 4m55s kube-system kube-apiserver-k8s-master03.xiaohui.cn 1/1 Running 0 4m43s kube-system kube-controller-manager-k8s-master01.xiaohui.cn 1/1 Running 0 6m34s kube-system kube-controller-manager-k8s-master02.xiaohui.cn 1/1 Running 0 4m55s kube-system kube-controller-manager-k8s-master03.xiaohui.cn 1/1 Running 0 4m43s kube-system kube-proxy-9dj5v 1/1 Running 0 4m50s kube-system kube-proxy-b675s 1/1 Running 0 4m56s kube-system kube-proxy-hsskk 1/1 Running 0 6m30s kube-system kube-scheduler-k8s-master01.xiaohui.cn 1/1 Running 0 6m34s kube-system kube-scheduler-k8s-master02.xiaohui.cn 1/1 Running 0 4m55s kube-system kube-scheduler-k8s-master03.xiaohui.cn 1/1 Running 0 4m43s
重新生成加入master命令 第一台初始化好24小时后,token什么的就会失效,24小时后加入更多master或者worker节点可以用下面的方式,需要注意的是,新加入的所有节点,都必须完成所有关于部署k8s的先决条件
加入master除了需要token之外,还需要certificate key
在现有的master上,生成token和certificate key
1 2 3 4 5 6 7 8 [root@k8s-master01 ~]# kubeadm token create --print-join-command kubeadm join 192.168.8.200:64433 --token 1bfxvf.x48fpyrbifxh39ot --discovery-token-ca-cert-hash sha256:2fd9a0461559cbb7a018ccb47350531d4d7ac4419458f4766694d9d05ba4af9c [root@k8s-master01 ~]# kubeadm init phase upload-certs --upload-certs [upload-certs] Storing the certificates in Secret "kubeadm-certs" in the "kube-system" Namespace [upload-certs] Using certificate key: ab028cb7a36bcbb6a7a9df0b74ab208d5d4825e2bb70d81fde2d8ba519c80159
把上面两个拼接起来,然后额外加入–certificate-key和–control-plane就行了,由于我们安装的是docker和cri-docker,所以还得加一个–cri-socket
1 2 3 4 5 kubeadm join 192.168.8.200:64433 --token 1bfxvf.x48fpyrbifxh39ot \ --discovery-token-ca-cert-hash sha256:2fd9a0461559cbb7a018ccb47350531d4d7ac4419458f4766694d9d05ba4af9c \ --certificate-key ab028cb7a36bcbb6a7a9df0b74ab208d5d4825e2bb70d81fde2d8ba519c80159 \ --control-plane \ --cri-socket unix:///var/run/cri-dockerd.sock
授予自己管理权限
1 2 3 mkdir -p $HOME /.kubesudo cp -i /etc/kubernetes/admin.conf $HOME /.kube/configsudo chown $(id -u):$(id -g) $HOME /.kube/config
查询节点
1 2 3 4 5 6 [root@k8s-master04 ~]# kubectl get nodes NAME STATUS ROLES AGE VERSION k8s-master01.xiaohui.cn Ready control-plane 59m v1.31.1 k8s-master02.xiaohui.cn Ready control-plane 57m v1.31.1 k8s-master03.xiaohui.cn Ready control-plane 57m v1.31.1 k8s-master04.xiaohui.cn Ready control-plane 37s v1.31.1
加入worker节点 需要注意的是,每个需要加入的worker,也必须完成本地hosts添加和部署Kubernetes中的所有先决条件,不然不能加入
开通Worker防火墙
完成以上所有在worker上的任务后,为worker节点开通防火墙
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 firewall-cmd --zone=public --add-port=10250/tcp --permanent firewall-cmd --zone=public --add-port=10256/tcp --permanent firewall-cmd --zone=public --add-port=30000-32767/tcp --permanent firewall-cmd --add-port=179/tcp --permanent firewall-cmd --add-port=4789/udp --permanent firewall-cmd --add-port=5473/tcp --permanent firewall-cmd --add-port=9099/tcp --permanent firewall-cmd --reload
在现有的master上,生成token,worker节点不需要certificate key
1 2 [root@k8s-master01 ~]# kubeadm token create --print-join-command kubeadm join 192.168.8.200:64433 --token r00nr4.bvjdc1lor5kuv5hh --discovery-token-ca-cert-hash sha256:2fd9a0461559cbb7a018ccb47350531d4d7ac4419458f4766694d9d05ba4af9c
在新的worker节点加入
1 2 3 kubeadm join 192.168.8.200:64433 --token r00nr4.bvjdc1lor5kuv5hh \ --discovery-token-ca-cert-hash sha256:2fd9a0461559cbb7a018ccb47350531d4d7ac4419458f4766694d9d05ba4af9c \ --cri-socket unix:///var/run/cri-dockerd.sock
获取节点列表
给worker节点打上worker标签,并查询节点和pod
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 root@k8s-master01:~# kubectl label nodes k8s-worker01.xiaohui.cn node-role.kubernetes.io/Worker= node/k8s-worker01.xiaohui.cn labeled root@k8s-master01:~# kubectl get nodes NAME STATUS ROLES AGE VERSION k8s-master01.xiaohui.cn Ready control-plane 87m v1.31.1 k8s-master02.xiaohui.cn Ready control-plane 83m v1.31.1 k8s-master03.xiaohui.cn Ready control-plane 83m v1.31.1 k8s-master04.xiaohui.cn Ready control-plane 27m v1.31.1 k8s-worker01.xiaohui.cn Ready Worker 104s v1.31.1 root@k8s-master01:~# kubectl get pod -A NAMESPACE NAME READY STATUS RESTARTS AGE kube-system calico-kube-controllers-564b9d64dd-gsdv6 1/1 Running 0 80m kube-system calico-node-8m4g7 1/1 Running 0 27m kube-system calico-node-hvn97 1/1 Running 0 80m kube-system calico-node-lh8q6 1/1 Running 0 80m kube-system calico-node-nrvzp 1/1 Running 0 80m kube-system calico-node-zwwfz 1/1 Running 0 109s kube-system coredns-fcd6c9c4-2qmdt 1/1 Running 0 87m kube-system coredns-fcd6c9c4-6gxq5 1/1 Running 0 87m kube-system etcd-k8s-master01.xiaohui.cn 1/1 Running 0 87m kube-system etcd-k8s-master02.xiaohui.cn 1/1 Running 0 83m kube-system etcd-k8s-master03.xiaohui.cn 1/1 Running 0 83m kube-system etcd-k8s-master04.xiaohui.cn 1/1 Running 0 10m kube-system kube-apiserver-k8s-master01.xiaohui.cn 1/1 Running 0 87m kube-system kube-apiserver-k8s-master02.xiaohui.cn 1/1 Running 0 83m kube-system kube-apiserver-k8s-master03.xiaohui.cn 1/1 Running 0 83m kube-system kube-apiserver-k8s-master04.xiaohui.cn 1/1 Running 0 10m kube-system kube-controller-manager-k8s-master01.xiaohui.cn 1/1 Running 0 87m kube-system kube-controller-manager-k8s-master02.xiaohui.cn 1/1 Running 0 83m kube-system kube-controller-manager-k8s-master03.xiaohui.cn 1/1 Running 0 83m kube-system kube-controller-manager-k8s-master04.xiaohui.cn 1/1 Running 0 10m kube-system kube-proxy-4vpqw 1/1 Running 0 83m kube-system kube-proxy-bfl4c 1/1 Running 0 27m kube-system kube-proxy-mx2nq 1/1 Running 0 109s kube-system kube-proxy-rwbvv 1/1 Running 0 87m kube-system kube-proxy-tst26 1/1 Running 0 83m kube-system kube-scheduler-k8s-master01.xiaohui.cn 1/1 Running 0 87m kube-system kube-scheduler-k8s-master02.xiaohui.cn 1/1 Running 0 83m kube-system kube-scheduler-k8s-master03.xiaohui.cn 1/1 Running 0 83m kube-system kube-scheduler-k8s-master04.xiaohui.cn 1/1 Running 0 10m