《空姐背后最新章节列表-空姐背后最新章节目录-笔趣阁》剧情介绍:阿米拉蒂自己都惊呆了旁观的观众更是目瞪口呆:这到底是个什么操作兄弟你这是在跳高不是在练习芭蕾舞好吧跟小弟弟杠上了空姐背后最新章节列表-空姐背后最新章节目录-笔趣阁洪杰说道正是如此听说上一次他们仅仅以一小片海墓残片进入其中受伤惨重有一名元婴长老都险些没出来怕是想要借这些招募的人的力量进入其中刚复出就被抵制这次口碑崩塌的蒙嘉慧连郑伊健也救不了
《空姐背后最新章节列表-空姐背后最新章节目录-笔趣阁》视频说明:说完自腰间摸出一个小木筒但听砰的一声大响五彩烟火洒向天空kubernetes(k8s)集群安装部署手册原创2023-09-09 17:24·运维生活日志目录一 集群部署过程 11 版本以及规划信息 12 集群部署 13异常节点处理: 13一 集群部署过程1 版本以及规划信息1.1 版本信息名称 版本号内核 3.10.0-1160.el7.x86_64系统 CentOS Linux release 7.9.2009 (Core)docker 20.10.9kubeadm 1.22.0kubelet 1.22.0kubectl 1.22.01.2 服务器信息规划IP 主机名192.168.10.128 master192.168.10.129 node-01192.168.10.130 node-022 集群部署说明: 第一~第八步 三个节点均操作;第九十步master节点操作第十一步骤node节点操作2.1 关闭防火墙操作systemctl stop firewalld;systemctl disable firewalld2.2 关闭selinuxsetenforce 0 #临时关闭sed -i '/SELINUX/s/enforcing/disabled/g' /etc/sysconfig/selinux #永久关闭2.3 关闭swapswapoff -a #临时关闭sed -ri 's/.*swap.*/#&/' /etc/fstab #永久关闭2.4 修改hosts文件cat >> /etc/hosts < /etc/sysctl.d/k8s.conf << EOFnet.bridge.bridge-nf-call-ip6tables = 1net.bridge.bridge-nf-call-iptables = 1EOFsysctl --system2.6 安装dockerwget https://mirrors.aliyun.com/docker-ce/linux/centos/docker-ce.repo -O /etc/yum.repos.d/docker-ce.repoyum list docker-ce --showduplicates | sort -r #查看docker相关版本yum install docker-ce-20.10.9 docker-ce-cli-20.10.9 #安装指定的版本systemctl start dockersystemctl enable dockerdocker info|grep cgroup此处可能存在的问题 docker版本和kubelete存在兼容性问题解决方案一{"exec-opts": ["native.cgroupdriver=systemd"]}解决方案二# docker.servicevi /usr/lib/systemd/system/docker.service# 添加ExecStart=/usr/bin/dockerd --exec-opt native.cgroupdriver=systemd2.7 添加阿里云yum软件源cat > /etc/yum.repos.d/kubernetes.repo << EOF[kubernetes]name=Kubernetesbaseurl=https://mirrors.aliyun.com/kubernetes/yum/repos/kubernetes-el7-x86_64enabled=1gpgcheck=0repo_gpgcheck=1gpgkey=https://mirrors.aliyun.com/kubernetes/yum/doc/yum-key.gpghttps://mirrors.aliyun.com/kubernetes/yum/doc/rpm-package-key.gpgEOF2.8 安装kubeadm、kubelet、kubectlyum install -y kubelet-1.18.0 kubectl-1.18.0 kubeadm-1.18.0systemctl enable kubelet2.9 初始化master节点kubeadm init \--image-repository=registry.cn-hangzhou.aliyuncs.com/google_containers \--apiserver-advertise-address=192.168.10.128 \--kubernetes-version=v1.18.0 \--service-cidr=10.96.0.0/12 \--pod-network-cidr=10.244.0.0/16 \--v=6mkdir -p $HOME/.kubesudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/configsudo chown $(id -u):$(id -g) $HOME/.kube/config# 第1条命令:表示创建隐藏文件夹.kube# 第2条命令:表示复制文件admin.conf到该文件夹下# 第3条命令:表示设置文件的所属用户、用户组[root@master ~]# kubeadm init \> --image-repository=registry.cn-hangzhou.aliyuncs.com/google_containers \> --apiserver-advertise-address=192.168.10.128 \> --kubernetes-version=v1.18.0 \> --service-cidr=10.96.0.0/12 \> --pod-network-cidr=10.244.0.0/16 \> --v=6I0730 00:35:34.465349 26797 initconfiguration.go:103] detected and using CRI socket: /var/run/dockershim.sockW0730 00:35:34.467144 26797 configset.go:202] WARNING: kubeadm cannot validate component configs for API groups [kubelet.config.k8s.io kubeproxy.config.k8s.io][init] Using Kubernetes version: v1.18.0[preflight] Running pre-flight checksI0730 00:35:34.468752 26797 checks.go:577] validating Kubernetes and kubeadm versionI0730 00:35:34.468817 26797 checks.go:166] validating if the firewall is enabled and activeI0730 00:35:34.495833 26797 checks.go:201] validating availability of port 6443I0730 00:35:34.498186 26797 checks.go:201] validating availability of port 10259I0730 00:35:34.498364 26797 checks.go:201] validating availability of port 10257I0730 00:35:34.498438 26797 checks.go:286] validating the existence of file /etc/kubernetes/manifests/kube-apiserver.yamlI0730 00:35:34.498478 26797 checks.go:286] validating the existence of file /etc/kubernetes/manifests/kube-controller-manager.yamlI0730 00:35:34.498495 26797 checks.go:286] validating the existence of file /etc/kubernetes/manifests/kube-scheduler.yamlI0730 00:35:34.498511 26797 checks.go:286] validating the existence of file /etc/kubernetes/manifests/etcd.yamlI0730 00:35:34.498542 26797 checks.go:432] validating if the connectivity type is via proxy or directI0730 00:35:34.498639 26797 checks.go:471] validating http connectivity to first IP address in the CIDRI0730 00:35:34.498692 26797 checks.go:471] validating http connectivity to first IP address in the CIDRI0730 00:35:34.498729 26797 checks.go:102] validating the container runtimeI0730 00:35:34.749587 26797 checks.go:128] validating if the service is enabled and active[WARNING IsDockerSystemdCheck]: detected "cgroupfs" as the Docker cgroup driver. The recommended driver is "systemd". Please follow the guide at https://kubernetes.io/docs/setup/cri/I0730 00:35:35.004904 26797 checks.go:335] validating the contents of file /proc/sys/net/bridge/bridge-nf-call-iptablesI0730 00:35:35.005062 26797 checks.go:335] validating the contents of file /proc/sys/net/ipv4/ip_forwardI0730 00:35:35.005204 26797 checks.go:649] validating whether swap is enabled or notI0730 00:35:35.005266 26797 checks.go:376] validating the presence of executable conntrackI0730 00:35:35.005761 26797 checks.go:376] validating the presence of executable ipI0730 00:35:35.006005 26797 checks.go:376] validating the presence of executable iptablesI0730 00:35:35.006037 26797 checks.go:376] validating the presence of executable mountI0730 00:35:35.006272 26797 checks.go:376] validating the presence of executable nsenterI0730 00:35:35.006587 26797 checks.go:376] validating the presence of executable ebtablesI0730 00:35:35.006621 26797 checks.go:376] validating the presence of executable ethtoolI0730 00:35:35.006642 26797 checks.go:376] validating the presence of executable socatI0730 00:35:35.006666 26797 checks.go:376] validating the presence of executable tcI0730 00:35:35.006684 26797 checks.go:376] validating the presence of executable touchI0730 00:35:35.006709 26797 checks.go:520] running all checks[WARNING SystemVerification]: this Docker version is not on the list of validated versions: 20.10.9. Latest validated version: 19.03I0730 00:35:35.243646 26797 checks.go:406] checking whether the given node name is reachable using net.LookupHostI0730 00:35:35.244443 26797 checks.go:618] validating kubelet versionI0730 00:35:35.388021 26797 checks.go:128] validating if the service is enabled and activeI0730 00:35:35.407445 26797 checks.go:201] validating availability of port 10250I0730 00:35:35.407861 26797 checks.go:201] validating availability of port 2379I0730 00:35:35.408055 26797 checks.go:201] validating availability of port 2380I0730 00:35:35.408214 26797 checks.go:249] validating the existence and emptiness of directory /var/lib/etcd[preflight] Pulling images required for setting up a Kubernetes cluster[preflight] This might take a minute or two, depending on the speed of your internet connection[preflight] You can also perform this action in beforehand using 'kubeadm config images pull'I0730 00:35:35.508189 26797 checks.go:844] pulling registry.cn-hangzhou.aliyuncs.com/google_containers/kube-apiserver:v1.18.0I0730 00:36:37.874415 26797 checks.go:844] pulling registry.cn-hangzhou.aliyuncs.com/google_containers/kube-controller-manager:v1.18.0I0730 00:37:22.204601 26797 checks.go:844] pulling registry.cn-hangzhou.aliyuncs.com/google_containers/kube-scheduler:v1.18.0I0730 00:37:54.572787 26797 checks.go:844] pulling registry.cn-hangzhou.aliyuncs.com/google_containers/kube-proxy:v1.18.0I0730 00:38:42.961883 26797 checks.go:844] pulling registry.cn-hangzhou.aliyuncs.com/google_containers/pause:3.2I0730 00:38:57.533578 26797 checks.go:844] pulling registry.cn-hangzhou.aliyuncs.com/google_containers/etcd:3.4.3-0I0730 00:40:27.427614 26797 checks.go:844] pulling registry.cn-hangzhou.aliyuncs.com/google_containers/coredns:1.6.7I0730 00:40:49.102866 26797 kubelet.go:64] Stopping the kubelet[kubelet-start] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env"[kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml"[kubelet-start] Starting the kubelet[certs] Using certificateDir folder "/etc/kubernetes/pki"I0730 00:40:49.560999 26797 certs.go:103] creating a new certificate authority for ca[certs] Generating "ca" certificate and key[certs] Generating "apiserver" certificate and key[certs] apiserver serving cert is signed for DNS names [master kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local] and IPs [10.96.0.1 192.168.10.128][certs] Generating "apiserver-kubelet-client" certificate and keyI0730 00:40:51.171593 26797 certs.go:103] creating a new certificate authority for front-proxy-ca[certs] Generating "front-proxy-ca" certificate and key[certs] Generating "front-proxy-client" certificate and keyI0730 00:40:52.331002 26797 certs.go:103] creating a new certificate authority for etcd-ca[certs] Generating "etcd/ca" certificate and key[certs] Generating "etcd/server" certificate and key[certs] etcd/server serving cert is signed for DNS names [master localhost] and IPs [192.168.10.128 127.0.0.1 ::1][certs] Generating "etcd/peer" certificate and key[certs] etcd/peer serving cert is signed for DNS names [master localhost] and IPs [192.168.10.128 127.0.0.1 ::1][certs] Generating "etcd/healthcheck-client" certificate and key[certs] Generating "apiserver-etcd-client" certificate and keyI0730 00:40:54.818549 26797 certs.go:69] creating new public/private key files for signing service account users[certs] Generating "sa" key and public key[kubeconfig] Using kubeconfig folder "/etc/kubernetes"I0730 00:40:55.968722 26797 kubeconfig.go:79] creating kubeconfig file for admin.conf[kubeconfig] Writing "admin.conf" kubeconfig fileI0730 00:40:56.245786 26797 kubeconfig.go:79] creating kubeconfig file for kubelet.conf[kubeconfig] Writing "kubelet.conf" kubeconfig fileI0730 00:40:56.621396 26797 kubeconfig.go:79] creating kubeconfig file for controller-manager.conf[kubeconfig] Writing "controller-manager.conf" kubeconfig fileI0730 00:40:56.818129 26797 kubeconfig.go:79] creating kubeconfig file for scheduler.conf[kubeconfig] Writing "scheduler.conf" kubeconfig file[control-plane] Using manifest folder "/etc/kubernetes/manifests"[control-plane] Creating static Pod manifest for "kube-apiserver"I0730 00:40:57.162267 26797 manifests.go:91] [control-plane] getting StaticPodSpecsI0730 00:40:57.163474 26797 manifests.go:104] [control-plane] adding volume "ca-certs" for component "kube-apiserver"I0730 00:40:57.163497 26797 manifests.go:104] [control-plane] adding volume "etc-pki" for component "kube-apiserver"I0730 00:40:57.163507 26797 manifests.go:104] [control-plane] adding volume "k8s-certs" for component "kube-apiserver"I0730 00:40:57.176186 26797 manifests.go:121] [control-plane] wrote static Pod manifest for component "kube-apiserver" to "/etc/kubernetes/manifests/kube-apiserver.yaml"[control-plane] Creating static Pod manifest for "kube-controller-manager"I0730 00:40:57.176232 26797 manifests.go:91] [control-plane] getting StaticPodSpecsW0730 00:40:57.176346 26797 manifests.go:225] the default kube-apiserver authorization-mode is "Node,RBAC"; using "Node,RBAC"I0730 00:40:57.176680 26797 manifests.go:104] [control-plane] adding volume "ca-certs" for component "kube-controller-manager"I0730 00:40:57.176688 26797 manifests.go:104] [control-plane] adding volume "etc-pki" for component "kube-controller-manager"I0730 00:40:57.176693 26797 manifests.go:104] [control-plane] adding volume "flexvolume-dir" for component "kube-controller-manager"I0730 00:40:57.176697 26797 manifests.go:104] [control-plane] adding volume "k8s-certs" for component "kube-controller-manager"I0730 00:40:57.176702 26797 manifests.go:104] [control-plane] adding volume "kubeconfig" for component "kube-controller-manager"I0730 00:40:57.177947 26797 manifests.go:121] [control-plane] wrote static Pod manifest for component "kube-controller-manager" to "/etc/kubernetes/manifests/kube-controller-manager.yaml"[control-plane] Creating static Pod manifest for "kube-scheduler"I0730 00:40:57.177980 26797 manifests.go:91] [control-plane] getting StaticPodSpecsW0730 00:40:57.178062 26797 manifests.go:225] the default kube-apiserver authorization-mode is "Node,RBAC"; using "Node,RBAC"I0730 00:40:57.178409 26797 manifests.go:104] [control-plane] adding volume "kubeconfig" for component "kube-scheduler"I0730 00:40:57.179219 26797 manifests.go:121] [control-plane] wrote static Pod manifest for component "kube-scheduler" to "/etc/kubernetes/manifests/kube-scheduler.yaml"[etcd] Creating static Pod manifest for local etcd in "/etc/kubernetes/manifests"I0730 00:40:57.180242 26797 local.go:72] [etcd] wrote Static Pod manifest for a local etcd member to "/etc/kubernetes/manifests/etcd.yaml"I0730 00:40:57.180284 26797 waitcontrolplane.go:87] [wait-control-plane] Waiting for the API server to be healthyI0730 00:40:57.181798 26797 loader.go:375] Config loaded from file: /etc/kubernetes/admin.conf[wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory "/etc/kubernetes/manifests". This can take up to 4m0sI0730 00:40:57.185824 26797 round_trippers.go:443] GET https://192.168.10.128:6443/healthz?timeout=10s in 0 millisecondsI0730 00:40:57.687883 26797 round_trippers.go:443] GET https://192.168.10.128:6443/healthz?timeout=10s in 0 millisecondsI0730 00:40:58.187689 26797 round_trippers.go:443] GET https://192.168.10.128:6443/healthz?timeout=10s in 0 millisecondsI0730 00:40:58.687505 26797 round_trippers.go:443] GET https://192.168.10.128:6443/healthz?timeout=10s in 0 millisecondsI0730 00:40:59.187708 26797 round_trippers.go:443] GET https://192.168.10.128:6443/healthz?timeout=10s in 0 millisecondsI0730 00:40:59.687437 26797 round_trippers.go:443] GET https://192.168.10.128:6443/healthz?timeout=10s in 0 millisecondsI0730 00:41:00.187753 26797 round_trippers.go:443] GET https://192.168.10.128:6443/healthz?timeout=10s in 0 millisecondsI0730 00:41:00.687830 26797 round_trippers.go:443] GET https://192.168.10.128:6443/healthz?timeout=10s in 0 millisecondsI0730 00:41:01.187695 26797 round_trippers.go:443] GET https://192.168.10.128:6443/healthz?timeout=10s in 0 millisecondsI0730 00:41:01.686916 26797 round_trippers.go:443] GET https://192.168.10.128:6443/healthz?timeout=10s in 0 millisecondsI0730 00:41:02.188601 26797 round_trippers.go:443] GET https://192.168.10.128:6443/healthz?timeout=10s in 0 millisecondsI0730 00:41:02.686665 26797 round_trippers.go:443] GET https://192.168.10.128:6443/healthz?timeout=10s in 0 millisecondsI0730 00:41:03.187222 26797 round_trippers.go:443] GET https://192.168.10.128:6443/healthz?timeout=10s in 0 millisecondsI0730 00:41:03.688352 26797 round_trippers.go:443] GET https://192.168.10.128:6443/healthz?timeout=10s in 0 millisecondsI0730 00:41:04.187550 26797 round_trippers.go:443] GET https://192.168.10.128:6443/healthz?timeout=10s in 1 millisecondsI0730 00:41:04.691266 26797 round_trippers.go:443] GET https://192.168.10.128:6443/healthz?timeout=10s in 1 millisecondsI0730 00:41:14.130375 26797 round_trippers.go:443] GET https://192.168.10.128:6443/healthz?timeout=10s 500 Internal Server Error in 8944 millisecondsI0730 00:41:14.202729 26797 round_trippers.go:443] GET https://192.168.10.128:6443/healthz?timeout=10s 500 Internal Server Error in 16 millisecondsI0730 00:41:14.689598 26797 round_trippers.go:443] GET https://192.168.10.128:6443/healthz?timeout=10s 500 Internal Server Error in 3 millisecondsI0730 00:41:15.196163 26797 round_trippers.go:443] GET https://192.168.10.128:6443/healthz?timeout=10s 500 Internal Server Error in 9 millisecondsI0730 00:41:15.689662 26797 round_trippers.go:443] GET https://192.168.10.128:6443/healthz?timeout=10s 500 Internal Server Error in 3 millisecondsI0730 00:41:16.190375 26797 round_trippers.go:443] GET https://192.168.10.128:6443/healthz?timeout=10s 200 OK in 3 milliseconds[apiclient] All control plane components are healthy after 19.005906 secondsI0730 00:41:16.190562 26797 uploadconfig.go:108] [upload-config] Uploading the kubeadm ClusterConfiguration to a ConfigMap[upload-config] Storing the configuration used in ConfigMap "kubeadm-config" in the "kube-system" NamespaceI0730 00:41:16.202115 26797 round_trippers.go:443] POST https://192.168.10.128:6443/api/v1/namespaces/kube-system/configmaps?timeout=10s 201 Created in 6 millisecondsI0730 00:41:16.210907 26797 round_trippers.go:443] POST https://192.168.10.128:6443/apis/rbac.authorization.k8s.io/v1/namespaces/kube-system/roles?timeout=10s 201 Created in 7 millisecondsI0730 00:41:16.219527 26797 round_trippers.go:443] POST https://192.168.10.128:6443/apis/rbac.authorization.k8s.io/v1/namespaces/kube-system/rolebindings?timeout=10s 201 Created in 6 millisecondsI0730 00:41:16.220662 26797 uploadconfig.go:122] [upload-config] Uploading the kubelet component config to a ConfigMap[kubelet] Creating a ConfigMap "kubelet-config-1.18" in namespace kube-system with the configuration for the kubelets in the clusterI0730 00:41:16.227796 26797 round_trippers.go:443] POST https://192.168.10.128:6443/api/v1/namespaces/kube-system/configmaps?timeout=10s 201 Created in 6 millisecondsI0730 00:41:16.234309 26797 round_trippers.go:443] POST https://192.168.10.128:6443/apis/rbac.authorization.k8s.io/v1/namespaces/kube-system/roles?timeout=10s 201 Created in 6 millisecondsI0730 00:41:16.241883 26797 round_trippers.go:443] POST https://192.168.10.128:6443/apis/rbac.authorization.k8s.io/v1/namespaces/kube-system/rolebindings?timeout=10s 201 Created in 7 millisecondsI0730 00:41:16.242124 26797 uploadconfig.go:127] [upload-config] Preserving the CRISocket information for the control-plane nodeI0730 00:41:16.242159 26797 patchnode.go:30] [patchnode] Uploading the CRI Socket information "/var/run/dockershim.sock" to the Node API object "master" as an annotationI0730 00:41:16.751565 26797 round_trippers.go:443] GET https://192.168.10.128:6443/api/v1/nodes/master?timeout=10s 200 OK in 4 millisecondsI0730 00:41:16.767568 26797 round_trippers.go:443] PATCH https://192.168.10.128:6443/api/v1/nodes/master?timeout=10s 200 OK in 11 milliseconds[upload-certs] Skipping phase. Please see --upload-certs[mark-control-plane] Marking the node master as control-plane by adding the label "node-role.kubernetes.io/master=''"[mark-control-plane] Marking the node master as control-plane by adding the taints [node-role.kubernetes.io/master:NoSchedule]I0730 00:41:17.273970 26797 round_trippers.go:443] GET https://192.168.10.128:6443/api/v1/nodes/master?timeout=10s 200 OK in 4 millisecondsI0730 00:41:17.285850 26797 round_trippers.go:443] PATCH https://192.168.10.128:6443/api/v1/nodes/master?timeout=10s 200 OK in 9 milliseconds[bootstrap-token] Using token: re6mt8.bwol9dej5nsqubxt[bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC RolesI0730 00:41:17.291184 26797 round_trippers.go:443] GET https://192.168.10.128:6443/api/v1/namespaces/kube-system/secrets/bootstrap-token-re6mt8?timeout=10s 404 Not Found in 4 millisecondsI0730 00:41:17.299410 26797 round_trippers.go:443] POST https://192.168.10.128:6443/api/v1/namespaces/kube-system/secrets?timeout=10s 201 Created in 6 milliseconds[bootstrap-token] configured RBAC rules to allow Node Bootstrap tokens to get nodesI0730 00:41:17.309247 26797 round_trippers.go:443] POST https://192.168.10.128:6443/apis/rbac.authorization.k8s.io/v1/clusterroles?timeout=10s 201 Created in 8 millisecondsI0730 00:41:17.318135 26797 round_trippers.go:443] POST https://192.168.10.128:6443/apis/rbac.authorization.k8s.io/v1/clusterrolebindings?timeout=10s 201 Created in 7 milliseconds[bootstrap-token] configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentialsI0730 00:41:17.325642 26797 round_trippers.go:443] POST https://192.168.10.128:6443/apis/rbac.authorization.k8s.io/v1/clusterrolebindings?timeout=10s 201 Created in 6 milliseconds[bootstrap-token] configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap TokenI0730 00:41:17.332197 26797 round_trippers.go:443] POST https://192.168.10.128:6443/apis/rbac.authorization.k8s.io/v1/clusterrolebindings?timeout=10s 201 Created in 6 milliseconds[bootstrap-token] configured RBAC rules to allow certificate rotation for all node client certificates in the clusterI0730 00:41:17.338051 26797 round_trippers.go:443] POST https://192.168.10.128:6443/apis/rbac.authorization.k8s.io/v1/clusterrolebindings?timeout=10s 201 Created in 5 milliseconds[bootstrap-token] Creating the "cluster-info" ConfigMap in the "kube-public" namespaceI0730 00:41:17.338274 26797 clusterinfo.go:45] [bootstrap-token] loading admin kubeconfigI0730 00:41:17.339582 26797 loader.go:375] Config loaded from file: /etc/kubernetes/admin.confI0730 00:41:17.339606 26797 clusterinfo.go:53] [bootstrap-token] copying the cluster from admin.conf to the bootstrap kubeconfigI0730 00:41:17.340061 26797 clusterinfo.go:65] [bootstrap-token] creating/updating ConfigMap in kube-public namespaceI0730 00:41:17.346320 26797 round_trippers.go:443] POST https://192.168.10.128:6443/api/v1/namespaces/kube-public/configmaps?timeout=10s 201 Created in 6 millisecondsI0730 00:41:17.346598 26797 clusterinfo.go:79] creating the RBAC rules for exposing the cluster-info ConfigMap in the kube-public namespaceI0730 00:41:17.354296 26797 round_trippers.go:443] POST https://192.168.10.128:6443/apis/rbac.authorization.k8s.io/v1/namespaces/kube-public/roles?timeout=10s 201 Created in 7 millisecondsI0730 00:41:17.360015 26797 round_trippers.go:443] POST https://192.168.10.128:6443/apis/rbac.authorization.k8s.io/v1/namespaces/kube-public/rolebindings?timeout=10s 201 Created in 5 millisecondsI0730 00:41:17.361663 26797 kubeletfinalize.go:88] [kubelet-finalize] Assuming that kubelet client certificate rotation is enabled: found "/var/lib/kubelet/pki/kubelet-client-current.pem"[kubelet-finalize] Updating "/etc/kubernetes/kubelet.conf" to point to a rotatable kubelet client certificate and keyI0730 00:41:17.362911 26797 loader.go:375] Config loaded from file: /etc/kubernetes/kubelet.confI0730 00:41:17.364376 26797 kubeletfinalize.go:132] [kubelet-finalize] Restarting the kubelet to enable client certificate rotationI0730 00:41:17.680009 26797 round_trippers.go:443] GET https://192.168.10.128:6443/apis/apps/v1/namespaces/kube-system/deployments?labelSelector=k8s-app%3Dkube-dns 200 OK in 11 millisecondsI0730 00:41:17.694805 26797 round_trippers.go:443] GET https://192.168.10.128:6443/api/v1/namespaces/kube-system/configmaps/kube-dns?timeout=10s 404 Not Found in 5 millisecondsI0730 00:41:17.700146 26797 round_trippers.go:443] GET https://192.168.10.128:6443/api/v1/namespaces/kube-system/configmaps/coredns?timeout=10s 404 Not Found in 4 millisecondsI0730 00:41:17.709319 26797 round_trippers.go:443] POST https://192.168.10.128:6443/api/v1/namespaces/kube-system/configmaps?timeout=10s 201 Created in 8 millisecondsI0730 00:41:17.716415 26797 round_trippers.go:443] POST https://192.168.10.128:6443/apis/rbac.authorization.k8s.io/v1/clusterroles?timeout=10s 201 Created in 6 millisecondsI0730 00:41:17.722154 26797 round_trippers.go:443] POST https://192.168.10.128:6443/apis/rbac.authorization.k8s.io/v1/clusterrolebindings?timeout=10s 201 Created in 5 millisecondsI0730 00:41:17.730289 26797 round_trippers.go:443] POST https://192.168.10.128:6443/api/v1/namespaces/kube-system/serviceaccounts?timeout=10s 201 Created in 6 millisecondsI0730 00:41:17.776066 26797 round_trippers.go:443] POST https://192.168.10.128:6443/apis/apps/v1/namespaces/kube-system/deployments?timeout=10s 201 Created in 34 millisecondsI0730 00:41:17.790010 26797 round_trippers.go:443] POST https://192.168.10.128:6443/api/v1/namespaces/kube-system/services?timeout=10s 201 Created in 11 milliseconds[addons] Applied essential addon: CoreDNSI0730 00:41:17.798512 26797 round_trippers.go:443] POST https://192.168.10.128:6443/api/v1/namespaces/kube-system/serviceaccounts?timeout=10s 201 Created in 8 millisecondsI0730 00:41:17.809712 26797 round_trippers.go:443] POST https://192.168.10.128:6443/api/v1/namespaces/kube-system/configmaps?timeout=10s 201 Created in 8 millisecondsI0730 00:41:17.861931 26797 round_trippers.go:443] POST https://192.168.10.128:6443/apis/apps/v1/namespaces/kube-system/daemonsets?timeout=10s 201 Created in 35 millisecondsI0730 00:41:17.872631 26797 round_trippers.go:443] POST https://192.168.10.128:6443/apis/rbac.authorization.k8s.io/v1/clusterrolebindings?timeout=10s 201 Created in 9 millisecondsI0730 00:41:17.878624 26797 round_trippers.go:443] POST https://192.168.10.128:6443/apis/rbac.authorization.k8s.io/v1/namespaces/kube-system/roles?timeout=10s 201 Created in 5 millisecondsI0730 00:41:17.884579 26797 round_trippers.go:443] POST https://192.168.10.128:6443/apis/rbac.authorization.k8s.io/v1/namespaces/kube-system/rolebindings?timeout=10s 201 Created in 5 milliseconds[addons] Applied essential addon: kube-proxyI0730 00:41:17.885611 26797 loader.go:375] Config loaded from file: /etc/kubernetes/admin.confI0730 00:41:17.886868 26797 loader.go:375] Config loaded from file: /etc/kubernetes/admin.confYour Kubernetes control-plane has initialized successfully!To start using your cluster, you need to run the following as a regular user:mkdir -p $HOME/.kubesudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/configsudo chown $(id -u):$(id -g) $HOME/.kube/configYou should now deploy a pod network to the cluster.Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at:https://kubernetes.io/docs/concepts/cluster-administration/addons/Then you can join any number of worker nodes by running the following on each as root:kubeadm join 192.168.10.128:6443 --token re6mt8.bwol9dej5nsqubxt \--discovery-token-ca-cert-hash sha256:21ec2e66f355b7d31652b2b790e4862ef2fb9b6069d7425d122007a7d2edce782.10 安装pod网络插件kubectl apply -f https://raw.githubusercontent.com/coreos/flannel/master/Documentation/kube-flannel.yaml" with one of the options listed at:https://kubernetes.io/docs/concepts/cluster-administration/addons/Then you can join any number of worker nodesbyttps://www.ipaddress.com进行查看将域名和IP对应关系写到hosts文件中2.11 node节点加入集群kubeadm join 192.168.10.128:6443 --token whxbm2.x70y9k1feop2604b \--discovery-token-ca-cert-hash sha256:d4306836b7ef32fd802e559cf717bbbe3af6dd75d3cb2fa4d3db7a16cc25a6573异常节点处理:如需要删除节点的话则需要执行如下操作主节点操作# 在主节点上驱逐该node节点上的Podkubectl drain node02 --delete-local-data --force --ignore-daemonsets# 在主节点上删除该node节点kubectl delete node node02从节点上操作# 重置k8skubeadm reset# 删除残留的文件rm -rf /etc/kubernetes/*# 清除iptables或者ipvs的配置iptables -F && iptables -t nat -F && iptables -t mangle -F && iptables -Xipvsadm --clear如果是master节点 还需要注意rm -rf ~/.kube随后两人在社交媒体上分享了这一喜讯配图是他们穿着情侣装
郭颖儿摇头上古种子现今所存极少只能在前辈先人的游记轶事中见到603696 安记食品
活动中民警在车站候车厅以及站外停车场人员聚集地点向乘客发放宣传资料讲解交通安全知识和日常乘车应注意的安全事项随后民警登上待发客车通过发放交通安全宣传单、粘贴佩戴乘车安全带提示标提醒乘客在乘坐客车时必须主动规范系好安全带自觉抵制交通违法行为并要求驾驶员自觉遵守交通法律法规严格杜绝超员、超载、超速行为严禁带病驾驶、疲劳驾驶夏季天气炎热发车前要对车辆灭火器、货舱物品、轮胎情况进行安全检查