安装报错Error: failed post-install: timed out waiting for the condition

export IP=172.27.xxx.xxx
export PORT=30000
chmod +x ./all_in_one_install_quickstart.sh
./all_in_one_install_quickstart.sh

请提供更多信息,如果不知道提供什么信息,参考如何有效的反馈问题

我也碰到了这个问题。开始以为系统问题,从ubuntu 20.04,切换成到centos7问题依旧。

全新系统all-in-one安装错误

系统版本:CentOS Linux release 7.8.2003 (Core)、Ubuntu Server 20.04

[root@kc-01 ~]# cat install.sh

# 服务器 IP 地址,用于访问 Zadig 系统
# 如果您有自己的域名,可以配置您可用的域名 export DOMAIN = <DOMAIN>
# 如果出现端口占用情况,换一个端口再尝试
# all_in_one_install_quickstart.sh 替换成实际下载的脚本

export IP=192.168.0.212
export PORT=30000
./all_in_one_install_quickstart.sh

[root@kc-01 ~]# sh install.sh
 _____          _ _
/ _  / __ _  __| (_) __ _
\// / / _. |/ _. | |/ _. |
 / //\ (_| | (_| | | (_| |
/____/\__,_|\__,_|_|\__, |
                    |___/
 Welcome to the Koderover Installer
[INFO] Checking system for requirements...
./all_in_one_install_quickstart.sh: line 198: dig: command not found
 The installer is incompatible with swap on, disable swap to continue installing? [y/N]
y
[INFO] executing swapoff --all
[INFO] => Commenting swap entries in /etc/fstab

[INFO] => A backup of /etc/fstab has been made at /etc/fstab.bak


[INFO]
Changes have been made to /etc/fstab. We recommend reviewing them after completing this installation to ensure mounts are correctly configured.


[ERROR] kubernetes is incompatible with selinux, disable it? [y/N]
y
 preflight check completed in 13 seconds
 install preparation completed in 13 seconds
[INFO] fetching docker...
######################################################################## 100.0%
warning: ./packages/docker/19.03.10/rhel-7/containerd.io-1.4.3-3.1.el7.x86_64.rpm: Header V4 RSA/SHA512 Signature, key ID 621e9f35: NOKEY
Created symlink from /etc/systemd/system/multi-user.target.wants/docker.service to /usr/lib/systemd/system/docker.service.

* Applying /usr/lib/sysctl.d/00-system.conf ...
net.bridge.bridge-nf-call-ip6tables = 0
net.bridge.bridge-nf-call-iptables = 0
net.bridge.bridge-nf-call-arptables = 0
* Applying /usr/lib/sysctl.d/10-default-yama-scope.conf ...
kernel.yama.ptrace_scope = 0
* Applying /usr/lib/sysctl.d/50-default.conf ...
kernel.sysrq = 16
kernel.core_uses_pid = 1
net.ipv4.conf.default.rp_filter = 1
net.ipv4.conf.all.rp_filter = 1
net.ipv4.conf.default.accept_source_route = 0
net.ipv4.conf.all.accept_source_route = 0
net.ipv4.conf.default.promote_secondaries = 1
net.ipv4.conf.all.promote_secondaries = 1
fs.protected_hardlinks = 1
fs.protected_symlinks = 1
* Applying /etc/sysctl.d/99-sysctl.conf ...
* Applying /etc/sysctl.d/k8s.conf ...
net.ipv4.conf.all.rp_filter = 1
net.ipv4.ip_forward = 1
net.bridge.bridge-nf-call-ip6tables = 1
net.bridge.bridge-nf-call-iptables = 1
net.ipv4.conf.all.forwarding = 1
* Applying /etc/sysctl.conf ...
[INFO] Downloading kubernetes binary.....
######################################################################## 100.0%
* Applying /usr/lib/sysctl.d/00-system.conf ...
net.bridge.bridge-nf-call-ip6tables = 0
net.bridge.bridge-nf-call-iptables = 0
net.bridge.bridge-nf-call-arptables = 0
* Applying /usr/lib/sysctl.d/10-default-yama-scope.conf ...
kernel.yama.ptrace_scope = 0
* Applying /usr/lib/sysctl.d/50-default.conf ...
kernel.sysrq = 16
kernel.core_uses_pid = 1
net.ipv4.conf.default.rp_filter = 1
net.ipv4.conf.all.rp_filter = 1
net.ipv4.conf.default.accept_source_route = 0
net.ipv4.conf.all.accept_source_route = 0
net.ipv4.conf.default.promote_secondaries = 1
net.ipv4.conf.all.promote_secondaries = 1
fs.protected_hardlinks = 1
fs.protected_symlinks = 1
* Applying /etc/sysctl.d/99-sysctl.conf ...
* Applying /etc/sysctl.d/k8s.conf ...
net.ipv4.conf.all.rp_filter = 1
net.ipv4.ip_forward = 1
net.bridge.bridge-nf-call-ip6tables = 1
net.bridge.bridge-nf-call-iptables = 1
net.ipv4.conf.all.forwarding = 1
* Applying /etc/sysctl.conf ...
[INFO] Installing kubelet, kubeadm, kubectl and cni host packages
 kubelet command missing
warning: ./packages/kubernetes/1.19.3/rhel-7/14bfe6e75a9efc8eca3f638eb22c7e2ce759c67f95b43b16fae4ebabde1549f3-cri-tools-1.13.0-0.x86_64.rpm: Header V4 RSA/SHA512 Signature, key ID 3e1ba8d5: NOKEY
Created symlink from /etc/systemd/system/multi-user.target.wants/kubelet.service to /usr/lib/systemd/system/kubelet.service.
 kubernetes host packages installed
79d541cda6cb: Loading layer [==================================================>]  3.041MB/3.041MB
c12e92a17b61: Loading layer [==================================================>]  1.734MB/1.734MB
f9b944e24088: Loading layer [==================================================>]  107.3MB/107.3MB
Loaded image: k8s.gcr.io/kube-controller-manager:v1.19.3
91e3a07063b3: Loading layer [==================================================>]  53.89MB/53.89MB
b4e54f331697: Loading layer [==================================================>]  21.78MB/21.78MB
b9b82a97c787: Loading layer [==================================================>]  5.168MB/5.168MB
1b55846906e8: Loading layer [==================================================>]  4.608kB/4.608kB
061bfb5cb861: Loading layer [==================================================>]  8.192kB/8.192kB
78dd6c0504a7: Loading layer [==================================================>]  8.704kB/8.704kB
f1b0b899d419: Loading layer [==================================================>]  38.81MB/38.81MB
Loaded image: k8s.gcr.io/kube-proxy:v1.19.3
ba0dae6243cc: Loading layer [==================================================>]  684.5kB/684.5kB
Loaded image: k8s.gcr.io/pause:3.2
225df95e717c: Loading layer [==================================================>]  336.4kB/336.4kB
96d17b0b58a7: Loading layer [==================================================>]  45.02MB/45.02MB
Loaded image: k8s.gcr.io/coredns:1.7.0
94bd98e8a8a9: Loading layer [==================================================>]  42.13MB/42.13MB
Loaded image: k8s.gcr.io/kube-scheduler:v1.19.3
4b5c08158bcb: Loading layer [==================================================>]  115.3MB/115.3MB
Loaded image: k8s.gcr.io/kube-apiserver:v1.19.3
d72a74c56330: Loading layer [==================================================>]  3.031MB/3.031MB
d61c79b29299: Loading layer [==================================================>]   2.13MB/2.13MB
1a4e46412eb0: Loading layer [==================================================>]  225.3MB/225.3MB
bfa5849f3d09: Loading layer [==================================================>]   2.19MB/2.19MB
bb63b9467928: Loading layer [==================================================>]  21.98MB/21.98MB
Loaded image: k8s.gcr.io/etcd:3.4.13-0

W0723 15:20:47.698558    2985 configset.go:348] WARNING: kubeadm cannot validate component configs for API groups [kubelet.config.k8s.io kubeproxy.config.k8s.io]
[init] Using Kubernetes version: v1.19.3
[preflight] Running pre-flight checks
	[WARNING Hostname]: hostname "kc-01" could not be reached
	[WARNING Hostname]: hostname "kc-01": lookup kc-01 on 192.168.0.1:53: no such host
[preflight] Pulling images required for setting up a Kubernetes cluster
[preflight] This might take a minute or two, depending on the speed of your internet connection
[preflight] You can also perform this action in beforehand using 'kubeadm config images pull'
[certs] Using certificateDir folder "/etc/kubernetes/pki"
[certs] Generating "ca" certificate and key
[certs] Generating "apiserver" certificate and key
[certs] apiserver serving cert is signed for DNS names [kc-01 kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local] and IPs [10.96.0.1 192.168.0.212]
[certs] Generating "apiserver-kubelet-client" certificate and key
[certs] Generating "front-proxy-ca" certificate and key
[certs] Generating "front-proxy-client" certificate and key
[certs] Generating "etcd/ca" certificate and key
[certs] Generating "etcd/server" certificate and key
[certs] etcd/server serving cert is signed for DNS names [kc-01 localhost] and IPs [192.168.0.212 127.0.0.1 ::1]
[certs] Generating "etcd/peer" certificate and key
[certs] etcd/peer serving cert is signed for DNS names [kc-01 localhost] and IPs [192.168.0.212 127.0.0.1 ::1]
[certs] Generating "etcd/healthcheck-client" certificate and key
[certs] Generating "apiserver-etcd-client" certificate and key
[certs] Generating "sa" key and public key
[kubeconfig] Using kubeconfig folder "/etc/kubernetes"
[kubeconfig] Writing "admin.conf" kubeconfig file
[kubeconfig] Writing "kubelet.conf" kubeconfig file
[kubeconfig] Writing "controller-manager.conf" kubeconfig file
[kubeconfig] Writing "scheduler.conf" kubeconfig file
[kubelet-start] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env"
[kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml"
[kubelet-start] Starting the kubelet
[control-plane] Using manifest folder "/etc/kubernetes/manifests"
[control-plane] Creating static Pod manifest for "kube-apiserver"
[control-plane] Creating static Pod manifest for "kube-controller-manager"
[control-plane] Creating static Pod manifest for "kube-scheduler"
[etcd] Creating static Pod manifest for local etcd in "/etc/kubernetes/manifests"
[wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory "/etc/kubernetes/manifests". This can take up to 4m0s
[apiclient] All control plane components are healthy after 15.005791 seconds

[upload-config] Storing the configuration used in ConfigMap "kubeadm-config" in the "kube-system" Namespace
[kubelet] Creating a ConfigMap "kubelet-config-1.19" in namespace kube-system with the configuration for the kubelets in the cluster
[upload-certs] Skipping phase. Please see --upload-certs
[mark-control-plane] Marking the node kc-01 as control-plane by adding the label "node-role.kubernetes.io/master=''"
[mark-control-plane] Marking the node kc-01 as control-plane by adding the taints [node-role.kubernetes.io/master:NoSchedule]
[bootstrap-token] Using token: 5bebr5.w2tgssogsaeczcrw
[bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles
[bootstrap-token] configured RBAC rules to allow Node Bootstrap tokens to get nodes
[bootstrap-token] configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials
[bootstrap-token] configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token
[bootstrap-token] configured RBAC rules to allow certificate rotation for all node client certificates in the cluster

[bootstrap-token] Creating the "cluster-info" ConfigMap in the "kube-public" namespace
[kubelet-finalize] Updating "/etc/kubernetes/kubelet.conf" to point to a rotatable kubelet client certificate and key
[addons] Applied essential addon: CoreDNS
[addons] Applied essential addon: kube-proxy

Your Kubernetes control-plane has initialized successfully!

To start using your cluster, you need to run the following as a regular user:

  mkdir -p $HOME/.kube
  sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
  sudo chown $(id -u):$(id -g) $HOME/.kube/config

You should now deploy a pod network to the cluster.
Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at:
  https://kubernetes.io/docs/concepts/cluster-administration/addons/

Then you can join any number of worker nodes by running the following on each as root:

kubeadm join 192.168.0.212:6443 --token 5bebr5.w2tgssogsaeczcrw \
    --discovery-token-ca-cert-hash sha256:cbee87899a656fe781d3e4d02e801dcdd97327f4130e62ce43dea2ac46bcdefc
node/kc-01 untainted
serviceaccount/weave-net created
clusterrole.rbac.authorization.k8s.io/weave-net created
clusterrolebinding.rbac.authorization.k8s.io/weave-net created
role.rbac.authorization.k8s.io/weave-net created
rolebinding.rbac.authorization.k8s.io/weave-net created
daemonset.apps/weave-net created
 infrastructure installation completed in 3 minute(s) and 31 second(s)
NO ENCRYPTION KEY PROVIDED, ZADIG HAS GENERATED AN ENCRYPTION KEY
A390341514A1619E58DFA781BD8DA46B
THIS KEY WILL BE USED FOR POSSIBLE FUTURE REINSTALLATION, PLEASE SAVE THIS KEY CAREFULLY
installing helm client...
succeed to install helm client: version.BuildInfo{Version:"v3.6.1", GitCommit:"61d8e8c4a6f95540c15c6a65f36a6dd0a45e7a2f", GitTreeState:"clean", GoVersion:"go1.16.5"}
installing zadig ...
"koderover-chart" has been added to your repositories
Hang tight while we grab the latest from your chart repositories...
...Successfully got an update from the "koderover-chart" chart repository
Update Complete. ⎈Happy Helming!⎈
Release "zadig-zadig" does not exist. Installing it now.
Error: failed post-install: timed out waiting for the condition
 *****************************************
 *    Koderover installer exit report    *
 *****************************************
 ✔ ROOT PRIVILEGE CHECK SUCCESS
 ✔ SYSTEM CHECK SUCCESS
 ✔ DEPENDENCY INSTALLATION SUCCESS
 ✔ KUBERNETES CLUSTER INITIALIZATION SUCCESS
[ERROR] ⚙ ZADIG INSTALLATION FAILED
 *****************************************
 *            END OF REPORT              *
 *****************************************