KubeCube 是一个开源的企业级容器平台,为企业提供 Kubernetes 资源可视化管理以及统一的多集群多租户管理功能。KubeCube 可以简化应用部署、管理应用的生命周期和提供丰富的监控界面和日志审计功能,帮助企业快速构建一个强大和功能丰富的容器云管理平台。
然后报错:
2022-11-24 09:15:42 INFO environment checking
|---------------------------------------------------------------------|
| sshpass | conntrack | unzip | libseccomp |
|---------------------------------------------------------------------|
| ✓ | ✓ | ✓ | ✓ |
|---------------------------------------------------------------------|
2022-11-24 09:15:42 INFO downloading manifests for kubecube
--2022-11-24 09:15:42-- https://kubecube.nos-eastchina1.126.net/kubecube-installer/v1.4/manifests.tar.gz
正在解析主机 kubecube.nos-eastchina1.126.net (kubecube.nos-eastchina1.126.net)... 59.111.35.1, 59.111.35.2
正在连接 kubecube.nos-eastchina1.126.net (kubecube.nos-eastchina1.126.net)|59.111.35.1|:443... 已连接。
错误: 无法验证 kubecube.nos-eastchina1.126.net 的由 “CN=GeoTrust RSA CN CA G2,O=DigiCert Inc,C=US” 颁发的证书:
无法本地校验颁发者的权限。
要以不安全的方式连接至 kubecube.nos-eastchina1.126.net,使用“--no-check-certificate”。
把文件下载下来,然后修改 70 行代码,wget 增加--no-check-certificate 参数。
66 if [ -e "./manifests" ]; then
67 echo -e "$(date +'%Y-%m-%d %H:%M:%S') \033[32mINFO\033[0m manifests already exist"
68 else
69 echo -e "$(date +'%Y-%m-%d %H:%M:%S') \033[32mINFO\033[0m downloading manifests for kubecube"
70 wget --no-check-certificate https://kubecube.nos-eastchina1.126.net/kubecube-installer/v1.4/manifests.tar.gz -O manifests.tar.gz
71
72 tar -xzvf manifests.tar.gz > /dev/null
73 fi
然后就可以下载安装了。顺序执行,从网易的服务器上面下载容器镜像。
# bash entry.sh
2022-11-24 09:35:59 INFO environment checking
|---------------------------------------------------------------------|
| sshpass | conntrack | unzip | libseccomp |
|---------------------------------------------------------------------|
| ✓ | ✓ | ✓ | ✓ |
|---------------------------------------------------------------------|
2022-11-24 09:35:59 INFO manifests already exist
-------------System Infomation-------------
System running time:11 days,19 hours, 19 minutes
IP: 192.168.1.110
CPU model: Intel(R) Core(TM) i5-7200U CPU @ 2.50GHz
CPU arch:x86_64
CPU cores: 4
CPU load: 0.31 0.55 0.75
--------------------------------------------
2022-11-24 09:35:59 WARN docker is already running.
2022-11-24 09:35:59 WARN docker is already running.
2022-11-24 09:35:59 WARN kubernetes binaries existed
2022-11-24 09:35:59 INFO downloading images
v3.19.1-m: Pulling from kubecube/calico/node
Digest: sha256:b70bd93db80365b4c57014bdc2ccaaac9d8c09edae94443c396f2174742650dd
Status: Image is up to date for hub.c.163.com/kubecube/calico/node:v3.19.1-m
hub.c.163.com/kubecube/calico/node:v3.19.1-m
v3.19.1-m: Pulling from kubecube/calico/cni
Digest: sha256:b6c282aca28c1da56d607af6d692394bcec3a34d569a71139cf0286a54f5ca69
Status: Image is up to date for hub.c.163.com/kubecube/calico/cni:v3.19.1-m
hub.c.163.com/kubecube/calico/cni:v3.19.1-m
v3.19.1-m: Pulling from kubecube/calico/pod2daemon-flexvol
Digest: sha256:cecf91b6c518bb25f46ff1c44d1c99c1b5ade250485ab9c116f7b4cc14ae59e9
Status: Image is up to date for hub.c.163.com/kubecube/calico/pod2daemon-flexvol:v3.19.1-m
hub.c.163.com/kubecube/calico/pod2daemon-flexvol:v3.19.1-m
v3.19.1-m: Pulling from kubecube/calico/kube-controllers
Digest: sha256:df8155aa54e5f72abe6d8618c9ea5246a0dd2cfa006df79a6e41cf8b4a7a3486
Status: Image is up to date for hub.c.163.com/kubecube/calico/kube-controllers:v3.19.1-m
hub.c.163.com/kubecube/calico/kube-controllers:v3.19.1-m
v1.23.5: Pulling from google_containers/kube-apiserver
Digest: sha256:ddf5bf7196eb534271f9e5d403f4da19838d5610bb5ca191001bde5f32b5492e
Status: Image is up to date for registry.cn-hangzhou.aliyuncs.com/google_containers/kube-apiserver:v1.23.5
registry.cn-hangzhou.aliyuncs.com/google_containers/kube-apiserver:v1.23.5
v1.23.5: Pulling from google_containers/kube-controller-manager
Digest: sha256:cca0fb3532abedcc95c5f64268d54da9ecc56cc4817ff08d0128941cf2b0e1a4
Status: Image is up to date for registry.cn-hangzhou.aliyuncs.com/google_containers/kube-controller-manager:v1.23.5
registry.cn-hangzhou.aliyuncs.com/google_containers/kube-controller-manager:v1.23.5
v1.23.5: Pulling from google_containers/kube-scheduler
Digest: sha256:489efb65da9edc40bf0911f3e6371e5bb6b8ad8fde1d55193a6cc84c2ef36626
Status: Image is up to date for registry.cn-hangzhou.aliyuncs.com/google_containers/kube-scheduler:v1.23.5
registry.cn-hangzhou.aliyuncs.com/google_containers/kube-scheduler:v1.23.5
v1.23.5: Pulling from google_containers/kube-proxy
Digest: sha256:c1f625d115fbd9a12eac615653fc81c0edb33b2b5a76d1e09d5daed11fa557c1
Status: Image is up to date for registry.cn-hangzhou.aliyuncs.com/google_containers/kube-proxy:v1.23.5
registry.cn-hangzhou.aliyuncs.com/google_containers/kube-proxy:v1.23.5
3.6: Pulling from google_containers/pause
Digest: sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db
Status: Image is up to date for registry.cn-hangzhou.aliyuncs.com/google_containers/pause:3.6
registry.cn-hangzhou.aliyuncs.com/google_containers/pause:3.6
3.5.1-0: Pulling from google_containers/etcd
Digest: sha256:64b9ea357325d5db9f8a723dcf503b5a449177b17ac87d69481e126bb724c263
Status: Image is up to date for registry.cn-hangzhou.aliyuncs.com/google_containers/etcd:3.5.1-0
registry.cn-hangzhou.aliyuncs.com/google_containers/etcd:3.5.1-0
v1.8.6: Pulling from google_containers/coredns
Digest: sha256:5b6ec0d6de9baaf3e92d0f66cd96a25b9edbce8716f5f15dcd1a616b3abd590e
Status: Image is up to date for registry.cn-hangzhou.aliyuncs.com/google_containers/coredns:v1.8.6
registry.cn-hangzhou.aliyuncs.com/google_containers/coredns:v1.8.6
v1.2.0: Pulling from kubecube/audit
Digest: sha256:a1e4352507ec7ef1d005c18645f6004077373848f942653e905a23a3808ae8a9
Status: Image is up to date for hub.c.163.com/kubecube/audit:v1.2.0
hub.c.163.com/kubecube/audit:v1.2.0
v1.2.1: Pulling from kubecube/cloudshell
Digest: sha256:8a8c335f34c142444d7d11b1b88af41e722634b02abe65fa2366a3f4e8f1b371
Status: Image is up to date for hub.c.163.com/kubecube/cloudshell:v1.2.1
hub.c.163.com/kubecube/cloudshell:v1.2.1
v1.2.4: Pulling from kubecube/webconsole
Digest: sha256:51618de04401628435a17fcc0dba450e0c97a94706cc23141b6b2614f54bbc45
Status: Image is up to date for hub.c.163.com/kubecube/webconsole:v1.2.4
hub.c.163.com/kubecube/webconsole:v1.2.4
v1.2.0: Pulling from kubecube/frontend
Digest: sha256:8102081b357e430dd3f9c12ebfa3ffb1d13942cb37f75513aa9a56defe1a4722
Status: Image is up to date for hub.c.163.com/kubecube/frontend:v1.2.0
hub.c.163.com/kubecube/frontend:v1.2.0
v1.4.1: Pulling from kubecube/cube
Digest: sha256:919b061f7a46f9f0e95d2d99c6ed8351ba9270e30b7ab22c51e1b1e728ac6e10
Status: Image is up to date for hub.c.163.com/kubecube/cube:v1.4.1
hub.c.163.com/kubecube/cube:v1.4.1
v1.4.1: Pulling from kubecube/warden
Digest: sha256:4ecfcdfefa81729cd46724c5f522ab2653ffe40e9450639e3cc2d2c6247e1f81
Status: Image is up to date for hub.c.163.com/kubecube/warden:v1.4.1
hub.c.163.com/kubecube/warden:v1.4.1
v1.4.1: Pulling from kubecube/warden-init
Digest: sha256:186205272e5d3c0b7fa7a7015db7a68246b081376ffc5c3141664b2c34454fd0
Status: Image is up to date for hub.c.163.com/kubecube/warden-init:v1.4.1
hub.c.163.com/kubecube/warden-init:v1.4.1
v1.2.0: Pulling from kubecube/warden-dependence
Digest: sha256:7ae4bd0008197ec141a5e305968d015725901561d708d0d178bfd7c5a38df2f0
Status: Image is up to date for hub.c.163.com/kubecube/warden-dependence:v1.2.0
hub.c.163.com/kubecube/warden-dependence:v1.2.0
v0.21.0: Pulling from kubecube/alertmanager
Digest: sha256:702e01b4c96e4721927a7ae93afefc9b7fa5bc1cd5dcfbd23e70e50cfff2795e
Status: Image is up to date for hub.c.163.com/kubecube/alertmanager:v0.21.0
hub.c.163.com/kubecube/alertmanager:v0.21.0
v0.47.0: Pulling from kubecube/prometheus-config-reloader
Digest: sha256:0029252e7cf8cf38fc58795928d4e1c746b9e609432a2ee5417a9cab4633b864
Status: Image is up to date for hub.c.163.com/kubecube/prometheus-config-reloader:v0.47.0
hub.c.163.com/kubecube/prometheus-config-reloader:v0.47.0
1.10.7: Pulling from kubecube/k8s-sidecar
Digest: sha256:ac60db5cfb11c84f23c81a717463d668c7db9134f9a2283d38e13455f8481a6c
Status: Image is up to date for hub.c.163.com/kubecube/k8s-sidecar:1.10.7
hub.c.163.com/kubecube/k8s-sidecar:1.10.7
7.5.5: Pulling from kubecube/grafana
Digest: sha256:58ea68c27090cee44872800fd15c55592905b1ab86daa8ffbb42fd6cbdfbe3e2
Status: Image is up to date for hub.c.163.com/kubecube/grafana:7.5.5
hub.c.163.com/kubecube/grafana:7.5.5
v1.9.8: Pulling from kubecube/kube-state-metrics
Digest: sha256:de088703b8faab6f293bb2a16931cd814b1e2ddfe786074457946004e81e6fa7
Status: Image is up to date for hub.c.163.com/kubecube/kube-state-metrics:v1.9.8
hub.c.163.com/kubecube/kube-state-metrics:v1.9.8
v0.47.0: Pulling from kubecube/prometheus-operator
Digest: sha256:89a2d121b1a8f9a4a45dd20fdcf081a4468a0a0ad4e0cbe1aa7fd289e5a85cb3
Status: Image is up to date for hub.c.163.com/kubecube/prometheus-operator:v0.47.0
hub.c.163.com/kubecube/prometheus-operator:v0.47.0
v1.1.2: Pulling from kubecube/node-exporter
Digest: sha256:4239af7a8cffcfa003ff624398ff6e78dfea81794f686186282d4ebe99d4a8a1
Status: Image is up to date for hub.c.163.com/kubecube/node-exporter:v1.1.2
hub.c.163.com/kubecube/node-exporter:v1.1.2
v2.26.1: Pulling from kubecube/prometheus
Digest: sha256:fb5ef0e43748499f9803a7806782a4fd358637216f1eab28e315b2f279d70331
Status: Image is up to date for hub.c.163.com/kubecube/prometheus:v2.26.1
hub.c.163.com/kubecube/prometheus:v2.26.1
v0.22.0: Pulling from kubecube/thanos
Digest: sha256:6680c5a66cf4228a961efa31594e25be6a80bda67901633534a918f457392597
Status: Image is up to date for hub.c.163.com/kubecube/thanos:v0.22.0
hub.c.163.com/kubecube/thanos:v0.22.0
1.21.0: Pulling from kubecube/kubectl-tools
Digest: sha256:a8841469c637b699a60ea12e9853dc29624fe3c36742ee00e7bf107ac67e0737
Status: Image is up to date for hub.c.163.com/kubecube/kubectl-tools:1.21.0
hub.c.163.com/kubecube/kubectl-tools:1.21.0
v1.0.0: Pulling from kubecube/hnc/hnc-manager
Digest: sha256:8ecd6af56dfd845257801f1fa3cea8119b78956d6e737352e7d874f1d80daa1f
Status: Image is up to date for hub.c.163.com/kubecube/hnc/hnc-manager:v1.0.0
hub.c.163.com/kubecube/hnc/hnc-manager:v1.0.0
v0.46.0-m: Pulling from kubecube/ingress-nginx/controller
Digest: sha256:710ad4d51a680011d48381fb5b9bb97f3dda0e45cc2c4c73358d86e4c23617a1
Status: Image is up to date for hub.c.163.com/kubecube/ingress-nginx/controller:v0.46.0-m
hub.c.163.com/kubecube/ingress-nginx/controller:v0.46.0-m
v1.2.0: Pulling from kubecube/ingress-nginx/controller
Digest: sha256:314435f9465a7b2973e3aa4f2edad7465cc7bcdc8304be5d146d70e4da136e51
Status: Image is up to date for hub.c.163.com/kubecube/ingress-nginx/controller:v1.2.0
hub.c.163.com/kubecube/ingress-nginx/controller:v1.2.0
v1.1.1: Pulling from kubecube/ingress-nginx/kube-webhook-certgen
Digest: sha256:78351fc9d9b5f835e0809921c029208faeb7fbb6dc2d3b0d1db0a6584195cfed
Status: Image is up to date for hub.c.163.com/kubecube/ingress-nginx/kube-webhook-certgen:v1.1.1
hub.c.163.com/kubecube/ingress-nginx/kube-webhook-certgen:v1.1.1
v1.5.1-m: Pulling from kubecube/jettech/kube-webhook-certgen
Digest: sha256:ead5a540eb86b8e6f82de7394902b427c2856224b5bb98f7335c9d03ce5dd38c
Status: Image is up to date for hub.c.163.com/kubecube/jettech/kube-webhook-certgen:v1.5.1-m
hub.c.163.com/kubecube/jettech/kube-webhook-certgen:v1.5.1-m
v1.5.1: Pulling from kubecube/jettech/kube-webhook-certgen
Digest: sha256:ead5a540eb86b8e6f82de7394902b427c2856224b5bb98f7335c9d03ce5dd38c
Status: Image is up to date for hub.c.163.com/kubecube/jettech/kube-webhook-certgen:v1.5.1
hub.c.163.com/kubecube/jettech/kube-webhook-certgen:v1.5.1
v0.0.19-m: Pulling from kubecube/rancher/local-path-provisioner
Digest: sha256:6bb91f85457463f733b2140ff4fe12afe1b443dc9abee7ca6a231c76ddd2d374
Status: Image is up to date for hub.c.163.com/kubecube/rancher/local-path-provisioner:v0.0.19-m
hub.c.163.com/kubecube/rancher/local-path-provisioner:v0.0.19-m
v0.4.1-m: Pulling from kubecube/rancher/metrics-server
Digest: sha256:fa30c9576d6545a193cd7fe97af450cdaf11f9eda31c76396af4a5e0737f92b8
Status: Image is up to date for hub.c.163.com/kubecube/rancher/metrics-server:v0.4.1-m
hub.c.163.com/kubecube/rancher/metrics-server:v0.4.1-m
latest-m: Pulling from kubecube/busybox
Digest: sha256:bacfbf3788dc26694339403484c710771635dd7c11472f652b31bb30b224b097
Status: Image is up to date for hub.c.163.com/kubecube/busybox:latest-m
hub.c.163.com/kubecube/busybox:latest-m
2022-11-24 09:36:46 INFO doing previous preparation
2022-11-24 09:36:46 DEBUG closing swap
2022-11-24 09:36:46 DEBUG config kernel params, passing bridge flow of IPv4 to iptables chain
net.bridge.bridge-nf-call-ip6tables = 1
net.bridge.bridge-nf-call-iptables = 1
2022-11-24 09:36:46 INFO enable kubelet service
2022-11-24 09:36:51 INFO make configuration for kubeadm
2022-11-24 09:36:51 DEBUG vip not be set, use node ip
2022-11-24 09:36:51 INFO installing node MODE: master
2022-11-24 09:36:51 INFO init kubernetes, version: 1.23.5
[init] Using Kubernetes version: v1.23.5
[preflight] Running pre-flight checks
[WARNING FileExisting-ebtables]: ebtables not found in system path
[WARNING FileExisting-ethtool]: ethtool not found in system path
[WARNING FileExisting-socat]: socat not found in system path
[preflight] Pulling images required for setting up a Kubernetes cluster
[preflight] This might take a minute or two, depending on the speed of your internet connection
[preflight] You can also perform this action in beforehand using 'kubeadm config images pull'
[certs] Using certificateDir folder "/etc/kubernetes/pki"
[certs] Using existing ca certificate authority
[certs] Using existing apiserver certificate and key on disk
[certs] Using existing apiserver-kubelet-client certificate and key on disk
[certs] Using existing front-proxy-ca certificate authority
[certs] Using existing front-proxy-client certificate and key on disk
[certs] Using existing etcd/ca certificate authority
[certs] Using existing etcd/server certificate and key on disk
[certs] Using existing etcd/peer certificate and key on disk
[certs] Using existing etcd/healthcheck-client certificate and key on disk
[certs] Using existing apiserver-etcd-client certificate and key on disk
[certs] Using the existing "sa" key
[kubeconfig] Using kubeconfig folder "/etc/kubernetes"
[kubeconfig] Using existing kubeconfig file: "/etc/kubernetes/admin.conf"
[kubeconfig] Using existing kubeconfig file: "/etc/kubernetes/kubelet.conf"
[kubeconfig] Using existing kubeconfig file: "/etc/kubernetes/controller-manager.conf"
[kubeconfig] Using existing kubeconfig file: "/etc/kubernetes/scheduler.conf"
[kubelet-start] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env"
[kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml"
[kubelet-start] Starting the kubelet
[control-plane] Using manifest folder "/etc/kubernetes/manifests"
[control-plane] Creating static Pod manifest for "kube-apiserver"
[control-plane] Creating static Pod manifest for "kube-controller-manager"
[control-plane] Creating static Pod manifest for "kube-scheduler"
[etcd] Creating static Pod manifest for local etcd in "/etc/kubernetes/manifests"
[wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory "/etc/kubernetes/manifests". This can take up to 4m0s
[kubelet-check] Initial timeout of 40s passed.
[kubelet-check] It seems like the kubelet isn't running or healthy.
[kubelet-check] The HTTP call equal to 'curl -sSL http://localhost:10248/healthz' failed with error: Get "http://localhost:10248/healthz": dial tcp 127.0.0.1:10248: connect: connection refused.
[kubelet-check] It seems like the kubelet isn't running or healthy.
[kubelet-check] The HTTP call equal to 'curl -sSL http://localhost:10248/healthz' failed with error: Get "http://localhost:10248/healthz": dial tcp 127.0.0.1:10248: connect: connection refused.
[kubelet-check] It seems like the kubelet isn't running or healthy.
[kubelet-check] The HTTP call equal to 'curl -sSL http://localhost:10248/healthz' failed with error: Get "http://localhost:10248/healthz": dial tcp 127.0.0.1:10248: connect: connection refused.
[kubelet-check] It seems like the kubelet isn't running or healthy.
[kubelet-check] The HTTP call equal to 'curl -sSL http://localhost:10248/healthz' failed with error: Get "http://localhost:10248/healthz": dial tcp 127.0.0.1:10248: connect: connection refused.
[kubelet-check] It seems like the kubelet isn't running or healthy.
[kubelet-check] The HTTP call equal to 'curl -sSL http://localhost:10248/healthz' failed with error: Get "http://localhost:10248/healthz": dial tcp 127.0.0.1:10248: connect: connection refused.
Unfortunately, an error has occurred:
timed out waiting for the condition
This error is likely caused by:
- The kubelet is not running
- The kubelet is unhealthy due to a misconfiguration of the node in some way (required cgroups disabled)
If you are on a systemd-powered system, you can try to troubleshoot the error with the following commands:
- 'systemctl status kubelet'
- 'journalctl -xeu kubelet'
Additionally, a control plane component may have crashed or exited when started by the container runtime.
To troubleshoot, list all containers using your preferred container runtimes CLI.
Here is one example how you may list all Kubernetes containers running in docker:
- 'docker ps -a | grep kube | grep -v pause'
Once you have found the failing container, you can inspect its logs with:
- 'docker logs CONTAINERID'
error execution phase wait-control-plane: couldn't initialize a Kubernetes cluster
To see the stack trace of this error execute with --v=5 or higher
2022-11-24 09:38:47 ERROR install kubernetes failed
没有启动成功,电脑的linux系统是 xubuntu 18 版本,找个22 版本的机器再试试。