系统配置:
Linux 3.10.0-327.36.3.el7.x86_64 #1 SMP Mon Oct 24 16:09:20 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux
配置前系统 操作:
Last login: Mon Dec 26 22:26:56 2016
[root@localhost ~]# systemctl disable firewalld
Removed symlink /etc/systemd/system/dbus-org.fedoraproject.FirewallD1.service.
Removed symlink /etc/systemd/system/basic.target.wants/firewalld.service.
[root@localhost ~]# systemctl stop firewalld
[root@localhost ~]#
配置主机名:
hostnamectl --static set-hostname centos-master
确定selinux是关闭的:
/etc/selinux/config
SELINUX=disabled
-----------------------------------------可选--------------------------
把下面两行加入到/etc/hosts中
61.91.161.217 gcr.io
61.91.161.217 www.gcr.io
--------------------------------------------------------------------------
以下在所有的节点上安装:
方法1:
来源于官网配置
在centos系统yum中加入:
cat < /etc/yum.repos.d/kubernetes.repo
[kubernetes]
name=Kubernetes
baseurl=http://yum.kubernetes.io/repos/kubernetes-el7-x86_64
enabled=1
gpgcheck=0
repo_gpgcheck=0
EOF
yum install -y socat kubelet kubeadm kubectl kubernetes-cni
---这里需要注意,根据自己的实际情况选择,如果在下一步下载镜 像的速度太慢,可以加上---
在docker的启动文件 /lib/systemd/system/docker.service中加入
--registry-mirror="http://b438f72b.m.daocloud.io"
详细如下:
[root@localhost ~]# vi /lib/systemd/system/docker.service
[Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
After=network.target
[Service]
Type=notify
# the default is not to use systemd for cgroups because the delegate issues still
# exists and systemd currently does not support the cgroup feature set required
# for containers run by docker
ExecStart=/usr/bin/dockerd --registry-mirror="http://b438f72b.m.daocloud.io"
ExecReload=/bin/kill -s HUP $MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
# Uncomment TasksMax if your systemd version supports it.
# Only systemd 226 and above support this version.
#TasksMax=infinity
TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
[Install]
WantedBy=multi-user.target
~
让docker开机启动
systemctl enable docker
启动docker
systemctl start docker
让kubelet开机启动
systemctl enable kubelet
启动kubelet
systemctl start kubelet
----------------------
systemctl enable docker;systemctl start docker;systemctl enable kubelet;systemctl start kubelet
---------------------------------------------
下载镜像:
images=(kube-proxy-amd64:v1.5.1 kube-discovery-amd64:1.0 kubedns-amd64:1.9 kube-scheduler-amd64:v1.5.1 kube-controller-manager-amd64:v1.5.1 kube-apiserver-amd64:v1.5.1 etcd-amd64:3.0.14-kubeadm kube-dnsmasq-amd64:1.4 exechealthz-amd64:1.2 pause-amd64:3.0 kubernetes-dashboard-amd64:v1.5.0 dnsmasq-metrics-amd64:1.0) for imageName in ${images[@]} ; do docker pull jicki/$imageName docker tag jicki/$imageName gcr.io/google_containers/$imageName docker rmi jicki/$imageName done
--------------------------------------------------
注意--
虽然我们在这里安装下载了weaveworks/weave-kube:1.8.2 但还是要注意安装weaveworks的yaml文件中对应的版本。
特别是在安装dns时候,kubeadm会自动安装,因此没有yaml,那么使用下面的命令进行查看:
kubectl --namespace=kube-system edit deployment kube-dns
--------------------------------------------------
这两个是网络
docker pull weaveworks/weave-kube:1.8.2
docker pull weaveworks/weave-npc:1.8.2
docker pull kubernetes/heapster:canary
docker pull kubernetes/heapster_influxdb:v0.6
docker pull gcr.io/google_containers/heapster_grafana:v3.1.1
以上的操作在每一个节点都要执行,镜像也要在每一个节点上下载。
接下来配置集群:
在你的master服务器上面运行:
kubeadm init --api-advertise-addresses=192.168.7.206 --pod-network-cidr 10.245.0.0/16
上面的192.168.7.206是我的master的地址。
这个命令不可以运行两回。 也就是只能运行一次。如果再次运行,需要执行 kubeadm reset.
输出 内容如下:
[root@centos-master ~]# kubeadm init
[kubeadm] WARNING: kubeadm is in alpha, please do not use it for production clusters.
[preflight] Running pre-flight checks
[init] Using Kubernetes version: v1.5.1
[tokens] Generated token: "60a95a.93c425347a1695ab"
[certificates] Generated Certificate Authority key and certificate.
[certificates] Generated API Server key and certificate
[certificates] Generated Service Account signing keys
[certificates] Created keys and certificates in "/etc/kubernetes/pki"
[kubeconfig] Wrote KubeConfig file to disk: "/etc/kubernetes/kubelet.conf"
[kubeconfig] Wrote KubeConfig file to disk: "/etc/kubernetes/admin.conf"
[apiclient] Created API client, waiting for the control plane to become ready
[apiclient] All control plane components are healthy after 81.803134 seconds
[apiclient] Waiting for at least one node to register and become ready
[apiclient] First node is ready after 2.002437 seconds
[apiclient] Creating a test deployment
[apiclient] Test deployment succeeded
[token-discovery] Created the kube-discovery deployment, waiting for it to become ready
[token-discovery] kube-discovery is ready after 22.002704 seconds
[addons] Created essential addon: kube-proxy
[addons] Created essential addon: kube-dns
Your Kubernetes master has initialized successfully!
You should now deploy a pod network to the cluster.
Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at:
http://kubernetes.io/docs/admin/addons/
You can now join any number of machines by running the following on each node:
kubeadm join --token=60a95a.93c425347a1695ab 192.168.7.206
在上面的操作中,已经初始化了master. 最后一行是token,后面将用来增加结点。
接下来安装结点:
在所有的结点运行下面的命令:
kubeadm join --token=60a95a.93c425347a1695ab 192.168.7.206
运行完输出 信息如下:
[root@centos-minion-1 kubelet]# kubeadm join --token=60a95a.93c425347a1695ab 192.168.7.206
[kubeadm] WARNING: kubeadm is in alpha, please do not use it for production clusters.
[preflight] Running pre-flight checks
[tokens] Validating provided token
[discovery] Created cluster info discovery client, requesting info from "http://192.168.7.206:9898/cluster-info/v1/?token-id=60a95a"
[discovery] Cluster info object received, verifying signature using given token
[discovery] Cluster info signature and contents are valid, will use API endpoints [https://192.168.7.206:6443]
[bootstrap] Trying to connect to endpoint https://192.168.7.206:6443
[bootstrap] Detected server version: v1.5.1
[bootstrap] Successfully established connection with endpoint "https://192.168.7.206:6443"
[csr] Created API client to obtain unique certificate for this node, generating keys and certificate signing request
[csr] Received signed certificate from the API server:
Issuer: CN=kubernetes | Subject: CN=system:node:centos-minion-1 | CA: false
Not before: 2016-12-23 07:06:00 +0000 UTC Not After: 2017-12-23 07:06:00 +0000 UTC
[csr] Generating kubelet configuration
[kubeconfig] Wrote KubeConfig file to disk: "/etc/kubernetes/kubelet.conf"
Node join complete:
* Certificate signing request sent to master and response
received.
* Kubelet informed of new secure connection details.
Run 'kubectl get nodes' on the master to see this machine join.
在master上面运行kubectl get nodes
[root@centos-master ~]# kubectl get nodes
NAME STATUS AGE
centos-master Ready,master 14m
centos-minion-1 Ready 5m
centos-minion-2 Ready 45s
网上有一个命令是:
kubectl taint nodes --all dedicated-
[root@centos-master ~]# kubectl taint nodes --all dedicated-
taint key="dedicated" and effect="" not found.
taint key="dedicated" and effect="" not found.
taint key="dedicated" and effect="" not found.
这个命令好像是可以让master上面也可以运行pod。
但执行没有效果。
kubectl get nodes
[root@centos-master ~]# kubectl get nodes
NAME STATUS AGE
centos-master Ready,master 21m
centos-minion-1 Ready 13m
centos-minion-2 Ready 8m
然后:
[root@centos-master new]# kubectl --namespace=kube-system get pod
NAME READY STATUS RESTARTS AGE
dummy-2088944543-9zfjl 1/1 Running 0 2d
etcd-centos-master 1/1 Running 0 2d
kube-apiserver-centos-master 1/1 Running 0 2d
kube-controller-manager-centos-master 1/1 Running 0 2d
kube-discovery-1769846148-6ldk1 1/1 Running 0 2d
kube-proxy-34q7p 1/1 Running 0 2d
kube-proxy-hqkkg 1/1 Running 1 2d
kube-proxy-nbgn3 1/1 Running 0 2d
kube-scheduler-centos-master 1/1 Running 0 2d
weave-net-kkdh9 2/2 Running 0 42m
weave-net-mtd83 2/2 Running 0 2m
weave-net-q91sr 2/2 Running 2 42m
现在,需要安装pod network.
执行下面命令:
kubectl apply -f https://git.io/weave-kube
这个命令里面的镜像可能无法下载,那么,按以下的方法来做。
[root@centos-master new]# vi weave-daemonset.yaml
apiVersion: extensions/v1beta1
kind: DaemonSet
metadata:
name: weave-net
namespace: kube-system
spec:
template:
metadata:
labels:
name: weave-net
annotations:
scheduler.alpha.kubernetes.io/tolerations: |
[
{
"key": "dedicated",
"operator": "Equal",
"value": "master",
"effect": "NoSchedule"
}
]
spec:
hostNetwork: true
hostPID: true
containers:
- name: weave
image: weaveworks/weave-kube:1.8.2
command:
- /home/weave/launch.sh
livenessProbe:
initialDelaySeconds: 30
httpGet:
host: 127.0.0.1
path: /status
port: 6784
securityContext:
privileged: true
volumeMounts:
- name: weavedb
mountPath: /weavedb
- name: cni-bin
mountPath: /opt
- name: cni-bin2
mountPath: /host_home
- name: cni-conf
mountPath: /etc
resources:
requests:
cpu: 10m
- name: weave-npc
image: weaveworks/weave-kube:1.8.2
resources:
requests:
cpu: 10m
securityContext:
privileged: true
restartPolicy: Always
volumes:
- name: weavedb
emptyDir: {}
- name: cni-bin
hostPath:
path: /opt
- name: cni-bin2
hostPath:
path: /home
- name: cni-conf
hostPath:
path: /etc
注意上面的镜像地址,如果前面下载的镜像不对,最好找好相应版本的镜像。
kubectl apply -f weave-daemonset.yaml
安装完后,显示pod:
[root@localhost ~]# kubectl --namespace=kube-system get pod
NAME READY STATUS RESTARTS AGE
dummy-2088944543-xjj21 1/1 Running 0 55m
etcd-centos-master 1/1 Running 0 55m
kube-apiserver-centos-master 1/1 Running 0 55m
kube-controller-manager-centos-master 1/1 Running 0 55m
kube-discovery-1769846148-c45gd 1/1 Running 0 55m
kube-dns-2924299975-96xms 4/4 Running 0 55m
kube-proxy-33lsn 1/1 Running 0 55m
kube-proxy-jnz6q 1/1 Running 0 55m
kube-proxy-vfql2 1/1 Running 0 20m
kube-scheduler-centos-master 1/1 Running 0 55m
weave-net-k5tlz 2/2 Running 0 19m
weave-net-q3n89 2/2 Running 0 19m
weave-net-x57k7 2/2 Running 0 19m
安装dashboard过程如下:
下载安装yaml文件:
wget https://rawgit.com/kubernetes/dashboard/master/src/deploy/kubernetes-dashboard.yaml
[root@centos-master new]# cat kubernetes-dashboard.yaml
# Copyright 2015 Google Inc. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
# Configuration to deploy release version of the Dashboard UI.
#
# Example usage: kubectl create -f
kind: Deployment
apiVersion: extensions/v1beta1
metadata:
labels:
app: kubernetes-dashboard
name: kubernetes-dashboard
namespace: kube-system
spec:
replicas: 1
selector:
matchLabels:
app: kubernetes-dashboard
template:
metadata:
labels:
app: kubernetes-dashboard
# Comment the following annotation if Dashboard must not be deployed on master
annotations:
scheduler.alpha.kubernetes.io/tolerations: |
[
{
"key": "dedicated",
"operator": "Equal",
"value": "master",
"effect": "NoSchedule"
}
]
spec:
containers:
- name: kubernetes-dashboard
image: gcr.io/google_containers/kubernetes-dashboard-amd64:v1.5.0
imagePullPolicy: Always
ports:
- containerPort: 9090
protocol: TCP
args:
# Uncomment the following line to manually specify Kubernetes API server Host
# If not specified, Dashboard will attempt to auto discover the API server and connect
# to it. Uncomment only if the default does not work.
# - --apiserver-host=http://my-address:port
livenessProbe:
httpGet:
path: /
port: 9090
initialDelaySeconds: 30
timeoutSeconds: 30
---
kind: Service
apiVersion: v1
metadata:
labels:
app: kubernetes-dashboard
name: kubernetes-dashboard
namespace: kube-system
spec:
type: NodePort
ports:
- port: 80
targetPort: 9090
selector:
app: kubernetes-dashboard
kubectl create -f https://rawgit.com/kubernetes/dashboard/master/src/deploy/kubernetes-dashboard.yaml
查看到
kubernetes-dashboard-3095304083-w3fjd
已经运行。
[root@centos-master new]# kubectl get pod --all-namespaces
NAMESPACE NAME READY STATUS RESTARTS AGE
kube-system dummy-2088944543-9zfjl 1/1 Running 0 2d
kube-system etcd-centos-master 1/1 Running 0 2d
kube-system kube-apiserver-centos-master 1/1 Running 0 2d
kube-system kube-controller-manager-centos-master 1/1 Running 0 2d
kube-system kube-discovery-1769846148-6ldk1 1/1 Running 0 2d
kube-system kube-proxy-34q7p 1/1 Running 0 2d
kube-system kube-proxy-hqkkg 1/1 Running 1 2d
kube-system kube-proxy-nbgn3 1/1 Running 0 2d
kube-system kube-scheduler-centos-master 1/1 Running 0 2d
kube-system kubernetes-dashboard-3095304083-w3fjd 1/1 Running 0 31m
kube-system weave-net-kkdh9 2/2 Running 0 1h
kube-system weave-net-mtd83 2/2 Running 0 48m
kube-system weave-net-q91sr 2/2 Running 2 1h
查看节点端口为:31551
[root@centos-master new]# kubectl describe svc kubernetes-dashboard --namespace=kube-system
Name: kubernetes-dashboard
Namespace: kube-system
Labels: app=kubernetes-dashboard
Selector: app=kubernetes-dashboard
Type: NodePort
IP: 10.96.35.20
Port: 80/TCP
NodePort: 31551/TCP
Endpoints: 10.40.0.1:9090
Session Affinity: None
No events.
安装heapster
yaml文件来自于github.
[root@localhost heapster]# cat grafana-deployment.yaml
apiVersion: extensions/v1beta1
kind: Deployment
metadata:
name: monitoring-grafana
namespace: kube-system
spec:
replicas: 1
template:
metadata:
labels:
task: monitoring
k8s-app: grafana
spec:
volumes:
- name: grafana-storage
emptyDir: {}
containers:
- name: grafana
image: gcr.io/google_containers/heapster_grafana:v3.1.1
ports:
- containerPort: 3000
protocol: TCP
volumeMounts:
- mountPath: /var
name: grafana-storage
env:
- name: INFLUXDB_HOST
value: monitoring-influxdb
- name: GRAFANA_PORT
value: "3000"
# The following env variables are required to make Grafana accessible via
# the kubernetes api-server proxy. On production clusters, we recommend
# removing these env variables, setup auth for grafana, and expose the grafana
# service using a LoadBalancer or a public IP.
- name: GF_AUTH_BASIC_ENABLED
value: "false"
- name: GF_AUTH_ANONYMOUS_ENABLED
value: "true"
- name: GF_AUTH_ANONYMOUS_ORG_ROLE
value: Admin
- name: GF_SERVER_ROOT_URL
# value: /api/v1/proxy/namespaces/kube-system/services/monitoring-grafana/
value: /
grafana-deployment.yaml 中的image我们在上面已经下载过
docker pull gcr.io/google_containers/heapster_grafana:v3.1.1
[root@localhost heapster]# cat grafana-service.yaml
apiVersion: v1
kind: Service
metadata:
labels:
# For use as a Cluster add-on (https://github.com/kubernetes/kubernetes/tree/master/cluster/addons)
# If you are NOT using this as an addon, you should comment out this line.
kubernetes.io/cluster-service: 'true'
kubernetes.io/name: monitoring-grafana
name: monitoring-grafana
namespace: kube-system
spec:
# In a production setup, we recommend accessing Grafana through an external Loadbalancer
# or through a public IP.
# type: LoadBalancer
type: NodePort
ports:
- port: 80
targetPort: 3000
selector:
k8s-app: grafana
[root@localhost heapster]# cat influxdb-deployment.yaml
apiVersion: extensions/v1beta1
kind: Deployment
metadata:
name: monitoring-influxdb
namespace: kube-system
spec:
replicas: 1
template:
metadata:
labels:
task: monitoring
k8s-app: influxdb
spec:
volumes:
- name: influxdb-storage
emptyDir: {}
containers:
- name: influxdb
image: kubernetes/heapster_influxdb:v0.6
volumeMounts:
- mountPath: /data
name: influxdb-storage
[root@localhost heapster]# cat influxdb-service.yaml
apiVersion: v1
kind: Service
metadata:
labels:
task: monitoring
# For use as a Cluster add-on (https://github.com/kubernetes/kubernetes/tree/master/cluster/addons)
# If you are NOT using this as an addon, you should comment out this line.
kubernetes.io/cluster-service: 'true'
kubernetes.io/name: monitoring-influxdb
name: monitoring-influxdb
namespace: kube-system
spec:
# type: NodePort
ports:
- name: api
port: 8086
targetPort: 8086
selector:
k8s-app: influxdb
heapster-deployment.yaml 文件内容
[root@localhost heapster]# cat heapster-deployment.yaml
apiVersion: extensions/v1beta1
kind: Deployment
metadata:
name: heapster
namespace: kube-system
spec:
replicas: 1
template:
metadata:
labels:
task: monitoring
k8s-app: heapster
version: v6
spec:
containers:
- name: heapster
image: kubernetes/heapster:canary
imagePullPolicy: Always
command:
- /heapster
- --source=kubernetes:https://kubernetes.default
- --sink=influxdb:http://monitoring-influxdb:8086
image也是下载过的。
[root@localhost heapster]# cat heapster-service.yaml
apiVersion: v1
kind: Service
metadata:
labels:
task: monitoring
# For use as a Cluster add-on (https://github.com/kubernetes/kubernetes/tree/master/cluster/addons)
# If you are NOT using this as an addon, you should comment out this line.
kubernetes.io/cluster-service: 'true'
kubernetes.io/name: Heapster
name: heapster
namespace: kube-system
spec:
ports:
- port: 80
targetPort: 8082
selector:
k8s-app: heapster
六个文件准备好了以后。
直接
kubectl create -f grafana-deployment.yaml -f grafana-service.yaml -f influxdb-deployment.yaml -f influxdb-service.yaml -f heapster-deployment.yaml -f heapster-service.yaml
[root@localhost heapster]# kubectl get pod --namespace=kube-system
NAME READY STATUS RESTARTS AGE
dummy-2088944543-xjj21 1/1 Running 0 2h
etcd-centos-master 1/1 Running 0 2h
heapster-2193675300-j1jxn 1/1 Running 0 1h
kube-apiserver-centos-master 1/1 Running 0 2h
kube-controller-manager-centos-master 1/1 Running 0 2h
kube-discovery-1769846148-c45gd 1/1 Running 0 1h
kube-dns-2924299975-96xms 4/4 Running 0 1h
kube-proxy-33lsn 1/1 Running 0 1h
kube-proxy-jnz6q 1/1 Running 0 1h
kube-proxy-vfql2 1/1 Running 0 1h
kube-scheduler-centos-master 1/1 Running 0 2h
kubernetes-dashboard-3000605155-8mxgz 1/1 Running 0 1h
monitoring-grafana-810108360-h92v7 1/1 Running 0 1h
monitoring-influxdb-3065341217-q2445 1/1 Running 0 1h
weave-net-k5tlz 2/2 Running 0 1h
weave-net-q3n89 2/2 Running 0 1h
weave-net-x57k7 2/2 Running 0 1h
查看服务:
[root@localhost heapster]# kubectl get svc --namespace=kube-system
NAME CLUSTER-IP EXTERNAL-IP PORT(S) AGE
heapster 10.98.45.1 80/TCP 1h
kube-dns 10.96.0.10 53/UDP,53/TCP 2h
kubernetes-dashboard 10.108.45.66 80:32155/TCP 1h
monitoring-grafana 10.97.110.225 80:30687/TCP 1h
monitoring-influxdb 10.96.175.67 8086/TCP 1h
查看grafana的详细信息:
[root@localhost heapster]# kubectl --namespace=kube-system describe svc monitoring-grafana
Name: monitoring-grafana
Namespace: kube-system
Labels: kubernetes.io/cluster-service=true
kubernetes.io/name=monitoring-grafana
Selector: k8s-app=grafana
Type: NodePort
IP: 10.97.110.225
Port: 80/TCP
NodePort: 30687/TCP
Endpoints: 10.32.0.2:3000
Session Affinity: None
No events.
看到开放端口为30687
通过节点IP加端口号访问:
点图标:
点默认;
然后确认是k8s就可以了:
接下来:
图形出来了: