千家信息网

Kubernetes V1.12 二进制 部署多Master

发表于:2025-02-03 作者:千家信息网编辑
千家信息网最后更新 2025年02月03日,环境:高可用环境架构:组件版本:软件版本Linux操作系统CentOS7.5_x64Kubernetes1.12Docker18.xx-ceEtcd3.xFlannel0.10服务器角色:角色IP组件
千家信息网最后更新 2025年02月03日Kubernetes V1.12 二进制 部署多Master

环境:

高可用环境架构:

组件版本:

软件版本
Linux操作系统CentOS7.5_x64
Kubernetes1.12
Docker18.xx-ce
Etcd3.x
Flannel0.10

服务器角色:

角色IP组件
master01192.168.1.43kube-apiserver,kube-controller-manager,kube-scheduler etcd
master02192.168.1.63kube-apiserver,kube-controller-manager,kube-scheduler etcd
node01192.168.1.30kubelet,kube-proxy,docker,flannel,etcd
node02192.168.1.51kubelet,kube-proxy,docker,flannel
node03192.168.1.141kubelet,kube-proxy,docker,flannel
Load Balancer (Master)192.168.1.31 192.168.1.230 (VIP)Nginx L4
Load Balancer (Backup)192.168.1.186Nginx L4

自签SSL证书:

组件使用的证书
etcdca.pem,server.pem,server-key.pem
flannelca.pem,server.pem,server-key.pem
kube-apiserverca.pem,server.pem,server-key.pem
kubeletca.pem,ca-key.pem
kube-proxyca.pem,kube-proxy.pem,kube-proxy-key.pem
kubectlca.pem,admin.pem,admin-key.pem

准备工作:

关闭防火墙:# systemctl stop firewalld && systemctl disable firewalld同步时间:(ssl验证时间)# yum -y install ntpdate && ntpdate time.windows.com

1.部署ETCD集群

1.1安装cfssl

curl -L https://pkg.cfssl.org/R1.2/cfssl_linux-amd64 -o /usr/local/bin/cfssl                #cfssl来生成证书curl -L https://pkg.cfssl.org/R1.2/cfssljson_linux-amd64 -o /usr/local/bin/cfssljson        #cfssljson传入json文件生成证书curl -L https://pkg.cfssl.org/R1.2/cfssl-certinfo_linux-amd64 -o /usr/local/bin/cfssl-certinfo      #cfssl-cetinfo查看生成证书信息chmod +x /usr/local/bin/cfssl /usr/local/bin/cfssljson /usr/local/bin/cfssl-certinfo

1.2生成证书

# mkdir ~/k8s/etcd-cert -p# cd ~/k8s/etcd-cert

ca根证书:

# cat > ca-config.json <

ca请求签名证书:

# cat > ca-csr.json <

为ETCD颁发ssl证书:(将etcd节点ip加入其中)

# cat > server-csr.json <

生成证书:

初始化ca根证书:        cfssl gencert -initca ca-csr.json | cfssljson -bare ca -    #会生成ca-key.pem,ca.pem生成证书:        cfssl gencert -ca=ca.pem -ca-key=ca-key.pem -config=ca-config.json -profile=www server-csr.json | cfssljson -bare server     #会生成server-key,server#说明:        #-ca=ca.pem 指定ca        #-ca-key=ca-key.pem 指定ca私钥        #-config=ca-config.json 指定ca配置文件        #-profile=www 应用配置文件中的www

1.3ETCD数据库集群部署

二进制包下载:https://github.com/etcd-io/etcd/releases

解压二进制包:

# cd ~/k8s# tar -zxvf etcd-v3.3.10-linux-amd64.tar.gz创建etcd目录:# mkdir /opt/etcd/{cfg,bin,ssl} -p     #配置,可执行,证书目录

移动可执行文件到etcd目录:

# cd ~/k8s/etcd-v3.3.10-linux-amd64# mv etcd etcdctl /opt/etcd/bin/# ls /opt/etcd/bin/    etcd  etcdctl

把刚生成的拷贝ssl文件到etc目录:

# cd ~/k8s/etcd-cert# cp *pem # ls /opt/etcd/ssl/    ca-key.pem  ca.pem  server-key.pem  server.pem

创建etcd配置文件:

# cat </opt/etcd/cfg/etcd#[Member]ETCD_NAME="etcd01"ETCD_DATA_DIR="/var/lib/etcd/default.etcd"ETCD_LISTEN_PEER_URLS="https://192.168.1.43:2380"ETCD_LISTEN_CLIENT_URLS="https://192.168.1.43:2379"#[Clustering]ETCD_INITIAL_ADVERTISE_PEER_URLS="https://192.168.1.43:2380"ETCD_ADVERTISE_CLIENT_URLS="https://192.168.1.43:2379"ETCD_INITIAL_CLUSTER="etcd01=https://192.168.1.43:2380,etcd02=https://192.168.1.30:2380,etcd03=https://192.168.1.51:2380"ETCD_INITIAL_CLUSTER_TOKEN="etcd-cluster"ETCD_INITIAL_CLUSTER_STATE="new"EOF
  • ETCD_NAME 节点名称
  • ETCD_DATA_DIR 数据目录
  • ETCD_LISTEN_PEER_URLS 集群通信监听地址
  • ETCD_LISTEN_CLIENT_URLS 客户端访问监听地址
  • ETCD_INITIAL_ADVERTISE_PEER_URLS 集群通告地址
  • ETCD_ADVERTISE_CLIENT_URLS 客户端通告地址
  • ETCD_INITIAL_CLUSTER 集群节点地址
  • ETCD_INITIAL_CLUSTER_TOKEN 集群Token
  • ETCD_INITIAL_CLUSTER_STATE 加入集群的当前状态,new是新集群,existing表示加入已有集群

创建systemctld管理文件:

# cat /usr/lib/systemd/system/etcd.service [Unit]Description=Etcd ServerAfter=network.targetAfter=network-online.targetWants=network-online.target[Service]Type=notifyEnvironmentFile=/opt/etcd/cfg/etcdExecStart=/opt/etcd/bin/etcd \--name=${ETCD_NAME} \--data-dir=${ETCD_DATA_DIR} \--listen-peer-urls=${ETCD_LISTEN_PEER_URLS} \--listen-client-urls=${ETCD_LISTEN_CLIENT_URLS},http://127.0.0.1:2379 \--advertise-client-urls=${ETCD_ADVERTISE_CLIENT_URLS} \--initial-advertise-peer-urls=${ETCD_INITIAL_ADVERTISE_PEER_URLS} \--initial-cluster=${ETCD_INITIAL_CLUSTER} \--initial-cluster-token=${ETCD_INITIAL_CLUSTER_TOKEN} \--initial-cluster-state=new \--cert-file=/opt/etcd/ssl/server.pem \--key-file=/opt/etcd/ssl/server-key.pem \--peer-cert-file=/opt/etcd/ssl/server.pem \--peer-key-file=/opt/etcd/ssl/server-key.pem \--trusted-ca-file=/opt/etcd/ssl/ca.pem \--peer-trusted-ca-file=/opt/etcd/ssl/ca.pemRestart=on-failureLimitNOFILE=65536[Install]WantedBy=multi-user.target

开机并启动etcd:

# systemctl daemon-reload && systemctl enable etcd && systemctl restart etcd

1.4其他节点安装etcd

拷贝etcd文件到node1,node2

# scp -r /opt/etcd/ root@192.168.1.30:/opt/# scp -r /opt/etcd/ root@192.168.1.51:/opt/# scp /usr/lib/systemd/system/etcd.service root@192.168.1.51:/usr/lib/systemd/system/# scp /usr/lib/systemd/system/etcd.service root@192.168.1.30:/usr/lib/systemd/system/

修改配置文件

node1:# cat /opt/etcd/cfg/etcd#[Member]ETCD_NAME="etcd02"ETCD_DATA_DIR="/var/lib/etcd/default.etcd"ETCD_LISTEN_PEER_URLS="https://192.168.1.30:2380"ETCD_LISTEN_CLIENT_URLS="https://192.168.1.30:2379"#[Clustering]ETCD_INITIAL_ADVERTISE_PEER_URLS="https://192.168.1.30:2380"ETCD_ADVERTISE_CLIENT_URLS="https://192.168.1.30:2379"ETCD_INITIAL_CLUSTER="etcd01=https://192.168.1.43:2380,etcd02=https://192.168.1.30:2380,etcd03=https://192.168.1.51:2380"ETCD_INITIAL_CLUSTER_TOKEN="etcd-cluster"ETCD_INITIAL_CLUSTER_STATE="new"node2:# cat /opt/etcd/cfg/etcd#[Member]ETCD_NAME="etcd03"ETCD_DATA_DIR="/var/lib/etcd/default.etcd"ETCD_LISTEN_PEER_URLS="https://192.168.1.51:2380"ETCD_LISTEN_CLIENT_URLS="https://192.168.1.51:2379"#[Clustering]ETCD_INITIAL_ADVERTISE_PEER_URLS="https://192.168.1.51:2380"ETCD_ADVERTISE_CLIENT_URLS="https://192.168.1.51:2379"ETCD_INITIAL_CLUSTER="etcd01=https://192.168.1.43:2380,etcd02=https://192.168.1.30:2380,etcd03=https://192.168.1.51:2380"ETCD_INITIAL_CLUSTER_TOKEN="etcd-cluster"ETCD_INITIAL_CLUSTER_STATE="new"开机并启动etcd:# systemctl daemon-reload && systemctl enable etcd && systemctl restart etcd

查看etcd集群状态:

# cd /root/k8s/etcd-cert#  /opt/etcd/bin/etcdctl \> --ca-file=ca.pem --cert-file=server.pem --key-file=server-key.pem \> --endpoints="https://192.168.1.43:2379,https://192.168.1.30:2379,https://192.168.1.51:2379" \> cluster-health    member 8da171dbef9ded69 is healthy: got healthy result from https://192.168.1.51:2379    member d250ef9d0d70c7c9 is healthy: got healthy result from https://192.168.1.30:2379    member f3b3c9aa5b97cee8 is healthy: got healthy result from https://192.168.1.43:2379    cluster is healthy

2.在node安装docker

# yum install -y yum-utils device-mapper-persistent-data lvm2     #安装依赖包 # yum-config-manager --add-repo https://download.docker.com/linux/centos/docker-ce.repo     #添加Docker软件包源 # yum install -y docker-ce     #安装Docker CE # systemctl start docker && systemctl enable docker     #启动Docker服务并设置开机启动

3.部署Flannel网络

工作原理:

3.1将flannel子网段写入etcd中,供flanneld使用:

# cd /root/k8s/etcd-cert# /opt/etcd/bin/etcdctl \--ca-file=ca.pem --cert-file=server.pem --key-file=server-key.pem \--endpoints="https://192.168.1.43:2379,https://192.168.1.30:2379,https://192.168.1.51:2379" \set /coreos.com/network/config  '{ "Network": "172.17.0.0/16", "Backend": {"Type": "vxlan"}}'    { "Network": "172.17.0.0/16", "Backend": {"Type": "vxlan"}}

3.2部署Flannel(只部署到node节点)

下载二进制包:https://github.com/coreos/flannel/releases

解压二进制包:# tar -zxvf flannel-v0.10.0-linux-amd64.tar.gz创建k8s目录# mkdir /opt/kubernetes/{cfg,bin,ssl} -p移动可执行文件到k8s目录# mv flanneld mk-docker-opts.sh /opt/kubernetes/bin

创建flannel配置文件:

# cat </opt/kubernetes/cfg/flanneldFLANNEL_OPTIONS="--etcd-endpoints=https://192.168.1.43:2379,https://192.168.1.30:2379,https://192.168.1.51:2379 \-etcd-cafile=/opt/etcd/ssl/ca.pem \-etcd-certfile=/opt/etcd/ssl/server.pem \-etcd-keyfile=/opt/etcd/ssl/server-key.pem"EOF

创建flannel system管理文件:

cat </usr/lib/systemd/system/flanneld.service[Unit]Description=Flanneld overlay address etcd agentAfter=network-online.target network.targetBefore=docker.service[Service]Type=notifyEnvironmentFile=/opt/kubernetes/cfg/flanneldExecStart=/opt/kubernetes/bin/flanneld --ip-masq \$FLANNEL_OPTIONSExecStartPost=/opt/kubernetes/bin/mk-docker-opts.sh -k DOCKER_NETWORK_OPTIONS -d /run/flannel/subnet.envRestart=on-failure[Install]WantedBy=multi-user.targetEOF

配置Docker启动指定子网段:

# vim /usr/lib/systemd/system/docker.serviceEnvironmentFile=/run/flannel/subnet.envExecStart=/usr/bin/dockerd $DOCKER_NETWORK_OPTIONSExecReload=/bin/kill -s HUP $MAINPID

重启docker和flannel

# systemctl daemon-reload && systemctl start flanneld && systemctl enable flanneld# systemctl restart docker

检查是否生效

# ps -ef |grep docker    root      42770      1  0 12:41 ?        00:00:00 /usr/bin/dockerd --bip=172.17.75.1/24 --ip-masq=false --mtu=1450# ip addr3: flannel.1:  mtu 1450 qdisc noqueue state UNKNOWN group   default     link/ether ce:e0:c4:9f:7b:64 brd ff:ff:ff:ff:ff:ff    inet 172.17.75.0/32 scope global flannel.1       valid_lft forever preferred_lft forever    inet6 fe80::cce0:c4ff:fe9f:7b64/64 scope link        valid_lft forever preferred_lft forever4: docker0:  mtu 1500 qdisc noqueue state DOWN group default     link/ether 02:42:41:6d:53:ce brd ff:ff:ff:ff:ff:ff    inet 172.17.75.1/24 brd 172.17.75.255 scope global docker0       valid_lft forever preferred_lft forever

拷贝文件到其他节点:

scp -r /opt/kubernetes/ root@192.168.1.51:/optscp -r /usr/lib/systemd/system/{flanneld,docker}.service root@192.168.1.51:/usr/lib/systemd/system/

最后保证全网互通。

# docker run -it busybox sh# ping 172.17.67.2

4.在Master节点部署组件

在部署Kubernetes之前一定要确保etcd、flannel、docker是正常工作的,否则先解决问题再继续。

4.1生成证书

创建ca证书:

创建目录:# cd ~/k8s# mkdir k8s-cert# cd k8s-cert# cat > ca-config.json < ca-csr.json <

生成api server证书(注意授权ip访问apiserver,高可用需要加入master ip,lb ip,VIP)

cat > server-csr.json <

生成kube-proxy证书

cat > kube-proxy-csr.json <

4.2部署apiserver组件

下载二进制包:https://github.com/kubernetes/kubernetes/blob/master/CHANGELOG-1.12.md
下载这个包(kubernetes-server-linux-amd64.tar.gz)就够了,包含了所需的所有组件。

# cd ~/k8s# tar -zxvf kubernetes-server-linux-amd64.tar.gz# cd ~/k8s/kubernetes/server/bin/# cp kube-apiserver kube-scheduler kube-controller-manager kubectl /opt/kubernetes/bin/

创建token文件:

生成token:# head -c 16 /dev/urandom | od -An -t x | tr -d ' '# vim /opt/kubernetes/cfg/token.csv    2f7a15198f7c0c3af3ba7f264b6885c2,kubelet-bootstrap,10001,"system:kubelet-bootstrap"

第一列:随机字符串,自己可生成
第二列:用户名
第三列:UID
第四列:用户组
创建apiserver配置文件:(注意修改master地址,etcd服务)

cat </opt/kubernetes/cfg/kube-apiserverKUBE_APISERVER_OPTS="--logtostderr=true \\--v=4 \\--etcd-servers=https://192.168.1.43:2379,https://192.168.1.30:2379,https://192.168.1.51:2379 \\--bind-address=192.168.1.43 \\--secure-port=6443 \\--advertise-address=192.168.1.43 \\--allow-privileged=true \\--service-cluster-ip-range=10.0.0.0/24 \\--enable-admission-plugins=NamespaceLifecycle,LimitRanger,ServiceAccount,ResourceQuota,NodeRestriction \\--authorization-mode=RBAC,Node \\--kubelet-https=true \\--enable-bootstrap-token-auth \\--token-auth-file=/opt/kubernetes/cfg/token.csv \\--service-node-port-range=30000-50000 \\--tls-cert-file=/opt/kubernetes/ssl/server.pem  \\--tls-private-key-file=/opt/kubernetes/ssl/server-key.pem \\--client-ca-file=/opt/kubernetes/ssl/ca.pem \\--service-account-key-file=/opt/kubernetes/ssl/ca-key.pem \\--etcd-cafile=/opt/etcd/ssl/ca.pem \\--etcd-certfile=/opt/etcd/ssl/server.pem \\--etcd-keyfile=/opt/etcd/ssl/server-key.pem"EOF

配置好前面生成的证书,确保能连接etcd。

参数说明:

  • --logtostderr 启用日志
  • ---v 日志等级
  • --etcd-servers etcd集群地址
  • --bind-address 监听地址
  • --secure-port https安全端口
  • --advertise-address 集群通告地址
  • --allow-privileged 启用授权
  • --service-cluster-ip-range Service虚拟IP地址段
  • --enable-admission-plugins 准入控制模块
  • --authorization-mode 认证授权,启用RBAC授权和节点自管理
  • --enable-bootstrap-token-auth 启用TLS bootstrap功能,后面会讲到
  • --token-auth-file token文件
  • --service-node-port-range Service Node类型默认分配端口范围

systemd管理apiserver:

cat </usr/lib/systemd/system/kube-apiserver.service[Unit]Description=Kubernetes API ServerDocumentation=https://github.com/kubernetes/kubernetes[Service]EnvironmentFile=-/opt/kubernetes/cfg/kube-apiserverExecStart=/opt/kubernetes/bin/kube-apiserver \$KUBE_APISERVER_OPTSRestart=on-failure[Install]WantedBy=multi-user.targetEOF

启动

# systemctl daemon-reload && systemctl enable kube-apiserver && systemctl restart kube-apiserver# ps -ef | grep kube-apiserver

4.3部署kube-scheduler

创建配置文件:

# cat </opt/kubernetes/cfg/kube-schedulerKUBE_SCHEDULER_OPTS="--logtostderr=true \\--v=4 \\--master=127.0.0.1:8080 \\--leader-elect"EOF

参数说明:

  • --master 连接本地apiserver
  • --leader-elect 当该组件启动多个时,自动选举(HA)

创建systemd管理文件:

# cat </usr/lib/systemd/system/kube-scheduler.service[Unit]Description=Kubernetes SchedulerDocumentation=https://github.com/kubernetes/kubernetes[Service]EnvironmentFile=-/opt/kubernetes/cfg/kube-schedulerExecStart=/opt/kubernetes/bin/kube-scheduler \$KUBE_SCHEDULER_OPTSRestart=on-failure[Install]WantedBy=multi-user.targetEOF

启动:

# systemctl daemon-reload  && systemctl enable kube-scheduler && systemctl restart kube-scheduler# ps -ef | grep kube-scheduler

4.4部署controller-manager组件

创建controller-manager配置文件:

cat </opt/kubernetes/cfg/kube-controller-managerKUBE_CONTROLLER_MANAGER_OPTS="--logtostderr=true \\--v=4 \\--master=127.0.0.1:8080 \\--leader-elect=true \\--address=127.0.0.1 \\--service-cluster-ip-range=10.0.0.0/24 \\--cluster-name=kubernetes \\--cluster-signing-cert-file=/opt/kubernetes/ssl/ca.pem \\--cluster-signing-key-file=/opt/kubernetes/ssl/ca-key.pem  \\--root-ca-file=/opt/kubernetes/ssl/ca.pem \\--service-account-private-key-file=/opt/kubernetes/ssl/ca-key.pem \\--experimental-cluster-signing-duration=87600h0m0s"EOF

systemd管理controller-manager组件:

cat </usr/lib/systemd/system/kube-controller-manager.service[Unit]Description=Kubernetes Controller ManagerDocumentation=https://github.com/kubernetes/kubernetes[Service]EnvironmentFile=-/opt/kubernetes/cfg/kube-controller-managerExecStart=/opt/kubernetes/bin/kube-controller-manager \$KUBE_CONTROLLER_MANAGER_OPTSRestart=on-failure[Install]WantedBy=multi-user.targetEOF

启动

# systemctl daemon-reload && systemctl enable kube-controller-manager && systemctl restart kube-controller-manager# ps -ef | grep kube-controller-manager

所有组件都已经启动成功,通过kubectl工具查看当前集群组件状态:

# /opt/kubernetes/bin/kubectl get csNAME                 STATUS    MESSAGE             ERRORscheduler            Healthy   ok                  controller-manager   Healthy   ok                  etcd-1               Healthy   {"health":"true"}   etcd-0               Healthy   {"health":"true"}   etcd-2               Healthy   {"health":"true"} 如上输出说明组件都正常。

5.部署node组件

Master apiserver启用TLS认证后,Node节点kubelet组件想要加入集群,必须使用CA签发的有效证书才能与apiserver通信,当Node节点很多时,签署证书是一件很繁琐的事情,因此有了TLS Bootstrapping机制,kubelet会以一个低权限用户自动向apiserver申请证书,kubelet的证书由apiserver动态签署。

认证大致工作流程如图所示:

5.1将kubelet-bootstrap用户绑定到系统集群角色(在master上)

kubectl create clusterrolebinding kubelet-bootstrap \  --clusterrole=system:node-bootstrapper \  --user=kubelet-bootstrap

5.2创建kubeconfig文件

创建kubelet bootstrapping kubeconfig(在master上)

# cd ~/k8s# mkdir kubeconfig# cd kubeconfig/设置kubectl环境变量:# vi /etc/profile# export PATH=$PATH:/opt/kubernetes/bin/# source /etc/profile# 设置集群参数kubectl config set-cluster kubernetes \  --certificate-authority=/root/k8s/k8s-cert/ca.pem \  --embed-certs=true \  --server=https://192.168.1.43:6443 \  --kubeconfig=bootstrap.kubeconfig# 设置客户端认证参数kubectl config set-credentials kubelet-bootstrap \  --token=2f7a15198f7c0c3af3ba7f264b6885c2 \  --kubeconfig=bootstrap.kubeconfig# 设置上下文参数kubectl config set-context default \  --cluster=kubernetes \  --user=kubelet-bootstrap \  --kubeconfig=bootstrap.kubeconfig# 设置默认上下文kubectl config use-context default --kubeconfig=bootstrap.kubeconfig

创建kube-proxy kubeconfig文件:(在master上)

kubectl config set-cluster kubernetes \  --certificate-authority=/root/k8s/k8s-cert/ca.pem \  --embed-certs=true \  --server=https://192.168.1.43:6443 \  --kubeconfig=kube-proxy.kubeconfigkubectl config set-credentials kube-proxy \  --client-certificate=/root/k8s/k8s-cert/kube-proxy.pem \  --client-key=/root/k8s/k8s-cert/kube-proxy-key.pem \  --embed-certs=true \  --kubeconfig=kube-proxy.kubeconfigkubectl config set-context default \  --cluster=kubernetes \  --user=kube-proxy \  --kubeconfig=kube-proxy.kubeconfigkubectl config use-context default --kubeconfig=kube-proxy.kubeconfig
# ls    bootstrap.kubeconfig  kube-proxy.kubeconfig

拷贝配置文件到node

# scp kube-proxy.kubeconfig bootstrap.kubeconfig root@192.168.1.30:/opt/kubernetes/cfg/# scp kube-proxy.kubeconfig bootstrap.kubeconfig root@192.168.1.51:/opt/kubernetes/cfg/

5.3部署kubelet组件

将前面下载的二进制包中的kubelet和kube-proxy拷贝到/opt/kubernetes/bin目录下。

# cd ~/k8s/kubernetes/server/bin# scp kubelet  kube-proxy root@192.168.1.30:/opt/kubernetes/bin/# scp kubelet  kube-proxy root@192.168.1.51:/opt/kubernetes/bin/

创建kubelet配置文件:

cat </opt/kubernetes/cfg/kubeletKUBELET_OPTS="--logtostderr=true \\--v=4 \\--hostname-override=192.168.1.30 \\--kubeconfig=/opt/kubernetes/cfg/kubelet.kubeconfig \\--bootstrap-kubeconfig=/opt/kubernetes/cfg/bootstrap.kubeconfig \\--config=/opt/kubernetes/cfg/kubelet.config \\--cert-dir=/opt/kubernetes/ssl \\--pod-infra-container-image=registry.cn-hangzhou.aliyuncs.com/google-containers/pause-amd64:3.0"EOF

参数说明:

  • --hostname-override 在集群中显示的主机名
  • --kubeconfig 指定kubeconfig文件位置,会自动生成
  • --bootstrap-kubeconfig 指定刚才生成的bootstrap.kubeconfig文件
  • --cert-dir 颁发证书存放位置
  • --pod-infra-container-image 管理Pod网络的镜像

其中/opt/kubernetes/cfg/kubelet.config配置文件如下:

cat </opt/kubernetes/cfg/kubelet.configkind: KubeletConfigurationapiVersion: kubelet.config.k8s.io/v1beta1address: 192.168.1.30 port: 10250readOnlyPort: 10255cgroupDriver: cgroupfsclusterDNS:- 10.0.0.2clusterDomain: cluster.local.failSwapOn: falseauthentication:  anonymous:    enabled: trueEOF

systemd管理kubelet组件:

cat </usr/lib/systemd/system/kubelet.service[Unit]Description=Kubernetes KubeletAfter=docker.serviceRequires=docker.service[Service]EnvironmentFile=/opt/kubernetes/cfg/kubeletExecStart=/opt/kubernetes/bin/kubelet \$KUBELET_OPTSRestart=on-failureKillMode=process[Install]WantedBy=multi-user.targetEOF

启动:

# systemctl daemon-reload && systemctl enable kubelet && systemctl restart kubelet# ps -ef | grep kubelet

在Master审批Node加入集群:
启动后还没加入到集群中,需要手动允许该节点才可以。
在Master节点查看请求签名的Node:

# kubectl get csr# kubectl certificate approve XXXXID# kubectl get node

5.4部署kube-proxy组件

创建kube-proxy配置文件:

cat </opt/kubernetes/cfg/kube-proxyKUBE_PROXY_OPTS="--logtostderr=true \\--v=4 \\--hostname-override=192.168.1.30 \\--cluster-cidr=10.0.0.0/24 \\--proxy-mode=ipvs \\--masquerade-all=true \\--kubeconfig=/opt/kubernetes/cfg/kube-proxy.kubeconfig"EOF

systemd管理kube-proxy组件:

cat </usr/lib/systemd/system/kube-proxy.service[Unit]Description=Kubernetes ProxyAfter=network.target[Service]EnvironmentFile=-/opt/kubernetes/cfg/kube-proxyExecStart=/opt/kubernetes/bin/kube-proxy \$KUBE_PROXY_OPTSRestart=on-failure[Install]WantedBy=multi-user.targetEOF

启动:

# systemctl daemon-reload && systemctl enable kube-proxy && systemctl restart kube-proxy# ps -ef | grep kube-proxy

拷贝配置文件到其他node:

配置文件:# scp -r /opt/kubernetes/ root@192.168.1.51:/opt/systemd管理文件:# scp /usr/lib/systemd/system/{kubelet,kube-proxy}.service root@192.168.1.51:/usr/lib/systemd/system/删除ssl文件(master颁发):# rm -f /opt/kubernetes/ssl/*修改配置文件(节点ip):# cd /opt/kubernetes/cfg    kubelet,kubelet.config,kube-proxy,

启动:

# systemctl daemon-reload && systemctl enable kube-proxy && systemctl restart kube-proxy# ps -ef | grep kube-proxy# systemctl daemon-reload && systemctl enable kubelet && systemctl restart kubelet# ps -ef | grep kubelet

在Master审批Node加入集群:
启动后还没加入到集群中,需要手动允许该节点才可以。
在Master节点查看请求签名的Node:

# kubectl get csr# kubectl certificate approve XXXXID# kubectl get node

6.查看集群状态

# kubectl get nodeNAME           STATUS   ROLES    AGE   VERSION192.168.1.30   Ready       14h   v1.12.7192.168.1.51   Ready       23s   v1.12.7# kubectl get csNAME                 STATUS    MESSAGE             ERRORcontroller-manager   Healthy   ok                  scheduler            Healthy   ok                  etcd-1               Healthy   {"health":"true"}   etcd-2               Healthy   {"health":"true"}   etcd-0               Healthy   {"health":"true"}

至此单master搭建完毕,下面拓展多master

7.master02部署

拷贝所有组件到master02:# scp -r /opt/kubernetes/ root@192.168.1.63:/opt拷贝systemd文件拷贝:# scp /usr/lib/systemd/system/{kube-apiserver,kube-controller-manager,kube-scheduler}.service root@192.168.1.63:/usr/lib/systemd/system/拷贝etcd文件:# scp -r /opt/etcd/ root@192.168.1.63:/opt/修改apiserver地址(address):# vi /opt/kubernetes/cfg/kube-apiserver

启动:

启动kube-apiserver:# systemctl daemon-reload && systemctl enable kube-apiserver && systemctl restart kube-apiserver启动kube-scheduler:# systemctl daemon-reload  && systemctl enable kube-scheduler && systemctl restart kube-scheduler启动kube-controller-manager:# systemctl daemon-reload && systemctl enable kube-controller-manager && systemctl restart kube-controller-manager查看三个组件启动:#ps -ef | grep kube

查看集群状态:

设置kubectl环境变量:# vi /etc/profile# export PATH=$PATH:/opt/kubernetes/bin/# source /etc/profile# kubectl get nodeNAME           STATUS   ROLES    AGE   VERSION192.168.1.30   Ready       15h   v1.12.7192.168.1.51   Ready       53m   v1.12.7# kubectl get csNAME                 STATUS    MESSAGE             ERRORcontroller-manager   Healthy   ok                  scheduler            Healthy   ok                  etcd-0               Healthy   {"health":"true"}   etcd-1               Healthy   {"health":"true"}   etcd-2               Healthy   {"health":"true"}

8.Load Balancer 部署

8.1部署nginx

nginx-master:

配置源:# vim /etc/yum.repos.d/nginx.repo[nginx]name=nginx repobaseurl=http://nginx.org/packages/centos/7/$basearch/gpgcheck=0安装nginx:# yum -y install nginx添加L4负载均衡:# vim /etc/nginx/nginx.confstream {   log_format  main  '$remote_addr $upstream_addr - [$time_local] $status $upstream_bytes_sent';    access_log  /var/log/nginx/k8s-access.log  main;    upstream k8s-apiserver {        server 192.168.1.43:6443;        server 192.168.1.63:6443;    }    server {                listen 6443;                proxy_pass k8s-apiserver;    }    }

启动:

关闭selinux:# setenforce 0# vi /etc/selinux/config将SELINUX=enforcing改为SELINUX=disabled#systemctl start nginx# netstat -anpt | grep 6443# echo "master" > /usr/share/nginx/html/index.html

nginx-backup:

配置源:# vim /etc/yum.repos.d/nginx.repo[nginx]name=nginx repobaseurl=http://nginx.org/packages/centos/7/$basearch/gpgcheck=0安装nginx:# yum -y install nginx拷贝到backup:# scp /etc/nginx/nginx.conf root@192.168.1.31:/etc/nginx/关闭selinux:# setenforce 0# vi /etc/selinux/config将SELINUX=enforcing改为SELINUX=disabled#systemctl start nginx# netstat -anpt | grep 6443# echo "backup" > /usr/share/nginx/html/index.html

8.2部署keepalived

master和backup安装keeplived:

# yum -y install keepalived

master的keeplived配置文件:

# vim /etc/keepalived/keepalived.conf! Configuration File for keepalivedglobal_defs {   # 接收邮件地址   notification_email {     acassen@firewall.loc     failover@firewall.loc     sysadmin@firewall.loc   }   # 邮件发送地址   notification_email_from Alexandre.Cassen@firewall.loc   smtp_server 127.0.0.1   smtp_connect_timeout 30   router_id NGINX_MASTER}vrrp_script check_nginx {    script "/etc/nginx/check_nginx.sh"}vrrp_instance VI_1 {    state MASTER    interface ens32    virtual_router_id 51 # VRRP 路由 ID实例,每个实例是唯一的    priority 100    # 优先级,备服务器设置 90    advert_int 1    # 指定VRRP 心跳包通告间隔时间,默认1秒    authentication {        auth_type PASS        auth_pass 1111    }    virtual_ipaddress {        192.168.1.230/24    }    track_script {        check_nginx    }}

backup的keeplived配置文件:

! Configuration File for keepalived global_defs {    # 接收邮件地址    notification_email {      acassen@firewall.loc      failover@firewall.loc      sysadmin@firewall.loc    }    # 邮件发送地址    notification_email_from Alexandre.Cassen@firewall.loc     smtp_server 127.0.0.1    smtp_connect_timeout 30    router_id NGINX_BACKUP} vrrp_script check_nginx {    script "/etc/nginx/check_nginx.sh"}vrrp_instance VI_1 {     state BACKUP    interface ens32    virtual_router_id 51 # VRRP 路由 ID实例,每个实例是唯一的     priority 90    # 优先级,备服务器设置 90     advert_int 1    # 指定VRRP 心跳包通告间隔时间,默认1秒     authentication {         auth_type PASS              auth_pass 1111     }      virtual_ipaddress {         192.168.1.230/24     }     track_script {        check_nginx    } }

nginx检查脚本:

# vim /etc/nginx/check_nginx.shcount=$(ps -ef |grep nginx |egrep -cv "grep|$$")if [ "$count" -eq 0 ];then    systemctl stop keepalivedfi

启动:

# systemctl start keepalived关闭master的nginx进行测试:# systemctl stop nginx

9.修改node节点apiserver为负载均衡ip

# cd /opt/kubernetes/cfg# vi bootstrap.kubeconfig# vi kubelet.kubeconfig# vi kube-proxy.kubeconfig# systemctl restart kubelet# systemctl restart kube-proxy

10.运行一个测试示例

10.1授权kubelet:

kubectl create clusterrolebinding system:anonymous --clusterrole=cluster-admin --user=system:anonymous

10.2创建一个Nginx Web,测试集群是否正常工作:

# kubectl run nginx --image=nginx --replicas=3# kubectl expose deployment nginx --port=88 --target-port=80 --type=NodePort

查看Pod,Service:

# kubectl get podNAME                    READY   STATUS    RESTARTS   AGEnginx-dbddb74b8-j4bjq   1/1     Running   0          19mnginx-dbddb74b8-kpqht   1/1     Running   0          19mnginx-dbddb74b8-xjn5k   1/1     Running   0          19m# kubectl get svcNAME         TYPE        CLUSTER-IP   EXTERNAL-IP   PORT(S)        AGEkubernetes   ClusterIP   10.0.0.1             443/TCP        16hnginx        NodePort    10.0.0.33            88:32694/TCP   20m

11.部署Web UI(Dashboard)

地址:https://github.com/kubernetes/kubernetes/tree/master/cluster/addons/dashboard

# cd /k8s/Dashboard# ls    dashboard-configmap.yaml  dashboard-controller.yaml  dashboard-rbac.yaml  dashboard-secret.yaml  dashboard-service.yaml  k8s-admin.yaml# kubectl apply -f .# kubectl get pod,svc -o wide --all-namespaces | grep dashboardkube-system   pod/kubernetes-dashboard-65f974f565-crvwj   1/1     Running   1          6m1s   172.17.75.2   192.168.1.30   kube-system   service/kubernetes-dashboard   NodePort    10.0.0.192           443:30001/TCP   6m      k8s-app=kubernetes-dashboard

访问(尽量用火狐):https://192.168.1.30:30001

查看token:

#  kubectl get secret --all-namespaces | grep dashboardkube-system   dashboard-admin-token-nrvzx        kubernetes.io/service-account-token   3      9m16skube-system   kubernetes-dashboard-certs         Opaque                                0      9m17skube-system   kubernetes-dashboard-key-holder    Opaque                                2      9m17skube-system   kubernetes-dashboard-token-cqqm8   kubernetes.io/service-account-token   3      9m17s# kubectl describe secret dashboard-admin-token-nrvzx -n kube-system
0