# K8s **Repository Path**: yujianzhao/K8s ## Basic Information - **Project Name**: K8s - **Description**: k8s集群二进制包一键化,支持centos7.3-7.7++++kubernetes v1.14.4集群一键离线安装,一键批量增删node节点,一键集成k8s持久化方案Heketi+GlusterFS+nfs+helm+动态存储+dns+ipvs+prometheus +grafan - **Primary Language**: Shell - **License**: Not specified - **Default Branch**: master - **Homepage**: None - **GVP Project**: No ## Statistics - **Stars**: 0 - **Forks**: 314 - **Created**: 2019-12-07 - **Last Updated**: 2022-05-31 ## Categories & Tags **Categories**: Uncategorized **Tags**: None ## README QQ交流群 微信交流群 ==== ![template.png](https://images.gitee.com/uploads/images/2019/1010/140405_7e20f2dd_525507.png) * Q群名称:K8s自动化部署交流 * Q群 号:893480182 ### - 实测单机版支持腾讯云服务器 ### - 实测集群版支持天翼云服务器 ### - 实测集群版支持阿里云服务器(需要预先设置中文字符集否则菜单界面乱码不显示中文,后续我会修复这个问题到脚本里面 暂时没有阿里云服务器调试) 2019-11-06最新内测版地址(升级监控,升级helm,升级所有软件依赖,升级docker-ce最新版兼容centos7.3-7.7) ==== http://117.27.146.72:42344/K8s_Beta/ ==== 高能警告:不要 yum update!! ==== 特点: 完全离线,不依赖互联网(系统新装配置好ip开机即可!其他都什么都不用安装) ==== 1. 真正原生centos7.3.-7.6Minimal新装系统(只需要统一集群的root密码即可)一键搭建k8s集群 1. 单机/集群任意服务器数量一键安装(目前一个节点对应一个etcd节点后续会分离可自定义) 1. 一键批量增删node节点(新增的服务器系统环境必须干净密码统一) 1. ipvs负载均衡,内网yum源共享页面端口42344 1. 单机版或集群版默认集成prometheus+grafan监控环境,默认端口30000,账号密码admin admin 1. 图形化向导菜单安装,web管理页面dashboar端口42345 1. Heketi+GlusterFS(分布式存储集群)+helm完全一键离线部署 1. 默认版本v1.14.4,可执行替换各版本软件包,集群版目前已测安装数量在1-30台一键安装正常 1. 集群数量超过4台及以上默认开启k8s数据持久化方案:glusterfs+Heketi 最低3台分布式存储 \n (全自动自动安装,启用k8s集群持久化方案务必保证存储节点均有一块空盘(例如/deb/sdb无需分区(默认40%用于k8s集群持久化60%挂载到本机的/data目录))) 如启用Heketi+GlusterFS,默认会创建一个pvc验证动态存储效果 一键安装 === #### 一键安装介绍任选通道进行安装 一键安装命令(要求centos7系统为新装系统无任何软件环境可联网) 不推荐git下来仓库大概1.5gb左右比较大,可以直接下载离线包 ##一键安装通道01(默认走家庭宽带普通通道---不稳定不推荐) ``` shell while [ true ]; do rm -f K8s_1.0.tar*;curl -o K8s_1.0.tar http://www.linuxtools.cn:42344/K8s_1.0.tar && break 1 ||sleep 5;echo 网络错误正在重试下载 ;done && tar -xzvf K8s_1.0.tar && cd K8s/ && sh install.sh ``` ##一键安装通道02(走群友无私赞助电信机房专线服务器--高速稳定下载----强烈推荐) ``` shell while [ true ]; do rm -f K8s_1.0.tar*;curl -o K8s_1.0.tar http://117.27.146.72:42344/K8s_1.0.tar && break 1 ||sleep 5;echo 网络错误正在重试下载 ;done && tar -xzvf K8s_1.0.tar && cd K8s/ && sh install.sh ``` ``` [root@k8s-master-db2 ~]# [root@k8s-master-db2 ~]# sh K8s/shell_01/Check02.sh ==============master节点健康检测 kube-apiserver kube-controller-manager kube-scheduler etcd kubelet kube-proxy docker================== 192.168.123.30 | CHANGED | rc=0 >> active active active active active active active ===============================================note节点监控检测 etcd kubelet kube-proxy docker=============================================== 192.168.123.33 | CHANGED | rc=0 >> active active active active 192.168.123.31 | CHANGED | rc=0 >> active active active active 192.168.123.34 | CHANGED | rc=0 >> active active active active 192.168.123.35 | CHANGED | rc=0 >> active active active active 192.168.123.32 | CHANGED | rc=0 >> active active active active 192.168.123.36 | CHANGED | rc=0 >> active active active active 192.168.123.37 | CHANGED | rc=0 >> active active active active 192.168.123.38 | CHANGED | rc=0 >> active active active active 192.168.123.39 | CHANGED | rc=0 >> active active active active ===============================================监测csr,cs,pvc,pv,storageclasses=============================================== NAME STATUS MESSAGE ERROR componentstatus/scheduler Healthy ok componentstatus/controller-manager Healthy ok componentstatus/etcd-8 Healthy {"health":"true"} componentstatus/etcd-2 Healthy {"health":"true"} componentstatus/etcd-4 Healthy {"health":"true"} componentstatus/etcd-1 Healthy {"health":"true"} componentstatus/etcd-9 Healthy {"health":"true"} componentstatus/etcd-3 Healthy {"health":"true"} componentstatus/etcd-5 Healthy {"health":"true"} componentstatus/etcd-0 Healthy {"health":"true"} componentstatus/etcd-7 Healthy {"health":"true"} componentstatus/etcd-6 Healthy {"health":"true"} NAME STATUS VOLUME CAPACITY ACCESS MODES STORAGECLASS AGE persistentvolumeclaim/gluster1-test Bound pvc-320524b5-e0db-11e9-b631-000c291569a0 1Gi RWX gluster-heketi 3d4h persistentvolumeclaim/my-grafana Bound pvc-3915480c-e0db-11e9-b631-000c291569a0 10Gi RWO gluster-heketi 3d4h persistentvolumeclaim/my-mysql Bound pvc-bde4543d-e0f5-11e9-b631-000c291569a0 8Gi RWO gluster-heketi 3d1h persistentvolumeclaim/my-prometheus-prometheus-alertmanager Bound pvc-376a9230-e0db-11e9-b631-000c291569a0 2Gi RWO gluster-heketi 3d4h persistentvolumeclaim/my-prometheus-prometheus-server Bound pvc-376b29bc-e0db-11e9-b631-000c291569a0 8Gi RWO gluster-heketi 3d4h persistentvolumeclaim/redis-data-my-redis-master-0 Pending 3d persistentvolumeclaim/redis-data-my-redis-slave-0 Pending 3d NAME CAPACITY ACCESS MODES RECLAIM POLICY STATUS CLAIM STORAGECLASS REASON AGE persistentvolume/pvc-320524b5-e0db-11e9-b631-000c291569a0 1Gi RWX Delete Bound default/gluster1-test gluster-heketi 3d4h persistentvolume/pvc-376a9230-e0db-11e9-b631-000c291569a0 2Gi RWO Delete Bound default/my-prometheus-prometheus-alertmanager gluster-heketi 3d4h persistentvolume/pvc-376b29bc-e0db-11e9-b631-000c291569a0 8Gi RWO Delete Bound default/my-prometheus-prometheus-server gluster-heketi 3d4h persistentvolume/pvc-3915480c-e0db-11e9-b631-000c291569a0 10Gi RWO Delete Bound default/my-grafana gluster-heketi 3d4h persistentvolume/pvc-bde4543d-e0f5-11e9-b631-000c291569a0 8Gi RWO Delete Bound default/my-mysql gluster-heketi 3d1h NAME PROVISIONER AGE storageclass.storage.k8s.io/gluster-heketi kubernetes.io/glusterfs 3d4h ===============================================监测node节点labels=============================================== NAME STATUS ROLES AGE VERSION LABELS 192.168.123.30 Ready master 3d4h v1.14.4 beta.kubernetes.io/arch=amd64,beta.kubernetes.io/os=linux,dashboard=master,kubernetes.io/arch=amd64,kubernetes.io/hostname=192.168.123.30,kubernetes.io/os=linux,node-role.kubernetes.io/master=master,storagenode=glusterfs 192.168.123.31 Ready node 3d4h v1.14.4 beta.kubernetes.io/arch=amd64,beta.kubernetes.io/os=linux,kubernetes.io/arch=amd64,kubernetes.io/hostname=192.168.123.31,kubernetes.io/os=linux,node-role.kubernetes.io/node=node,storagenode=glusterfs 192.168.123.32 Ready node 3d4h v1.14.4 beta.kubernetes.io/arch=amd64,beta.kubernetes.io/os=linux,kubernetes.io/arch=amd64,kubernetes.io/hostname=192.168.123.32,kubernetes.io/os=linux,node-role.kubernetes.io/node=node,storagenode=glusterfs 192.168.123.33 Ready node 3d4h v1.14.4 beta.kubernetes.io/arch=amd64,beta.kubernetes.io/os=linux,kubernetes.io/arch=amd64,kubernetes.io/hostname=192.168.123.33,kubernetes.io/os=linux,node-role.kubernetes.io/node=node,storagenode=glusterfs 192.168.123.34 Ready node 3d4h v1.14.4 beta.kubernetes.io/arch=amd64,beta.kubernetes.io/os=linux,kubernetes.io/arch=amd64,kubernetes.io/hostname=192.168.123.34,kubernetes.io/os=linux,node-role.kubernetes.io/node=node,storagenode=glusterfs 192.168.123.35 Ready node 3d4h v1.14.4 beta.kubernetes.io/arch=amd64,beta.kubernetes.io/os=linux,kubernetes.io/arch=amd64,kubernetes.io/hostname=192.168.123.35,kubernetes.io/os=linux,node-role.kubernetes.io/node=node,storagenode=glusterfs 192.168.123.36 Ready node 3d4h v1.14.4 beta.kubernetes.io/arch=amd64,beta.kubernetes.io/os=linux,kubernetes.io/arch=amd64,kubernetes.io/hostname=192.168.123.36,kubernetes.io/os=linux,node-role.kubernetes.io/node=node,storagenode=glusterfs 192.168.123.37 Ready node 3d4h v1.14.4 beta.kubernetes.io/arch=amd64,beta.kubernetes.io/os=linux,kubernetes.io/arch=amd64,kubernetes.io/hostname=192.168.123.37,kubernetes.io/os=linux,node-role.kubernetes.io/node=node,storagenode=glusterfs 192.168.123.38 Ready node 3d4h v1.14.4 beta.kubernetes.io/arch=amd64,beta.kubernetes.io/os=linux,kubernetes.io/arch=amd64,kubernetes.io/hostname=192.168.123.38,kubernetes.io/os=linux,node-role.kubernetes.io/node=node,storagenode=glusterfs 192.168.123.39 Ready node 3d4h v1.14.4 beta.kubernetes.io/arch=amd64,beta.kubernetes.io/os=linux,kubernetes.io/arch=amd64,kubernetes.io/hostname=192.168.123.39,kubernetes.io/os=linux,node-role.kubernetes.io/node=node,storagenode=glusterfs ===============================================监测coredns是否正常工作=============================================== coredns-57656b67bb-s6bcj 1/1 Running 0 3d4h Server: 10.0.0.2 Address 1: 10.0.0.2 kube-dns.kube-system.svc.cluster.local Name: kubernetes Address 1: 10.0.0.1 kubernetes.default.svc.cluster.local pod "dns-test" deleted ===============================================监测,pods状态=============================================== NAMESPACE NAME READY STATUS RESTARTS AGE IP NODE NOMINATED NODE READINESS GATES default my-grafana-986448dfb-rvgw5 1/1 Running 0 3d4h 172.17.71.2 192.168.123.32 default my-mysql-6bdb66578-5p6j2 2/2 Running 0 3d1h 172.17.71.3 192.168.123.32 default my-prometheus-prometheus-alertmanager-79dfbddd64-x4l4h 2/2 Running 0 3d4h 172.17.40.2 192.168.123.33 default my-prometheus-prometheus-kube-state-metrics-64dcd5d669-zwmvj 1/1 Running 0 3d4h 172.17.46.2 192.168.123.34 default my-prometheus-prometheus-node-exporter-78q8k 1/1 Running 0 3d4h 192.168.123.31 192.168.123.31 default my-prometheus-prometheus-node-exporter-7gkn4 1/1 Running 0 3d4h 192.168.123.38 192.168.123.38 default my-prometheus-prometheus-node-exporter-bmr8b 1/1 Running 0 3d4h 192.168.123.33 192.168.123.33 default my-prometheus-prometheus-node-exporter-f8q5g 1/1 Running 0 3d4h 192.168.123.36 192.168.123.36 default my-prometheus-prometheus-node-exporter-jtg5f 1/1 Running 0 3d4h 192.168.123.35 192.168.123.35 default my-prometheus-prometheus-node-exporter-mh2j2 1/1 Running 0 3d4h 192.168.123.34 192.168.123.34 default my-prometheus-prometheus-node-exporter-mm5jf 1/1 Running 0 3d4h 192.168.123.37 192.168.123.37 default my-prometheus-prometheus-node-exporter-rjjtf 1/1 Running 0 3d4h 192.168.123.32 192.168.123.32 default my-prometheus-prometheus-node-exporter-v7ppl 1/1 Running 0 3d4h 192.168.123.39 192.168.123.39 default my-prometheus-prometheus-node-exporter-vztqv 1/1 Running 0 3d4h 192.168.123.30 192.168.123.30 default my-prometheus-prometheus-pushgateway-76d96d955d-6cpjj 1/1 Running 0 3d4h 172.17.88.4 192.168.123.31 default my-prometheus-prometheus-server-558dc894b5-fzw95 2/2 Running 0 3d4h 172.17.101.3 192.168.123.30 kube-system coredns-57656b67bb-s6bcj 1/1 Running 0 3d4h 172.17.88.3 192.168.123.31 kube-system kubernetes-dashboard-5b5697d4-txgfw 1/1 Running 0 3d4h 172.17.101.2 192.168.123.30 kube-system tiller-deploy-7f4d76c4b6-vcqgw 1/1 Running 0 3d4h 172.17.88.2 192.168.123.31 ===============================================监测node节点状态=============================================== NAME STATUS ROLES AGE VERSION INTERNAL-IP EXTERNAL-IP OS-IMAGE KERNEL-VERSION CONTAINER-RUNTIME 192.168.123.30 Ready master 3d4h v1.14.4 192.168.123.30 CentOS Linux 7 (Core) 3.10.0-957.el7.x86_64 docker://18.9.7 192.168.123.31 Ready node 3d4h v1.14.4 192.168.123.31 CentOS Linux 7 (Core) 3.10.0-957.el7.x86_64 docker://18.9.7 192.168.123.32 Ready node 3d4h v1.14.4 192.168.123.32 CentOS Linux 7 (Core) 3.10.0-957.el7.x86_64 docker://18.9.7 192.168.123.33 Ready node 3d4h v1.14.4 192.168.123.33 CentOS Linux 7 (Core) 3.10.0-957.el7.x86_64 docker://18.9.7 192.168.123.34 Ready node 3d4h v1.14.4 192.168.123.34 CentOS Linux 7 (Core) 3.10.0-957.el7.x86_64 docker://18.9.7 192.168.123.35 Ready node 3d4h v1.14.4 192.168.123.35 CentOS Linux 7 (Core) 3.10.0-957.el7.x86_64 docker://18.9.7 192.168.123.36 Ready node 3d4h v1.14.4 192.168.123.36 CentOS Linux 7 (Core) 3.10.0-957.el7.x86_64 docker://18.9.7 192.168.123.37 Ready node 3d4h v1.14.4 192.168.123.37 CentOS Linux 7 (Core) 3.10.0-957.el7.x86_64 docker://18.9.7 192.168.123.38 Ready node 3d4h v1.14.4 192.168.123.38 CentOS Linux 7 (Core) 3.10.0-957.el7.x86_64 docker://18.9.7 192.168.123.39 Ready node 3d4h v1.14.4 192.168.123.39 CentOS Linux 7 (Core) 3.10.0-957.el7.x86_64 docker://18.9.7 ================================================监测helm版本================================================ Client: &version.Version{SemVer:"v2.14.3", GitCommit:"0e7f3b6637f7af8fcfddb3d2941fcc7cbebb0085", GitTreeState:"clean"} Server: &version.Version{SemVer:"v2.14.3", GitCommit:"0e7f3b6637f7af8fcfddb3d2941fcc7cbebb0085", GitTreeState:"clean"} [root@k8s-master-db2 ~]# ``` * ps:目前是单master,后期会上多master高可用 * ps:近期提交代码过于频繁有时候可能会有=一些bug,欢迎到群随时提出 ==== ### [高能警告] 系统只能存在一个固定ip地址 一个网卡一个ip 切记美分系统不能多个ip多个网卡 ### [高能警告] 暂仅支持centos7.3-centos7.6, “不支持Centos7.2及其以下版本” ### [高能警告] 系统ip不能使用 10.0.0.0网段,尽量避开系统使用172.17.x.x 10.0.0.x网段(否则安装会有问题) ** # K8s升级替换v1.14.0 v1.15.0 #如果不需要使用v1.14.0 v1.15.0直接默认一键安装即可。master分支默认的是v1.14.4 ## 默认版本为v1.14.4,提供升级软件包v14 v15自行下载后放到 K8s/Software_package 目录即可(务必删除原有的) 链接:https://pan.baidu.com/s/1Sb8WH_z-dUI8z2vLEYWa_w 提取码:0eyz ![输入图片说明](https://images.gitee.com/uploads/images/2019/0629/223656_83724e63_525507.png "屏幕截图.png") ![输入图片说明](https://images.gitee.com/uploads/images/2019/0629/223707_c0937c7b_525507.png "屏幕截图.png") 放入前务必执行以下操作 ``` shell rm -fv K8s/Software_package/kubernetes-server-linux-amd64.tar.a* ``` ``` shell #可选执行-----替换第三方yum源 rm -fv rm -f /etc/yum.repos.d/* while [ true ]; do curl -o /etc/yum.repos.d/CentOS-Base.repo http://mirrors.aliyun.com/repo/Centos-7.repo && break 1 ;done while [ true ]; do curl -o /etc/yum.repos.d/epel.repo http://mirrors.aliyun.com/repo/epel-7.repo && break 1 ;done ``` 华丽分界线。。。。。。。。。。。。。。 一键安装 === #### 一键安装介绍任选通道进行安装 一键安装命令(要求centos7系统为新装系统无任何软件环境可联网) 不推荐git下来仓库大概1.5gb左右比较大,可以直接下载离线包 ##一键安装通道01(走私有服务器高速通道) ``` shell while [ true ]; do rm -f K8s_1.0.tar*;curl -o K8s_1.0.tar http://www.linuxtools.cn:42344/K8s_1.0.tar && break 1 ||sleep 5;echo 网络错误正在重试下载 ;done && tar -xzvf K8s_1.0.tar && cd K8s/ && sh install.sh ``` ##一键安装通道02(走码云服务器) ``` shell #零时弃用 ``` 视频演示地址 === https://www.bilibili.com/video/av57242055?from=search&seid=4003077921686184728 #### 测试环境 * VMware15虚拟化平台,所有服务器节点2核2G * 已测1-50节点安装正常 * 建议新装centos7.6系统,环境干净(不需要提前安装任何软件不需要提前安装docker).集群功能至少2台服务器节点 网络 | 系统 | 内核版本 | IP获取方式 | docker版本 | Kubernetes版本 |K8s集群安装方式 | ---- | ----- | ------ | ---- | ---- | ---- | ---- | 桥接模式 | 新装CentOS7.6.1810 (Core) | 3.10.0-957.el7.x86_64 | 手动设置固定IP(不能dhcp获取所有节点) | 18.06.1-ce | v1.14.4 | 二进制包安装 | #### 安装教程 ``` shell rm -f K8s_1.0.tar* ``` #下载通道01 走普通家庭宽带下载点 ``` shell curl -o K8s_1.0.tar http://www.linuxtools.cn:42344/K8s_1.0.tar ``` #下载通道02 走群友无私赞助电信机房专线服务器--高速稳定下载----强烈推荐 ``` shell curl -o K8s_1.0.tar http://117.27.146.72:42344/K8s_1.0.tar ``` ``` shell tar -xzvf K8s_1.0.tar cd K8s/ && sh install.sh ``` #### 使用说明 1. xxxx 2. xxxx 3. xxxx #### 参与贡献 #### 使用截图 ![QQ图片20190930164529.png](https://images.gitee.com/uploads/images/2019/1010/141320_a6151a6c_525507.png) ![2.png](https://images.gitee.com/uploads/images/2019/1010/141514_c8050b45_525507.png) ![4.png](https://images.gitee.com/uploads/images/2019/1010/141324_a4c45807_525507.png) ![6.png](https://images.gitee.com/uploads/images/2019/1010/141323_07f3c27d_525507.png) ![5.png](https://images.gitee.com/uploads/images/2019/1010/141554_679fef88_525507.png) ![9.png](https://images.gitee.com/uploads/images/2019/1010/141321_d7cb5691_525507.png) ![8.png](https://images.gitee.com/uploads/images/2019/1010/141328_3251a548_525507.png) ![1.png](https://images.gitee.com/uploads/images/2019/1010/141326_f9284af8_525507.png) ![3.png](https://images.gitee.com/uploads/images/2019/1010/141325_261123d2_525507.png) ![7.png](https://images.gitee.com/uploads/images/2019/1010/141331_f66eeeca_525507.png) ![QQ图片20190930164536.png](https://images.gitee.com/uploads/images/2019/1010/141552_1979527b_525507.png) ![QQ图片20190930164546.png](https://i.loli.net/2019/10/10/FukwDrqNHXERo2n.png) 监控图片一揽 === ![pod监控.png](https://images.gitee.com/uploads/images/2019/1010/140405_877c78cc_525507.png) ![QQ图片20190930163116.png](https://images.gitee.com/uploads/images/2019/1010/140406_cb19eb2c_525507.png) ![QQ图片20190930163124.png](https://images.gitee.com/uploads/images/2019/1010/140406_eb094842_525507.png) ![QQ图片20190930163136.jpg](https://images.gitee.com/uploads/images/2019/1010/140409_bb8858ab_525507.jpeg) ![QQ图片20190930163127.png](https://images.gitee.com/uploads/images/2019/1010/140406_19fda9ce_525507.png) ![QQ图片20190930163130.png](https://images.gitee.com/uploads/images/2019/1010/140410_6b7ec42a_525507.png) ![QQ图片20190930163133.png](https://images.gitee.com/uploads/images/2019/1010/140412_d50c060d_525507.png) Dashboard图片 === ![QQ截图20190930163923.png](https://images.gitee.com/uploads/images/2019/1010/140413_efecab0e_525507.png) ![QQ截图20190930164028.png](https://images.gitee.com/uploads/images/2019/1010/140413_030649b9_525507.png) ![QQ截图20190930163252.png](https://images.gitee.com/uploads/images/2019/1010/140413_03e292bd_525507.png) ![QQ截图20190930163943.png](https://images.gitee.com/uploads/images/2019/1010/140412_34dc7963_525507.png) ![QQ截图20190930164058.png](https://images.gitee.com/uploads/images/2019/1010/140415_04dd224b_525507.png) ![QQ截图20190930164008.png](https://images.gitee.com/uploads/images/2019/1010/140416_95d597bf_525507.png) ![QQ截图20190930164000.png](https://images.gitee.com/uploads/images/2019/1010/140416_1e7c906d_525507.png) 更新日志 === === ### ----------------- 2019-10-19 修复一些bug+内核优化 ### ----------------- ### ----------------- 2019-10-10 修复时区问题,所有pod默认中国上海时区 ### ----------------- ### ----------------- 2019-9-16 1. 集群版新增prometheus +grafan集群监控环境(被监控端根据集群数量自动弹性扩展) 默认端口30000,默认账户密码admin admin 1. 必须在启用数据持久化功能的基础上才会开启 1. grafana已内置了k8s集群pod监控模板,集群基础信息监控模板,已内置饼图插件,开箱即可用 ### ----------------- 2019-9-27 1. 单机版新增nfs_k8s动态存储,单机版k8s也能愉快的玩helm了,后续增加集群版nfs云端存储/本地存储持久化方案。 1. 单机版新增prometheus+grafan监控环境 1. 单机版默认storageclasses均为gluster-heketi ### ----------------- ### ----------------- 2019-9-25 1. 优化一些参数,集群版持久化功能支持最低2个节点起 1. 单机版新增helm 1. 修复一些bug ### ----------------- 2019-9-13 1. 集群版新增coredns 感谢群内dockercore大佬的指导 1. 优化集群版部署脚本,新增集群重要功能监测脚本 1. 新增内置busybox镜像测试dns功能 ### ----------------- 2019-8-26 1 新增node节点批量增删 2 新增glusterfs分布式复制卷---持久化存储(集群版4台及以上自动内置部署) ### ----------------- 2019-7-11 修复部分环境IP取值不精确导致etcd安装失败的问题 ### ----------------- 2019-7-10 1. 新增集群版 web图形化控制台dashboard 2. 更新docker-ce版本为 Version: 18.09.7 K8s集群版安装完毕,web控制界面dashboard地址为 http://IP:42345 ### ----------------- 2019-7-1 新增单机版 web图形化控制台dashboard K8s单机版安装完毕,web控制界面dashboard地址为 http://IP:42345