Kubernates集群搭建

1、Kubernates集群组件

  - etcd:在Master节点上,一个高可用的K/V键值对存储和服务发现系统

  - kube-apiserver:在Master节点上,实现跨主机的容器网络通信

  - kube-controller-manager:在Master节点上,确保集群服务

  - kube-scheduler:在Master节点上,调度容器,分配到Node节点上

  

  - kube-proxy:在Node节点上,提供网络代理服务

  - kubelet:在Node节点上,按照配置文件中定义的容器规格启动容器

  - flannel:在Node节点上,实现跨主机的容器网络的通信

2、集群示意图

3、环境安装

  • 安装epel-release源
yum install -y epel-release
  • 关闭防火墙和selinx(与docker容器的放火请规则冲突)
systemctl stop firewalld
systemctl disable firewalld
setenforce 0

4、安装Master节点

  • 安装etcd和kubernates-master
yum -y install etcd kubernetes-master
  • 编辑/etc/etcd/etcd.conf文件,设置etcd监听客户端的URLS
ETCD_NAME=default
ETCD_DATA_DIR="/var/lib/etcd/default.etcd"
ETCD_LISTEN_CLIENT_URLS="http://0.0.0.0:2379"
ETCD_ADVERTISE_CLIENT_URLS="http://localhost:2379"
  • 配置ServiceAccount,用于Master使用yaml文件在Node上生成Pod
openssl genrsa -out /etc/kubernetes/serviceaccount.key 2048
  • 编辑/etc/kubernetes/apiserver文件,设置kube-api的地址和端口和ServiceAccount
KUBE_API_ADDRESS="--insecure-bind-address=0.0.0.0"
KUBE_API_PORT="--port=8080"
KUBELET_PORT="--kubelet-port=10250"
KUBE_ETCD_SERVERS="--etcd-servers=http://127.0.0.1:2379"
KUBE_SERVICE_ADDRESSES="--service-cluster-ip-range=10.254.0.0/16"
KUBE_ADMISSION_CONTROL="--admission-control=NamespaceLifecycle,NamespaceExists,LimitRanger,SecurityContextDeny,ResourceQuota"
KUBE_API_ARGS="--service_account_key_file=/etc/kubernetes/serviceaccount.key"
  • 编辑/etc/kubernetes/controller-manager文件,配置ServiceAccount
KUBE_CONTROLLER_MANAGER_ARGS="--service_account_private_key_file=/etc/kubernetes/serviceaccount.key"
  • 启动etcd、kube-apiserver、kube-controller-manager、kube-scheduler等服务,并设置开机启动
for SERVICES in etcd kube-apiserver kube-controller-manager kube-scheduler; 
    do 
        systemctl restart $SERVICES;
        systemctl enable $SERVICES;
        systemctl status $SERVICES ; 
    done
  • 在etcd中定义flannel网络
etcdctl mk /atomic.io/network/config '{"Network":"172.17.0.0/16"}'

5、安装Node节点

  • 安装flannel和kubernetes-node
yum -y install flannel kubernetes-node
  • 修改/etc/sysconfig/flanneld文件,为flannel网络指定etcd服务
FLANNEL_ETCD="http://192.168.30.20:2379"
FLANNEL_ETCD_KEY="/atomic.io/network"
  • 修改/etc/kubernetes/config文件,指定kube master地址
KUBE_LOGTOSTDERR="--logtostderr=true"
KUBE_LOG_LEVEL="--v=0"
KUBE_ALLOW_PRIV="--allow-privileged=false"
KUBE_MASTER="--master=http://192.168.30.20:8080"
  • 修改/etc/kubernetes/kubelet文件,配置相应Node节点的信息

 

    • node1
KUBELET_ADDRESS="--address=0.0.0.0"
KUBELET_PORT="--port=10250"
KUBELET_HOSTNAME="--hostname-override=192.168.30.21" #修改成对应Node的IP
KUBELET_API_SERVER="--api-servers=http://192.168.30.20:8080" #指定Master节点的API Server
KUBELET_POD_INFRA_CONTAINER="--pod-infra-container-image=registry.access.redhat.com/rhel7/pod-infrastructure:latest"
KUBELET_ARGS=""
    • node2
KUBELET_ADDRESS="--address=0.0.0.0"
KUBELET_PORT="--port=10250"
KUBELET_HOSTNAME="--hostname-override=192.168.30.22" 
KUBELET_API_SERVER="--api-servers=http://192.168.30.20:8080"
KUBELET_POD_INFRA_CONTAINER="--pod-infra-container-image=registry.access.redhat.com/rhel7/pod-infrastructure:latest"
KUBELET_ARGS=""
    • node3
KUBELET_ADDRESS="--address=0.0.0.0"
KUBELET_PORT="--port=10250"
KUBELET_HOSTNAME="--hostname-override=192.168.30.23"
KUBELET_API_SERVER="--api-servers=http://192.168.30.20:8080"
KUBELET_POD_INFRA_CONTAINER="--pod-infra-container-image=registry.access.redhat.com/rhel7/pod-infrastructure:latest"
KUBELET_ARGS=""
    • node4
KUBELET_ADDRESS="--address=0.0.0.0"
KUBELET_PORT="--port=10250"
KUBELET_HOSTNAME="--hostname-override=192.168.30.24"
KUBELET_API_SERVER="--api-servers=http://192.168.30.20:8080"
KUBELET_POD_INFRA_CONTAINER="--pod-infra-container-image=registry.access.redhat.com/rhel7/pod-infrastructure:latest"
KUBELET_ARGS=""
  • 启动kube-proxy,kubelet,docker,flanneld等服务,并设置开机启动
for SERVICES in kube-proxy kubelet docker flanneld;
    do
        systemctl restart $SERVICES;
        systemctl enable $SERVICES;
        systemctl status $SERVICES;
    done
  • 安装rhsm,解决Pod一直处于ContainerCreating状态
wget http://mirror.centos.org/centos/7/os/x86_64/Packages/python-rhsm-certificates-1.19.10-1.el7_4.x86_64.rpm

 

rpm2cpio python-rhsm-certificates-1.19.10-1.el7_4.x86_64.rpm | cpio -iv --to-stdout ./etc/rhsm/ca/redhat-uep.pem | tee /etc/rhsm/ca/redhat-uep.pem

 

6、验证集群是否安装成功(在Master上执行)

kubectl get node

 文章来自于:韩德田Tivens-k8s入门系列之集群安装篇

posted @ 2019-03-09 10:55  张界  阅读(801)  评论(0编辑  收藏  举报