SUSE CaaS Platform 4 - 安装部署
SUSE CaaS Platform 相关文章
(2)SUSE CaaS Platform 4 - 安装部署
(3)SUSE CaaS Platform 4 - 安装技巧
(4)SUSE CaaS Platform 4 - Ceph RBD 作为 Pod 存储卷
(5)SUSE CaaS Platform 4 - 使用 Ceph RBD 作为持久存储(静态)
(6)SUSE CaaS Platform 4 - 使用 Ceph RBD 作为持久存储(动态)
一、前提条件和环境设置
1、搭建环境
- 操作系统版本: SLES15 SP1,无需安装 swap
- 内核版本:4.12.14-197.18-default
- Kubernetes版本:v1.15.2
- VMware Workstation 14
2、虚拟化环境搭建和系统安装参考:
3、网络环境
- 管理网络:带外访问,管理容器平台
- 外部网络:访问互联网,便于从镜像仓库上下载镜像,架构图中使用的是 VMnet8,通过地址转换访问互联网
- 存储网络:对接 SUSE Storage6 分布式存储的 Public 网络,提供给容器作为持久存储使用
4、CaaSP4 环境架构图
二、安装部署
系统初始化安装
1、编辑 Hosts 文件
# vim /etc/hosts 172.200.50.70 master01.example.com master01 172.200.50.71 master02.example.com master02 172.200.50.72 master03.example.com master03 172.200.50.73 worker01.example.com worker01 172.200.50.74 worker02.example.com worker02 172.200.50.75 worker03.example.com worker03 172.200.50.76 worker04.example.com worker04
2、添加仓库
## Pool zypper ar http://172.200.50.19/repo/SUSE/Products/SUSE-CAASP/4.0/x86_64/product/ SUSE-CAASP-4.0-Pool zypper ar http://172.200.50.19/repo/SUSE/Products/SLE-Module-Containers/15-SP1/x86_64/product/ SLE-Module-Containers-15-SP1-Pool zypper ar http://172.200.50.19/repo/SUSE/Products/SLE-Product-SLES/15-SP1/x86_64/product/ SLE-Product-SLES15-SP1-Pool zypper ar http://172.200.50.19/repo/SUSE/Products/SLE-Module-Basesystem/15-SP1/x86_64/product/ SLE-Module-Basesystem-SLES15-SP1-Pool zypper ar http://172.200.50.19/repo/SUSE/Products/SLE-Module-Server-Applications/15-SP1/x86_64/product/ SLE-Module-Server-Applications-SLES15-SP1-Pool zypper ar http://172.200.50.19/repo/SUSE/Products/SLE-Module-Legacy/15-SP1/x86_64/product/ SLE-Module-Legacy-SLES15-SP1-Pool ## Update zypper ar http://172.200.50.19/repo/SUSE/Updates/SLE-Product-SLES/15-SP1/x86_64/update/ SLE-Product-SLES15-SP1-Updates zypper ar http://172.200.50.19/repo/SUSE/Updates/SLE-Module-Basesystem/15-SP1/x86_64/update/ SLE-Module-Basesystem-SLES15-SP1-Upadates zypper ar http://172.200.50.19/repo/SUSE/Updates/SLE-Module-Server-Applications/15-SP1/x86_64/update/ SLE-Module-Server-Applications-SLES15-SP1-Upadates zypper ar http://172.200.50.19/repo/SUSE/Updates/SLE-Module-Legacy/15-SP1/x86_64/update/ SLE-Module-Legacy-SLES15-SP1-Updates zypper ar http://172.200.50.19/repo/SUSE/Updates/SUSE-CAASP/4.0/x86_64/update/ SUSE-CAASP-4.0-Updates zypper ar http://172.200.50.19/repo/SUSE/Updates/SLE-Module-Containers/15-SP1/x86_64/update/ SLE-Module-Containers-15-SP1-Updates
3、安装基本软件
# zypper in -y -t pattern yast2_basis base # zypper in -y net-tools vim man sudo tuned irqbalance # zypper in -y ethtool rsyslog iputils less supportutils-plugin-ses # zypper in -y net-tools-deprecated tree wget
4、关闭IPV6 (所有节点和admin)和 关闭 swap
# vim /etc/sysctl.conf net.ipv6.conf.all.disable_ipv6 = 1 net.ipv6.conf.default.disable_ipv6 = 1 net.ipv6.conf.lo.disable_ipv6 = 1
5、创建和配置 after.local 文件 (可选)
注意:如系统已配置使用swap,请选择关闭,并在开机启动时执行脚本
# touch /etc/init.d/after.local # chmod 744 /etc/init.d/after.local
cat >> /etc/init.d/after.local << EOF #! /bin/sh # # Copyright (c) 2010 SuSE LINUX Products GmbH, Germany. All rights reserved. # # Author: Werner Fink, 2010 # # /etc/init.d/after.local # # script with local commands to be executed from init after all scripts # of a runlevel have been executed. # # Here you should add things, that should happen directly after # runlevel has been reached. # swapoff -a EOF
6、执行 after.local,关闭 swap
# source /etc/init.d/after.local
CaaSP4 集群搭建
1、安装组件 (所有节点 )
# zypper -n in -t pattern SUSE-CaaSP-Management
2、SSH 互信 ,并启用ssh-agent
(1)master01 主机生成密钥对,并把公钥复制给woker01和woker02主机
# ssh-keygen # cd .ssh/ # ssh-copy-id master01 # ssh-copy-id worker01 # ssh-copy-id worker02
(2)master01主机,启动ssh-agent
# eval `ssh-agent` $SHELL Agent pid 13701
# ps -ef | grep ssh-agent root 13701 1 0 09:35 ? 00:00:00 ssh-agent root 13719 13702 0 09:35 pts/0 00:00:00 grep --color=auto ssh-agent
(3)将私钥添加到ssh-agent中
# ssh-add /root/.ssh/id_rsa Identity added: /root/.ssh/id_rsa (/root/.ssh/id_rsa)
# ssh-add -l 2048 SHA256:7sn4CfA7qk7/q38CA6Oz3tgvVShY9+fpOXZRdjAM1MM root@master01 (RSA)
3、初始化 Master 节点
语法:skuba cluster init --control-plane <LB IP/FQDN> my-cluster
# skuba cluster init --control-plane master01 my-cluster [init] configuration files written to /root/.ssh/my-cluster
初始化会产生 my-cluster目录,并且该目录具有kubeconfig配置文件
# ll /root/.ssh/my-cluster/ total 4 drwx------ 8 root root 78 Sep 15 13:31 addons -rw------- 1 root root 804 Sep 15 13:31 kubeadm-init.conf drwx------ 2 root root 62 Sep 15 13:31 kubeadm-join.conf.d
4、集群 Booststrap
# cd /root/.ssh/my-cluster/ # skuba node bootstrap --target master01 master01 -v3
由于使用 -v3 参数,因此终端会产生安装详细信息,直到最后显示 successfully bootstrapped
I0915 13:58:55.780337 4730 files.go:29] uploading to remote file "/tmp/gangway.d/gangway.yaml" with contents
I0915 13:58:55.867266 4730 ssh.go:167] running command: "kubectl --kubeconfig=/etc/kubernetes/admin.conf apply -f /tmp/gangway.d"
I0915 13:58:56.136908 4730 ssh.go:190] stdout | configmap/oidc-gangway-config created
I0915 13:58:56.143730 4730 ssh.go:190] stdout | deployment.apps/oidc-gangway created
I0915 13:58:56.201805 4730 ssh.go:190] stdout | service/oidc-gangway created
I0915 13:58:56.213596 4730 ssh.go:190] stdout | serviceaccount/oidc-gangway created
I0915 13:58:56.213946 4730 ssh.go:167] running command: "rm -rf /tmp/gangway.d"
I0915 13:58:56.255647 4730 states.go:40] === state gangway.deploy applied successfully ===
[bootstrap] successfully bootstrapped node "master01" with Kubernetes: "1.15.2" <=== 安装成功
5、加入 worker 节点
语法:
skuba node join --role <master/worker> --user <user-name> --sudo --target <IP/FQDN> <node-name>
- master01节点并用root账户执行:
# cd ~/.ssh/my-cluster # skuba node join --role worker --target worker01 worker01 -v3 # skuba node join --role worker --target worker02 worker02 -v3
I0915 14:42:41.768349 46412 states.go:35] === applying state skuba-update.start === I0915 14:42:41.769311 46412 ssh.go:167] running command: "systemctl enable --now skuba-update.timer" I0915 14:42:41.828002 46412 ssh.go:190] stderr | Created symlink /etc/systemd/system/timers.target.wants/skuba-update.timer → /usr/lib/systemd/system/skuba-update.timer. I0915 14:42:42.096214 46412 states.go:40] === state skuba-update.start applied successfully === [join] node successfully joined the cluster
检查集群状态
1、Master 节点
(1) Master 节点下载的镜像
master01:~ # crictl images IMAGE TAG IMAGE ID SIZE registry.suse.com/caasp/v4/cilium-init 1.5.3 726a5a2fd9f8e 121MB registry.suse.com/caasp/v4/cilium 1.5.3 15b448bea31e7 889MB registry.suse.com/caasp/v4/coredns 1.3.1 2f747b3021d99 163MB registry.suse.com/caasp/v4/etcd 3.3.11 b90b8b149824d 223MB registry.suse.com/caasp/v4/hyperkube v1.15.2 e19d9ad312c9d 672MB registry.suse.com/caasp/v4/kured 1.2.0 9d0512cb475d8 358MB registry.suse.com/caasp/v4/pause 3.1 247351923b047 936kB
(2)Master 节点运行的容器
# crictl ps -a CONTAINER IMAGE CREATED STATE NAME ATTEMPT POD ID 98cc214ea02e6 e19d9ad312c9d698740a587857fdebc77c17bd7fc05214a74658b0b6f779c597 3 minutes ago Running kube-proxy 0 21617c8adaa8d adb894ff5bd1e b90b8b149824d799c958d04fd16df356ccf254944467e5666e3abb7e6789ffa5 3 minutes ago Running etcd 0 0629bbad3a0f8 cfbab432bd3dd e19d9ad312c9d698740a587857fdebc77c17bd7fc05214a74658b0b6f779c597 3 minutes ago Running kube-scheduler 0 d463bbd4dbefd 1cf5a0829a1ff e19d9ad312c9d698740a587857fdebc77c17bd7fc05214a74658b0b6f779c597 3 minutes ago Running kube-controller-manager 0 7c045016665d2 597c1feedc3dd e19d9ad312c9d698740a587857fdebc77c17bd7fc05214a74658b0b6f779c597 3 minutes ago Running kube-apiserver 0 8f258d290806a
(3)Master 节点运行的pods
# crictl pods POD ID CREATED STATE NAME NAMESPACE ATTEMPT e897316a364c7 4 minutes ago Ready cilium-shddp kube-system 0 21617c8adaa8d 5 minutes ago Ready kube-proxy-2f57p kube-system 0 0629bbad3a0f8 6 minutes ago Ready etcd-master01 kube-system 0 d463bbd4dbefd 6 minutes ago Ready kube-scheduler-master01 kube-system 0 7c045016665d2 6 minutes ago Ready kube-controller-manager-master01 kube-system 0 8f258d290806a 6 minutes ago Ready kube-apiserver-master01 kube-system 0
2、woker 节点
(1)worker 节点下载镜像
worker01:~ # crictl images IMAGE TAG IMAGE ID SIZE registry.suse.com/caasp/v4/caasp-dex 2.16.0 32d64343dba35 146MB registry.suse.com/caasp/v4/cilium-init 1.5.3 726a5a2fd9f8e 121MB registry.suse.com/caasp/v4/cilium-operator 1.5.3 918bd75777bb1 196MB registry.suse.com/caasp/v4/cilium 1.5.3 15b448bea31e7 889MB registry.suse.com/caasp/v4/gangway 3.1.0 227070f1f75d6 136MB registry.suse.com/caasp/v4/hyperkube v1.15.2 e19d9ad312c9d 672MB registry.suse.com/caasp/v4/kured 1.2.0 9d0512cb475d8 358MB registry.suse.com/caasp/v4/pause 3.1 247351923b047 936kB
(2)worker 节点运行的容器
worker01:~ # crictl ps -a CONTAINER IMAGE CREATED STATE NAME ATTEMPT POD ID b55f388e47cda 32d64343dba35eed1b365322f3f60ffe2dc42b7b327c7cc2870c881a2efdbf1b 37 minutes ago Running oidc-dex 1 4a536a9499080 a0c58bf84de32 registry.suse.com/caasp/v4/caasp-dex@sha256:0542dfb58a97e2de8d25128b673aaa8c1e2dc2680b0db31872eb47b8fe6ce409 37 minutes ago Exited oidc-dex 0 4a536a9499080 fd585e82ec6eb registry.suse.com/caasp/v4/gangway@sha256:85922202aaa071e5e13084c58c1375ae7029af77bfe7d99d63e22e98ca574bc3 37 minutes ago Running oidc-gangway 0 c865bd3ac446c cdfbabee16375 registry.suse.com/caasp/v4/kured@sha256:52d0bb956a4a012131436263cdc8126f86042af1d6f0c93bd461d229743f5bf3 38 minutes ago Running kured 0 27af5c286c721 7f9d9d80f52ed registry.suse.com/caasp/v4/cilium-operator@sha256:02ba730698cba66dbd71d66b68aa3fd9ce9a78ece55e246bd837a0a444ebaaed 38 minutes ago Running cilium-operator 0 9a90ff209bb50 427c05258e7b8 registry.suse.com/caasp/v4/hyperkube@sha256:6a7baa891d90d68bc856d321dcf4cbdd2b86b25204043276f44b4f81a270a515 39 minutes ago Running kube-proxy 0 ed62ccc311d4e c5ac07f3c68ea 15b448bea31e7ff401acac76f6518986be200667c00869582b3cd3beb30dd69b 39 minutes ago Running cilium-agent 1 54c628d7909f9 4b432bb3a085f 15b448bea31e7ff401acac76f6518986be200667c00869582b3cd3beb30dd69b 41 minutes ago Exited cilium-agent 0 54c628d7909f9 27be548f473f5 registry.suse.com/caasp/v4/cilium-init@sha256:77d7101f0af12c3f6f1f1fd2062a0efc93d058b8358b3b9f2a02a69ba321dff4 41 minutes ago Exited clean-cilium-state 0 54c628d7909f9 c393915545030 15b448bea31e7ff401acac76f6518986be200667c00869582b3cd3beb30dd69b 42 minutes ago Exited install-cni-bin 0 54c628d7909f9 8c0a5fd0fa598 15b448bea31e7ff401acac76f6518986be200667c00869582b3cd3beb30dd69b 42 minutes ago Exited install-cni-conf 0 54c628d7909f9
(3)Worker 节点运行的pods
# crictl pods POD ID CREATED STATE NAME NAMESPACE ATTEMPT c865bd3ac446c 40 minutes ago Ready oidc-gangway-7b7fbbdbdf-s94h5 kube-system 0 9a90ff209bb50 40 minutes ago Ready cilium-operator-7d6ddddbf5-8clw8 kube-system 0 27af5c286c721 40 minutes ago Ready kured-n2zz4 kube-system 0 4a536a9499080 40 minutes ago Ready oidc-dex-55fc689dc-g6rbq kube-system 0 54c628d7909f9 43 minutes ago Ready cilium-4fp22 kube-system 0 ed62ccc311d4e About an hour ago Ready kube-proxy-fwt2t kube-system 0
3、CaaSP4 集群
# cd /root/.ssh/my-cluster/ # skuba cluster status NAME OS-IMAGE KERNEL-VERSION KUBELET-VERSION CONTAINER-RUNTIME HAS-UPDATES HAS-DISRUPTIVE-UPDATES master01 SUSE Linux Enterprise Server 15 SP1 4.12.14-197.18-default v1.15.2 cri-o://1.15.0 <none> <none> worker01 SUSE Linux Enterprise Server 15 SP1 4.12.14-197.18-default v1.15.2 cri-o://1.15.0 <none> <none> worker02 SUSE Linux Enterprise Server 15 SP1 4.12.14-197.18-default v1.15.2 cri-o://1.15.0 <none> <none>
CaaSP4 集群使用 kubectl
1、安装软件包(可选)
# zypper in kubernetes-client (默认已安装)
2、建立软连接
# ln -s /root/.ssh/my-cluster/admin.conf ~/.kube/config
3、集群信息
# kubectl cluster-info Kubernetes master is running at https://master01:6443 KubeDNS is running at https://master01:6443/api/v1/namespaces/kube-system/services/kube-dns:dns/proxy To further debug and diagnose cluster problems, use 'kubectl cluster-info dump'.
4、集群节点信息
# kubectl get nodes # kubectl get nodes -o wide NAME STATUS ROLES AGE VERSION master01 Ready master 72m v1.15.2 worker01 Ready <none> 29m v1.15.2 worker02 Ready <none> 28m v1.15.2
5、集群运行的 Pods 状态
# kubectl get pods --all-namespaces NAMESPACE NAME READY STATUS RESTARTS AGE kube-system cilium-dph24 1/1 Running 0 28m kube-system cilium-fjs6c 1/1 Running 0 28m kube-system cilium-operator-7d6ddddbf5-xd6p9 1/1 Running 0 72m kube-system cilium-s5l4k 1/1 Running 0 28m kube-system coredns-69c4947958-25rsk 1/1 Running 4 72m kube-system coredns-69c4947958-ntn7w 1/1 Running 6 72m kube-system etcd-master01 1/1 Running 0 71m kube-system kube-apiserver-master01 1/1 Running 0 71m kube-system kube-controller-manager-master01 1/1 Running 0 71m kube-system kube-proxy-4hk7q 1/1 Running 0 29m kube-system kube-proxy-8jgd7 1/1 Running 0 72m kube-system kube-proxy-mwrzx 1/1 Running 0 28m kube-system kube-scheduler-master01 1/1 Running 0 71m kube-system kured-26h25 1/1 Running 0 26m kube-system kured-hvhdm 1/1 Running 0 28m kube-system kured-tpgvh 1/1 Running 0 70m kube-system oidc-dex-55fc689dc-pxgv2 1/1 Running 1 72m kube-system oidc-gangway-7b7fbbdbdf-jvfdk 1/1 Running 0 72m
6、集群版本
# kubectl version --short Client Version: v1.15.2 Server Version: v1.15.2