[ ceph ] CEPH 部署完整版(CentOS 7 + luminous)
1. 前言
拜读了 胖哥的(el7+jewel)完整部署 受益匪浅,目前 CEPH 已经更新到 M 版本,配置方面或多或少都有了变动,本博文就做一个 ceph luminous 版本完整的配置安装。
提示:本文使用了大量的 ansible 命令,需有基础的 ansible 相关知识。
2. 环境介绍
主机配置
3 台装有 CentOS 7 的主机,每台主机有 5 块磁盘(虚拟机磁盘要大于30G),本次使用的是 Vmware Wrokstation ,创建虚拟机及安装操作系统不再描述。
每台主机配备 3 个 2T 硬盘,1 个 240G 硬盘 及 一个 800G 硬盘,其中 240G 硬盘假装是一个SSD,用来存储 RocksDB 的 block.wal 和 block.db 而 800G 的硬盘用作 OSD 的 SSD 。
每台主机配备两张网卡,网卡1 - 连接公网, 网卡2 - 集群网络,禁止访问外网。
网卡1网络:192.168.118.0/24 网卡2网络:192.168.61.0/24
3. 系统详细信息及集群配置介绍
集群介绍
4. 系统初始化
分为以下几个步骤:
(1)修改主机名并设置 ssh 互信访问;
(2)升级系统内核、关闭 Selinux 和 firewalld、修改系统文件句柄;
(3)配置 ntp 服务
修改主机名及设置ssh互信
修改主机名: #hostnamectl set-hostname ceph-node1 #hostnamectl set-hostname ceph-node2 #hostnamectl set-hostname ceph-node3 #hostnamectl set-hostname ceph-client #cat /etc/hosts 127.0.0.1 localhost localhost.localdomain localhost4 localhost4.localdomain4 ::1 localhost localhost.localdomain localhost6 localhost6.localdomain6 # 将如下内容添加到 /etc/hosts, 每台主机都需要添加 192.168.118.11 ceph-node1 192.168.118.12 ceph-node2 192.168.118.13 ceph-node3 192.168.118.14 ceph-client 在 ceph-deploy主机上设置互信访问: [root@ceph-node1 ~]# ssh-keygen -t rsa -P '' # 一路回车 [root@ceph-node1 ~]# ssh-copy-id ceph-node1 [root@ceph-node1 ~]# ssh-copy-id ceph-node2 [root@ceph-node1 ~]# ssh-copy-id ceph-node3 配置 yum 并安装 ansible #mkdir /tmp/bak ; mv /etc/yum.repos.d/* /tmp/bak -- 此命令在每台主机上执行,避免产生多余的yum配置文件 下载 163 镜像yum #wget -O /etc/yum.repos.d/CentOS7-Base-163.repo http://mirrors.163.com/.help/CentOS7-Base-163.repo 安装 ansible #yum install ansible -y 将ceph 主机添加到ansible群组ceph里 [root@ceph-node1 ~]# cat /etc/ansible/hosts [ceph] ceph-node1 ansible_ssh_user=root ceph-node2 ansible_ssh_user=root ceph-node3 ansible_ssh_user=root 测试: [root@ceph-node1 ~]# ansible ceph -m ping ceph-node1 | SUCCESS => { "changed": false, "ping": "pong" } ceph-node3 | SUCCESS => { "changed": false, "ping": "pong" } ceph-node2 | SUCCESS => { "changed": false, "ping": "pong" } 拷贝 hosts 文件到每台主机 [root@ceph-node1 ~]# ansible ceph -m copy -a 'src=/etc/hosts dest=/etc/'
注意:每台主机要修改对应的主机名
192.168.118.11 ceph-node1 192.168.118.12 ceph-node2 192.168.118.13 ceph-node3
升级系统内核、关闭 Selinux 和 firewalld、修改系统文件句柄
修改yum配置文件: [root@ceph-node1 ~]# ansible ceph -m get_url -a 'url=http://mirrors.163.com/.help/CentOS7-Base-163.repo dest=/etc/yum.repos.d/' 升级系统程序包: [root@ceph-node1 ~]# ansible ceph -m yum -a 'name=* state=latest' 升级内核: ansible ceph -m shell -a 'rpm -ivhU http://www.elrepo.org/elrepo-release-7.0-3.el7.elrepo.noarch.rpm' 安装 kernel-lt(长期维护版本) [root@ceph-node1 ~]# ansible ceph -m yum -a 'name=kernel-lt enablerepo="elrepo-kernel"' 查看当前系统内核包: [root@ceph-node1 ~]# ansible ceph -m shell -a 'rpm -qa | egrep kernel' [WARNING]: Consider using yum, dnf or zypper module rather than running rpm ceph-node3 | SUCCESS | rc=0 >> kernel-3.10.0-1062.el7.x86_64 kernel-3.10.0-1062.4.3.el7.x86_64 kernel-tools-3.10.0-1062.4.3.el7.x86_64 kernel-lt-4.4.204-1.el7.elrepo.x86_64 kernel-tools-libs-3.10.0-1062.4.3.el7.x86_64 kernel-headers-3.10.0-1062.4.3.el7.x86_64 ceph-node2 | SUCCESS | rc=0 >> kernel-3.10.0-1062.el7.x86_64 kernel-3.10.0-1062.4.3.el7.x86_64 kernel-tools-3.10.0-1062.4.3.el7.x86_64 kernel-lt-4.4.204-1.el7.elrepo.x86_64 kernel-tools-libs-3.10.0-1062.4.3.el7.x86_64 kernel-headers-3.10.0-1062.4.3.el7.x86_64 ceph-node1 | SUCCESS | rc=0 >> kernel-3.10.0-1062.el7.x86_64 kernel-tools-3.10.0-1062.4.3.el7.x86_64 kernel-lt-4.4.204-1.el7.elrepo.x86_64 kernel-tools-libs-3.10.0-1062.4.3.el7.x86_64 kernel-3.10.0-1062.4.3.el7.x86_64 kernel-headers-3.10.0-1062.4.3.el7.x86_64
内核包安装成功,安装的包就是:kernel-lt-4.4.204-1.el7.elrepo.x86_64
修改grub 启动为最新的内核
[root@ceph-node1 ~]# ansible ceph -m replace -a 'path=/etc/default/grub regexp="saved" replace=0' [root@ceph-node1 ~]# ansible ceph -m shell -a 'grub2-mkconfig -o /boot/grub2/grub.cfg'
在重启之前记得修改 selinux 和 firewalld
[root@ceph-node1 ~]# ansible ceph -m selinux -a 'conf=/etc/selinux/config state=disabled' [root@ceph-node1 ~]# ansible ceph -m systemd -a 'name=firewalld enabled=no' 设置最大文件句柄: [root@ceph-node1 ~]# ansible ceph -m lineinfile -a 'dest=/etc/security/limits.conf line="* \ soft nproc 65535\n* hard nproc 65535\n* soft nofile 65535\n* hard nofile 65535"' 设置开启执行: [root@ceph-node1 ~]# ansible ceph -m lineinfile -a 'dest=/etc/rc.local line="ulimit -SHn 65535"' [root@ceph-node1 ~]# ansible ceph -m file -a 'path=/etc/rc.d/rc.local mode=0744' 重启主机: [root@ceph-node1 ~]# ansible ceph -m shell -a 'reboot -f'
最后,配置 NTP服务
[root@ceph-node1 ~]# ansible ceph -m yum -a 'name=ntp state=latest' [root@ceph-node1 ~]# ansible ceph -m shell -a 'sed -i "/^server/d" /etc/ntp.conf' [root@ceph-node1 ~]# ansible ceph -m lineinfile -a 'dest=/etc/ntp.conf line="server tiger.sina.com.cn\nserver ntp1.aliyun.com"' [root@ceph-node1 ~]# ansible ceph -m systemd -a 'name=ntpd state=started' 查看是否配置成功 [root@ceph-node1 ~]# ansible ceph -m shell -a 'ntpq -p'
OK,初始化工作总算是完成了,上面应用了大量的 ansible 来实现,避免重复的操作。接下来还会用到 ansible
再次强调下 ceph 的架构:
240G SSD:用于3个2T SATA 和 800G SSD 的 block.wal 和 block.db 800G SSD:用作OSD,三台机器共 3个构成 ssd-pool 2T SATA:用作 OSD,三台机器共9个构成 sata-pool
磁盘分区
将 240 G 的盘分为 8 个分区,分别作为 block.wal 和 block.data 如果是在生成环境,这里建议使用 RAID 1 组一个冗余,然后在分区。假如 240G 已经是使用 RAID 1 组成的逻辑磁盘。
[root@ceph-node1 ~]# ansible ceph -m shell -a 'parted /dev/sde mklabel gpt' [root@ceph-node1 ~]# ansible ceph -m shell -a 'parted /dev/sde mkpart primary 2048s 12%' [root@ceph-node1 ~]# ansible ceph -m shell -a 'parted /dev/sde mkpart primary 12% 24%' [root@ceph-node1 ~]# ansible ceph -m shell -a 'parted /dev/sde mkpart primary 24% 36%' [root@ceph-node1 ~]# ansible ceph -m shell -a 'parted /dev/sde mkpart primary 36% 48%' [root@ceph-node1 ~]# ansible ceph -m shell -a 'parted /dev/sde mkpart primary 48% 60%' [root@ceph-node1 ~]# ansible ceph -m shell -a 'parted /dev/sde mkpart primary 60% 72%' [root@ceph-node1 ~]# ansible ceph -m shell -a 'parted /dev/sde mkpart primary 72% 84%' [root@ceph-node1 ~]# ansible ceph -m shell -a 'parted /dev/sde mkpart primary 84% 100%' [root@ceph-node1 ~]# lsblk NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT sda 8:0 0 20G 0 disk ├─sda1 8:1 0 1G 0 part /boot └─sda2 8:2 0 19G 0 part ├─centos-root 253:0 0 17G 0 lvm / └─centos-swap 253:1 0 2G 0 lvm [SWAP] sdb 8:16 0 2T 0 disk sdc 8:32 0 2T 0 disk sdd 8:48 0 2T 0 disk sde 8:64 0 240G 0 disk ├─sde1 8:65 0 28.8G 0 part ├─sde2 8:66 0 28.8G 0 part ├─sde3 8:67 0 28.8G 0 part ├─sde4 8:68 0 28.8G 0 part ├─sde5 8:69 0 28.8G 0 part ├─sde6 8:70 0 28.8G 0 part ├─sde7 8:71 0 28.8G 0 part └─sde8 8:72 0 38.4G 0 part sdf 8:80 0 800G 0 disk sr0 11:0 1 4.4G 0 rom /mnt/centos7
分区成功。
在 luminous 版本中,是用 ceph-volume 管理 OSD ,官方也推荐使用 lvm 管理磁盘。设置 LVM
设置 OSD 的 LVM
VGS [root@ceph-node1 ~]# ansible ceph -m shell -a 'vgcreate datavg1 /dev/sdb' [root@ceph-node1 ~]# ansible ceph -m shell -a 'vgcreate datavg2 /dev/sdc' [root@ceph-node1 ~]# ansible ceph -m shell -a 'vgcreate datavg3 /dev/sdd' [root@ceph-node1 ~]# ansible ceph -m shell -a 'vgcreate datavg4 /dev/sdf' LVS ansible ceph -m shell -a 'lvcreate -n datalv1 -l 100%Free datavg1' ansible ceph -m shell -a 'lvcreate -n datalv2 -l 100%Free datavg2' ansible ceph -m shell -a 'lvcreate -n datalv3 -l 100%Free datavg3' ansible ceph -m shell -a 'lvcreate -n datalv4 -l 100%Free datavg4'
设置 wal/db 的 LVM
VGS - wal [root@ceph-node1 ~]# ansible ceph -m shell -a 'vgcreate block_wal_vg1 /dev/sde1' [root@ceph-node1 ~]# ansible ceph -m shell -a 'vgcreate block_wal_vg2 /dev/sde2' [root@ceph-node1 ~]# ansible ceph -m shell -a 'vgcreate block_wal_vg3 /dev/sde3' [root@ceph-node1 ~]# ansible ceph -m shell -a 'vgcreate block_wal_vg4 /dev/sde4' VGS - db [root@ceph-node1 ~]# ansible ceph -m shell -a 'vgcreate block_db_vg1 /dev/sde5' [root@ceph-node1 ~]# ansible ceph -m shell -a 'vgcreate block_db_vg2 /dev/sde6' [root@ceph-node1 ~]# ansible ceph -m shell -a 'vgcreate block_db_vg3 /dev/sde7' [root@ceph-node1 ~]# ansible ceph -m shell -a 'vgcreate block_db_vg4 /dev/sde8' LVS - wal [root@ceph-node1 ~]# ansible ceph -m shell -a 'lvcreate -n wallv1 -l 100%Free block_wal_vg1' [root@ceph-node1 ~]# ansible ceph -m shell -a 'lvcreate -n wallv2 -l 100%Free block_wal_vg2' [root@ceph-node1 ~]# ansible ceph -m shell -a 'lvcreate -n wallv3 -l 100%Free block_wal_vg3' [root@ceph-node1 ~]# ansible ceph -m shell -a 'lvcreate -n wallv4 -l 100%Free block_wal_vg4' LVS - db [root@ceph-node1 ~]# ansible ceph -m shell -a 'lvcreate -n dblv1 -l 100%Free block_db_vg1' [root@ceph-node1 ~]# ansible ceph -m shell -a 'lvcreate -n dblv2 -l 100%Free block_db_vg2' [root@ceph-node1 ~]# ansible ceph -m shell -a 'lvcreate -n dblv3 -l 100%Free block_db_vg3' [root@ceph-node1 ~]# ansible ceph -m shell -a 'lvcreate -n dblv4 -l 100%Free block_db_vg4' [root@ceph-node1 ~]# lsblk NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT sda 8:0 0 20G 0 disk ├─sda1 8:1 0 1G 0 part /boot └─sda2 8:2 0 19G 0 part ├─centos-root 253:0 0 17G 0 lvm / └─centos-swap 253:1 0 2G 0 lvm [SWAP] sdb 8:16 0 2T 0 disk └─datavg1-datalv1 253:2 0 2T 0 lvm sdc 8:32 0 2T 0 disk └─datavg2-datalv2 253:3 0 2T 0 lvm sdd 8:48 0 2T 0 disk └─datavg3-datalv3 253:4 0 2T 0 lvm sde 8:64 0 240G 0 disk ├─sde1 8:65 0 28.8G 0 part │ └─block_wal_vg1-wallv1 253:6 0 28.8G 0 lvm ├─sde2 8:66 0 28.8G 0 part │ └─block_wal_vg2-wallv2 253:7 0 28.8G 0 lvm ├─sde3 8:67 0 28.8G 0 part │ └─block_wal_vg3-wallv3 253:8 0 28.8G 0 lvm ├─sde4 8:68 0 28.8G 0 part │ └─block_wal_vg4-wallv4 253:9 0 28.8G 0 lvm ├─sde5 8:69 0 28.8G 0 part │ └─block_db_vg1-dblv1 253:10 0 28.8G 0 lvm ├─sde6 8:70 0 28.8G 0 part │ └─block_db_vg2-dblv2 253:11 0 28.8G 0 lvm ├─sde7 8:71 0 28.8G 0 part │ └─block_db_vg3-dblv3 253:12 0 28.8G 0 lvm └─sde8 8:72 0 38.4G 0 part └─block_db_vg4-dblv4 253:13 0 38.4G 0 lvm sdf 8:80 0 800G 0 disk └─datavg4-datalv4 253:5 0 800G 0 lvm sr0 11:0 1 4.4G 0 rom /mnt/centos7
LVM 设置完成。
5. 部署 Ceph
首先要配置 yum 源,使用国内的yum 源会快很快,地址:http://mirrors.163.com/ceph/rpm-luminous/
[root@ceph-node1 ~]# cat /etc/yum.repos.d/ceph.repo [ceph] name = ceph gpgcheck = 0 baseurl = http://mirrors.163.com/ceph/rpm-luminous/el7/x86_64/ [ceph-deploy] name = ceph-deploy gpgcheck = 0 baseurl = https://download.ceph.com/rpm-luminous/el7/noarch/ [root@ceph-node1 ~]# yum repolist 将 ceph.repo 拷贝到另外两个节点 [root@ceph-node1 ~]# ansible ceph -m copy -a 'src=/etc/yum.repos.d/ceph.repo dest=/etc/yum.repos.d/' -l ceph-node2,ceph-node3 每个节点都需要安装 ceph [root@ceph-node1 ~]# ansible ceph -m yum -a 'name=ceph,ceph-radosgw state=latest'
注意:如果是虚拟机做测试,建议此处创建 快照。
6. 部署 Ceph-deploy
回顾下部署结构:
ceph-deploy 需要部署到 ceph-node1 节点就好。
[root@ceph-node1 ~]# yum install ceph-deploy -y [root@ceph-node1 ~]# ceph-deploy --version 2.0.1 [root@ceph-node1 ~]# ceph -v ceph version 12.2.12 (1436006594665279fe734b4c15d7e08c13ebd777) luminous (stable)
准备工作完全完成,接下来就开始创建集群。
7. 创建集群
新建 mycluster目录,所有操作的在此目录中执行,这样生成的文件都在此目录中
[root@ceph-node1 ~]# mkdir mycluster [root@ceph-node1 ~]# cd mycluster/ [root@ceph-node1 mycluster]# ceph-deploy new ceph-node1 ceph-node2 ceph-node3 \ --public-network=192.168.118.0/24 --cluster-network=192.168.61.0/24 [root@ceph-node1 mycluster]# ceph-deploy new ceph-node1 ceph-node2 ceph-node3 --public-network=192.168.118.0/24 --cluster-network=192.168.61.0/24 执行完上面的命令后,mycluster 目录中会生成三个文件: [root@ceph-node1 mycluster]# ls ceph.conf ceph.log ceph.mon.keyring 开始部署 Monitor [root@ceph-node1 mycluster]# ceph-deploy mon create-initial 将配置文件及密钥拷贝到其他 monitor 节点 [root@ceph-node1 mycluster]# ceph-deploy admin ceph-node1 ceph-node2 ceph-node3 执行完毕以后,可通过 ceph -s 查看集群状态: [root@ceph-node1 mycluster]# ceph -s cluster: id: 6770a64a-b474-409c-bdf1-85a47397ad6e health: HEALTH_OK services: mon: 3 daemons, quorum ceph-node1,ceph-node2,ceph-node3 mgr: no daemons active osd: 0 osds: 0 up, 0 in data: pools: 0 pools, 0 pgs objects: 0 objects, 0B usage: 0B used, 0B / 0B avail pgs:
出现如上信息,表明集群配置成功。
8. 部署 OSD
部署OSD ,block.wal 和 block.db 对应关系
需要部署两组 OSD 一组为 SATA 另一组为 SSD,首先创建 SATA组:
#sata ceph-deploy osd create ceph-node1 --bluestore --block-wal block_wal_vg1/wallv1 --block-db block_db_vg1/dblv1 --data datavg1/datalv1 ceph-deploy osd create ceph-node1 --bluestore --block-wal block_wal_vg2/wallv2 --block-db block_db_vg2/dblv2 --data datavg2/datalv2 ceph-deploy osd create ceph-node1 --bluestore --block-wal block_wal_vg3/wallv3 --block-db block_db_vg3/dblv3 --data datavg3/datalv3 ceph-deploy osd create ceph-node2 --bluestore --block-wal block_wal_vg1/wallv1 --block-db block_db_vg1/dblv1 --data datavg1/datalv1 ceph-deploy osd create ceph-node2 --bluestore --block-wal block_wal_vg2/wallv2 --block-db block_db_vg2/dblv2 --data datavg2/datalv2 ceph-deploy osd create ceph-node2 --bluestore --block-wal block_wal_vg3/wallv3 --block-db block_db_vg3/dblv3 --data datavg3/datalv3 ceph-deploy osd create ceph-node3 --bluestore --block-wal block_wal_vg1/wallv1 --block-db block_db_vg1/dblv1 --data datavg1/datalv1 ceph-deploy osd create ceph-node3 --bluestore --block-wal block_wal_vg2/wallv2 --block-db block_db_vg2/dblv2 --data datavg2/datalv2 ceph-deploy osd create ceph-node3 --bluestore --block-wal block_wal_vg3/wallv3 --block-db block_db_vg3/dblv3 --data datavg3/datalv3 #ssd ceph-deploy osd create ceph-node1 --bluestore --block-wal block_wal_vg4/wallv4 --block-db block_db_vg4/dblv4 --data datavg4/datalv4 ceph-deploy osd create ceph-node2 --bluestore --block-wal block_wal_vg4/wallv4 --block-db block_db_vg4/dblv4 --data datavg4/datalv4 ceph-deploy osd create ceph-node3 --bluestore --block-wal block_wal_vg4/wallv4 --block-db block_db_vg4/dblv4 --data datavg4/datalv4
提醒:建议一条一条执行,如果有问题能及时发现解决。
执行完毕,查看:
[root@ceph-node1 mycluster]# ceph osd tree ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF -1 20.34357 root default -3 6.78119 host ceph-node1 0 hdd 2.00000 osd.0 up 1.00000 1.00000 1 hdd 2.00000 osd.1 up 1.00000 1.00000 2 hdd 2.00000 osd.2 up 1.00000 1.00000 9 hdd 0.78119 osd.9 up 1.00000 1.00000 -5 6.78119 host ceph-node2 3 hdd 2.00000 osd.3 up 1.00000 1.00000 4 hdd 2.00000 osd.4 up 1.00000 1.00000 5 hdd 2.00000 osd.5 up 1.00000 1.00000 10 hdd 0.78119 osd.10 up 1.00000 1.00000 -7 6.78119 host ceph-node3 6 hdd 2.00000 osd.6 up 1.00000 1.00000 7 hdd 2.00000 osd.7 up 1.00000 1.00000 8 hdd 2.00000 osd.8 up 1.00000 1.00000 11 hdd 0.78119 osd.11 up 1.00000 1.00000
所有 OSD 都添加成功。检测下 ceph 集群的健康状态
[root@ceph-node1 ~]# ceph -s cluster: id: fe899ca1-b3c0-46f6-aa4a-42591760860f health: HEALTH_WARN no active mgr clock skew detected on mon.ceph-node2, mon.ceph-node3 services: mon: 3 daemons, quorum ceph-node1,ceph-node2,ceph-node3 mgr: no daemons active osd: 12 osds: 12 up, 12 in data: pools: 0 pools, 0 pgs objects: 0 objects, 0B usage: 0B used, 0B / 0B avail pgs:
提示:no active mgr 这个 manager 是在 J 版本以后加入的,此时需要开启这个功能模块:
[root@ceph-node1 ~]# cd mycluster/ [root@ceph-node1 mycluster]# ceph-deploy mgr create ceph-node1 ceph-node2 ceph-node3
注意:只要是牵扯到要执行 ceph-deploy 都需要进入到 mycluster 否则会报错。
再次查看集群健康
[root@ceph-node1 mycluster]# ceph -s cluster: id: fe899ca1-b3c0-46f6-aa4a-42591760860f health: HEALTH_WARN clock skew detected on mon.ceph-node3 services: mon: 3 daemons, quorum ceph-node1,ceph-node2,ceph-node3 mgr: ceph-node1(active), standbys: ceph-node3, ceph-node2 osd: 12 osds: 12 up, 12 in data: pools: 0 pools, 0 pgs objects: 0 objects, 0B usage: 12.1GiB used, 20.3TiB / 20.3TiB avail pgs:
clock skew detected 时钟偏移问题可以通过修改配置文件实现,也就是加大群集节点的时间偏移量,避免告警信息:
[root@ceph-node1 ~]# cd mycluster/ [root@ceph-node1 mycluster]# vim ceph.conf # 最后两行追加内容 ... mon clock drift allowed = 2 mon clock drift warn backoff = 30 # 将配置文件推送到集群的各个节点 [root@ceph-node1 mycluster]# ceph-deploy --overwrite-conf config push ceph-node{1..3} # 重启服务 [root@ceph-node1 mycluster]# ansible ceph -m systemd -a 'name=ceph-mon.target state=restarted' # 告警消除 [root@ceph-node1 mycluster]# ceph -s cluster: id: fe899ca1-b3c0-46f6-aa4a-42591760860f health: HEALTH_OK services: mon: 3 daemons, quorum ceph-node1,ceph-node2,ceph-node3 mgr: ceph-node1(active), standbys: ceph-node3, ceph-node2 osd: 12 osds: 12 up, 12 in data: pools: 0 pools, 0 pgs objects: 0 objects, 0B usage: 12.1GiB used, 20.3TiB / 20.3TiB avail pgs:
接下来就需要对 OSD 进行分组,在分组之前,需要介绍一个新功能。
Ceph 从 L 版本开始新增了一个功能叫 crush class,又被称为 磁盘智能分组。因为这个功能就是根据磁盘类型自动进行属性关联,然后进行分类减少了很多人为的操作。在这个功能之前,如果需要对ssd和hdd进行分组的时候,需要大量的修改 crushmap,然后绑定不同的存储池到不同的 crush树上面,而这个功能简化了这种逻辑。
配置 crush class
默认情况下,所有 osd 都会 class 的类型是 hdd:
[root@ceph-node1 mycluster]# ceph osd crush class ls [ "hdd" ]
查看当前OSD
[root@ceph-node1 mycluster]# ceph osd tree ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF -1 20.34357 root default -3 6.78119 host ceph-node1 0 hdd 2.00000 osd.0 up 1.00000 1.00000 1 hdd 2.00000 osd.1 up 1.00000 1.00000 2 hdd 2.00000 osd.2 up 1.00000 1.00000 9 hdd 0.78119 osd.9 up 1.00000 1.00000 -5 6.78119 host ceph-node2 3 hdd 2.00000 osd.3 up 1.00000 1.00000 4 hdd 2.00000 osd.4 up 1.00000 1.00000 5 hdd 2.00000 osd.5 up 1.00000 1.00000 10 hdd 0.78119 osd.10 up 1.00000 1.00000 -7 6.78119 host ceph-node3 6 hdd 2.00000 osd.6 up 1.00000 1.00000 7 hdd 2.00000 osd.7 up 1.00000 1.00000 8 hdd 2.00000 osd.8 up 1.00000 1.00000 11 hdd 0.78119 osd.11 up 1.00000 1.00000
可以看到,当前有3个节点,每个节点上有4个OSD,根据之前的分组,一组为 SATA 另一组为 SSD ,而且根据上面创建 OSD 的顺序,可以做以下总结:
既然默认都为 hdd组,那SATA就不用在创建了,osd9-11 创建为 ssd 组即可。首先,需要将 osd9-11 从 hdd 组中去除掉:
[root@ceph-node1 mycluster]# for i in {9..11};do ceph osd crush rm-device-class osd.$i;done done removing class of osd(s): 9 done removing class of osd(s): 10 done removing class of osd(s): 11
查看 osd
[root@ceph-node1 mycluster]# ceph osd tree ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF -1 20.34357 root default -3 6.78119 host ceph-node1 9 0.78119 osd.9 up 1.00000 1.00000 0 hdd 2.00000 osd.0 up 1.00000 1.00000 1 hdd 2.00000 osd.1 up 1.00000 1.00000 2 hdd 2.00000 osd.2 up 1.00000 1.00000 -5 6.78119 host ceph-node2 10 0.78119 osd.10 up 1.00000 1.00000 3 hdd 2.00000 osd.3 up 1.00000 1.00000 4 hdd 2.00000 osd.4 up 1.00000 1.00000 5 hdd 2.00000 osd.5 up 1.00000 1.00000 -7 6.78119 host ceph-node3 11 0.78119 osd.11 up 1.00000 1.00000 6 hdd 2.00000 osd.6 up 1.00000 1.00000 7 hdd 2.00000 osd.7 up 1.00000 1.00000 8 hdd 2.00000 osd.8 up 1.00000 1.00000
可以看到 osd9-11 class 列已经没有 hdd 标识了。此时就可以通过命令将osd9-11添加到 ssd 组了,如下:
# 将 osd9-11添加到 ssd 组 [root@ceph-node1 mycluster]# for i in {9..11}; do ceph osd crush set-device-class ssd osd.$i;done set osd(s) 9 to class 'ssd' set osd(s) 10 to class 'ssd' set osd(s) 11 to class 'ssd' # 查看 osd [root@ceph-node1 mycluster]# ceph osd tree ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF -1 20.34357 root default -3 6.78119 host ceph-node1 0 hdd 2.00000 osd.0 up 1.00000 1.00000 1 hdd 2.00000 osd.1 up 1.00000 1.00000 2 hdd 2.00000 osd.2 up 1.00000 1.00000 9 ssd 0.78119 osd.9 up 1.00000 1.00000 -5 6.78119 host ceph-node2 3 hdd 2.00000 osd.3 up 1.00000 1.00000 4 hdd 2.00000 osd.4 up 1.00000 1.00000 5 hdd 2.00000 osd.5 up 1.00000 1.00000 10 ssd 0.78119 osd.10 up 1.00000 1.00000 -7 6.78119 host ceph-node3 6 hdd 2.00000 osd.6 up 1.00000 1.00000 7 hdd 2.00000 osd.7 up 1.00000 1.00000 8 hdd 2.00000 osd.8 up 1.00000 1.00000 11 ssd 0.78119 osd.11 up 1.00000 1.00000 # 查看 class [root@ceph-node1 mycluster]# ceph osd crush class ls [ "hdd", "ssd" ]
可以发现 osd9-11 的 class 列都变为 ssd,查看 crush class 也多出一个 ssd 的组,接下来就需要创建ssd 的规则
创建基于 ssd 的 class rule
创建一个 class rule,取名为 root-ssd,使用 ssd的osd:
[root@ceph-node1 mycluster]# ceph osd crush rule create-replicated root_ssd default host ssd # 查看创建的 rule [root@ceph-node1 mycluster]# ceph osd crush rule ls replicated_rule root_ssd
通过以下操作可以查看详细的 crushmap信息:
[root@ceph-node1 mycluster]# ceph osd getcrushmap -o /tmp/crushmap 32 [root@ceph-node1 mycluster]# crushtool -d /tmp/crushmap -o /tmp/crushmap.txt [root@ceph-node1 mycluster]# cat /tmp/crushmap.txt # begin crush map tunable choose_local_tries 0 tunable choose_local_fallback_tries 0 tunable choose_total_tries 50 tunable chooseleaf_descend_once 1 tunable chooseleaf_vary_r 1 tunable chooseleaf_stable 1 tunable straw_calc_version 1 tunable allowed_bucket_algs 54 # devices device 0 osd.0 class hdd device 1 osd.1 class hdd device 2 osd.2 class hdd device 3 osd.3 class hdd device 4 osd.4 class hdd device 5 osd.5 class hdd device 6 osd.6 class hdd device 7 osd.7 class hdd device 8 osd.8 class hdd device 9 osd.9 class ssd device 10 osd.10 class ssd device 11 osd.11 class ssd # types type 0 osd type 1 host type 2 chassis type 3 rack type 4 row type 5 pdu type 6 pod type 7 room type 8 datacenter type 9 region type 10 root # buckets host ceph-node1 { id -3 # do not change unnecessarily id -4 class hdd # do not change unnecessarily id -9 class ssd # do not change unnecessarily # weight 6.781 alg straw2 hash 0 # rjenkins1 item osd.0 weight 2.000 item osd.1 weight 2.000 item osd.2 weight 2.000 item osd.9 weight 0.781 } host ceph-node2 { id -5 # do not change unnecessarily id -6 class hdd # do not change unnecessarily id -10 class ssd # do not change unnecessarily # weight 6.781 alg straw2 hash 0 # rjenkins1 item osd.3 weight 2.000 item osd.4 weight 2.000 item osd.5 weight 2.000 item osd.10 weight 0.781 } host ceph-node3 { id -7 # do not change unnecessarily id -8 class hdd # do not change unnecessarily id -11 class ssd # do not change unnecessarily # weight 6.781 alg straw2 hash 0 # rjenkins1 item osd.6 weight 2.000 item osd.7 weight 2.000 item osd.8 weight 2.000 item osd.11 weight 0.781 } root default { id -1 # do not change unnecessarily id -2 class hdd # do not change unnecessarily id -12 class ssd # do not change unnecessarily # weight 20.344 alg straw2 hash 0 # rjenkins1 item ceph-node1 weight 6.781 item ceph-node2 weight 6.781 item ceph-node3 weight 6.781 } # rules rule replicated_rule { id 0 type replicated min_size 1 max_size 10 step take default step chooseleaf firstn 0 type host step emit } rule root_ssd { id 1 type replicated min_size 1 max_size 10 step take default class ssd step chooseleaf firstn 0 type host step emit } # end crush map
到此为止,使用智能分组的功能,实现了 SATA 和 SSD 的分离。
测试 ssd分组
将 osd9-11 作为 ssd分组使用,来测试下是否真的将数据存入了 osd9-11
(1)创建一个基于root_ssd 规则的存储池
# 查看当前的 rule [root@ceph-node1 mycluster]# ceph osd crush rule ls replicated_rule root_ssd # 通过 root_ssd 创建 [root@ceph-node1 mycluster]# ceph osd pool create ssd_test 64 64 root_ssd pool 'ssd_test' created
(2)测试基于ssd的池
# 创建测试文件 [root@ceph-node1 mycluster]# echo 'test' > test.txt # 将测试文件以对象名test 写入到 ssd_test 池 [root@ceph-node1 mycluster]# rados -p ssd_test put test test.txt [root@ceph-node1 mycluster]# rados -p ssd_test ls test # 查看ssd_test 中对象名为 test 的文件分布在 OSD 的位置 [root@ceph-node1 mycluster]# ceph osd map ssd_test test osdmap e63 pool 'ssd_test' (1) object 'test' -> pg 1.40e8aab5 (1.35) -> up ([10,11,9], p10) acting ([10,11,9], p10)
可以看到,文件 test.txt 存储在 osd9-11 中,符合预期设定。
9. 总结
本文也算是把 CentOS 7 结合 Ceph L版本的配置完整的走了一边,可用于生产环境的部署。后续会记录一些运维中遇到的操作。